CodeV-SVA-no-think-8B / trainer_state.json
wyt2000's picture
Upload folder using huggingface_hub
703d322 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 1300,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0015384615384615385,
"grad_norm": 60.89158644985447,
"learning_rate": 2e-05,
"loss": 1.2107,
"step": 1
},
{
"epoch": 0.003076923076923077,
"grad_norm": 31.499919740759644,
"learning_rate": 1.9999992699997636e-05,
"loss": 0.7837,
"step": 2
},
{
"epoch": 0.004615384615384616,
"grad_norm": 6.103692995379824,
"learning_rate": 1.999997080000119e-05,
"loss": 0.3218,
"step": 3
},
{
"epoch": 0.006153846153846154,
"grad_norm": 2.6468306725602764,
"learning_rate": 1.9999934300042646e-05,
"loss": 0.2287,
"step": 4
},
{
"epoch": 0.007692307692307693,
"grad_norm": 3.4867739541054443,
"learning_rate": 1.9999883200175286e-05,
"loss": 0.2423,
"step": 5
},
{
"epoch": 0.009230769230769232,
"grad_norm": 2.641228316079499,
"learning_rate": 1.9999817500473724e-05,
"loss": 0.2253,
"step": 6
},
{
"epoch": 0.010769230769230769,
"grad_norm": 1.205202032246029,
"learning_rate": 1.9999737201033877e-05,
"loss": 0.1668,
"step": 7
},
{
"epoch": 0.012307692307692308,
"grad_norm": 1.9244405042319512,
"learning_rate": 1.999964230197298e-05,
"loss": 0.1847,
"step": 8
},
{
"epoch": 0.013846153846153847,
"grad_norm": 1.4464951947932236,
"learning_rate": 1.999953280342959e-05,
"loss": 0.1606,
"step": 9
},
{
"epoch": 0.015384615384615385,
"grad_norm": 1.4255296128290331,
"learning_rate": 1.999940870556357e-05,
"loss": 0.1471,
"step": 10
},
{
"epoch": 0.016923076923076923,
"grad_norm": 1.3096222246733347,
"learning_rate": 1.9999270008556108e-05,
"loss": 0.1463,
"step": 11
},
{
"epoch": 0.018461538461538463,
"grad_norm": 1.018551951164187,
"learning_rate": 1.99991167126097e-05,
"loss": 0.1405,
"step": 12
},
{
"epoch": 0.02,
"grad_norm": 1.1849805458454175,
"learning_rate": 1.9998948817948157e-05,
"loss": 0.1381,
"step": 13
},
{
"epoch": 0.021538461538461538,
"grad_norm": 1.01325599101221,
"learning_rate": 1.9998766324816606e-05,
"loss": 0.1453,
"step": 14
},
{
"epoch": 0.023076923076923078,
"grad_norm": 1.0760344920602294,
"learning_rate": 1.999856923348149e-05,
"loss": 0.119,
"step": 15
},
{
"epoch": 0.024615384615384615,
"grad_norm": 1.2012958039069457,
"learning_rate": 1.9998357544230558e-05,
"loss": 0.1507,
"step": 16
},
{
"epoch": 0.026153846153846153,
"grad_norm": 0.8734600178036379,
"learning_rate": 1.9998131257372878e-05,
"loss": 0.1334,
"step": 17
},
{
"epoch": 0.027692307692307693,
"grad_norm": 0.7321244232592351,
"learning_rate": 1.9997890373238827e-05,
"loss": 0.1259,
"step": 18
},
{
"epoch": 0.02923076923076923,
"grad_norm": 0.8736141545923927,
"learning_rate": 1.99976348921801e-05,
"loss": 0.1328,
"step": 19
},
{
"epoch": 0.03076923076923077,
"grad_norm": 0.7433251995484863,
"learning_rate": 1.9997364814569696e-05,
"loss": 0.1155,
"step": 20
},
{
"epoch": 0.03230769230769231,
"grad_norm": 0.7421987879217785,
"learning_rate": 1.9997080140801932e-05,
"loss": 0.1321,
"step": 21
},
{
"epoch": 0.033846153846153845,
"grad_norm": 0.7417777234873172,
"learning_rate": 1.9996780871292428e-05,
"loss": 0.1082,
"step": 22
},
{
"epoch": 0.03538461538461538,
"grad_norm": 1.2517479737195332,
"learning_rate": 1.999646700647812e-05,
"loss": 0.1219,
"step": 23
},
{
"epoch": 0.036923076923076927,
"grad_norm": 0.9094077693848758,
"learning_rate": 1.9996138546817248e-05,
"loss": 0.1314,
"step": 24
},
{
"epoch": 0.038461538461538464,
"grad_norm": 0.7797984417608248,
"learning_rate": 1.9995795492789368e-05,
"loss": 0.1331,
"step": 25
},
{
"epoch": 0.04,
"grad_norm": 0.7976816314521917,
"learning_rate": 1.9995437844895337e-05,
"loss": 0.114,
"step": 26
},
{
"epoch": 0.04153846153846154,
"grad_norm": 0.8888193123533839,
"learning_rate": 1.9995065603657317e-05,
"loss": 0.1399,
"step": 27
},
{
"epoch": 0.043076923076923075,
"grad_norm": 0.8169350017845466,
"learning_rate": 1.9994678769618784e-05,
"loss": 0.1148,
"step": 28
},
{
"epoch": 0.04461538461538461,
"grad_norm": 0.6804970735056192,
"learning_rate": 1.999427734334452e-05,
"loss": 0.1223,
"step": 29
},
{
"epoch": 0.046153846153846156,
"grad_norm": 0.7358036673244082,
"learning_rate": 1.99938613254206e-05,
"loss": 0.1281,
"step": 30
},
{
"epoch": 0.047692307692307694,
"grad_norm": 0.726915700239593,
"learning_rate": 1.9993430716454415e-05,
"loss": 0.123,
"step": 31
},
{
"epoch": 0.04923076923076923,
"grad_norm": 0.8370086699996789,
"learning_rate": 1.9992985517074653e-05,
"loss": 0.1121,
"step": 32
},
{
"epoch": 0.05076923076923077,
"grad_norm": 0.870233844578574,
"learning_rate": 1.9992525727931303e-05,
"loss": 0.1242,
"step": 33
},
{
"epoch": 0.052307692307692305,
"grad_norm": 0.6572319680505859,
"learning_rate": 1.9992051349695662e-05,
"loss": 0.115,
"step": 34
},
{
"epoch": 0.05384615384615385,
"grad_norm": 0.8992858978866627,
"learning_rate": 1.9991562383060316e-05,
"loss": 0.1173,
"step": 35
},
{
"epoch": 0.055384615384615386,
"grad_norm": 0.6238895374463513,
"learning_rate": 1.9991058828739164e-05,
"loss": 0.1188,
"step": 36
},
{
"epoch": 0.05692307692307692,
"grad_norm": 0.5905879008819883,
"learning_rate": 1.9990540687467394e-05,
"loss": 0.1162,
"step": 37
},
{
"epoch": 0.05846153846153846,
"grad_norm": 0.7475485922157608,
"learning_rate": 1.9990007960001487e-05,
"loss": 0.1341,
"step": 38
},
{
"epoch": 0.06,
"grad_norm": 0.7062185012543412,
"learning_rate": 1.9989460647119232e-05,
"loss": 0.1276,
"step": 39
},
{
"epoch": 0.06153846153846154,
"grad_norm": 0.6940436820752224,
"learning_rate": 1.9988898749619702e-05,
"loss": 0.1173,
"step": 40
},
{
"epoch": 0.06307692307692307,
"grad_norm": 0.6162083864347133,
"learning_rate": 1.998832226832327e-05,
"loss": 0.1148,
"step": 41
},
{
"epoch": 0.06461538461538462,
"grad_norm": 0.5010819273602969,
"learning_rate": 1.9987731204071596e-05,
"loss": 0.0971,
"step": 42
},
{
"epoch": 0.06615384615384616,
"grad_norm": 0.6715601368175806,
"learning_rate": 1.9987125557727633e-05,
"loss": 0.1279,
"step": 43
},
{
"epoch": 0.06769230769230769,
"grad_norm": 0.6360471415422658,
"learning_rate": 1.998650533017563e-05,
"loss": 0.1106,
"step": 44
},
{
"epoch": 0.06923076923076923,
"grad_norm": 0.6293809806468533,
"learning_rate": 1.9985870522321118e-05,
"loss": 0.1367,
"step": 45
},
{
"epoch": 0.07076923076923076,
"grad_norm": 0.6592095131341471,
"learning_rate": 1.9985221135090917e-05,
"loss": 0.1149,
"step": 46
},
{
"epoch": 0.07230769230769231,
"grad_norm": 0.7078516914096971,
"learning_rate": 1.9984557169433126e-05,
"loss": 0.1115,
"step": 47
},
{
"epoch": 0.07384615384615385,
"grad_norm": 0.5877725742836197,
"learning_rate": 1.9983878626317147e-05,
"loss": 0.1236,
"step": 48
},
{
"epoch": 0.07538461538461538,
"grad_norm": 0.5884104828878249,
"learning_rate": 1.9983185506733643e-05,
"loss": 0.12,
"step": 49
},
{
"epoch": 0.07692307692307693,
"grad_norm": 0.549670934870528,
"learning_rate": 1.9982477811694578e-05,
"loss": 0.1104,
"step": 50
},
{
"epoch": 0.07846153846153846,
"grad_norm": 0.8220615421808383,
"learning_rate": 1.9981755542233175e-05,
"loss": 0.1219,
"step": 51
},
{
"epoch": 0.08,
"grad_norm": 0.6798315000266852,
"learning_rate": 1.998101869940396e-05,
"loss": 0.1182,
"step": 52
},
{
"epoch": 0.08153846153846153,
"grad_norm": 0.5366855615637781,
"learning_rate": 1.9980267284282718e-05,
"loss": 0.1017,
"step": 53
},
{
"epoch": 0.08307692307692308,
"grad_norm": 0.7397417525631895,
"learning_rate": 1.9979501297966516e-05,
"loss": 0.114,
"step": 54
},
{
"epoch": 0.08461538461538462,
"grad_norm": 0.7086921001674501,
"learning_rate": 1.9978720741573693e-05,
"loss": 0.1179,
"step": 55
},
{
"epoch": 0.08615384615384615,
"grad_norm": 0.7075396570589779,
"learning_rate": 1.9977925616243865e-05,
"loss": 0.1143,
"step": 56
},
{
"epoch": 0.0876923076923077,
"grad_norm": 0.5477986445198524,
"learning_rate": 1.9977115923137912e-05,
"loss": 0.1115,
"step": 57
},
{
"epoch": 0.08923076923076922,
"grad_norm": 0.5949900627411874,
"learning_rate": 1.997629166343799e-05,
"loss": 0.1056,
"step": 58
},
{
"epoch": 0.09076923076923077,
"grad_norm": 0.60111264237703,
"learning_rate": 1.9975452838347513e-05,
"loss": 0.1097,
"step": 59
},
{
"epoch": 0.09230769230769231,
"grad_norm": 0.6683497211941376,
"learning_rate": 1.997459944909117e-05,
"loss": 0.1164,
"step": 60
},
{
"epoch": 0.09384615384615384,
"grad_norm": 0.7704030456569986,
"learning_rate": 1.9973731496914914e-05,
"loss": 0.1164,
"step": 61
},
{
"epoch": 0.09538461538461539,
"grad_norm": 0.707210267374331,
"learning_rate": 1.9972848983085945e-05,
"loss": 0.1192,
"step": 62
},
{
"epoch": 0.09692307692307692,
"grad_norm": 0.6016618867618871,
"learning_rate": 1.9971951908892743e-05,
"loss": 0.111,
"step": 63
},
{
"epoch": 0.09846153846153846,
"grad_norm": 0.662240507889871,
"learning_rate": 1.997104027564503e-05,
"loss": 0.1163,
"step": 64
},
{
"epoch": 0.1,
"grad_norm": 0.5798279422985586,
"learning_rate": 1.9970114084673796e-05,
"loss": 0.1102,
"step": 65
},
{
"epoch": 0.10153846153846154,
"grad_norm": 0.5611805453593508,
"learning_rate": 1.9969173337331283e-05,
"loss": 0.1104,
"step": 66
},
{
"epoch": 0.10307692307692308,
"grad_norm": 0.7514312817329308,
"learning_rate": 1.996821803499097e-05,
"loss": 0.1239,
"step": 67
},
{
"epoch": 0.10461538461538461,
"grad_norm": 0.5863699065118245,
"learning_rate": 1.9967248179047612e-05,
"loss": 0.1132,
"step": 68
},
{
"epoch": 0.10615384615384615,
"grad_norm": 0.5464510643573779,
"learning_rate": 1.9966263770917192e-05,
"loss": 0.1211,
"step": 69
},
{
"epoch": 0.1076923076923077,
"grad_norm": 0.566891174856317,
"learning_rate": 1.996526481203695e-05,
"loss": 0.1077,
"step": 70
},
{
"epoch": 0.10923076923076923,
"grad_norm": 0.6246113449127777,
"learning_rate": 1.9964251303865362e-05,
"loss": 0.1074,
"step": 71
},
{
"epoch": 0.11076923076923077,
"grad_norm": 0.6001393125416704,
"learning_rate": 1.9963223247882154e-05,
"loss": 0.1086,
"step": 72
},
{
"epoch": 0.1123076923076923,
"grad_norm": 0.5021027088495942,
"learning_rate": 1.996218064558829e-05,
"loss": 0.0983,
"step": 73
},
{
"epoch": 0.11384615384615385,
"grad_norm": 0.5707671002037145,
"learning_rate": 1.9961123498505965e-05,
"loss": 0.1076,
"step": 74
},
{
"epoch": 0.11538461538461539,
"grad_norm": 0.6742358069569517,
"learning_rate": 1.9960051808178616e-05,
"loss": 0.1148,
"step": 75
},
{
"epoch": 0.11692307692307692,
"grad_norm": 0.48574494522940415,
"learning_rate": 1.995896557617091e-05,
"loss": 0.1104,
"step": 76
},
{
"epoch": 0.11846153846153847,
"grad_norm": 0.5943862604998493,
"learning_rate": 1.9957864804068752e-05,
"loss": 0.1076,
"step": 77
},
{
"epoch": 0.12,
"grad_norm": 0.6097303023134425,
"learning_rate": 1.9956749493479263e-05,
"loss": 0.1115,
"step": 78
},
{
"epoch": 0.12153846153846154,
"grad_norm": 0.6195963924681878,
"learning_rate": 1.99556196460308e-05,
"loss": 0.1212,
"step": 79
},
{
"epoch": 0.12307692307692308,
"grad_norm": 0.5930349140433203,
"learning_rate": 1.9954475263372943e-05,
"loss": 0.1074,
"step": 80
},
{
"epoch": 0.12461538461538461,
"grad_norm": 0.5556371192720652,
"learning_rate": 1.995331634717649e-05,
"loss": 0.1064,
"step": 81
},
{
"epoch": 0.12615384615384614,
"grad_norm": 0.6530827388442122,
"learning_rate": 1.9952142899133452e-05,
"loss": 0.1121,
"step": 82
},
{
"epoch": 0.1276923076923077,
"grad_norm": 0.8859927355734215,
"learning_rate": 1.9950954920957074e-05,
"loss": 0.1152,
"step": 83
},
{
"epoch": 0.12923076923076923,
"grad_norm": 0.5935612351835159,
"learning_rate": 1.99497524143818e-05,
"loss": 0.1179,
"step": 84
},
{
"epoch": 0.13076923076923078,
"grad_norm": 0.4991766795445498,
"learning_rate": 1.994853538116329e-05,
"loss": 0.1149,
"step": 85
},
{
"epoch": 0.13230769230769232,
"grad_norm": 0.5555671407736567,
"learning_rate": 1.9947303823078418e-05,
"loss": 0.1012,
"step": 86
},
{
"epoch": 0.13384615384615384,
"grad_norm": 0.8909613249571459,
"learning_rate": 1.994605774192525e-05,
"loss": 0.0979,
"step": 87
},
{
"epoch": 0.13538461538461538,
"grad_norm": 0.5310150106650096,
"learning_rate": 1.9944797139523068e-05,
"loss": 0.1088,
"step": 88
},
{
"epoch": 0.13692307692307693,
"grad_norm": 0.749673935234374,
"learning_rate": 1.994352201771236e-05,
"loss": 0.1034,
"step": 89
},
{
"epoch": 0.13846153846153847,
"grad_norm": 0.5378188928120664,
"learning_rate": 1.99422323783548e-05,
"loss": 0.1057,
"step": 90
},
{
"epoch": 0.14,
"grad_norm": 0.511920881530378,
"learning_rate": 1.9940928223333254e-05,
"loss": 0.1121,
"step": 91
},
{
"epoch": 0.14153846153846153,
"grad_norm": 0.5822824054135747,
"learning_rate": 1.99396095545518e-05,
"loss": 0.1148,
"step": 92
},
{
"epoch": 0.14307692307692307,
"grad_norm": 0.6672810329410515,
"learning_rate": 1.9938276373935688e-05,
"loss": 0.1158,
"step": 93
},
{
"epoch": 0.14461538461538462,
"grad_norm": 0.4952193512759965,
"learning_rate": 1.9936928683431368e-05,
"loss": 0.1076,
"step": 94
},
{
"epoch": 0.14615384615384616,
"grad_norm": 0.625543647549158,
"learning_rate": 1.9935566485006464e-05,
"loss": 0.1125,
"step": 95
},
{
"epoch": 0.1476923076923077,
"grad_norm": 0.6035686191037584,
"learning_rate": 1.993418978064979e-05,
"loss": 0.109,
"step": 96
},
{
"epoch": 0.14923076923076922,
"grad_norm": 0.5236314046610593,
"learning_rate": 1.993279857237133e-05,
"loss": 0.1152,
"step": 97
},
{
"epoch": 0.15076923076923077,
"grad_norm": 0.6169915505799958,
"learning_rate": 1.9931392862202255e-05,
"loss": 0.1187,
"step": 98
},
{
"epoch": 0.1523076923076923,
"grad_norm": 0.5079224872644268,
"learning_rate": 1.99299726521949e-05,
"loss": 0.1066,
"step": 99
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.6279665388236755,
"learning_rate": 1.992853794442277e-05,
"loss": 0.1077,
"step": 100
},
{
"epoch": 0.15538461538461537,
"grad_norm": 0.5259964466933738,
"learning_rate": 1.992708874098054e-05,
"loss": 0.1107,
"step": 101
},
{
"epoch": 0.15692307692307692,
"grad_norm": 0.5221587453615723,
"learning_rate": 1.9925625043984052e-05,
"loss": 0.103,
"step": 102
},
{
"epoch": 0.15846153846153846,
"grad_norm": 0.5299978643337,
"learning_rate": 1.9924146855570298e-05,
"loss": 0.1083,
"step": 103
},
{
"epoch": 0.16,
"grad_norm": 0.5822846737863101,
"learning_rate": 1.9922654177897436e-05,
"loss": 0.1069,
"step": 104
},
{
"epoch": 0.16153846153846155,
"grad_norm": 0.5319364893865596,
"learning_rate": 1.9921147013144782e-05,
"loss": 0.0997,
"step": 105
},
{
"epoch": 0.16307692307692306,
"grad_norm": 0.6977409875795292,
"learning_rate": 1.9919625363512788e-05,
"loss": 0.118,
"step": 106
},
{
"epoch": 0.1646153846153846,
"grad_norm": 0.48032168376624,
"learning_rate": 1.9918089231223066e-05,
"loss": 0.1048,
"step": 107
},
{
"epoch": 0.16615384615384615,
"grad_norm": 0.5258925153786688,
"learning_rate": 1.991653861851837e-05,
"loss": 0.1125,
"step": 108
},
{
"epoch": 0.1676923076923077,
"grad_norm": 0.7141043684744723,
"learning_rate": 1.99149735276626e-05,
"loss": 0.1181,
"step": 109
},
{
"epoch": 0.16923076923076924,
"grad_norm": 0.5038196623101785,
"learning_rate": 1.9913393960940785e-05,
"loss": 0.0971,
"step": 110
},
{
"epoch": 0.17076923076923076,
"grad_norm": 0.48472163979718835,
"learning_rate": 1.9911799920659093e-05,
"loss": 0.1039,
"step": 111
},
{
"epoch": 0.1723076923076923,
"grad_norm": 0.669381566002829,
"learning_rate": 1.9910191409144825e-05,
"loss": 0.1073,
"step": 112
},
{
"epoch": 0.17384615384615384,
"grad_norm": 0.57115456078115,
"learning_rate": 1.9908568428746408e-05,
"loss": 0.0954,
"step": 113
},
{
"epoch": 0.1753846153846154,
"grad_norm": 0.5395647646350689,
"learning_rate": 1.9906930981833392e-05,
"loss": 0.1088,
"step": 114
},
{
"epoch": 0.17692307692307693,
"grad_norm": 0.5912868508214638,
"learning_rate": 1.9905279070796454e-05,
"loss": 0.1035,
"step": 115
},
{
"epoch": 0.17846153846153845,
"grad_norm": 0.56400423327266,
"learning_rate": 1.9903612698047387e-05,
"loss": 0.1223,
"step": 116
},
{
"epoch": 0.18,
"grad_norm": 0.6509772627591756,
"learning_rate": 1.9901931866019087e-05,
"loss": 0.1165,
"step": 117
},
{
"epoch": 0.18153846153846154,
"grad_norm": 0.4920607096781014,
"learning_rate": 1.990023657716558e-05,
"loss": 0.1036,
"step": 118
},
{
"epoch": 0.18307692307692308,
"grad_norm": 1.003150053450431,
"learning_rate": 1.989852683396198e-05,
"loss": 0.109,
"step": 119
},
{
"epoch": 0.18461538461538463,
"grad_norm": 0.5598317796088923,
"learning_rate": 1.9896802638904512e-05,
"loss": 0.1099,
"step": 120
},
{
"epoch": 0.18615384615384614,
"grad_norm": 1.5739160303036723,
"learning_rate": 1.9895063994510512e-05,
"loss": 0.1149,
"step": 121
},
{
"epoch": 0.18769230769230769,
"grad_norm": 0.6575202377156634,
"learning_rate": 1.9893310903318394e-05,
"loss": 0.1115,
"step": 122
},
{
"epoch": 0.18923076923076923,
"grad_norm": 0.6474947347376268,
"learning_rate": 1.9891543367887675e-05,
"loss": 0.1105,
"step": 123
},
{
"epoch": 0.19076923076923077,
"grad_norm": 0.6273901725857466,
"learning_rate": 1.9889761390798952e-05,
"loss": 0.1137,
"step": 124
},
{
"epoch": 0.19230769230769232,
"grad_norm": 0.5821482969182745,
"learning_rate": 1.988796497465392e-05,
"loss": 0.0932,
"step": 125
},
{
"epoch": 0.19384615384615383,
"grad_norm": 0.6213235073882689,
"learning_rate": 1.9886154122075344e-05,
"loss": 0.1111,
"step": 126
},
{
"epoch": 0.19538461538461538,
"grad_norm": 0.6063602698529751,
"learning_rate": 1.988432883570707e-05,
"loss": 0.1102,
"step": 127
},
{
"epoch": 0.19692307692307692,
"grad_norm": 0.5820333042475602,
"learning_rate": 1.9882489118214013e-05,
"loss": 0.0904,
"step": 128
},
{
"epoch": 0.19846153846153847,
"grad_norm": 0.5911684289948976,
"learning_rate": 1.9880634972282168e-05,
"loss": 0.1168,
"step": 129
},
{
"epoch": 0.2,
"grad_norm": 0.6156260168090836,
"learning_rate": 1.987876640061858e-05,
"loss": 0.1037,
"step": 130
},
{
"epoch": 0.20153846153846153,
"grad_norm": 0.6339706412660758,
"learning_rate": 1.9876883405951378e-05,
"loss": 0.1023,
"step": 131
},
{
"epoch": 0.20307692307692307,
"grad_norm": 0.5925980576093635,
"learning_rate": 1.9874985991029725e-05,
"loss": 0.1058,
"step": 132
},
{
"epoch": 0.20461538461538462,
"grad_norm": 0.44542298631854954,
"learning_rate": 1.987307415862385e-05,
"loss": 0.0947,
"step": 133
},
{
"epoch": 0.20615384615384616,
"grad_norm": 0.6583404135837835,
"learning_rate": 1.987114791152503e-05,
"loss": 0.1079,
"step": 134
},
{
"epoch": 0.2076923076923077,
"grad_norm": 0.6888868393971377,
"learning_rate": 1.9869207252545582e-05,
"loss": 0.1046,
"step": 135
},
{
"epoch": 0.20923076923076922,
"grad_norm": 0.5864137998381262,
"learning_rate": 1.9867252184518878e-05,
"loss": 0.1128,
"step": 136
},
{
"epoch": 0.21076923076923076,
"grad_norm": 0.6371765472826588,
"learning_rate": 1.986528271029931e-05,
"loss": 0.1108,
"step": 137
},
{
"epoch": 0.2123076923076923,
"grad_norm": 0.6990532523375615,
"learning_rate": 1.9863298832762317e-05,
"loss": 0.1121,
"step": 138
},
{
"epoch": 0.21384615384615385,
"grad_norm": 0.5777908223862595,
"learning_rate": 1.9861300554804357e-05,
"loss": 0.1041,
"step": 139
},
{
"epoch": 0.2153846153846154,
"grad_norm": 0.5437019574342667,
"learning_rate": 1.985928787934292e-05,
"loss": 0.1041,
"step": 140
},
{
"epoch": 0.2169230769230769,
"grad_norm": 0.5652145969922582,
"learning_rate": 1.985726080931651e-05,
"loss": 0.0977,
"step": 141
},
{
"epoch": 0.21846153846153846,
"grad_norm": 0.6831475109006377,
"learning_rate": 1.9855219347684654e-05,
"loss": 0.1136,
"step": 142
},
{
"epoch": 0.22,
"grad_norm": 0.5819561900456608,
"learning_rate": 1.9853163497427885e-05,
"loss": 0.1022,
"step": 143
},
{
"epoch": 0.22153846153846155,
"grad_norm": 0.5243548380003794,
"learning_rate": 1.985109326154774e-05,
"loss": 0.1068,
"step": 144
},
{
"epoch": 0.2230769230769231,
"grad_norm": 0.6756018205387149,
"learning_rate": 1.9849008643066774e-05,
"loss": 0.1116,
"step": 145
},
{
"epoch": 0.2246153846153846,
"grad_norm": 0.5271647520035011,
"learning_rate": 1.9846909645028524e-05,
"loss": 0.1004,
"step": 146
},
{
"epoch": 0.22615384615384615,
"grad_norm": 0.49376009484771377,
"learning_rate": 1.984479627049753e-05,
"loss": 0.0986,
"step": 147
},
{
"epoch": 0.2276923076923077,
"grad_norm": 0.5009074295953662,
"learning_rate": 1.9842668522559326e-05,
"loss": 0.1083,
"step": 148
},
{
"epoch": 0.22923076923076924,
"grad_norm": 0.5120110822425591,
"learning_rate": 1.9840526404320415e-05,
"loss": 0.1035,
"step": 149
},
{
"epoch": 0.23076923076923078,
"grad_norm": 0.4393813675110799,
"learning_rate": 1.9838369918908295e-05,
"loss": 0.1116,
"step": 150
},
{
"epoch": 0.2323076923076923,
"grad_norm": 0.6224825279714632,
"learning_rate": 1.983619906947144e-05,
"loss": 0.1067,
"step": 151
},
{
"epoch": 0.23384615384615384,
"grad_norm": 0.5038143439477663,
"learning_rate": 1.9834013859179284e-05,
"loss": 0.1104,
"step": 152
},
{
"epoch": 0.2353846153846154,
"grad_norm": 0.600725559118982,
"learning_rate": 1.9831814291222233e-05,
"loss": 0.1114,
"step": 153
},
{
"epoch": 0.23692307692307693,
"grad_norm": 0.561148124391653,
"learning_rate": 1.982960036881167e-05,
"loss": 0.1129,
"step": 154
},
{
"epoch": 0.23846153846153847,
"grad_norm": 0.4254379063861859,
"learning_rate": 1.982737209517991e-05,
"loss": 0.1018,
"step": 155
},
{
"epoch": 0.24,
"grad_norm": 0.5302552159073871,
"learning_rate": 1.982512947358024e-05,
"loss": 0.1103,
"step": 156
},
{
"epoch": 0.24153846153846154,
"grad_norm": 0.4768982697935648,
"learning_rate": 1.982287250728689e-05,
"loss": 0.0947,
"step": 157
},
{
"epoch": 0.24307692307692308,
"grad_norm": 0.4859749472919824,
"learning_rate": 1.9820601199595027e-05,
"loss": 0.1049,
"step": 158
},
{
"epoch": 0.24461538461538462,
"grad_norm": 0.49517524710561983,
"learning_rate": 1.981831555382076e-05,
"loss": 0.104,
"step": 159
},
{
"epoch": 0.24615384615384617,
"grad_norm": 0.462632720362608,
"learning_rate": 1.981601557330114e-05,
"loss": 0.101,
"step": 160
},
{
"epoch": 0.24769230769230768,
"grad_norm": 0.46154906358075987,
"learning_rate": 1.9813701261394136e-05,
"loss": 0.0971,
"step": 161
},
{
"epoch": 0.24923076923076923,
"grad_norm": 0.47520236421542805,
"learning_rate": 1.9811372621478643e-05,
"loss": 0.0887,
"step": 162
},
{
"epoch": 0.25076923076923074,
"grad_norm": 0.4386838659657528,
"learning_rate": 1.980902965695448e-05,
"loss": 0.1054,
"step": 163
},
{
"epoch": 0.2523076923076923,
"grad_norm": 0.5157191022451466,
"learning_rate": 1.9806672371242372e-05,
"loss": 0.1031,
"step": 164
},
{
"epoch": 0.25384615384615383,
"grad_norm": 0.49794944103500205,
"learning_rate": 1.9804300767783958e-05,
"loss": 0.1058,
"step": 165
},
{
"epoch": 0.2553846153846154,
"grad_norm": 0.5125213345791613,
"learning_rate": 1.9801914850041787e-05,
"loss": 0.1012,
"step": 166
},
{
"epoch": 0.2569230769230769,
"grad_norm": 0.5380352018124325,
"learning_rate": 1.979951462149929e-05,
"loss": 0.1096,
"step": 167
},
{
"epoch": 0.25846153846153846,
"grad_norm": 0.5283064485685516,
"learning_rate": 1.979710008566081e-05,
"loss": 0.0985,
"step": 168
},
{
"epoch": 0.26,
"grad_norm": 0.5187932463326723,
"learning_rate": 1.979467124605156e-05,
"loss": 0.0908,
"step": 169
},
{
"epoch": 0.26153846153846155,
"grad_norm": 0.5436736115984068,
"learning_rate": 1.979222810621766e-05,
"loss": 0.1033,
"step": 170
},
{
"epoch": 0.2630769230769231,
"grad_norm": 0.4474415033552116,
"learning_rate": 1.9789770669726088e-05,
"loss": 0.1053,
"step": 171
},
{
"epoch": 0.26461538461538464,
"grad_norm": 0.544303710798621,
"learning_rate": 1.97872989401647e-05,
"loss": 0.1159,
"step": 172
},
{
"epoch": 0.26615384615384613,
"grad_norm": 0.6201173307395067,
"learning_rate": 1.9784812921142232e-05,
"loss": 0.098,
"step": 173
},
{
"epoch": 0.2676923076923077,
"grad_norm": 0.447026396377833,
"learning_rate": 1.9782312616288262e-05,
"loss": 0.106,
"step": 174
},
{
"epoch": 0.2692307692307692,
"grad_norm": 0.5016997942574009,
"learning_rate": 1.977979802925324e-05,
"loss": 0.1059,
"step": 175
},
{
"epoch": 0.27076923076923076,
"grad_norm": 0.5883579853747641,
"learning_rate": 1.977726916370847e-05,
"loss": 0.1049,
"step": 176
},
{
"epoch": 0.2723076923076923,
"grad_norm": 0.43307778470945424,
"learning_rate": 1.977472602334609e-05,
"loss": 0.0917,
"step": 177
},
{
"epoch": 0.27384615384615385,
"grad_norm": 0.4108595764973751,
"learning_rate": 1.977216861187909e-05,
"loss": 0.0892,
"step": 178
},
{
"epoch": 0.2753846153846154,
"grad_norm": 0.5320809885623358,
"learning_rate": 1.976959693304129e-05,
"loss": 0.0994,
"step": 179
},
{
"epoch": 0.27692307692307694,
"grad_norm": 0.46480787256802003,
"learning_rate": 1.9767010990587342e-05,
"loss": 0.1013,
"step": 180
},
{
"epoch": 0.2784615384615385,
"grad_norm": 0.545604832179799,
"learning_rate": 1.9764410788292724e-05,
"loss": 0.0989,
"step": 181
},
{
"epoch": 0.28,
"grad_norm": 0.43112552357630096,
"learning_rate": 1.976179632995373e-05,
"loss": 0.1045,
"step": 182
},
{
"epoch": 0.2815384615384615,
"grad_norm": 0.6658366935016972,
"learning_rate": 1.9759167619387474e-05,
"loss": 0.1233,
"step": 183
},
{
"epoch": 0.28307692307692306,
"grad_norm": 0.46835535422995606,
"learning_rate": 1.9756524660431876e-05,
"loss": 0.0956,
"step": 184
},
{
"epoch": 0.2846153846153846,
"grad_norm": 0.4607296339259721,
"learning_rate": 1.9753867456945653e-05,
"loss": 0.1161,
"step": 185
},
{
"epoch": 0.28615384615384615,
"grad_norm": 0.41625713094163824,
"learning_rate": 1.9751196012808328e-05,
"loss": 0.0896,
"step": 186
},
{
"epoch": 0.2876923076923077,
"grad_norm": 0.46438501415946737,
"learning_rate": 1.9748510331920204e-05,
"loss": 0.1016,
"step": 187
},
{
"epoch": 0.28923076923076924,
"grad_norm": 0.5188858027802394,
"learning_rate": 1.9745810418202383e-05,
"loss": 0.1199,
"step": 188
},
{
"epoch": 0.2907692307692308,
"grad_norm": 0.4563595166293062,
"learning_rate": 1.9743096275596735e-05,
"loss": 0.1028,
"step": 189
},
{
"epoch": 0.2923076923076923,
"grad_norm": 0.48359296387625755,
"learning_rate": 1.9740367908065914e-05,
"loss": 0.0925,
"step": 190
},
{
"epoch": 0.29384615384615387,
"grad_norm": 0.5276131684607785,
"learning_rate": 1.9737625319593338e-05,
"loss": 0.1083,
"step": 191
},
{
"epoch": 0.2953846153846154,
"grad_norm": 0.38771222370278047,
"learning_rate": 1.973486851418318e-05,
"loss": 0.0904,
"step": 192
},
{
"epoch": 0.2969230769230769,
"grad_norm": 0.5429417039030707,
"learning_rate": 1.9732097495860388e-05,
"loss": 0.0963,
"step": 193
},
{
"epoch": 0.29846153846153844,
"grad_norm": 0.5257696459782908,
"learning_rate": 1.9729312268670642e-05,
"loss": 0.0976,
"step": 194
},
{
"epoch": 0.3,
"grad_norm": 0.4550685781931343,
"learning_rate": 1.972651283668038e-05,
"loss": 0.0992,
"step": 195
},
{
"epoch": 0.30153846153846153,
"grad_norm": 0.49086953834293273,
"learning_rate": 1.9723699203976768e-05,
"loss": 0.0994,
"step": 196
},
{
"epoch": 0.3030769230769231,
"grad_norm": 0.5264210943330065,
"learning_rate": 1.9720871374667714e-05,
"loss": 0.1021,
"step": 197
},
{
"epoch": 0.3046153846153846,
"grad_norm": 0.507327301576836,
"learning_rate": 1.9718029352881857e-05,
"loss": 0.0972,
"step": 198
},
{
"epoch": 0.30615384615384617,
"grad_norm": 0.4732432615412584,
"learning_rate": 1.971517314276854e-05,
"loss": 0.0918,
"step": 199
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.6042723204316705,
"learning_rate": 1.9712302748497838e-05,
"loss": 0.1064,
"step": 200
},
{
"epoch": 0.30923076923076925,
"grad_norm": 0.46540271774907827,
"learning_rate": 1.9709418174260523e-05,
"loss": 0.0906,
"step": 201
},
{
"epoch": 0.31076923076923074,
"grad_norm": 0.5214921994645452,
"learning_rate": 1.9706519424268077e-05,
"loss": 0.093,
"step": 202
},
{
"epoch": 0.3123076923076923,
"grad_norm": 0.4249410321063521,
"learning_rate": 1.9703606502752674e-05,
"loss": 0.1011,
"step": 203
},
{
"epoch": 0.31384615384615383,
"grad_norm": 0.5120931365008091,
"learning_rate": 1.970067941396719e-05,
"loss": 0.1093,
"step": 204
},
{
"epoch": 0.3153846153846154,
"grad_norm": 0.46019038450225697,
"learning_rate": 1.9697738162185163e-05,
"loss": 0.1094,
"step": 205
},
{
"epoch": 0.3169230769230769,
"grad_norm": 0.43539468765926526,
"learning_rate": 1.969478275170083e-05,
"loss": 0.0867,
"step": 206
},
{
"epoch": 0.31846153846153846,
"grad_norm": 0.499590646424274,
"learning_rate": 1.969181318682909e-05,
"loss": 0.0908,
"step": 207
},
{
"epoch": 0.32,
"grad_norm": 0.540818973654555,
"learning_rate": 1.9688829471905507e-05,
"loss": 0.1023,
"step": 208
},
{
"epoch": 0.32153846153846155,
"grad_norm": 0.4730876526778854,
"learning_rate": 1.9685831611286312e-05,
"loss": 0.1026,
"step": 209
},
{
"epoch": 0.3230769230769231,
"grad_norm": 0.49776463757579537,
"learning_rate": 1.968281960934838e-05,
"loss": 0.0996,
"step": 210
},
{
"epoch": 0.32461538461538464,
"grad_norm": 0.4311409270544766,
"learning_rate": 1.967979347048923e-05,
"loss": 0.099,
"step": 211
},
{
"epoch": 0.3261538461538461,
"grad_norm": 0.437076731735488,
"learning_rate": 1.9676753199127033e-05,
"loss": 0.0947,
"step": 212
},
{
"epoch": 0.32769230769230767,
"grad_norm": 0.4670077669146603,
"learning_rate": 1.9673698799700582e-05,
"loss": 0.0933,
"step": 213
},
{
"epoch": 0.3292307692307692,
"grad_norm": 0.4919569174836587,
"learning_rate": 1.9670630276669305e-05,
"loss": 0.1001,
"step": 214
},
{
"epoch": 0.33076923076923076,
"grad_norm": 0.47210284604663544,
"learning_rate": 1.9667547634513248e-05,
"loss": 0.1009,
"step": 215
},
{
"epoch": 0.3323076923076923,
"grad_norm": 0.37651995607019445,
"learning_rate": 1.9664450877733065e-05,
"loss": 0.0898,
"step": 216
},
{
"epoch": 0.33384615384615385,
"grad_norm": 0.534190751602756,
"learning_rate": 1.9661340010850025e-05,
"loss": 0.1104,
"step": 217
},
{
"epoch": 0.3353846153846154,
"grad_norm": 0.4951807767134038,
"learning_rate": 1.9658215038405997e-05,
"loss": 0.1068,
"step": 218
},
{
"epoch": 0.33692307692307694,
"grad_norm": 0.5728927862246131,
"learning_rate": 1.9655075964963443e-05,
"loss": 0.1055,
"step": 219
},
{
"epoch": 0.3384615384615385,
"grad_norm": 0.47984244005432625,
"learning_rate": 1.9651922795105404e-05,
"loss": 0.1021,
"step": 220
},
{
"epoch": 0.34,
"grad_norm": 0.5413827050893782,
"learning_rate": 1.9648755533435517e-05,
"loss": 0.1075,
"step": 221
},
{
"epoch": 0.3415384615384615,
"grad_norm": 0.4794837943323609,
"learning_rate": 1.9645574184577982e-05,
"loss": 0.1042,
"step": 222
},
{
"epoch": 0.34307692307692306,
"grad_norm": 0.47022722324174265,
"learning_rate": 1.9642378753177573e-05,
"loss": 0.0931,
"step": 223
},
{
"epoch": 0.3446153846153846,
"grad_norm": 0.43452426712372355,
"learning_rate": 1.963916924389962e-05,
"loss": 0.0918,
"step": 224
},
{
"epoch": 0.34615384615384615,
"grad_norm": 0.4951463295701747,
"learning_rate": 1.9635945661430006e-05,
"loss": 0.0958,
"step": 225
},
{
"epoch": 0.3476923076923077,
"grad_norm": 0.5231519816188233,
"learning_rate": 1.9632708010475166e-05,
"loss": 0.0985,
"step": 226
},
{
"epoch": 0.34923076923076923,
"grad_norm": 0.4125636678302975,
"learning_rate": 1.9629456295762067e-05,
"loss": 0.0922,
"step": 227
},
{
"epoch": 0.3507692307692308,
"grad_norm": 0.45904306816999607,
"learning_rate": 1.962619052203822e-05,
"loss": 0.0992,
"step": 228
},
{
"epoch": 0.3523076923076923,
"grad_norm": 0.4970133845701183,
"learning_rate": 1.9622910694071654e-05,
"loss": 0.0932,
"step": 229
},
{
"epoch": 0.35384615384615387,
"grad_norm": 0.37664569358821015,
"learning_rate": 1.961961681665092e-05,
"loss": 0.0906,
"step": 230
},
{
"epoch": 0.3553846153846154,
"grad_norm": 0.4876745281374299,
"learning_rate": 1.9616308894585078e-05,
"loss": 0.0994,
"step": 231
},
{
"epoch": 0.3569230769230769,
"grad_norm": 0.4680368478168109,
"learning_rate": 1.9612986932703698e-05,
"loss": 0.1076,
"step": 232
},
{
"epoch": 0.35846153846153844,
"grad_norm": 0.4857848795913984,
"learning_rate": 1.9609650935856847e-05,
"loss": 0.1032,
"step": 233
},
{
"epoch": 0.36,
"grad_norm": 0.7291002862027431,
"learning_rate": 1.9606300908915076e-05,
"loss": 0.1071,
"step": 234
},
{
"epoch": 0.36153846153846153,
"grad_norm": 0.6588220126790021,
"learning_rate": 1.9602936856769432e-05,
"loss": 0.0943,
"step": 235
},
{
"epoch": 0.3630769230769231,
"grad_norm": 0.5848274657760204,
"learning_rate": 1.959955878433143e-05,
"loss": 0.1132,
"step": 236
},
{
"epoch": 0.3646153846153846,
"grad_norm": 0.49893024052580587,
"learning_rate": 1.9596166696533062e-05,
"loss": 0.1029,
"step": 237
},
{
"epoch": 0.36615384615384616,
"grad_norm": 0.5662642891333689,
"learning_rate": 1.959276059832677e-05,
"loss": 0.1079,
"step": 238
},
{
"epoch": 0.3676923076923077,
"grad_norm": 0.5411680818094676,
"learning_rate": 1.9589340494685464e-05,
"loss": 0.0942,
"step": 239
},
{
"epoch": 0.36923076923076925,
"grad_norm": 0.49716800756716995,
"learning_rate": 1.95859063906025e-05,
"loss": 0.1028,
"step": 240
},
{
"epoch": 0.3707692307692308,
"grad_norm": 0.5605318150208776,
"learning_rate": 1.9582458291091664e-05,
"loss": 0.0976,
"step": 241
},
{
"epoch": 0.3723076923076923,
"grad_norm": 0.6030924939164457,
"learning_rate": 1.9578996201187187e-05,
"loss": 0.1022,
"step": 242
},
{
"epoch": 0.37384615384615383,
"grad_norm": 0.4416558734944546,
"learning_rate": 1.957552012594372e-05,
"loss": 0.0903,
"step": 243
},
{
"epoch": 0.37538461538461537,
"grad_norm": 0.5329784217079914,
"learning_rate": 1.957203007043634e-05,
"loss": 0.1164,
"step": 244
},
{
"epoch": 0.3769230769230769,
"grad_norm": 0.4644145144804026,
"learning_rate": 1.956852603976052e-05,
"loss": 0.0995,
"step": 245
},
{
"epoch": 0.37846153846153846,
"grad_norm": 0.43308288576075765,
"learning_rate": 1.9565008039032158e-05,
"loss": 0.0866,
"step": 246
},
{
"epoch": 0.38,
"grad_norm": 0.556662892440361,
"learning_rate": 1.9561476073387527e-05,
"loss": 0.0977,
"step": 247
},
{
"epoch": 0.38153846153846155,
"grad_norm": 0.5153442689897639,
"learning_rate": 1.9557930147983303e-05,
"loss": 0.0979,
"step": 248
},
{
"epoch": 0.3830769230769231,
"grad_norm": 0.4731621488928602,
"learning_rate": 1.9554370267996537e-05,
"loss": 0.1052,
"step": 249
},
{
"epoch": 0.38461538461538464,
"grad_norm": 0.4461201241494072,
"learning_rate": 1.9550796438624656e-05,
"loss": 0.0947,
"step": 250
},
{
"epoch": 0.3861538461538462,
"grad_norm": 0.47389146156943096,
"learning_rate": 1.954720866508546e-05,
"loss": 0.0995,
"step": 251
},
{
"epoch": 0.38769230769230767,
"grad_norm": 0.54566585639894,
"learning_rate": 1.9543606952617088e-05,
"loss": 0.1016,
"step": 252
},
{
"epoch": 0.3892307692307692,
"grad_norm": 0.582991146196512,
"learning_rate": 1.9539991306478046e-05,
"loss": 0.1188,
"step": 253
},
{
"epoch": 0.39076923076923076,
"grad_norm": 0.4578903975000862,
"learning_rate": 1.9536361731947182e-05,
"loss": 0.0928,
"step": 254
},
{
"epoch": 0.3923076923076923,
"grad_norm": 0.590776299920465,
"learning_rate": 1.953271823432367e-05,
"loss": 0.1132,
"step": 255
},
{
"epoch": 0.39384615384615385,
"grad_norm": 0.6253009080591825,
"learning_rate": 1.9529060818927032e-05,
"loss": 0.0979,
"step": 256
},
{
"epoch": 0.3953846153846154,
"grad_norm": 0.531397871124118,
"learning_rate": 1.952538949109708e-05,
"loss": 0.0959,
"step": 257
},
{
"epoch": 0.39692307692307693,
"grad_norm": 0.46705935199941867,
"learning_rate": 1.9521704256193962e-05,
"loss": 0.104,
"step": 258
},
{
"epoch": 0.3984615384615385,
"grad_norm": 0.5705361981014778,
"learning_rate": 1.9518005119598124e-05,
"loss": 0.1083,
"step": 259
},
{
"epoch": 0.4,
"grad_norm": 0.5695832340146562,
"learning_rate": 1.9514292086710307e-05,
"loss": 0.0992,
"step": 260
},
{
"epoch": 0.4015384615384615,
"grad_norm": 0.4464529106040974,
"learning_rate": 1.9510565162951538e-05,
"loss": 0.0987,
"step": 261
},
{
"epoch": 0.40307692307692305,
"grad_norm": 0.43620342439803284,
"learning_rate": 1.9506824353763127e-05,
"loss": 0.103,
"step": 262
},
{
"epoch": 0.4046153846153846,
"grad_norm": 0.48936436282657725,
"learning_rate": 1.9503069664606663e-05,
"loss": 0.1125,
"step": 263
},
{
"epoch": 0.40615384615384614,
"grad_norm": 0.5935399529172605,
"learning_rate": 1.9499301100963987e-05,
"loss": 0.1065,
"step": 264
},
{
"epoch": 0.4076923076923077,
"grad_norm": 0.5489632101646817,
"learning_rate": 1.9495518668337204e-05,
"loss": 0.0874,
"step": 265
},
{
"epoch": 0.40923076923076923,
"grad_norm": 0.4872571558922733,
"learning_rate": 1.949172237224867e-05,
"loss": 0.1067,
"step": 266
},
{
"epoch": 0.4107692307692308,
"grad_norm": 0.495735492589365,
"learning_rate": 1.9487912218240983e-05,
"loss": 0.1025,
"step": 267
},
{
"epoch": 0.4123076923076923,
"grad_norm": 0.520835006964255,
"learning_rate": 1.9484088211876963e-05,
"loss": 0.0926,
"step": 268
},
{
"epoch": 0.41384615384615386,
"grad_norm": 0.4979596342187618,
"learning_rate": 1.9480250358739667e-05,
"loss": 0.1001,
"step": 269
},
{
"epoch": 0.4153846153846154,
"grad_norm": 0.5398277717258808,
"learning_rate": 1.9476398664432356e-05,
"loss": 0.1038,
"step": 270
},
{
"epoch": 0.4169230769230769,
"grad_norm": 0.4620486050085766,
"learning_rate": 1.947253313457851e-05,
"loss": 0.0909,
"step": 271
},
{
"epoch": 0.41846153846153844,
"grad_norm": 0.5088637188704777,
"learning_rate": 1.9468653774821803e-05,
"loss": 0.1038,
"step": 272
},
{
"epoch": 0.42,
"grad_norm": 0.5312467309003265,
"learning_rate": 1.94647605908261e-05,
"loss": 0.1003,
"step": 273
},
{
"epoch": 0.42153846153846153,
"grad_norm": 0.5353118330387302,
"learning_rate": 1.9460853588275454e-05,
"loss": 0.0915,
"step": 274
},
{
"epoch": 0.4230769230769231,
"grad_norm": 0.5883396050759481,
"learning_rate": 1.9456932772874092e-05,
"loss": 0.0964,
"step": 275
},
{
"epoch": 0.4246153846153846,
"grad_norm": 0.3782557300574429,
"learning_rate": 1.9452998150346403e-05,
"loss": 0.0941,
"step": 276
},
{
"epoch": 0.42615384615384616,
"grad_norm": 0.5038727416567336,
"learning_rate": 1.944904972643694e-05,
"loss": 0.0966,
"step": 277
},
{
"epoch": 0.4276923076923077,
"grad_norm": 0.6044080637491093,
"learning_rate": 1.9445087506910403e-05,
"loss": 0.0945,
"step": 278
},
{
"epoch": 0.42923076923076925,
"grad_norm": 0.4283489498603955,
"learning_rate": 1.944111149755164e-05,
"loss": 0.0923,
"step": 279
},
{
"epoch": 0.4307692307692308,
"grad_norm": 0.4755286664345765,
"learning_rate": 1.9437121704165612e-05,
"loss": 0.0981,
"step": 280
},
{
"epoch": 0.4323076923076923,
"grad_norm": 0.4034836375835124,
"learning_rate": 1.9433118132577432e-05,
"loss": 0.0956,
"step": 281
},
{
"epoch": 0.4338461538461538,
"grad_norm": 0.5243724118676012,
"learning_rate": 1.9429100788632313e-05,
"loss": 0.0968,
"step": 282
},
{
"epoch": 0.43538461538461537,
"grad_norm": 0.4909950952163068,
"learning_rate": 1.9425069678195577e-05,
"loss": 0.0836,
"step": 283
},
{
"epoch": 0.4369230769230769,
"grad_norm": 0.6197661444062966,
"learning_rate": 1.9421024807152652e-05,
"loss": 0.1062,
"step": 284
},
{
"epoch": 0.43846153846153846,
"grad_norm": 0.4641188532073514,
"learning_rate": 1.9416966181409047e-05,
"loss": 0.0907,
"step": 285
},
{
"epoch": 0.44,
"grad_norm": 0.581115105716189,
"learning_rate": 1.9412893806890358e-05,
"loss": 0.1136,
"step": 286
},
{
"epoch": 0.44153846153846155,
"grad_norm": 0.5532740369160736,
"learning_rate": 1.9408807689542257e-05,
"loss": 0.095,
"step": 287
},
{
"epoch": 0.4430769230769231,
"grad_norm": 0.610460188491936,
"learning_rate": 1.9404707835330475e-05,
"loss": 0.102,
"step": 288
},
{
"epoch": 0.44461538461538463,
"grad_norm": 0.48489275085594646,
"learning_rate": 1.94005942502408e-05,
"loss": 0.1086,
"step": 289
},
{
"epoch": 0.4461538461538462,
"grad_norm": 0.4350745012942025,
"learning_rate": 1.9396466940279067e-05,
"loss": 0.1027,
"step": 290
},
{
"epoch": 0.44769230769230767,
"grad_norm": 0.4950020730573976,
"learning_rate": 1.9392325911471154e-05,
"loss": 0.1055,
"step": 291
},
{
"epoch": 0.4492307692307692,
"grad_norm": 0.4788608653907359,
"learning_rate": 1.9388171169862967e-05,
"loss": 0.1056,
"step": 292
},
{
"epoch": 0.45076923076923076,
"grad_norm": 0.451391245157289,
"learning_rate": 1.9384002721520423e-05,
"loss": 0.1034,
"step": 293
},
{
"epoch": 0.4523076923076923,
"grad_norm": 0.46706566593370247,
"learning_rate": 1.9379820572529463e-05,
"loss": 0.1032,
"step": 294
},
{
"epoch": 0.45384615384615384,
"grad_norm": 0.5208831601136724,
"learning_rate": 1.937562472899603e-05,
"loss": 0.1048,
"step": 295
},
{
"epoch": 0.4553846153846154,
"grad_norm": 0.4746886544547485,
"learning_rate": 1.9371415197046054e-05,
"loss": 0.0962,
"step": 296
},
{
"epoch": 0.45692307692307693,
"grad_norm": 0.5099090103562837,
"learning_rate": 1.936719198282545e-05,
"loss": 0.0975,
"step": 297
},
{
"epoch": 0.4584615384615385,
"grad_norm": 0.5371114216801979,
"learning_rate": 1.936295509250012e-05,
"loss": 0.0995,
"step": 298
},
{
"epoch": 0.46,
"grad_norm": 0.4583000967152501,
"learning_rate": 1.935870453225592e-05,
"loss": 0.0867,
"step": 299
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.4824466157164947,
"learning_rate": 1.9354440308298676e-05,
"loss": 0.1027,
"step": 300
},
{
"epoch": 0.46307692307692305,
"grad_norm": 0.3731332197286788,
"learning_rate": 1.9350162426854152e-05,
"loss": 0.0937,
"step": 301
},
{
"epoch": 0.4646153846153846,
"grad_norm": 0.5904943219369095,
"learning_rate": 1.9345870894168056e-05,
"loss": 0.1018,
"step": 302
},
{
"epoch": 0.46615384615384614,
"grad_norm": 0.4937106647701358,
"learning_rate": 1.934156571650603e-05,
"loss": 0.099,
"step": 303
},
{
"epoch": 0.4676923076923077,
"grad_norm": 0.3616397861267969,
"learning_rate": 1.9337246900153637e-05,
"loss": 0.0899,
"step": 304
},
{
"epoch": 0.46923076923076923,
"grad_norm": 0.4337547379197528,
"learning_rate": 1.933291445141635e-05,
"loss": 0.1094,
"step": 305
},
{
"epoch": 0.4707692307692308,
"grad_norm": 0.47132758397572894,
"learning_rate": 1.932856837661954e-05,
"loss": 0.0932,
"step": 306
},
{
"epoch": 0.4723076923076923,
"grad_norm": 0.5249116456685259,
"learning_rate": 1.9324208682108493e-05,
"loss": 0.0955,
"step": 307
},
{
"epoch": 0.47384615384615386,
"grad_norm": 0.4099876886395493,
"learning_rate": 1.931983537424835e-05,
"loss": 0.0925,
"step": 308
},
{
"epoch": 0.4753846153846154,
"grad_norm": 0.4571620412260989,
"learning_rate": 1.931544845942415e-05,
"loss": 0.0974,
"step": 309
},
{
"epoch": 0.47692307692307695,
"grad_norm": 0.5347160775026656,
"learning_rate": 1.9311047944040792e-05,
"loss": 0.1021,
"step": 310
},
{
"epoch": 0.47846153846153844,
"grad_norm": 0.45731698473766913,
"learning_rate": 1.9306633834523022e-05,
"loss": 0.0995,
"step": 311
},
{
"epoch": 0.48,
"grad_norm": 0.4463396489272035,
"learning_rate": 1.930220613731545e-05,
"loss": 0.0936,
"step": 312
},
{
"epoch": 0.4815384615384615,
"grad_norm": 0.449418485015138,
"learning_rate": 1.9297764858882516e-05,
"loss": 0.0883,
"step": 313
},
{
"epoch": 0.48307692307692307,
"grad_norm": 0.5562810223949127,
"learning_rate": 1.9293310005708485e-05,
"loss": 0.1089,
"step": 314
},
{
"epoch": 0.4846153846153846,
"grad_norm": 0.40867491037166837,
"learning_rate": 1.9288841584297445e-05,
"loss": 0.087,
"step": 315
},
{
"epoch": 0.48615384615384616,
"grad_norm": 0.6068014941268159,
"learning_rate": 1.9284359601173295e-05,
"loss": 0.1002,
"step": 316
},
{
"epoch": 0.4876923076923077,
"grad_norm": 0.5485840263407294,
"learning_rate": 1.927986406287973e-05,
"loss": 0.1071,
"step": 317
},
{
"epoch": 0.48923076923076925,
"grad_norm": 0.4326599983402119,
"learning_rate": 1.9275354975980245e-05,
"loss": 0.0972,
"step": 318
},
{
"epoch": 0.4907692307692308,
"grad_norm": 0.48038555828580165,
"learning_rate": 1.92708323470581e-05,
"loss": 0.0962,
"step": 319
},
{
"epoch": 0.49230769230769234,
"grad_norm": 0.5096604190301004,
"learning_rate": 1.926629618271634e-05,
"loss": 0.1089,
"step": 320
},
{
"epoch": 0.4938461538461538,
"grad_norm": 0.5074786000084778,
"learning_rate": 1.9261746489577767e-05,
"loss": 0.0953,
"step": 321
},
{
"epoch": 0.49538461538461537,
"grad_norm": 0.3551898568112085,
"learning_rate": 1.9257183274284934e-05,
"loss": 0.0812,
"step": 322
},
{
"epoch": 0.4969230769230769,
"grad_norm": 0.4154869588023741,
"learning_rate": 1.925260654350014e-05,
"loss": 0.0878,
"step": 323
},
{
"epoch": 0.49846153846153846,
"grad_norm": 0.46068163561485903,
"learning_rate": 1.924801630390541e-05,
"loss": 0.0994,
"step": 324
},
{
"epoch": 0.5,
"grad_norm": 0.49072829349035685,
"learning_rate": 1.92434125622025e-05,
"loss": 0.1022,
"step": 325
},
{
"epoch": 0.5015384615384615,
"grad_norm": 0.43272682508224936,
"learning_rate": 1.9238795325112867e-05,
"loss": 0.1084,
"step": 326
},
{
"epoch": 0.5030769230769231,
"grad_norm": 0.4497283552516036,
"learning_rate": 1.9234164599377692e-05,
"loss": 0.0967,
"step": 327
},
{
"epoch": 0.5046153846153846,
"grad_norm": 0.4492106413880127,
"learning_rate": 1.9229520391757828e-05,
"loss": 0.1007,
"step": 328
},
{
"epoch": 0.5061538461538462,
"grad_norm": 0.4890385196278267,
"learning_rate": 1.9224862709033823e-05,
"loss": 0.09,
"step": 329
},
{
"epoch": 0.5076923076923077,
"grad_norm": 0.4508644606981008,
"learning_rate": 1.9220191558005897e-05,
"loss": 0.0969,
"step": 330
},
{
"epoch": 0.5092307692307693,
"grad_norm": 0.5761631636154984,
"learning_rate": 1.9215506945493933e-05,
"loss": 0.1051,
"step": 331
},
{
"epoch": 0.5107692307692308,
"grad_norm": 0.4285970063917521,
"learning_rate": 1.921080887833746e-05,
"loss": 0.0952,
"step": 332
},
{
"epoch": 0.5123076923076924,
"grad_norm": 0.46226415222748873,
"learning_rate": 1.9206097363395668e-05,
"loss": 0.099,
"step": 333
},
{
"epoch": 0.5138461538461538,
"grad_norm": 0.44278457207986294,
"learning_rate": 1.9201372407547367e-05,
"loss": 0.1025,
"step": 334
},
{
"epoch": 0.5153846153846153,
"grad_norm": 0.5672748490771847,
"learning_rate": 1.9196634017690993e-05,
"loss": 0.1013,
"step": 335
},
{
"epoch": 0.5169230769230769,
"grad_norm": 0.4566472958118853,
"learning_rate": 1.9191882200744602e-05,
"loss": 0.1158,
"step": 336
},
{
"epoch": 0.5184615384615384,
"grad_norm": 0.5248322529451815,
"learning_rate": 1.9187116963645845e-05,
"loss": 0.0957,
"step": 337
},
{
"epoch": 0.52,
"grad_norm": 0.44219344780447695,
"learning_rate": 1.918233831335197e-05,
"loss": 0.0975,
"step": 338
},
{
"epoch": 0.5215384615384615,
"grad_norm": 0.378145742243331,
"learning_rate": 1.9177546256839814e-05,
"loss": 0.0855,
"step": 339
},
{
"epoch": 0.5230769230769231,
"grad_norm": 0.4543314495704193,
"learning_rate": 1.9172740801105777e-05,
"loss": 0.101,
"step": 340
},
{
"epoch": 0.5246153846153846,
"grad_norm": 0.5388495866373283,
"learning_rate": 1.9167921953165827e-05,
"loss": 0.1077,
"step": 341
},
{
"epoch": 0.5261538461538462,
"grad_norm": 0.4931391711463696,
"learning_rate": 1.9163089720055484e-05,
"loss": 0.0937,
"step": 342
},
{
"epoch": 0.5276923076923077,
"grad_norm": 0.5226419648010249,
"learning_rate": 1.9158244108829815e-05,
"loss": 0.0978,
"step": 343
},
{
"epoch": 0.5292307692307693,
"grad_norm": 0.514073044628901,
"learning_rate": 1.915338512656341e-05,
"loss": 0.1116,
"step": 344
},
{
"epoch": 0.5307692307692308,
"grad_norm": 0.536463160406073,
"learning_rate": 1.9148512780350384e-05,
"loss": 0.09,
"step": 345
},
{
"epoch": 0.5323076923076923,
"grad_norm": 0.40472290941801015,
"learning_rate": 1.914362707730437e-05,
"loss": 0.096,
"step": 346
},
{
"epoch": 0.5338461538461539,
"grad_norm": 0.4419558495038164,
"learning_rate": 1.9138728024558494e-05,
"loss": 0.1043,
"step": 347
},
{
"epoch": 0.5353846153846153,
"grad_norm": 0.45250393937417616,
"learning_rate": 1.913381562926538e-05,
"loss": 0.1044,
"step": 348
},
{
"epoch": 0.536923076923077,
"grad_norm": 0.4970429188190532,
"learning_rate": 1.9128889898597117e-05,
"loss": 0.1002,
"step": 349
},
{
"epoch": 0.5384615384615384,
"grad_norm": 0.5382027157131241,
"learning_rate": 1.912395083974528e-05,
"loss": 0.0992,
"step": 350
},
{
"epoch": 0.54,
"grad_norm": 0.4173235180068536,
"learning_rate": 1.91189984599209e-05,
"loss": 0.0873,
"step": 351
},
{
"epoch": 0.5415384615384615,
"grad_norm": 0.42760297701096356,
"learning_rate": 1.9114032766354453e-05,
"loss": 0.096,
"step": 352
},
{
"epoch": 0.5430769230769231,
"grad_norm": 0.524358521462744,
"learning_rate": 1.910905376629585e-05,
"loss": 0.0959,
"step": 353
},
{
"epoch": 0.5446153846153846,
"grad_norm": 0.47250449647238696,
"learning_rate": 1.910406146701444e-05,
"loss": 0.0849,
"step": 354
},
{
"epoch": 0.5461538461538461,
"grad_norm": 0.44398596927938827,
"learning_rate": 1.9099055875798974e-05,
"loss": 0.1024,
"step": 355
},
{
"epoch": 0.5476923076923077,
"grad_norm": 0.46499606497240575,
"learning_rate": 1.9094036999957623e-05,
"loss": 0.0925,
"step": 356
},
{
"epoch": 0.5492307692307692,
"grad_norm": 0.3886634015877156,
"learning_rate": 1.9089004846817947e-05,
"loss": 0.0807,
"step": 357
},
{
"epoch": 0.5507692307692308,
"grad_norm": 0.4118047391602366,
"learning_rate": 1.908395942372689e-05,
"loss": 0.0988,
"step": 358
},
{
"epoch": 0.5523076923076923,
"grad_norm": 0.434600405160762,
"learning_rate": 1.9078900738050776e-05,
"loss": 0.1069,
"step": 359
},
{
"epoch": 0.5538461538461539,
"grad_norm": 0.4148807319841889,
"learning_rate": 1.9073828797175284e-05,
"loss": 0.1028,
"step": 360
},
{
"epoch": 0.5553846153846154,
"grad_norm": 0.42028520435641853,
"learning_rate": 1.9068743608505454e-05,
"loss": 0.1056,
"step": 361
},
{
"epoch": 0.556923076923077,
"grad_norm": 0.43510420212696854,
"learning_rate": 1.9063645179465663e-05,
"loss": 0.1009,
"step": 362
},
{
"epoch": 0.5584615384615385,
"grad_norm": 0.4936074681803371,
"learning_rate": 1.905853351749962e-05,
"loss": 0.1,
"step": 363
},
{
"epoch": 0.56,
"grad_norm": 0.3632144105495107,
"learning_rate": 1.9053408630070352e-05,
"loss": 0.0916,
"step": 364
},
{
"epoch": 0.5615384615384615,
"grad_norm": 0.3536771592078361,
"learning_rate": 1.9048270524660197e-05,
"loss": 0.0913,
"step": 365
},
{
"epoch": 0.563076923076923,
"grad_norm": 0.483138657389253,
"learning_rate": 1.904311920877079e-05,
"loss": 0.0992,
"step": 366
},
{
"epoch": 0.5646153846153846,
"grad_norm": 0.4057661988297129,
"learning_rate": 1.903795468992306e-05,
"loss": 0.1027,
"step": 367
},
{
"epoch": 0.5661538461538461,
"grad_norm": 0.40666719198061757,
"learning_rate": 1.9032776975657207e-05,
"loss": 0.0981,
"step": 368
},
{
"epoch": 0.5676923076923077,
"grad_norm": 0.44070081776133335,
"learning_rate": 1.902758607353269e-05,
"loss": 0.1104,
"step": 369
},
{
"epoch": 0.5692307692307692,
"grad_norm": 0.5047821437081584,
"learning_rate": 1.9022381991128235e-05,
"loss": 0.0963,
"step": 370
},
{
"epoch": 0.5707692307692308,
"grad_norm": 0.42260385596144445,
"learning_rate": 1.9017164736041795e-05,
"loss": 0.1029,
"step": 371
},
{
"epoch": 0.5723076923076923,
"grad_norm": 0.4244718881899699,
"learning_rate": 1.9011934315890576e-05,
"loss": 0.0966,
"step": 372
},
{
"epoch": 0.5738461538461539,
"grad_norm": 0.4953235016748773,
"learning_rate": 1.9006690738310988e-05,
"loss": 0.1037,
"step": 373
},
{
"epoch": 0.5753846153846154,
"grad_norm": 0.3770254505903871,
"learning_rate": 1.900143401095866e-05,
"loss": 0.086,
"step": 374
},
{
"epoch": 0.5769230769230769,
"grad_norm": 0.41771963874965723,
"learning_rate": 1.8996164141508412e-05,
"loss": 0.0858,
"step": 375
},
{
"epoch": 0.5784615384615385,
"grad_norm": 0.4061141404696647,
"learning_rate": 1.899088113765426e-05,
"loss": 0.1021,
"step": 376
},
{
"epoch": 0.58,
"grad_norm": 0.45946326931814974,
"learning_rate": 1.898558500710939e-05,
"loss": 0.0979,
"step": 377
},
{
"epoch": 0.5815384615384616,
"grad_norm": 0.4143960427519077,
"learning_rate": 1.8980275757606157e-05,
"loss": 0.0965,
"step": 378
},
{
"epoch": 0.583076923076923,
"grad_norm": 0.40587432220084,
"learning_rate": 1.8974953396896066e-05,
"loss": 0.0953,
"step": 379
},
{
"epoch": 0.5846153846153846,
"grad_norm": 0.4036824003812063,
"learning_rate": 1.8969617932749766e-05,
"loss": 0.0947,
"step": 380
},
{
"epoch": 0.5861538461538461,
"grad_norm": 0.431070678494207,
"learning_rate": 1.896426937295704e-05,
"loss": 0.103,
"step": 381
},
{
"epoch": 0.5876923076923077,
"grad_norm": 0.3955536485889764,
"learning_rate": 1.8958907725326783e-05,
"loss": 0.0928,
"step": 382
},
{
"epoch": 0.5892307692307692,
"grad_norm": 0.4357741837489556,
"learning_rate": 1.8953532997687008e-05,
"loss": 0.0997,
"step": 383
},
{
"epoch": 0.5907692307692308,
"grad_norm": 0.39754441486925746,
"learning_rate": 1.8948145197884815e-05,
"loss": 0.0859,
"step": 384
},
{
"epoch": 0.5923076923076923,
"grad_norm": 0.43549014378990397,
"learning_rate": 1.89427443337864e-05,
"loss": 0.0934,
"step": 385
},
{
"epoch": 0.5938461538461538,
"grad_norm": 0.42300645280983395,
"learning_rate": 1.893733041327702e-05,
"loss": 0.0951,
"step": 386
},
{
"epoch": 0.5953846153846154,
"grad_norm": 0.41282699722561766,
"learning_rate": 1.8931903444261007e-05,
"loss": 0.0941,
"step": 387
},
{
"epoch": 0.5969230769230769,
"grad_norm": 0.4480032982295742,
"learning_rate": 1.8926463434661738e-05,
"loss": 0.0905,
"step": 388
},
{
"epoch": 0.5984615384615385,
"grad_norm": 0.44297004392731376,
"learning_rate": 1.8921010392421628e-05,
"loss": 0.1016,
"step": 389
},
{
"epoch": 0.6,
"grad_norm": 0.5433110545398998,
"learning_rate": 1.8915544325502123e-05,
"loss": 0.0935,
"step": 390
},
{
"epoch": 0.6015384615384616,
"grad_norm": 0.40571632988210427,
"learning_rate": 1.891006524188368e-05,
"loss": 0.0937,
"step": 391
},
{
"epoch": 0.6030769230769231,
"grad_norm": 0.43732620714659554,
"learning_rate": 1.8904573149565766e-05,
"loss": 0.0947,
"step": 392
},
{
"epoch": 0.6046153846153847,
"grad_norm": 0.4600862732120384,
"learning_rate": 1.889906805656684e-05,
"loss": 0.095,
"step": 393
},
{
"epoch": 0.6061538461538462,
"grad_norm": 0.44446566484906674,
"learning_rate": 1.8893549970924335e-05,
"loss": 0.1005,
"step": 394
},
{
"epoch": 0.6076923076923076,
"grad_norm": 0.4372133849372649,
"learning_rate": 1.888801890069467e-05,
"loss": 0.0926,
"step": 395
},
{
"epoch": 0.6092307692307692,
"grad_norm": 0.4496197566055402,
"learning_rate": 1.8882474853953193e-05,
"loss": 0.0889,
"step": 396
},
{
"epoch": 0.6107692307692307,
"grad_norm": 0.43381954647371906,
"learning_rate": 1.8876917838794226e-05,
"loss": 0.0892,
"step": 397
},
{
"epoch": 0.6123076923076923,
"grad_norm": 0.5656036739453132,
"learning_rate": 1.8871347863331015e-05,
"loss": 0.087,
"step": 398
},
{
"epoch": 0.6138461538461538,
"grad_norm": 0.42869081240488516,
"learning_rate": 1.886576493569572e-05,
"loss": 0.0999,
"step": 399
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.9645884517575197,
"learning_rate": 1.8860169064039422e-05,
"loss": 0.0877,
"step": 400
},
{
"epoch": 0.6169230769230769,
"grad_norm": 0.3909376825178073,
"learning_rate": 1.8854560256532098e-05,
"loss": 0.09,
"step": 401
},
{
"epoch": 0.6184615384615385,
"grad_norm": 0.41607837043434165,
"learning_rate": 1.884893852136261e-05,
"loss": 0.0962,
"step": 402
},
{
"epoch": 0.62,
"grad_norm": 0.4118435975798478,
"learning_rate": 1.884330386673869e-05,
"loss": 0.0914,
"step": 403
},
{
"epoch": 0.6215384615384615,
"grad_norm": 0.41468438143981523,
"learning_rate": 1.8837656300886937e-05,
"loss": 0.0932,
"step": 404
},
{
"epoch": 0.6230769230769231,
"grad_norm": 0.42739734031323684,
"learning_rate": 1.8831995832052802e-05,
"loss": 0.1006,
"step": 405
},
{
"epoch": 0.6246153846153846,
"grad_norm": 0.40676045777564296,
"learning_rate": 1.8826322468500567e-05,
"loss": 0.0911,
"step": 406
},
{
"epoch": 0.6261538461538462,
"grad_norm": 0.4400751065379498,
"learning_rate": 1.8820636218513354e-05,
"loss": 0.099,
"step": 407
},
{
"epoch": 0.6276923076923077,
"grad_norm": 0.42437410068559034,
"learning_rate": 1.8814937090393082e-05,
"loss": 0.0968,
"step": 408
},
{
"epoch": 0.6292307692307693,
"grad_norm": 0.3919701002838343,
"learning_rate": 1.8809225092460488e-05,
"loss": 0.0936,
"step": 409
},
{
"epoch": 0.6307692307692307,
"grad_norm": 0.5028949538298951,
"learning_rate": 1.880350023305509e-05,
"loss": 0.0948,
"step": 410
},
{
"epoch": 0.6323076923076923,
"grad_norm": 0.4075578032238594,
"learning_rate": 1.8797762520535178e-05,
"loss": 0.0981,
"step": 411
},
{
"epoch": 0.6338461538461538,
"grad_norm": 0.34153544828671295,
"learning_rate": 1.8792011963277827e-05,
"loss": 0.0887,
"step": 412
},
{
"epoch": 0.6353846153846154,
"grad_norm": 0.49338561568395417,
"learning_rate": 1.8786248569678847e-05,
"loss": 0.1008,
"step": 413
},
{
"epoch": 0.6369230769230769,
"grad_norm": 0.3911057155983341,
"learning_rate": 1.8780472348152792e-05,
"loss": 0.0957,
"step": 414
},
{
"epoch": 0.6384615384615384,
"grad_norm": 0.4863828952857313,
"learning_rate": 1.8774683307132956e-05,
"loss": 0.0962,
"step": 415
},
{
"epoch": 0.64,
"grad_norm": 0.4959721632631627,
"learning_rate": 1.876888145507133e-05,
"loss": 0.0974,
"step": 416
},
{
"epoch": 0.6415384615384615,
"grad_norm": 0.44751359386684436,
"learning_rate": 1.8763066800438638e-05,
"loss": 0.1072,
"step": 417
},
{
"epoch": 0.6430769230769231,
"grad_norm": 0.4168309844234487,
"learning_rate": 1.8757239351724262e-05,
"loss": 0.0986,
"step": 418
},
{
"epoch": 0.6446153846153846,
"grad_norm": 0.48992479490511887,
"learning_rate": 1.8751399117436292e-05,
"loss": 0.1142,
"step": 419
},
{
"epoch": 0.6461538461538462,
"grad_norm": 0.5250869322124576,
"learning_rate": 1.8745546106101466e-05,
"loss": 0.0966,
"step": 420
},
{
"epoch": 0.6476923076923077,
"grad_norm": 0.41428994308159467,
"learning_rate": 1.873968032626518e-05,
"loss": 0.0964,
"step": 421
},
{
"epoch": 0.6492307692307693,
"grad_norm": 0.40863286499944806,
"learning_rate": 1.8733801786491487e-05,
"loss": 0.0982,
"step": 422
},
{
"epoch": 0.6507692307692308,
"grad_norm": 0.40530569813116685,
"learning_rate": 1.8727910495363043e-05,
"loss": 0.0893,
"step": 423
},
{
"epoch": 0.6523076923076923,
"grad_norm": 0.3994309755298727,
"learning_rate": 1.872200646148115e-05,
"loss": 0.0917,
"step": 424
},
{
"epoch": 0.6538461538461539,
"grad_norm": 0.41310852897533573,
"learning_rate": 1.8716089693465696e-05,
"loss": 0.0869,
"step": 425
},
{
"epoch": 0.6553846153846153,
"grad_norm": 0.4477653763330846,
"learning_rate": 1.8710160199955158e-05,
"loss": 0.1001,
"step": 426
},
{
"epoch": 0.6569230769230769,
"grad_norm": 0.4749180294295519,
"learning_rate": 1.8704217989606606e-05,
"loss": 0.0989,
"step": 427
},
{
"epoch": 0.6584615384615384,
"grad_norm": 0.4374196609501067,
"learning_rate": 1.869826307109567e-05,
"loss": 0.0955,
"step": 428
},
{
"epoch": 0.66,
"grad_norm": 0.39438024250010495,
"learning_rate": 1.869229545311653e-05,
"loss": 0.0939,
"step": 429
},
{
"epoch": 0.6615384615384615,
"grad_norm": 0.4203742463904514,
"learning_rate": 1.8686315144381914e-05,
"loss": 0.0964,
"step": 430
},
{
"epoch": 0.6630769230769231,
"grad_norm": 0.4761695840193704,
"learning_rate": 1.8680322153623077e-05,
"loss": 0.1037,
"step": 431
},
{
"epoch": 0.6646153846153846,
"grad_norm": 0.3677131847676501,
"learning_rate": 1.8674316489589782e-05,
"loss": 0.09,
"step": 432
},
{
"epoch": 0.6661538461538462,
"grad_norm": 0.4356656851311741,
"learning_rate": 1.8668298161050308e-05,
"loss": 0.1005,
"step": 433
},
{
"epoch": 0.6676923076923077,
"grad_norm": 0.5173595615830835,
"learning_rate": 1.8662267176791418e-05,
"loss": 0.104,
"step": 434
},
{
"epoch": 0.6692307692307692,
"grad_norm": 0.565640191679134,
"learning_rate": 1.8656223545618345e-05,
"loss": 0.1059,
"step": 435
},
{
"epoch": 0.6707692307692308,
"grad_norm": 0.5304970770470983,
"learning_rate": 1.8650167276354802e-05,
"loss": 0.0969,
"step": 436
},
{
"epoch": 0.6723076923076923,
"grad_norm": 0.3582385963482857,
"learning_rate": 1.8644098377842934e-05,
"loss": 0.0901,
"step": 437
},
{
"epoch": 0.6738461538461539,
"grad_norm": 0.4778942975736773,
"learning_rate": 1.863801685894335e-05,
"loss": 0.0997,
"step": 438
},
{
"epoch": 0.6753846153846154,
"grad_norm": 0.6029016365712476,
"learning_rate": 1.8631922728535054e-05,
"loss": 0.0938,
"step": 439
},
{
"epoch": 0.676923076923077,
"grad_norm": 0.6376765327998692,
"learning_rate": 1.8625815995515493e-05,
"loss": 0.1069,
"step": 440
},
{
"epoch": 0.6784615384615384,
"grad_norm": 0.4435527425755955,
"learning_rate": 1.8619696668800494e-05,
"loss": 0.0934,
"step": 441
},
{
"epoch": 0.68,
"grad_norm": 0.5412282043268205,
"learning_rate": 1.8613564757324276e-05,
"loss": 0.099,
"step": 442
},
{
"epoch": 0.6815384615384615,
"grad_norm": 0.5181149152675278,
"learning_rate": 1.860742027003944e-05,
"loss": 0.1045,
"step": 443
},
{
"epoch": 0.683076923076923,
"grad_norm": 0.40925583148990097,
"learning_rate": 1.8601263215916928e-05,
"loss": 0.0893,
"step": 444
},
{
"epoch": 0.6846153846153846,
"grad_norm": 0.42471412248636775,
"learning_rate": 1.8595093603946053e-05,
"loss": 0.0986,
"step": 445
},
{
"epoch": 0.6861538461538461,
"grad_norm": 0.35148548657598855,
"learning_rate": 1.858891144313445e-05,
"loss": 0.0911,
"step": 446
},
{
"epoch": 0.6876923076923077,
"grad_norm": 0.40331605360650485,
"learning_rate": 1.8582716742508066e-05,
"loss": 0.0967,
"step": 447
},
{
"epoch": 0.6892307692307692,
"grad_norm": 0.44814576432850384,
"learning_rate": 1.8576509511111182e-05,
"loss": 0.0917,
"step": 448
},
{
"epoch": 0.6907692307692308,
"grad_norm": 0.4489452721920411,
"learning_rate": 1.8570289758006346e-05,
"loss": 0.0989,
"step": 449
},
{
"epoch": 0.6923076923076923,
"grad_norm": 0.3822084964419905,
"learning_rate": 1.8564057492274407e-05,
"loss": 0.0923,
"step": 450
},
{
"epoch": 0.6938461538461539,
"grad_norm": 0.47514026675128,
"learning_rate": 1.8557812723014476e-05,
"loss": 0.1076,
"step": 451
},
{
"epoch": 0.6953846153846154,
"grad_norm": 0.5418245481869945,
"learning_rate": 1.8551555459343918e-05,
"loss": 0.1116,
"step": 452
},
{
"epoch": 0.696923076923077,
"grad_norm": 0.5023521612267582,
"learning_rate": 1.8545285710398343e-05,
"loss": 0.1078,
"step": 453
},
{
"epoch": 0.6984615384615385,
"grad_norm": 0.5662436210170848,
"learning_rate": 1.8539003485331584e-05,
"loss": 0.0951,
"step": 454
},
{
"epoch": 0.7,
"grad_norm": 0.42546907920042626,
"learning_rate": 1.853270879331569e-05,
"loss": 0.0942,
"step": 455
},
{
"epoch": 0.7015384615384616,
"grad_norm": 0.41449390666943065,
"learning_rate": 1.8526401643540924e-05,
"loss": 0.0994,
"step": 456
},
{
"epoch": 0.703076923076923,
"grad_norm": 0.48839987100466653,
"learning_rate": 1.852008204521572e-05,
"loss": 0.1008,
"step": 457
},
{
"epoch": 0.7046153846153846,
"grad_norm": 0.43738519055543357,
"learning_rate": 1.8513750007566696e-05,
"loss": 0.0954,
"step": 458
},
{
"epoch": 0.7061538461538461,
"grad_norm": 0.41775219787174245,
"learning_rate": 1.850740553983863e-05,
"loss": 0.0875,
"step": 459
},
{
"epoch": 0.7076923076923077,
"grad_norm": 0.7358518540594707,
"learning_rate": 1.8501048651294447e-05,
"loss": 0.0933,
"step": 460
},
{
"epoch": 0.7092307692307692,
"grad_norm": 0.4241607505141305,
"learning_rate": 1.8494679351215212e-05,
"loss": 0.1001,
"step": 461
},
{
"epoch": 0.7107692307692308,
"grad_norm": 0.4683808796780716,
"learning_rate": 1.84882976489001e-05,
"loss": 0.0853,
"step": 462
},
{
"epoch": 0.7123076923076923,
"grad_norm": 0.44346576858606745,
"learning_rate": 1.8481903553666405e-05,
"loss": 0.0995,
"step": 463
},
{
"epoch": 0.7138461538461538,
"grad_norm": 0.45180799988411885,
"learning_rate": 1.84754970748495e-05,
"loss": 0.0848,
"step": 464
},
{
"epoch": 0.7153846153846154,
"grad_norm": 0.4173368978655163,
"learning_rate": 1.846907822180286e-05,
"loss": 0.0954,
"step": 465
},
{
"epoch": 0.7169230769230769,
"grad_norm": 0.42068772817462785,
"learning_rate": 1.8462647003898005e-05,
"loss": 0.0951,
"step": 466
},
{
"epoch": 0.7184615384615385,
"grad_norm": 0.49661800554688595,
"learning_rate": 1.845620343052452e-05,
"loss": 0.0949,
"step": 467
},
{
"epoch": 0.72,
"grad_norm": 0.48646267743412663,
"learning_rate": 1.844974751109002e-05,
"loss": 0.0964,
"step": 468
},
{
"epoch": 0.7215384615384616,
"grad_norm": 0.4106987772360316,
"learning_rate": 1.8443279255020153e-05,
"loss": 0.0974,
"step": 469
},
{
"epoch": 0.7230769230769231,
"grad_norm": 0.3921781146014997,
"learning_rate": 1.843679867175858e-05,
"loss": 0.0848,
"step": 470
},
{
"epoch": 0.7246153846153847,
"grad_norm": 0.44000195303506684,
"learning_rate": 1.8430305770766947e-05,
"loss": 0.1023,
"step": 471
},
{
"epoch": 0.7261538461538461,
"grad_norm": 0.5186352660566815,
"learning_rate": 1.84238005615249e-05,
"loss": 0.0943,
"step": 472
},
{
"epoch": 0.7276923076923076,
"grad_norm": 0.4684858223587338,
"learning_rate": 1.8417283053530047e-05,
"loss": 0.1027,
"step": 473
},
{
"epoch": 0.7292307692307692,
"grad_norm": 0.39874288545350156,
"learning_rate": 1.8410753256297948e-05,
"loss": 0.1023,
"step": 474
},
{
"epoch": 0.7307692307692307,
"grad_norm": 0.3687327920324425,
"learning_rate": 1.8404211179362116e-05,
"loss": 0.102,
"step": 475
},
{
"epoch": 0.7323076923076923,
"grad_norm": 0.37907564191738674,
"learning_rate": 1.8397656832273982e-05,
"loss": 0.0902,
"step": 476
},
{
"epoch": 0.7338461538461538,
"grad_norm": 0.5844119747750028,
"learning_rate": 1.8391090224602895e-05,
"loss": 0.1024,
"step": 477
},
{
"epoch": 0.7353846153846154,
"grad_norm": 0.4973700871932812,
"learning_rate": 1.8384511365936112e-05,
"loss": 0.0986,
"step": 478
},
{
"epoch": 0.7369230769230769,
"grad_norm": 0.4220278832402647,
"learning_rate": 1.837792026587876e-05,
"loss": 0.0947,
"step": 479
},
{
"epoch": 0.7384615384615385,
"grad_norm": 0.48741372675693995,
"learning_rate": 1.837131693405386e-05,
"loss": 0.1039,
"step": 480
},
{
"epoch": 0.74,
"grad_norm": 0.5095101140536407,
"learning_rate": 1.8364701380102267e-05,
"loss": 0.1014,
"step": 481
},
{
"epoch": 0.7415384615384616,
"grad_norm": 0.3686730605957879,
"learning_rate": 1.8358073613682705e-05,
"loss": 0.0879,
"step": 482
},
{
"epoch": 0.7430769230769231,
"grad_norm": 0.4018358216960023,
"learning_rate": 1.8351433644471708e-05,
"loss": 0.0951,
"step": 483
},
{
"epoch": 0.7446153846153846,
"grad_norm": 0.42272315942515715,
"learning_rate": 1.8344781482163635e-05,
"loss": 0.1006,
"step": 484
},
{
"epoch": 0.7461538461538462,
"grad_norm": 0.36114348383132194,
"learning_rate": 1.8338117136470648e-05,
"loss": 0.0876,
"step": 485
},
{
"epoch": 0.7476923076923077,
"grad_norm": 0.44109486857223745,
"learning_rate": 1.8331440617122694e-05,
"loss": 0.1014,
"step": 486
},
{
"epoch": 0.7492307692307693,
"grad_norm": 0.4735198416376007,
"learning_rate": 1.8324751933867496e-05,
"loss": 0.101,
"step": 487
},
{
"epoch": 0.7507692307692307,
"grad_norm": 0.5212542112337067,
"learning_rate": 1.831805109647053e-05,
"loss": 0.0965,
"step": 488
},
{
"epoch": 0.7523076923076923,
"grad_norm": 0.3867194456131576,
"learning_rate": 1.831133811471503e-05,
"loss": 0.0923,
"step": 489
},
{
"epoch": 0.7538461538461538,
"grad_norm": 0.4232158553279915,
"learning_rate": 1.8304612998401947e-05,
"loss": 0.086,
"step": 490
},
{
"epoch": 0.7553846153846154,
"grad_norm": 0.4882561711422476,
"learning_rate": 1.829787575734995e-05,
"loss": 0.084,
"step": 491
},
{
"epoch": 0.7569230769230769,
"grad_norm": 0.49255164116648453,
"learning_rate": 1.8291126401395425e-05,
"loss": 0.1011,
"step": 492
},
{
"epoch": 0.7584615384615384,
"grad_norm": 0.39462697940845287,
"learning_rate": 1.8284364940392426e-05,
"loss": 0.0981,
"step": 493
},
{
"epoch": 0.76,
"grad_norm": 0.42965433064236386,
"learning_rate": 1.8277591384212693e-05,
"loss": 0.1049,
"step": 494
},
{
"epoch": 0.7615384615384615,
"grad_norm": 0.42153666897549313,
"learning_rate": 1.827080574274562e-05,
"loss": 0.1015,
"step": 495
},
{
"epoch": 0.7630769230769231,
"grad_norm": 0.5244873385591531,
"learning_rate": 1.8264008025898248e-05,
"loss": 0.1021,
"step": 496
},
{
"epoch": 0.7646153846153846,
"grad_norm": 0.48573346836145637,
"learning_rate": 1.825719824359524e-05,
"loss": 0.1012,
"step": 497
},
{
"epoch": 0.7661538461538462,
"grad_norm": 0.4201750596778976,
"learning_rate": 1.8250376405778897e-05,
"loss": 0.0914,
"step": 498
},
{
"epoch": 0.7676923076923077,
"grad_norm": 0.3689138791311251,
"learning_rate": 1.824354252240909e-05,
"loss": 0.0971,
"step": 499
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.446354377870851,
"learning_rate": 1.8236696603463297e-05,
"loss": 0.1104,
"step": 500
},
{
"epoch": 0.7707692307692308,
"grad_norm": 0.5302794302913245,
"learning_rate": 1.8229838658936566e-05,
"loss": 0.0913,
"step": 501
},
{
"epoch": 0.7723076923076924,
"grad_norm": 0.6151683521942346,
"learning_rate": 1.8222968698841495e-05,
"loss": 0.0842,
"step": 502
},
{
"epoch": 0.7738461538461539,
"grad_norm": 0.4018248722446391,
"learning_rate": 1.821608673320823e-05,
"loss": 0.0896,
"step": 503
},
{
"epoch": 0.7753846153846153,
"grad_norm": 0.36582381181570395,
"learning_rate": 1.8209192772084446e-05,
"loss": 0.0899,
"step": 504
},
{
"epoch": 0.7769230769230769,
"grad_norm": 0.42007534297254007,
"learning_rate": 1.820228682553533e-05,
"loss": 0.1012,
"step": 505
},
{
"epoch": 0.7784615384615384,
"grad_norm": 0.3706728601304087,
"learning_rate": 1.8195368903643565e-05,
"loss": 0.0937,
"step": 506
},
{
"epoch": 0.78,
"grad_norm": 0.3386897127268892,
"learning_rate": 1.818843901650932e-05,
"loss": 0.0864,
"step": 507
},
{
"epoch": 0.7815384615384615,
"grad_norm": 0.4604244232600737,
"learning_rate": 1.8181497174250236e-05,
"loss": 0.0886,
"step": 508
},
{
"epoch": 0.7830769230769231,
"grad_norm": 0.46651975178638583,
"learning_rate": 1.8174543387001403e-05,
"loss": 0.0942,
"step": 509
},
{
"epoch": 0.7846153846153846,
"grad_norm": 0.37252137712706535,
"learning_rate": 1.8167577664915354e-05,
"loss": 0.098,
"step": 510
},
{
"epoch": 0.7861538461538462,
"grad_norm": 0.4549144644337013,
"learning_rate": 1.816060001816205e-05,
"loss": 0.1043,
"step": 511
},
{
"epoch": 0.7876923076923077,
"grad_norm": 0.49452287761067387,
"learning_rate": 1.8153610456928853e-05,
"loss": 0.0976,
"step": 512
},
{
"epoch": 0.7892307692307692,
"grad_norm": 0.5123177646707533,
"learning_rate": 1.8146608991420533e-05,
"loss": 0.0974,
"step": 513
},
{
"epoch": 0.7907692307692308,
"grad_norm": 0.39618956334101985,
"learning_rate": 1.8139595631859228e-05,
"loss": 0.0891,
"step": 514
},
{
"epoch": 0.7923076923076923,
"grad_norm": 0.3864877721629785,
"learning_rate": 1.8132570388484442e-05,
"loss": 0.0849,
"step": 515
},
{
"epoch": 0.7938461538461539,
"grad_norm": 0.48621924311976095,
"learning_rate": 1.8125533271553045e-05,
"loss": 0.1014,
"step": 516
},
{
"epoch": 0.7953846153846154,
"grad_norm": 0.46058902465123464,
"learning_rate": 1.811848429133922e-05,
"loss": 0.1114,
"step": 517
},
{
"epoch": 0.796923076923077,
"grad_norm": 0.40700166379805536,
"learning_rate": 1.811142345813449e-05,
"loss": 0.0903,
"step": 518
},
{
"epoch": 0.7984615384615384,
"grad_norm": 0.38144448483836707,
"learning_rate": 1.810435078224767e-05,
"loss": 0.0999,
"step": 519
},
{
"epoch": 0.8,
"grad_norm": 0.4087914237321906,
"learning_rate": 1.809726627400487e-05,
"loss": 0.0905,
"step": 520
},
{
"epoch": 0.8015384615384615,
"grad_norm": 0.454530116045963,
"learning_rate": 1.8090169943749477e-05,
"loss": 0.0981,
"step": 521
},
{
"epoch": 0.803076923076923,
"grad_norm": 0.387856092368395,
"learning_rate": 1.8083061801842133e-05,
"loss": 0.0914,
"step": 522
},
{
"epoch": 0.8046153846153846,
"grad_norm": 0.4175456777206446,
"learning_rate": 1.8075941858660737e-05,
"loss": 0.0996,
"step": 523
},
{
"epoch": 0.8061538461538461,
"grad_norm": 0.7864656011521015,
"learning_rate": 1.8068810124600403e-05,
"loss": 0.0899,
"step": 524
},
{
"epoch": 0.8076923076923077,
"grad_norm": 0.48921911883769525,
"learning_rate": 1.8061666610073465e-05,
"loss": 0.0979,
"step": 525
},
{
"epoch": 0.8092307692307692,
"grad_norm": 0.48154421315499524,
"learning_rate": 1.805451132550946e-05,
"loss": 0.1026,
"step": 526
},
{
"epoch": 0.8107692307692308,
"grad_norm": 0.45822965917426073,
"learning_rate": 1.8047344281355112e-05,
"loss": 0.1056,
"step": 527
},
{
"epoch": 0.8123076923076923,
"grad_norm": 0.4699180830359611,
"learning_rate": 1.8040165488074294e-05,
"loss": 0.0978,
"step": 528
},
{
"epoch": 0.8138461538461539,
"grad_norm": 0.4386694887627527,
"learning_rate": 1.8032974956148064e-05,
"loss": 0.0852,
"step": 529
},
{
"epoch": 0.8153846153846154,
"grad_norm": 0.3982089506558907,
"learning_rate": 1.8025772696074593e-05,
"loss": 0.0847,
"step": 530
},
{
"epoch": 0.816923076923077,
"grad_norm": 0.4451526496740161,
"learning_rate": 1.8018558718369187e-05,
"loss": 0.0949,
"step": 531
},
{
"epoch": 0.8184615384615385,
"grad_norm": 0.42797389669933644,
"learning_rate": 1.8011333033564255e-05,
"loss": 0.096,
"step": 532
},
{
"epoch": 0.82,
"grad_norm": 0.5816907168348651,
"learning_rate": 1.8004095652209304e-05,
"loss": 0.0979,
"step": 533
},
{
"epoch": 0.8215384615384616,
"grad_norm": 0.4209426200782058,
"learning_rate": 1.799684658487091e-05,
"loss": 0.0912,
"step": 534
},
{
"epoch": 0.823076923076923,
"grad_norm": 0.4459278908996159,
"learning_rate": 1.7989585842132713e-05,
"loss": 0.0932,
"step": 535
},
{
"epoch": 0.8246153846153846,
"grad_norm": 0.3996692934376165,
"learning_rate": 1.7982313434595405e-05,
"loss": 0.0846,
"step": 536
},
{
"epoch": 0.8261538461538461,
"grad_norm": 0.46838655788136374,
"learning_rate": 1.7975029372876706e-05,
"loss": 0.0931,
"step": 537
},
{
"epoch": 0.8276923076923077,
"grad_norm": 0.44272759560569114,
"learning_rate": 1.7967733667611346e-05,
"loss": 0.0991,
"step": 538
},
{
"epoch": 0.8292307692307692,
"grad_norm": 0.6005247882009117,
"learning_rate": 1.7960426329451062e-05,
"loss": 0.0947,
"step": 539
},
{
"epoch": 0.8307692307692308,
"grad_norm": 0.43973649591876807,
"learning_rate": 1.7953107369064563e-05,
"loss": 0.0946,
"step": 540
},
{
"epoch": 0.8323076923076923,
"grad_norm": 0.36356195409947895,
"learning_rate": 1.7945776797137544e-05,
"loss": 0.0829,
"step": 541
},
{
"epoch": 0.8338461538461538,
"grad_norm": 0.4199497642090339,
"learning_rate": 1.7938434624372638e-05,
"loss": 0.0925,
"step": 542
},
{
"epoch": 0.8353846153846154,
"grad_norm": 0.5142377507204998,
"learning_rate": 1.7931080861489425e-05,
"loss": 0.0973,
"step": 543
},
{
"epoch": 0.8369230769230769,
"grad_norm": 0.5226363888586346,
"learning_rate": 1.7923715519224397e-05,
"loss": 0.1065,
"step": 544
},
{
"epoch": 0.8384615384615385,
"grad_norm": 0.4729813679434869,
"learning_rate": 1.791633860833096e-05,
"loss": 0.0895,
"step": 545
},
{
"epoch": 0.84,
"grad_norm": 0.3606695909284096,
"learning_rate": 1.7908950139579406e-05,
"loss": 0.0832,
"step": 546
},
{
"epoch": 0.8415384615384616,
"grad_norm": 0.570980416724036,
"learning_rate": 1.7901550123756906e-05,
"loss": 0.1002,
"step": 547
},
{
"epoch": 0.8430769230769231,
"grad_norm": 0.5594313800490134,
"learning_rate": 1.7894138571667482e-05,
"loss": 0.1008,
"step": 548
},
{
"epoch": 0.8446153846153847,
"grad_norm": 0.4315817673384466,
"learning_rate": 1.7886715494132008e-05,
"loss": 0.0888,
"step": 549
},
{
"epoch": 0.8461538461538461,
"grad_norm": 0.3916992320727455,
"learning_rate": 1.787928090198818e-05,
"loss": 0.0858,
"step": 550
},
{
"epoch": 0.8476923076923077,
"grad_norm": 0.5103149586075123,
"learning_rate": 1.7871834806090502e-05,
"loss": 0.1078,
"step": 551
},
{
"epoch": 0.8492307692307692,
"grad_norm": 0.5179913026426116,
"learning_rate": 1.7864377217310282e-05,
"loss": 0.0941,
"step": 552
},
{
"epoch": 0.8507692307692307,
"grad_norm": 0.5462964124972205,
"learning_rate": 1.7856908146535602e-05,
"loss": 0.0844,
"step": 553
},
{
"epoch": 0.8523076923076923,
"grad_norm": 0.40593162957333356,
"learning_rate": 1.784942760467131e-05,
"loss": 0.0943,
"step": 554
},
{
"epoch": 0.8538461538461538,
"grad_norm": 0.38375960879766213,
"learning_rate": 1.7841935602638997e-05,
"loss": 0.0804,
"step": 555
},
{
"epoch": 0.8553846153846154,
"grad_norm": 0.38136544658194893,
"learning_rate": 1.7834432151376992e-05,
"loss": 0.0945,
"step": 556
},
{
"epoch": 0.8569230769230769,
"grad_norm": 0.4194877597337682,
"learning_rate": 1.7826917261840337e-05,
"loss": 0.092,
"step": 557
},
{
"epoch": 0.8584615384615385,
"grad_norm": 0.4923272058939714,
"learning_rate": 1.7819390945000775e-05,
"loss": 0.0986,
"step": 558
},
{
"epoch": 0.86,
"grad_norm": 0.4875318696461418,
"learning_rate": 1.781185321184673e-05,
"loss": 0.1017,
"step": 559
},
{
"epoch": 0.8615384615384616,
"grad_norm": 0.39052926422478096,
"learning_rate": 1.7804304073383298e-05,
"loss": 0.1091,
"step": 560
},
{
"epoch": 0.8630769230769231,
"grad_norm": 0.38763635529234175,
"learning_rate": 1.7796743540632226e-05,
"loss": 0.0825,
"step": 561
},
{
"epoch": 0.8646153846153846,
"grad_norm": 0.3905770386460791,
"learning_rate": 1.778917162463189e-05,
"loss": 0.0875,
"step": 562
},
{
"epoch": 0.8661538461538462,
"grad_norm": 0.5703432509432513,
"learning_rate": 1.77815883364373e-05,
"loss": 0.1001,
"step": 563
},
{
"epoch": 0.8676923076923077,
"grad_norm": 0.4597959384906591,
"learning_rate": 1.777399368712005e-05,
"loss": 0.0882,
"step": 564
},
{
"epoch": 0.8692307692307693,
"grad_norm": 0.45607761156419957,
"learning_rate": 1.7766387687768338e-05,
"loss": 0.0991,
"step": 565
},
{
"epoch": 0.8707692307692307,
"grad_norm": 0.39506065545533703,
"learning_rate": 1.7758770349486924e-05,
"loss": 0.0943,
"step": 566
},
{
"epoch": 0.8723076923076923,
"grad_norm": 0.4913700102929278,
"learning_rate": 1.7751141683397128e-05,
"loss": 0.1026,
"step": 567
},
{
"epoch": 0.8738461538461538,
"grad_norm": 0.4481916097965894,
"learning_rate": 1.7743501700636804e-05,
"loss": 0.1035,
"step": 568
},
{
"epoch": 0.8753846153846154,
"grad_norm": 0.40743252425944065,
"learning_rate": 1.7735850412360332e-05,
"loss": 0.0915,
"step": 569
},
{
"epoch": 0.8769230769230769,
"grad_norm": 0.36201263572487186,
"learning_rate": 1.7728187829738596e-05,
"loss": 0.096,
"step": 570
},
{
"epoch": 0.8784615384615385,
"grad_norm": 0.41877789233929563,
"learning_rate": 1.772051396395897e-05,
"loss": 0.0921,
"step": 571
},
{
"epoch": 0.88,
"grad_norm": 0.43388638709693894,
"learning_rate": 1.7712828826225303e-05,
"loss": 0.1016,
"step": 572
},
{
"epoch": 0.8815384615384615,
"grad_norm": 0.4621448068884329,
"learning_rate": 1.7705132427757895e-05,
"loss": 0.0957,
"step": 573
},
{
"epoch": 0.8830769230769231,
"grad_norm": 0.384208019183389,
"learning_rate": 1.7697424779793497e-05,
"loss": 0.0812,
"step": 574
},
{
"epoch": 0.8846153846153846,
"grad_norm": 0.35252484159436787,
"learning_rate": 1.7689705893585273e-05,
"loss": 0.0918,
"step": 575
},
{
"epoch": 0.8861538461538462,
"grad_norm": 0.40143829209617715,
"learning_rate": 1.7681975780402807e-05,
"loss": 0.0843,
"step": 576
},
{
"epoch": 0.8876923076923077,
"grad_norm": 0.4307373370195823,
"learning_rate": 1.7674234451532065e-05,
"loss": 0.0961,
"step": 577
},
{
"epoch": 0.8892307692307693,
"grad_norm": 0.46711054218809045,
"learning_rate": 1.766648191827539e-05,
"loss": 0.0918,
"step": 578
},
{
"epoch": 0.8907692307692308,
"grad_norm": 0.43239124932710876,
"learning_rate": 1.7658718191951483e-05,
"loss": 0.0865,
"step": 579
},
{
"epoch": 0.8923076923076924,
"grad_norm": 0.36077221637299095,
"learning_rate": 1.7650943283895393e-05,
"loss": 0.0791,
"step": 580
},
{
"epoch": 0.8938461538461538,
"grad_norm": 0.452062987622938,
"learning_rate": 1.7643157205458483e-05,
"loss": 0.0977,
"step": 581
},
{
"epoch": 0.8953846153846153,
"grad_norm": 0.3729118607027113,
"learning_rate": 1.7635359968008438e-05,
"loss": 0.0964,
"step": 582
},
{
"epoch": 0.8969230769230769,
"grad_norm": 0.31864024958361536,
"learning_rate": 1.7627551582929223e-05,
"loss": 0.0706,
"step": 583
},
{
"epoch": 0.8984615384615384,
"grad_norm": 0.4897692560311474,
"learning_rate": 1.761973206162109e-05,
"loss": 0.095,
"step": 584
},
{
"epoch": 0.9,
"grad_norm": 0.43634370121716065,
"learning_rate": 1.7611901415500536e-05,
"loss": 0.1012,
"step": 585
},
{
"epoch": 0.9015384615384615,
"grad_norm": 0.4303303314789484,
"learning_rate": 1.7604059656000313e-05,
"loss": 0.0991,
"step": 586
},
{
"epoch": 0.9030769230769231,
"grad_norm": 0.3965278873786633,
"learning_rate": 1.759620679456939e-05,
"loss": 0.0889,
"step": 587
},
{
"epoch": 0.9046153846153846,
"grad_norm": 0.3471384513881275,
"learning_rate": 1.758834284267295e-05,
"loss": 0.0901,
"step": 588
},
{
"epoch": 0.9061538461538462,
"grad_norm": 0.35615891292502644,
"learning_rate": 1.7580467811792374e-05,
"loss": 0.084,
"step": 589
},
{
"epoch": 0.9076923076923077,
"grad_norm": 0.43993009413917056,
"learning_rate": 1.7572581713425195e-05,
"loss": 0.0986,
"step": 590
},
{
"epoch": 0.9092307692307692,
"grad_norm": 0.41369164253274765,
"learning_rate": 1.7564684559085138e-05,
"loss": 0.0888,
"step": 591
},
{
"epoch": 0.9107692307692308,
"grad_norm": 0.3683031999775181,
"learning_rate": 1.7556776360302038e-05,
"loss": 0.0912,
"step": 592
},
{
"epoch": 0.9123076923076923,
"grad_norm": 0.4537934276100818,
"learning_rate": 1.7548857128621878e-05,
"loss": 0.0921,
"step": 593
},
{
"epoch": 0.9138461538461539,
"grad_norm": 0.43435742031494934,
"learning_rate": 1.7540926875606734e-05,
"loss": 0.0869,
"step": 594
},
{
"epoch": 0.9153846153846154,
"grad_norm": 0.38341724227303553,
"learning_rate": 1.753298561283478e-05,
"loss": 0.0903,
"step": 595
},
{
"epoch": 0.916923076923077,
"grad_norm": 0.41746460815399494,
"learning_rate": 1.7525033351900268e-05,
"loss": 0.1006,
"step": 596
},
{
"epoch": 0.9184615384615384,
"grad_norm": 0.3777485553214645,
"learning_rate": 1.7517070104413497e-05,
"loss": 0.0947,
"step": 597
},
{
"epoch": 0.92,
"grad_norm": 0.41830568154844494,
"learning_rate": 1.7509095882000823e-05,
"loss": 0.1039,
"step": 598
},
{
"epoch": 0.9215384615384615,
"grad_norm": 0.36147731651312526,
"learning_rate": 1.7501110696304598e-05,
"loss": 0.0875,
"step": 599
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.3427586374547824,
"learning_rate": 1.7493114558983207e-05,
"loss": 0.0816,
"step": 600
},
{
"epoch": 0.9246153846153846,
"grad_norm": 0.3939804203050907,
"learning_rate": 1.7485107481711014e-05,
"loss": 0.0946,
"step": 601
},
{
"epoch": 0.9261538461538461,
"grad_norm": 0.381608432957127,
"learning_rate": 1.7477089476178354e-05,
"loss": 0.0929,
"step": 602
},
{
"epoch": 0.9276923076923077,
"grad_norm": 0.4292122557448878,
"learning_rate": 1.7469060554091518e-05,
"loss": 0.0965,
"step": 603
},
{
"epoch": 0.9292307692307692,
"grad_norm": 0.3762116025408573,
"learning_rate": 1.7461020727172736e-05,
"loss": 0.0938,
"step": 604
},
{
"epoch": 0.9307692307692308,
"grad_norm": 0.3455952506447102,
"learning_rate": 1.745297000716016e-05,
"loss": 0.0741,
"step": 605
},
{
"epoch": 0.9323076923076923,
"grad_norm": 0.38631458311410155,
"learning_rate": 1.7444908405807845e-05,
"loss": 0.0865,
"step": 606
},
{
"epoch": 0.9338461538461539,
"grad_norm": 0.37275711289343844,
"learning_rate": 1.7436835934885735e-05,
"loss": 0.0886,
"step": 607
},
{
"epoch": 0.9353846153846154,
"grad_norm": 0.4401140100719851,
"learning_rate": 1.742875260617964e-05,
"loss": 0.0976,
"step": 608
},
{
"epoch": 0.936923076923077,
"grad_norm": 0.3921855404257059,
"learning_rate": 1.7420658431491224e-05,
"loss": 0.0928,
"step": 609
},
{
"epoch": 0.9384615384615385,
"grad_norm": 0.4113951217410347,
"learning_rate": 1.741255342263798e-05,
"loss": 0.0885,
"step": 610
},
{
"epoch": 0.94,
"grad_norm": 0.3262775097841378,
"learning_rate": 1.7404437591453237e-05,
"loss": 0.0873,
"step": 611
},
{
"epoch": 0.9415384615384615,
"grad_norm": 0.42350913769649273,
"learning_rate": 1.73963109497861e-05,
"loss": 0.0958,
"step": 612
},
{
"epoch": 0.943076923076923,
"grad_norm": 0.36988134430557396,
"learning_rate": 1.7388173509501475e-05,
"loss": 0.0928,
"step": 613
},
{
"epoch": 0.9446153846153846,
"grad_norm": 0.39410658459241166,
"learning_rate": 1.7380025282480028e-05,
"loss": 0.0868,
"step": 614
},
{
"epoch": 0.9461538461538461,
"grad_norm": 0.41180225877084126,
"learning_rate": 1.7371866280618176e-05,
"loss": 0.0911,
"step": 615
},
{
"epoch": 0.9476923076923077,
"grad_norm": 0.35890123689534553,
"learning_rate": 1.7363696515828062e-05,
"loss": 0.0899,
"step": 616
},
{
"epoch": 0.9492307692307692,
"grad_norm": 0.2987459340450158,
"learning_rate": 1.7355516000037555e-05,
"loss": 0.0769,
"step": 617
},
{
"epoch": 0.9507692307692308,
"grad_norm": 0.43301969400641127,
"learning_rate": 1.73473247451902e-05,
"loss": 0.0892,
"step": 618
},
{
"epoch": 0.9523076923076923,
"grad_norm": 0.379899779524929,
"learning_rate": 1.733912276324524e-05,
"loss": 0.0751,
"step": 619
},
{
"epoch": 0.9538461538461539,
"grad_norm": 0.35284361586327184,
"learning_rate": 1.7330910066177574e-05,
"loss": 0.0941,
"step": 620
},
{
"epoch": 0.9553846153846154,
"grad_norm": 0.4459109096619344,
"learning_rate": 1.7322686665977738e-05,
"loss": 0.086,
"step": 621
},
{
"epoch": 0.9569230769230769,
"grad_norm": 0.4302144299712948,
"learning_rate": 1.7314452574651902e-05,
"loss": 0.0865,
"step": 622
},
{
"epoch": 0.9584615384615385,
"grad_norm": 0.4053685427552524,
"learning_rate": 1.7306207804221845e-05,
"loss": 0.092,
"step": 623
},
{
"epoch": 0.96,
"grad_norm": 0.46831862180021,
"learning_rate": 1.7297952366724935e-05,
"loss": 0.1046,
"step": 624
},
{
"epoch": 0.9615384615384616,
"grad_norm": 0.4538044661350598,
"learning_rate": 1.7289686274214116e-05,
"loss": 0.1076,
"step": 625
},
{
"epoch": 0.963076923076923,
"grad_norm": 0.48389404664226576,
"learning_rate": 1.7281409538757886e-05,
"loss": 0.0923,
"step": 626
},
{
"epoch": 0.9646153846153847,
"grad_norm": 0.3865282157362884,
"learning_rate": 1.727312217244028e-05,
"loss": 0.0881,
"step": 627
},
{
"epoch": 0.9661538461538461,
"grad_norm": 0.41052873518942035,
"learning_rate": 1.726482418736086e-05,
"loss": 0.0982,
"step": 628
},
{
"epoch": 0.9676923076923077,
"grad_norm": 0.4633845480863643,
"learning_rate": 1.7256515595634688e-05,
"loss": 0.0963,
"step": 629
},
{
"epoch": 0.9692307692307692,
"grad_norm": 0.49146329005893413,
"learning_rate": 1.7248196409392312e-05,
"loss": 0.0924,
"step": 630
},
{
"epoch": 0.9707692307692307,
"grad_norm": 0.35983268721595085,
"learning_rate": 1.7239866640779745e-05,
"loss": 0.0884,
"step": 631
},
{
"epoch": 0.9723076923076923,
"grad_norm": 0.490761583695885,
"learning_rate": 1.7231526301958454e-05,
"loss": 0.0973,
"step": 632
},
{
"epoch": 0.9738461538461538,
"grad_norm": 0.3594410933659223,
"learning_rate": 1.722317540510534e-05,
"loss": 0.0929,
"step": 633
},
{
"epoch": 0.9753846153846154,
"grad_norm": 0.32824532953007957,
"learning_rate": 1.7214813962412715e-05,
"loss": 0.0886,
"step": 634
},
{
"epoch": 0.9769230769230769,
"grad_norm": 0.3557861935512496,
"learning_rate": 1.720644198608829e-05,
"loss": 0.0885,
"step": 635
},
{
"epoch": 0.9784615384615385,
"grad_norm": 0.45266556902976496,
"learning_rate": 1.7198059488355153e-05,
"loss": 0.0933,
"step": 636
},
{
"epoch": 0.98,
"grad_norm": 0.40557942924254287,
"learning_rate": 1.7189666481451755e-05,
"loss": 0.0951,
"step": 637
},
{
"epoch": 0.9815384615384616,
"grad_norm": 0.35736570345844687,
"learning_rate": 1.718126297763189e-05,
"loss": 0.0888,
"step": 638
},
{
"epoch": 0.9830769230769231,
"grad_norm": 0.4631511259456779,
"learning_rate": 1.717284898916468e-05,
"loss": 0.091,
"step": 639
},
{
"epoch": 0.9846153846153847,
"grad_norm": 0.43267790292943314,
"learning_rate": 1.7164424528334548e-05,
"loss": 0.0961,
"step": 640
},
{
"epoch": 0.9861538461538462,
"grad_norm": 0.46968347009435635,
"learning_rate": 1.715598960744121e-05,
"loss": 0.0887,
"step": 641
},
{
"epoch": 0.9876923076923076,
"grad_norm": 0.447356115085785,
"learning_rate": 1.7147544238799664e-05,
"loss": 0.0908,
"step": 642
},
{
"epoch": 0.9892307692307692,
"grad_norm": 0.32693771260343074,
"learning_rate": 1.7139088434740142e-05,
"loss": 0.0872,
"step": 643
},
{
"epoch": 0.9907692307692307,
"grad_norm": 0.422322021584036,
"learning_rate": 1.7130622207608126e-05,
"loss": 0.0975,
"step": 644
},
{
"epoch": 0.9923076923076923,
"grad_norm": 0.34862552621490467,
"learning_rate": 1.712214556976431e-05,
"loss": 0.0944,
"step": 645
},
{
"epoch": 0.9938461538461538,
"grad_norm": 0.3800495940989224,
"learning_rate": 1.7113658533584594e-05,
"loss": 0.0793,
"step": 646
},
{
"epoch": 0.9953846153846154,
"grad_norm": 0.5118689124842419,
"learning_rate": 1.7105161111460046e-05,
"loss": 0.0949,
"step": 647
},
{
"epoch": 0.9969230769230769,
"grad_norm": 0.412157009432569,
"learning_rate": 1.7096653315796915e-05,
"loss": 0.1011,
"step": 648
},
{
"epoch": 0.9984615384615385,
"grad_norm": 0.3780510510068702,
"learning_rate": 1.7088135159016584e-05,
"loss": 0.0982,
"step": 649
},
{
"epoch": 1.0,
"grad_norm": 0.41945932684457304,
"learning_rate": 1.7079606653555563e-05,
"loss": 0.0953,
"step": 650
},
{
"epoch": 1.0015384615384615,
"grad_norm": 0.43622049064952106,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.0749,
"step": 651
},
{
"epoch": 1.003076923076923,
"grad_norm": 0.38132209199435835,
"learning_rate": 1.706251864641304e-05,
"loss": 0.0726,
"step": 652
},
{
"epoch": 1.0046153846153847,
"grad_norm": 0.30052022007987506,
"learning_rate": 1.7053959169680033e-05,
"loss": 0.0715,
"step": 653
},
{
"epoch": 1.0061538461538462,
"grad_norm": 0.3185644660405221,
"learning_rate": 1.7045389394163297e-05,
"loss": 0.0688,
"step": 654
},
{
"epoch": 1.0076923076923077,
"grad_norm": 0.35208492988636686,
"learning_rate": 1.7036809332374713e-05,
"loss": 0.0689,
"step": 655
},
{
"epoch": 1.0092307692307692,
"grad_norm": 0.44091806848570947,
"learning_rate": 1.7028218996841173e-05,
"loss": 0.0709,
"step": 656
},
{
"epoch": 1.0107692307692309,
"grad_norm": 0.37173547317449596,
"learning_rate": 1.7019618400104572e-05,
"loss": 0.0663,
"step": 657
},
{
"epoch": 1.0123076923076924,
"grad_norm": 0.3752761957439137,
"learning_rate": 1.7011007554721778e-05,
"loss": 0.0696,
"step": 658
},
{
"epoch": 1.0138461538461538,
"grad_norm": 0.3413943290234514,
"learning_rate": 1.700238647326464e-05,
"loss": 0.0675,
"step": 659
},
{
"epoch": 1.0153846153846153,
"grad_norm": 0.37509528395315783,
"learning_rate": 1.6993755168319934e-05,
"loss": 0.0718,
"step": 660
},
{
"epoch": 1.0169230769230768,
"grad_norm": 0.39194820102366207,
"learning_rate": 1.6985113652489374e-05,
"loss": 0.062,
"step": 661
},
{
"epoch": 1.0184615384615385,
"grad_norm": 0.296028112394497,
"learning_rate": 1.697646193838957e-05,
"loss": 0.0658,
"step": 662
},
{
"epoch": 1.02,
"grad_norm": 0.3737919319488901,
"learning_rate": 1.6967800038652035e-05,
"loss": 0.0615,
"step": 663
},
{
"epoch": 1.0215384615384615,
"grad_norm": 0.36005445736837366,
"learning_rate": 1.6959127965923144e-05,
"loss": 0.0691,
"step": 664
},
{
"epoch": 1.023076923076923,
"grad_norm": 0.3514084769551997,
"learning_rate": 1.695044573286413e-05,
"loss": 0.0724,
"step": 665
},
{
"epoch": 1.0246153846153847,
"grad_norm": 0.37783560046213566,
"learning_rate": 1.6941753352151057e-05,
"loss": 0.063,
"step": 666
},
{
"epoch": 1.0261538461538462,
"grad_norm": 0.3639679221324666,
"learning_rate": 1.69330508364748e-05,
"loss": 0.0682,
"step": 667
},
{
"epoch": 1.0276923076923077,
"grad_norm": 0.42252833886280206,
"learning_rate": 1.692433819854104e-05,
"loss": 0.0694,
"step": 668
},
{
"epoch": 1.0292307692307692,
"grad_norm": 0.44692488489777815,
"learning_rate": 1.6915615451070234e-05,
"loss": 0.0705,
"step": 669
},
{
"epoch": 1.0307692307692307,
"grad_norm": 0.3678986564514892,
"learning_rate": 1.6906882606797595e-05,
"loss": 0.0657,
"step": 670
},
{
"epoch": 1.0323076923076924,
"grad_norm": 0.43913282105160406,
"learning_rate": 1.689813967847308e-05,
"loss": 0.071,
"step": 671
},
{
"epoch": 1.0338461538461539,
"grad_norm": 0.32313030736032045,
"learning_rate": 1.6889386678861365e-05,
"loss": 0.0665,
"step": 672
},
{
"epoch": 1.0353846153846153,
"grad_norm": 0.3913357186511293,
"learning_rate": 1.6880623620741843e-05,
"loss": 0.0677,
"step": 673
},
{
"epoch": 1.0369230769230768,
"grad_norm": 0.3502824876065054,
"learning_rate": 1.6871850516908575e-05,
"loss": 0.0651,
"step": 674
},
{
"epoch": 1.0384615384615385,
"grad_norm": 0.5615618864576786,
"learning_rate": 1.68630673801703e-05,
"loss": 0.0669,
"step": 675
},
{
"epoch": 1.04,
"grad_norm": 0.45343396226547117,
"learning_rate": 1.68542742233504e-05,
"loss": 0.0817,
"step": 676
},
{
"epoch": 1.0415384615384615,
"grad_norm": 0.40223580906946954,
"learning_rate": 1.684547105928689e-05,
"loss": 0.0691,
"step": 677
},
{
"epoch": 1.043076923076923,
"grad_norm": 0.39525415008396314,
"learning_rate": 1.683665790083239e-05,
"loss": 0.0709,
"step": 678
},
{
"epoch": 1.0446153846153847,
"grad_norm": 0.29528931632742,
"learning_rate": 1.682783476085412e-05,
"loss": 0.0562,
"step": 679
},
{
"epoch": 1.0461538461538462,
"grad_norm": 0.4540404912733847,
"learning_rate": 1.6819001652233867e-05,
"loss": 0.0668,
"step": 680
},
{
"epoch": 1.0476923076923077,
"grad_norm": 0.4901824401666557,
"learning_rate": 1.6810158587867973e-05,
"loss": 0.0735,
"step": 681
},
{
"epoch": 1.0492307692307692,
"grad_norm": 0.35657036998216646,
"learning_rate": 1.6801305580667318e-05,
"loss": 0.0659,
"step": 682
},
{
"epoch": 1.0507692307692307,
"grad_norm": 0.34213142704883004,
"learning_rate": 1.679244264355729e-05,
"loss": 0.0631,
"step": 683
},
{
"epoch": 1.0523076923076924,
"grad_norm": 0.39311143967753803,
"learning_rate": 1.6783569789477795e-05,
"loss": 0.0652,
"step": 684
},
{
"epoch": 1.0538461538461539,
"grad_norm": 0.3501379606305712,
"learning_rate": 1.677468703138319e-05,
"loss": 0.0699,
"step": 685
},
{
"epoch": 1.0553846153846154,
"grad_norm": 0.359427195373609,
"learning_rate": 1.6765794382242315e-05,
"loss": 0.0657,
"step": 686
},
{
"epoch": 1.0569230769230769,
"grad_norm": 0.3900165610051957,
"learning_rate": 1.6756891855038436e-05,
"loss": 0.0648,
"step": 687
},
{
"epoch": 1.0584615384615386,
"grad_norm": 0.3768191856115309,
"learning_rate": 1.6747979462769253e-05,
"loss": 0.0661,
"step": 688
},
{
"epoch": 1.06,
"grad_norm": 0.3788900739237935,
"learning_rate": 1.673905721844686e-05,
"loss": 0.0689,
"step": 689
},
{
"epoch": 1.0615384615384615,
"grad_norm": 0.38294691852219787,
"learning_rate": 1.6730125135097736e-05,
"loss": 0.0733,
"step": 690
},
{
"epoch": 1.063076923076923,
"grad_norm": 0.35955788229196967,
"learning_rate": 1.6721183225762726e-05,
"loss": 0.0653,
"step": 691
},
{
"epoch": 1.0646153846153845,
"grad_norm": 0.4314633145507115,
"learning_rate": 1.6712231503497028e-05,
"loss": 0.0693,
"step": 692
},
{
"epoch": 1.0661538461538462,
"grad_norm": 0.3819072156771492,
"learning_rate": 1.670326998137016e-05,
"loss": 0.0686,
"step": 693
},
{
"epoch": 1.0676923076923077,
"grad_norm": 0.32895823745294955,
"learning_rate": 1.669429867246594e-05,
"loss": 0.0696,
"step": 694
},
{
"epoch": 1.0692307692307692,
"grad_norm": 0.4310560863117649,
"learning_rate": 1.668531758988249e-05,
"loss": 0.0715,
"step": 695
},
{
"epoch": 1.0707692307692307,
"grad_norm": 0.49530944777177144,
"learning_rate": 1.6676326746732197e-05,
"loss": 0.0684,
"step": 696
},
{
"epoch": 1.0723076923076924,
"grad_norm": 0.3399836957762591,
"learning_rate": 1.666732615614169e-05,
"loss": 0.0772,
"step": 697
},
{
"epoch": 1.073846153846154,
"grad_norm": 0.3717915125573905,
"learning_rate": 1.665831583125184e-05,
"loss": 0.0735,
"step": 698
},
{
"epoch": 1.0753846153846154,
"grad_norm": 0.4045082908654806,
"learning_rate": 1.6649295785217722e-05,
"loss": 0.0712,
"step": 699
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.38922594233503904,
"learning_rate": 1.664026603120861e-05,
"loss": 0.0709,
"step": 700
},
{
"epoch": 1.0784615384615384,
"grad_norm": 0.5013858686476608,
"learning_rate": 1.6631226582407954e-05,
"loss": 0.0764,
"step": 701
},
{
"epoch": 1.08,
"grad_norm": 0.3717981315840534,
"learning_rate": 1.6622177452013347e-05,
"loss": 0.0705,
"step": 702
},
{
"epoch": 1.0815384615384616,
"grad_norm": 0.38788483529072765,
"learning_rate": 1.661311865323652e-05,
"loss": 0.0719,
"step": 703
},
{
"epoch": 1.083076923076923,
"grad_norm": 1.053330169816148,
"learning_rate": 1.660405019930333e-05,
"loss": 0.0735,
"step": 704
},
{
"epoch": 1.0846153846153845,
"grad_norm": 0.6972695498767989,
"learning_rate": 1.6594972103453727e-05,
"loss": 0.0718,
"step": 705
},
{
"epoch": 1.0861538461538462,
"grad_norm": 0.37342568285958827,
"learning_rate": 1.6585884378941727e-05,
"loss": 0.0753,
"step": 706
},
{
"epoch": 1.0876923076923077,
"grad_norm": 0.4267000057052577,
"learning_rate": 1.6576787039035417e-05,
"loss": 0.0649,
"step": 707
},
{
"epoch": 1.0892307692307692,
"grad_norm": 0.36310651208043226,
"learning_rate": 1.6567680097016917e-05,
"loss": 0.0686,
"step": 708
},
{
"epoch": 1.0907692307692307,
"grad_norm": 0.42763138085861285,
"learning_rate": 1.6558563566182365e-05,
"loss": 0.0657,
"step": 709
},
{
"epoch": 1.0923076923076924,
"grad_norm": 0.4251635596635269,
"learning_rate": 1.65494374598419e-05,
"loss": 0.077,
"step": 710
},
{
"epoch": 1.093846153846154,
"grad_norm": 0.6823950567290125,
"learning_rate": 1.6540301791319647e-05,
"loss": 0.0642,
"step": 711
},
{
"epoch": 1.0953846153846154,
"grad_norm": 0.30086732326600235,
"learning_rate": 1.6531156573953677e-05,
"loss": 0.0635,
"step": 712
},
{
"epoch": 1.096923076923077,
"grad_norm": 0.3272124590429498,
"learning_rate": 1.652200182109602e-05,
"loss": 0.0711,
"step": 713
},
{
"epoch": 1.0984615384615384,
"grad_norm": 0.3783056835805405,
"learning_rate": 1.6512837546112617e-05,
"loss": 0.0702,
"step": 714
},
{
"epoch": 1.1,
"grad_norm": 0.34618660310859284,
"learning_rate": 1.6503663762383312e-05,
"loss": 0.066,
"step": 715
},
{
"epoch": 1.1015384615384616,
"grad_norm": 0.3509817308378501,
"learning_rate": 1.6494480483301836e-05,
"loss": 0.0681,
"step": 716
},
{
"epoch": 1.103076923076923,
"grad_norm": 0.435238684258864,
"learning_rate": 1.6485287722275783e-05,
"loss": 0.0654,
"step": 717
},
{
"epoch": 1.1046153846153846,
"grad_norm": 0.4858221064894468,
"learning_rate": 1.6476085492726582e-05,
"loss": 0.0621,
"step": 718
},
{
"epoch": 1.106153846153846,
"grad_norm": 0.5547851396481795,
"learning_rate": 1.6466873808089496e-05,
"loss": 0.065,
"step": 719
},
{
"epoch": 1.1076923076923078,
"grad_norm": 0.3551740998729382,
"learning_rate": 1.645765268181359e-05,
"loss": 0.0695,
"step": 720
},
{
"epoch": 1.1092307692307692,
"grad_norm": 0.5426959849480466,
"learning_rate": 1.6448422127361707e-05,
"loss": 0.0679,
"step": 721
},
{
"epoch": 1.1107692307692307,
"grad_norm": 0.4014572952029352,
"learning_rate": 1.6439182158210468e-05,
"loss": 0.0682,
"step": 722
},
{
"epoch": 1.1123076923076922,
"grad_norm": 0.37264714876106975,
"learning_rate": 1.642993278785023e-05,
"loss": 0.0622,
"step": 723
},
{
"epoch": 1.113846153846154,
"grad_norm": 0.4085985468005736,
"learning_rate": 1.642067402978508e-05,
"loss": 0.0708,
"step": 724
},
{
"epoch": 1.1153846153846154,
"grad_norm": 0.3741294371749341,
"learning_rate": 1.64114058975328e-05,
"loss": 0.0631,
"step": 725
},
{
"epoch": 1.116923076923077,
"grad_norm": 0.45876871469434627,
"learning_rate": 1.640212840462488e-05,
"loss": 0.0663,
"step": 726
},
{
"epoch": 1.1184615384615384,
"grad_norm": 0.3422472176749258,
"learning_rate": 1.639284156460646e-05,
"loss": 0.0613,
"step": 727
},
{
"epoch": 1.12,
"grad_norm": 0.39030560028876443,
"learning_rate": 1.6383545391036327e-05,
"loss": 0.0659,
"step": 728
},
{
"epoch": 1.1215384615384616,
"grad_norm": 0.32978927481973175,
"learning_rate": 1.63742398974869e-05,
"loss": 0.0583,
"step": 729
},
{
"epoch": 1.123076923076923,
"grad_norm": 0.3669713631154274,
"learning_rate": 1.63649250975442e-05,
"loss": 0.0723,
"step": 730
},
{
"epoch": 1.1246153846153846,
"grad_norm": 0.35663226905770323,
"learning_rate": 1.6355601004807856e-05,
"loss": 0.0687,
"step": 731
},
{
"epoch": 1.126153846153846,
"grad_norm": 0.39492116903405033,
"learning_rate": 1.6346267632891027e-05,
"loss": 0.0705,
"step": 732
},
{
"epoch": 1.1276923076923078,
"grad_norm": 0.36069535113212814,
"learning_rate": 1.6336924995420453e-05,
"loss": 0.0634,
"step": 733
},
{
"epoch": 1.1292307692307693,
"grad_norm": 0.4776466189937346,
"learning_rate": 1.6327573106036384e-05,
"loss": 0.0772,
"step": 734
},
{
"epoch": 1.1307692307692307,
"grad_norm": 0.3872373026437133,
"learning_rate": 1.6318211978392588e-05,
"loss": 0.0729,
"step": 735
},
{
"epoch": 1.1323076923076922,
"grad_norm": 0.3782947210886841,
"learning_rate": 1.630884162615631e-05,
"loss": 0.0698,
"step": 736
},
{
"epoch": 1.1338461538461537,
"grad_norm": 0.38313732578124937,
"learning_rate": 1.6299462063008272e-05,
"loss": 0.0659,
"step": 737
},
{
"epoch": 1.1353846153846154,
"grad_norm": 0.32705882716559326,
"learning_rate": 1.6290073302642637e-05,
"loss": 0.0649,
"step": 738
},
{
"epoch": 1.136923076923077,
"grad_norm": 0.3322452475653334,
"learning_rate": 1.6280675358767005e-05,
"loss": 0.0675,
"step": 739
},
{
"epoch": 1.1384615384615384,
"grad_norm": 0.3711248581770431,
"learning_rate": 1.6271268245102377e-05,
"loss": 0.0737,
"step": 740
},
{
"epoch": 1.1400000000000001,
"grad_norm": 0.39494242888459563,
"learning_rate": 1.626185197538314e-05,
"loss": 0.0706,
"step": 741
},
{
"epoch": 1.1415384615384616,
"grad_norm": 0.3116998586738848,
"learning_rate": 1.6252426563357054e-05,
"loss": 0.0641,
"step": 742
},
{
"epoch": 1.143076923076923,
"grad_norm": 0.42963979037379485,
"learning_rate": 1.6242992022785225e-05,
"loss": 0.0683,
"step": 743
},
{
"epoch": 1.1446153846153846,
"grad_norm": 0.36479367771942905,
"learning_rate": 1.623354836744209e-05,
"loss": 0.0713,
"step": 744
},
{
"epoch": 1.146153846153846,
"grad_norm": 0.6468927692309728,
"learning_rate": 1.6224095611115385e-05,
"loss": 0.0652,
"step": 745
},
{
"epoch": 1.1476923076923078,
"grad_norm": 0.6063113124725473,
"learning_rate": 1.6214633767606142e-05,
"loss": 0.0746,
"step": 746
},
{
"epoch": 1.1492307692307693,
"grad_norm": 0.4304453405042257,
"learning_rate": 1.620516285072866e-05,
"loss": 0.0698,
"step": 747
},
{
"epoch": 1.1507692307692308,
"grad_norm": 0.317379420605374,
"learning_rate": 1.6195682874310473e-05,
"loss": 0.0647,
"step": 748
},
{
"epoch": 1.1523076923076923,
"grad_norm": 0.39131831324165883,
"learning_rate": 1.6186193852192356e-05,
"loss": 0.0755,
"step": 749
},
{
"epoch": 1.1538461538461537,
"grad_norm": 0.35798460402065585,
"learning_rate": 1.617669579822829e-05,
"loss": 0.0746,
"step": 750
},
{
"epoch": 1.1553846153846155,
"grad_norm": 0.3502631169305045,
"learning_rate": 1.6167188726285433e-05,
"loss": 0.0697,
"step": 751
},
{
"epoch": 1.156923076923077,
"grad_norm": 0.40987521557285755,
"learning_rate": 1.6157672650244113e-05,
"loss": 0.0798,
"step": 752
},
{
"epoch": 1.1584615384615384,
"grad_norm": 0.352033824405812,
"learning_rate": 1.6148147583997813e-05,
"loss": 0.0678,
"step": 753
},
{
"epoch": 1.16,
"grad_norm": 0.4345817539730366,
"learning_rate": 1.6138613541453127e-05,
"loss": 0.0728,
"step": 754
},
{
"epoch": 1.1615384615384616,
"grad_norm": 0.35074866147767675,
"learning_rate": 1.6129070536529767e-05,
"loss": 0.0687,
"step": 755
},
{
"epoch": 1.1630769230769231,
"grad_norm": 0.36887909746623887,
"learning_rate": 1.611951858316052e-05,
"loss": 0.072,
"step": 756
},
{
"epoch": 1.1646153846153846,
"grad_norm": 0.2983907917380261,
"learning_rate": 1.6109957695291246e-05,
"loss": 0.0685,
"step": 757
},
{
"epoch": 1.166153846153846,
"grad_norm": 0.3934269192468973,
"learning_rate": 1.610038788688084e-05,
"loss": 0.0677,
"step": 758
},
{
"epoch": 1.1676923076923078,
"grad_norm": 0.37318389733402674,
"learning_rate": 1.6090809171901237e-05,
"loss": 0.0718,
"step": 759
},
{
"epoch": 1.1692307692307693,
"grad_norm": 0.38645652176143314,
"learning_rate": 1.6081221564337356e-05,
"loss": 0.0687,
"step": 760
},
{
"epoch": 1.1707692307692308,
"grad_norm": 0.38329410093067623,
"learning_rate": 1.6071625078187113e-05,
"loss": 0.0666,
"step": 761
},
{
"epoch": 1.1723076923076923,
"grad_norm": 0.30301582024440277,
"learning_rate": 1.6062019727461384e-05,
"loss": 0.0671,
"step": 762
},
{
"epoch": 1.1738461538461538,
"grad_norm": 0.32424463849184526,
"learning_rate": 1.605240552618398e-05,
"loss": 0.0715,
"step": 763
},
{
"epoch": 1.1753846153846155,
"grad_norm": 0.35922202384394764,
"learning_rate": 1.6042782488391644e-05,
"loss": 0.0732,
"step": 764
},
{
"epoch": 1.176923076923077,
"grad_norm": 0.3647659403794086,
"learning_rate": 1.603315062813401e-05,
"loss": 0.0707,
"step": 765
},
{
"epoch": 1.1784615384615384,
"grad_norm": 0.34982427128139343,
"learning_rate": 1.6023509959473608e-05,
"loss": 0.0725,
"step": 766
},
{
"epoch": 1.18,
"grad_norm": 0.427718797984176,
"learning_rate": 1.601386049648581e-05,
"loss": 0.0737,
"step": 767
},
{
"epoch": 1.1815384615384614,
"grad_norm": 0.3924968355298374,
"learning_rate": 1.6004202253258844e-05,
"loss": 0.0649,
"step": 768
},
{
"epoch": 1.1830769230769231,
"grad_norm": 0.30260539928386654,
"learning_rate": 1.5994535243893742e-05,
"loss": 0.0695,
"step": 769
},
{
"epoch": 1.1846153846153846,
"grad_norm": 0.35596304966031656,
"learning_rate": 1.5984859482504347e-05,
"loss": 0.067,
"step": 770
},
{
"epoch": 1.1861538461538461,
"grad_norm": 0.3129839541191524,
"learning_rate": 1.5975174983217273e-05,
"loss": 0.069,
"step": 771
},
{
"epoch": 1.1876923076923076,
"grad_norm": 0.3674547487152307,
"learning_rate": 1.5965481760171897e-05,
"loss": 0.0631,
"step": 772
},
{
"epoch": 1.1892307692307693,
"grad_norm": 0.3335296888041543,
"learning_rate": 1.5955779827520327e-05,
"loss": 0.0641,
"step": 773
},
{
"epoch": 1.1907692307692308,
"grad_norm": 0.39407159089071714,
"learning_rate": 1.5946069199427387e-05,
"loss": 0.071,
"step": 774
},
{
"epoch": 1.1923076923076923,
"grad_norm": 0.37085530721330084,
"learning_rate": 1.5936349890070602e-05,
"loss": 0.0733,
"step": 775
},
{
"epoch": 1.1938461538461538,
"grad_norm": 0.3065421660271998,
"learning_rate": 1.592662191364017e-05,
"loss": 0.0646,
"step": 776
},
{
"epoch": 1.1953846153846155,
"grad_norm": 0.4157113450091342,
"learning_rate": 1.5916885284338937e-05,
"loss": 0.0719,
"step": 777
},
{
"epoch": 1.196923076923077,
"grad_norm": 0.37188603187355035,
"learning_rate": 1.5907140016382385e-05,
"loss": 0.0682,
"step": 778
},
{
"epoch": 1.1984615384615385,
"grad_norm": 0.34699200375559375,
"learning_rate": 1.5897386123998613e-05,
"loss": 0.0745,
"step": 779
},
{
"epoch": 1.2,
"grad_norm": 0.31961705854482836,
"learning_rate": 1.588762362142831e-05,
"loss": 0.0647,
"step": 780
},
{
"epoch": 1.2015384615384614,
"grad_norm": 0.35665468708944836,
"learning_rate": 1.5877852522924733e-05,
"loss": 0.0657,
"step": 781
},
{
"epoch": 1.2030769230769232,
"grad_norm": 0.3379980896855412,
"learning_rate": 1.586807284275369e-05,
"loss": 0.0631,
"step": 782
},
{
"epoch": 1.2046153846153846,
"grad_norm": 0.398246969459009,
"learning_rate": 1.5858284595193514e-05,
"loss": 0.0732,
"step": 783
},
{
"epoch": 1.2061538461538461,
"grad_norm": 0.34921906662522734,
"learning_rate": 1.584848779453506e-05,
"loss": 0.0627,
"step": 784
},
{
"epoch": 1.2076923076923076,
"grad_norm": 0.37101620781077305,
"learning_rate": 1.5838682455081657e-05,
"loss": 0.0621,
"step": 785
},
{
"epoch": 1.209230769230769,
"grad_norm": 0.341332526505076,
"learning_rate": 1.5828868591149104e-05,
"loss": 0.063,
"step": 786
},
{
"epoch": 1.2107692307692308,
"grad_norm": 0.3041915924214054,
"learning_rate": 1.581904621706565e-05,
"loss": 0.0678,
"step": 787
},
{
"epoch": 1.2123076923076923,
"grad_norm": 0.36665714928540116,
"learning_rate": 1.580921534717196e-05,
"loss": 0.0743,
"step": 788
},
{
"epoch": 1.2138461538461538,
"grad_norm": 0.3377775940167995,
"learning_rate": 1.5799375995821116e-05,
"loss": 0.064,
"step": 789
},
{
"epoch": 1.2153846153846155,
"grad_norm": 0.35353779842042976,
"learning_rate": 1.5789528177378574e-05,
"loss": 0.079,
"step": 790
},
{
"epoch": 1.216923076923077,
"grad_norm": 0.3232778165659884,
"learning_rate": 1.577967190622215e-05,
"loss": 0.066,
"step": 791
},
{
"epoch": 1.2184615384615385,
"grad_norm": 0.3577677354201108,
"learning_rate": 1.5769807196742008e-05,
"loss": 0.0653,
"step": 792
},
{
"epoch": 1.22,
"grad_norm": 0.31783041296834197,
"learning_rate": 1.5759934063340627e-05,
"loss": 0.0706,
"step": 793
},
{
"epoch": 1.2215384615384615,
"grad_norm": 0.40205477740857387,
"learning_rate": 1.575005252043279e-05,
"loss": 0.0725,
"step": 794
},
{
"epoch": 1.2230769230769232,
"grad_norm": 0.3925627479595986,
"learning_rate": 1.5740162582445545e-05,
"loss": 0.07,
"step": 795
},
{
"epoch": 1.2246153846153847,
"grad_norm": 0.3311986658898681,
"learning_rate": 1.5730264263818212e-05,
"loss": 0.0762,
"step": 796
},
{
"epoch": 1.2261538461538461,
"grad_norm": 0.35012400117324377,
"learning_rate": 1.5720357579002346e-05,
"loss": 0.0621,
"step": 797
},
{
"epoch": 1.2276923076923076,
"grad_norm": 0.351604049632638,
"learning_rate": 1.5710442542461705e-05,
"loss": 0.0638,
"step": 798
},
{
"epoch": 1.2292307692307691,
"grad_norm": 0.356583134009754,
"learning_rate": 1.5700519168672248e-05,
"loss": 0.0713,
"step": 799
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.3285074874868023,
"learning_rate": 1.5690587472122104e-05,
"loss": 0.0643,
"step": 800
},
{
"epoch": 1.2323076923076923,
"grad_norm": 0.3450768181316114,
"learning_rate": 1.568064746731156e-05,
"loss": 0.0633,
"step": 801
},
{
"epoch": 1.2338461538461538,
"grad_norm": 0.3539947251115169,
"learning_rate": 1.5670699168753022e-05,
"loss": 0.0662,
"step": 802
},
{
"epoch": 1.2353846153846153,
"grad_norm": 0.32288987773337546,
"learning_rate": 1.5660742590971014e-05,
"loss": 0.0679,
"step": 803
},
{
"epoch": 1.236923076923077,
"grad_norm": 0.5408106149836094,
"learning_rate": 1.5650777748502144e-05,
"loss": 0.0739,
"step": 804
},
{
"epoch": 1.2384615384615385,
"grad_norm": 0.3333851374522989,
"learning_rate": 1.5640804655895086e-05,
"loss": 0.0666,
"step": 805
},
{
"epoch": 1.24,
"grad_norm": 0.3514156125809535,
"learning_rate": 1.5630823327710558e-05,
"loss": 0.0609,
"step": 806
},
{
"epoch": 1.2415384615384615,
"grad_norm": 0.4355584652770312,
"learning_rate": 1.5620833778521306e-05,
"loss": 0.0698,
"step": 807
},
{
"epoch": 1.2430769230769232,
"grad_norm": 0.3863399662303979,
"learning_rate": 1.561083602291208e-05,
"loss": 0.0747,
"step": 808
},
{
"epoch": 1.2446153846153847,
"grad_norm": 0.4272070354943623,
"learning_rate": 1.5600830075479604e-05,
"loss": 0.0625,
"step": 809
},
{
"epoch": 1.2461538461538462,
"grad_norm": 0.34034969408428656,
"learning_rate": 1.559081595083256e-05,
"loss": 0.0611,
"step": 810
},
{
"epoch": 1.2476923076923077,
"grad_norm": 0.42893862551923334,
"learning_rate": 1.5580793663591583e-05,
"loss": 0.0669,
"step": 811
},
{
"epoch": 1.2492307692307691,
"grad_norm": 0.37930257242842513,
"learning_rate": 1.557076322838922e-05,
"loss": 0.0699,
"step": 812
},
{
"epoch": 1.2507692307692309,
"grad_norm": 0.33145455304026866,
"learning_rate": 1.5560724659869905e-05,
"loss": 0.0681,
"step": 813
},
{
"epoch": 1.2523076923076923,
"grad_norm": 0.4059667858998905,
"learning_rate": 1.555067797268995e-05,
"loss": 0.068,
"step": 814
},
{
"epoch": 1.2538461538461538,
"grad_norm": 0.3467956370682722,
"learning_rate": 1.5540623181517532e-05,
"loss": 0.0659,
"step": 815
},
{
"epoch": 1.2553846153846153,
"grad_norm": 0.5088686209429053,
"learning_rate": 1.5530560301032644e-05,
"loss": 0.0795,
"step": 816
},
{
"epoch": 1.2569230769230768,
"grad_norm": 0.3799873270111144,
"learning_rate": 1.5520489345927095e-05,
"loss": 0.0691,
"step": 817
},
{
"epoch": 1.2584615384615385,
"grad_norm": 0.41391636034513485,
"learning_rate": 1.551041033090449e-05,
"loss": 0.0721,
"step": 818
},
{
"epoch": 1.26,
"grad_norm": 0.40772591130685176,
"learning_rate": 1.5500323270680194e-05,
"loss": 0.0681,
"step": 819
},
{
"epoch": 1.2615384615384615,
"grad_norm": 0.3349423110799659,
"learning_rate": 1.549022817998132e-05,
"loss": 0.0623,
"step": 820
},
{
"epoch": 1.2630769230769232,
"grad_norm": 0.5203080319006015,
"learning_rate": 1.5480125073546705e-05,
"loss": 0.0659,
"step": 821
},
{
"epoch": 1.2646153846153847,
"grad_norm": 0.42351756936892165,
"learning_rate": 1.5470013966126886e-05,
"loss": 0.0702,
"step": 822
},
{
"epoch": 1.2661538461538462,
"grad_norm": 0.3483945108051493,
"learning_rate": 1.5459894872484083e-05,
"loss": 0.0636,
"step": 823
},
{
"epoch": 1.2676923076923077,
"grad_norm": 0.37839615755712913,
"learning_rate": 1.5449767807392184e-05,
"loss": 0.0676,
"step": 824
},
{
"epoch": 1.2692307692307692,
"grad_norm": 0.4465419577943629,
"learning_rate": 1.5439632785636707e-05,
"loss": 0.0631,
"step": 825
},
{
"epoch": 1.2707692307692309,
"grad_norm": 0.39462290024075386,
"learning_rate": 1.542948982201479e-05,
"loss": 0.0634,
"step": 826
},
{
"epoch": 1.2723076923076924,
"grad_norm": 0.3797179172041511,
"learning_rate": 1.5419338931335155e-05,
"loss": 0.0699,
"step": 827
},
{
"epoch": 1.2738461538461539,
"grad_norm": 0.3416577637686262,
"learning_rate": 1.5409180128418123e-05,
"loss": 0.0636,
"step": 828
},
{
"epoch": 1.2753846153846153,
"grad_norm": 0.4613004722112054,
"learning_rate": 1.539901342809554e-05,
"loss": 0.0684,
"step": 829
},
{
"epoch": 1.2769230769230768,
"grad_norm": 0.42702108586314186,
"learning_rate": 1.5388838845210798e-05,
"loss": 0.0792,
"step": 830
},
{
"epoch": 1.2784615384615385,
"grad_norm": 0.3530542594699597,
"learning_rate": 1.5378656394618788e-05,
"loss": 0.0647,
"step": 831
},
{
"epoch": 1.28,
"grad_norm": 0.43241017204879106,
"learning_rate": 1.5368466091185893e-05,
"loss": 0.0781,
"step": 832
},
{
"epoch": 1.2815384615384615,
"grad_norm": 0.5132567987415939,
"learning_rate": 1.5358267949789968e-05,
"loss": 0.0728,
"step": 833
},
{
"epoch": 1.283076923076923,
"grad_norm": 0.36683064622627476,
"learning_rate": 1.5348061985320298e-05,
"loss": 0.0725,
"step": 834
},
{
"epoch": 1.2846153846153845,
"grad_norm": 0.33492481470642155,
"learning_rate": 1.53378482126776e-05,
"loss": 0.0644,
"step": 835
},
{
"epoch": 1.2861538461538462,
"grad_norm": 0.3210179795426996,
"learning_rate": 1.5327626646773975e-05,
"loss": 0.0694,
"step": 836
},
{
"epoch": 1.2876923076923077,
"grad_norm": 0.3454605703445639,
"learning_rate": 1.5317397302532933e-05,
"loss": 0.063,
"step": 837
},
{
"epoch": 1.2892307692307692,
"grad_norm": 0.36111805478759756,
"learning_rate": 1.530716019488931e-05,
"loss": 0.0729,
"step": 838
},
{
"epoch": 1.290769230769231,
"grad_norm": 0.4150741361759226,
"learning_rate": 1.529691533878929e-05,
"loss": 0.0692,
"step": 839
},
{
"epoch": 1.2923076923076924,
"grad_norm": 0.44151832030522525,
"learning_rate": 1.528666274919037e-05,
"loss": 0.0668,
"step": 840
},
{
"epoch": 1.2938461538461539,
"grad_norm": 0.39281808968582055,
"learning_rate": 1.527640244106133e-05,
"loss": 0.0614,
"step": 841
},
{
"epoch": 1.2953846153846154,
"grad_norm": 0.3355371699856118,
"learning_rate": 1.526613442938223e-05,
"loss": 0.066,
"step": 842
},
{
"epoch": 1.2969230769230768,
"grad_norm": 0.3343752965517015,
"learning_rate": 1.5255858729144368e-05,
"loss": 0.0618,
"step": 843
},
{
"epoch": 1.2984615384615386,
"grad_norm": 0.4219851770072625,
"learning_rate": 1.5245575355350273e-05,
"loss": 0.0706,
"step": 844
},
{
"epoch": 1.3,
"grad_norm": 0.47236267011925603,
"learning_rate": 1.5235284323013674e-05,
"loss": 0.0768,
"step": 845
},
{
"epoch": 1.3015384615384615,
"grad_norm": 0.3160781099772521,
"learning_rate": 1.5224985647159489e-05,
"loss": 0.0677,
"step": 846
},
{
"epoch": 1.303076923076923,
"grad_norm": 0.4211840324694599,
"learning_rate": 1.5214679342823786e-05,
"loss": 0.0703,
"step": 847
},
{
"epoch": 1.3046153846153845,
"grad_norm": 0.4382171224080807,
"learning_rate": 1.5204365425053773e-05,
"loss": 0.0728,
"step": 848
},
{
"epoch": 1.3061538461538462,
"grad_norm": 0.3730723749029246,
"learning_rate": 1.5194043908907774e-05,
"loss": 0.0644,
"step": 849
},
{
"epoch": 1.3076923076923077,
"grad_norm": 0.43669340037201293,
"learning_rate": 1.518371480945521e-05,
"loss": 0.0725,
"step": 850
},
{
"epoch": 1.3092307692307692,
"grad_norm": 0.29956287148615507,
"learning_rate": 1.5173378141776569e-05,
"loss": 0.0595,
"step": 851
},
{
"epoch": 1.3107692307692307,
"grad_norm": 0.4010201519686623,
"learning_rate": 1.5163033920963393e-05,
"loss": 0.0715,
"step": 852
},
{
"epoch": 1.3123076923076922,
"grad_norm": 0.3284401287146176,
"learning_rate": 1.515268216211825e-05,
"loss": 0.0649,
"step": 853
},
{
"epoch": 1.3138461538461539,
"grad_norm": 0.36834718137576217,
"learning_rate": 1.5142322880354706e-05,
"loss": 0.0761,
"step": 854
},
{
"epoch": 1.3153846153846154,
"grad_norm": 0.4038185340311018,
"learning_rate": 1.5131956090797326e-05,
"loss": 0.0713,
"step": 855
},
{
"epoch": 1.3169230769230769,
"grad_norm": 0.3485378538805922,
"learning_rate": 1.5121581808581623e-05,
"loss": 0.0721,
"step": 856
},
{
"epoch": 1.3184615384615386,
"grad_norm": 0.324099392540781,
"learning_rate": 1.5111200048854055e-05,
"loss": 0.0704,
"step": 857
},
{
"epoch": 1.32,
"grad_norm": 0.3958228511004757,
"learning_rate": 1.5100810826771997e-05,
"loss": 0.0665,
"step": 858
},
{
"epoch": 1.3215384615384616,
"grad_norm": 0.4382455185601089,
"learning_rate": 1.5090414157503715e-05,
"loss": 0.076,
"step": 859
},
{
"epoch": 1.323076923076923,
"grad_norm": 0.3260556756343216,
"learning_rate": 1.5080010056228353e-05,
"loss": 0.0671,
"step": 860
},
{
"epoch": 1.3246153846153845,
"grad_norm": 0.3038142480938741,
"learning_rate": 1.5069598538135905e-05,
"loss": 0.0681,
"step": 861
},
{
"epoch": 1.3261538461538462,
"grad_norm": 0.3817393937929787,
"learning_rate": 1.505917961842719e-05,
"loss": 0.07,
"step": 862
},
{
"epoch": 1.3276923076923077,
"grad_norm": 0.43653178960329847,
"learning_rate": 1.504875331231384e-05,
"loss": 0.0682,
"step": 863
},
{
"epoch": 1.3292307692307692,
"grad_norm": 0.4191601359367802,
"learning_rate": 1.5038319635018264e-05,
"loss": 0.0709,
"step": 864
},
{
"epoch": 1.3307692307692307,
"grad_norm": 0.40482174127743453,
"learning_rate": 1.5027878601773633e-05,
"loss": 0.0681,
"step": 865
},
{
"epoch": 1.3323076923076922,
"grad_norm": 0.39987356115193523,
"learning_rate": 1.5017430227823867e-05,
"loss": 0.0653,
"step": 866
},
{
"epoch": 1.333846153846154,
"grad_norm": 0.36706667375847013,
"learning_rate": 1.5006974528423585e-05,
"loss": 0.0631,
"step": 867
},
{
"epoch": 1.3353846153846154,
"grad_norm": 0.3710865070928732,
"learning_rate": 1.4996511518838129e-05,
"loss": 0.0608,
"step": 868
},
{
"epoch": 1.3369230769230769,
"grad_norm": 0.5238263977692204,
"learning_rate": 1.4986041214343487e-05,
"loss": 0.0739,
"step": 869
},
{
"epoch": 1.3384615384615386,
"grad_norm": 0.38666104678955204,
"learning_rate": 1.4975563630226311e-05,
"loss": 0.0515,
"step": 870
},
{
"epoch": 1.34,
"grad_norm": 0.31886781420723,
"learning_rate": 1.4965078781783882e-05,
"loss": 0.0649,
"step": 871
},
{
"epoch": 1.3415384615384616,
"grad_norm": 0.32582039756760706,
"learning_rate": 1.4954586684324077e-05,
"loss": 0.0682,
"step": 872
},
{
"epoch": 1.343076923076923,
"grad_norm": 0.4323139162984884,
"learning_rate": 1.494408735316537e-05,
"loss": 0.0722,
"step": 873
},
{
"epoch": 1.3446153846153845,
"grad_norm": 0.4522628240441855,
"learning_rate": 1.4933580803636787e-05,
"loss": 0.0641,
"step": 874
},
{
"epoch": 1.3461538461538463,
"grad_norm": 0.38114193229980947,
"learning_rate": 1.4923067051077893e-05,
"loss": 0.0724,
"step": 875
},
{
"epoch": 1.3476923076923077,
"grad_norm": 0.4247141871033444,
"learning_rate": 1.4912546110838775e-05,
"loss": 0.0713,
"step": 876
},
{
"epoch": 1.3492307692307692,
"grad_norm": 0.4058780791287694,
"learning_rate": 1.490201799828001e-05,
"loss": 0.0741,
"step": 877
},
{
"epoch": 1.3507692307692307,
"grad_norm": 0.43398667126598395,
"learning_rate": 1.4891482728772645e-05,
"loss": 0.0694,
"step": 878
},
{
"epoch": 1.3523076923076922,
"grad_norm": 0.4177065232740013,
"learning_rate": 1.4880940317698182e-05,
"loss": 0.0731,
"step": 879
},
{
"epoch": 1.353846153846154,
"grad_norm": 0.3284355368343791,
"learning_rate": 1.4870390780448545e-05,
"loss": 0.0747,
"step": 880
},
{
"epoch": 1.3553846153846154,
"grad_norm": 0.44954781903955254,
"learning_rate": 1.485983413242606e-05,
"loss": 0.0628,
"step": 881
},
{
"epoch": 1.356923076923077,
"grad_norm": 0.5129727422741225,
"learning_rate": 1.4849270389043444e-05,
"loss": 0.0687,
"step": 882
},
{
"epoch": 1.3584615384615384,
"grad_norm": 0.4467443171166444,
"learning_rate": 1.4838699565723764e-05,
"loss": 0.0652,
"step": 883
},
{
"epoch": 1.3599999999999999,
"grad_norm": 0.4496929006760978,
"learning_rate": 1.4828121677900427e-05,
"loss": 0.0744,
"step": 884
},
{
"epoch": 1.3615384615384616,
"grad_norm": 0.3414293713875311,
"learning_rate": 1.4817536741017153e-05,
"loss": 0.0728,
"step": 885
},
{
"epoch": 1.363076923076923,
"grad_norm": 0.316120367911126,
"learning_rate": 1.4806944770527958e-05,
"loss": 0.0666,
"step": 886
},
{
"epoch": 1.3646153846153846,
"grad_norm": 0.40460300362699714,
"learning_rate": 1.479634578189712e-05,
"loss": 0.0727,
"step": 887
},
{
"epoch": 1.3661538461538463,
"grad_norm": 0.45659447717946505,
"learning_rate": 1.4785739790599174e-05,
"loss": 0.0782,
"step": 888
},
{
"epoch": 1.3676923076923078,
"grad_norm": 0.40368455430062866,
"learning_rate": 1.4775126812118865e-05,
"loss": 0.0704,
"step": 889
},
{
"epoch": 1.3692307692307693,
"grad_norm": 0.35547903664705943,
"learning_rate": 1.4764506861951151e-05,
"loss": 0.0774,
"step": 890
},
{
"epoch": 1.3707692307692307,
"grad_norm": 0.3259048929714102,
"learning_rate": 1.4753879955601162e-05,
"loss": 0.0568,
"step": 891
},
{
"epoch": 1.3723076923076922,
"grad_norm": 0.33398368713521703,
"learning_rate": 1.474324610858419e-05,
"loss": 0.0696,
"step": 892
},
{
"epoch": 1.373846153846154,
"grad_norm": 0.3975414141870634,
"learning_rate": 1.4732605336425651e-05,
"loss": 0.0653,
"step": 893
},
{
"epoch": 1.3753846153846154,
"grad_norm": 0.3695667545172178,
"learning_rate": 1.472195765466108e-05,
"loss": 0.0657,
"step": 894
},
{
"epoch": 1.376923076923077,
"grad_norm": 0.3070295149212851,
"learning_rate": 1.4711303078836098e-05,
"loss": 0.0607,
"step": 895
},
{
"epoch": 1.3784615384615384,
"grad_norm": 0.43373009062750495,
"learning_rate": 1.4700641624506392e-05,
"loss": 0.077,
"step": 896
},
{
"epoch": 1.38,
"grad_norm": 0.3207986109640768,
"learning_rate": 1.4689973307237687e-05,
"loss": 0.0663,
"step": 897
},
{
"epoch": 1.3815384615384616,
"grad_norm": 0.33029490975452497,
"learning_rate": 1.4679298142605735e-05,
"loss": 0.0661,
"step": 898
},
{
"epoch": 1.383076923076923,
"grad_norm": 0.39797201272606875,
"learning_rate": 1.466861614619628e-05,
"loss": 0.0683,
"step": 899
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.3569032293188448,
"learning_rate": 1.465792733360504e-05,
"loss": 0.0712,
"step": 900
},
{
"epoch": 1.3861538461538463,
"grad_norm": 0.33179043133017183,
"learning_rate": 1.4647231720437687e-05,
"loss": 0.076,
"step": 901
},
{
"epoch": 1.3876923076923076,
"grad_norm": 0.3185134355526598,
"learning_rate": 1.4636529322309825e-05,
"loss": 0.0716,
"step": 902
},
{
"epoch": 1.3892307692307693,
"grad_norm": 0.38927875322344885,
"learning_rate": 1.4625820154846953e-05,
"loss": 0.0607,
"step": 903
},
{
"epoch": 1.3907692307692308,
"grad_norm": 0.3592258244516518,
"learning_rate": 1.4615104233684467e-05,
"loss": 0.0636,
"step": 904
},
{
"epoch": 1.3923076923076922,
"grad_norm": 0.440027348689472,
"learning_rate": 1.4604381574467616e-05,
"loss": 0.0812,
"step": 905
},
{
"epoch": 1.393846153846154,
"grad_norm": 0.31797170088925414,
"learning_rate": 1.4593652192851487e-05,
"loss": 0.0624,
"step": 906
},
{
"epoch": 1.3953846153846154,
"grad_norm": 0.44417520359272944,
"learning_rate": 1.4582916104500977e-05,
"loss": 0.0734,
"step": 907
},
{
"epoch": 1.396923076923077,
"grad_norm": 0.34291352445561796,
"learning_rate": 1.457217332509079e-05,
"loss": 0.0684,
"step": 908
},
{
"epoch": 1.3984615384615384,
"grad_norm": 0.3078336428229946,
"learning_rate": 1.4561423870305383e-05,
"loss": 0.0594,
"step": 909
},
{
"epoch": 1.4,
"grad_norm": 0.45773936133687876,
"learning_rate": 1.4550667755838965e-05,
"loss": 0.0735,
"step": 910
},
{
"epoch": 1.4015384615384616,
"grad_norm": 0.3093789239621972,
"learning_rate": 1.4539904997395468e-05,
"loss": 0.0665,
"step": 911
},
{
"epoch": 1.403076923076923,
"grad_norm": 0.33556510387782423,
"learning_rate": 1.4529135610688529e-05,
"loss": 0.067,
"step": 912
},
{
"epoch": 1.4046153846153846,
"grad_norm": 0.3821254219317531,
"learning_rate": 1.4518359611441452e-05,
"loss": 0.0795,
"step": 913
},
{
"epoch": 1.406153846153846,
"grad_norm": 0.3968625825470183,
"learning_rate": 1.4507577015387204e-05,
"loss": 0.0694,
"step": 914
},
{
"epoch": 1.4076923076923076,
"grad_norm": 0.40920318095123365,
"learning_rate": 1.4496787838268378e-05,
"loss": 0.0751,
"step": 915
},
{
"epoch": 1.4092307692307693,
"grad_norm": 0.333518429842856,
"learning_rate": 1.4485992095837178e-05,
"loss": 0.0678,
"step": 916
},
{
"epoch": 1.4107692307692308,
"grad_norm": 0.4180262330660976,
"learning_rate": 1.4475189803855399e-05,
"loss": 0.0669,
"step": 917
},
{
"epoch": 1.4123076923076923,
"grad_norm": 0.3270671609911376,
"learning_rate": 1.4464380978094386e-05,
"loss": 0.0628,
"step": 918
},
{
"epoch": 1.413846153846154,
"grad_norm": 0.3321568177478066,
"learning_rate": 1.445356563433503e-05,
"loss": 0.0675,
"step": 919
},
{
"epoch": 1.4153846153846155,
"grad_norm": 0.4278414076070176,
"learning_rate": 1.4442743788367741e-05,
"loss": 0.0756,
"step": 920
},
{
"epoch": 1.416923076923077,
"grad_norm": 0.3573839526102382,
"learning_rate": 1.4431915455992416e-05,
"loss": 0.0743,
"step": 921
},
{
"epoch": 1.4184615384615384,
"grad_norm": 0.3965761887911091,
"learning_rate": 1.4421080653018426e-05,
"loss": 0.0686,
"step": 922
},
{
"epoch": 1.42,
"grad_norm": 0.3277382293498253,
"learning_rate": 1.4410239395264594e-05,
"loss": 0.0593,
"step": 923
},
{
"epoch": 1.4215384615384616,
"grad_norm": 0.413916370460171,
"learning_rate": 1.4399391698559153e-05,
"loss": 0.057,
"step": 924
},
{
"epoch": 1.4230769230769231,
"grad_norm": 0.40814754638707723,
"learning_rate": 1.438853757873975e-05,
"loss": 0.0673,
"step": 925
},
{
"epoch": 1.4246153846153846,
"grad_norm": 0.37821667270068476,
"learning_rate": 1.4377677051653404e-05,
"loss": 0.0703,
"step": 926
},
{
"epoch": 1.426153846153846,
"grad_norm": 0.5102274833966777,
"learning_rate": 1.4366810133156495e-05,
"loss": 0.076,
"step": 927
},
{
"epoch": 1.4276923076923076,
"grad_norm": 0.44547602217095333,
"learning_rate": 1.4355936839114718e-05,
"loss": 0.0747,
"step": 928
},
{
"epoch": 1.4292307692307693,
"grad_norm": 0.3911040675941954,
"learning_rate": 1.43450571854031e-05,
"loss": 0.0715,
"step": 929
},
{
"epoch": 1.4307692307692308,
"grad_norm": 0.3796479456399161,
"learning_rate": 1.4334171187905928e-05,
"loss": 0.0732,
"step": 930
},
{
"epoch": 1.4323076923076923,
"grad_norm": 0.33940059068204437,
"learning_rate": 1.4323278862516774e-05,
"loss": 0.0702,
"step": 931
},
{
"epoch": 1.4338461538461538,
"grad_norm": 0.3509046699774803,
"learning_rate": 1.431238022513843e-05,
"loss": 0.0657,
"step": 932
},
{
"epoch": 1.4353846153846153,
"grad_norm": 0.36668044712623865,
"learning_rate": 1.430147529168292e-05,
"loss": 0.0625,
"step": 933
},
{
"epoch": 1.436923076923077,
"grad_norm": 0.4281432928035457,
"learning_rate": 1.4290564078071445e-05,
"loss": 0.0813,
"step": 934
},
{
"epoch": 1.4384615384615385,
"grad_norm": 0.33710258376419955,
"learning_rate": 1.4279646600234388e-05,
"loss": 0.068,
"step": 935
},
{
"epoch": 1.44,
"grad_norm": 0.32354838701141914,
"learning_rate": 1.4268722874111265e-05,
"loss": 0.0648,
"step": 936
},
{
"epoch": 1.4415384615384617,
"grad_norm": 0.3021595123301058,
"learning_rate": 1.4257792915650728e-05,
"loss": 0.0653,
"step": 937
},
{
"epoch": 1.4430769230769231,
"grad_norm": 0.44802198690971745,
"learning_rate": 1.4246856740810517e-05,
"loss": 0.0768,
"step": 938
},
{
"epoch": 1.4446153846153846,
"grad_norm": 0.3593940216420963,
"learning_rate": 1.4235914365557455e-05,
"loss": 0.0638,
"step": 939
},
{
"epoch": 1.4461538461538461,
"grad_norm": 0.34159209475332464,
"learning_rate": 1.4224965805867413e-05,
"loss": 0.0643,
"step": 940
},
{
"epoch": 1.4476923076923076,
"grad_norm": 0.35984267550149907,
"learning_rate": 1.4214011077725293e-05,
"loss": 0.0723,
"step": 941
},
{
"epoch": 1.4492307692307693,
"grad_norm": 0.39317799185440994,
"learning_rate": 1.4203050197125005e-05,
"loss": 0.0643,
"step": 942
},
{
"epoch": 1.4507692307692308,
"grad_norm": 0.336857468826351,
"learning_rate": 1.4192083180069441e-05,
"loss": 0.0618,
"step": 943
},
{
"epoch": 1.4523076923076923,
"grad_norm": 0.35186803870389044,
"learning_rate": 1.4181110042570447e-05,
"loss": 0.0683,
"step": 944
},
{
"epoch": 1.4538461538461538,
"grad_norm": 0.29942739165233795,
"learning_rate": 1.4170130800648814e-05,
"loss": 0.0652,
"step": 945
},
{
"epoch": 1.4553846153846153,
"grad_norm": 0.34891798063795276,
"learning_rate": 1.4159145470334237e-05,
"loss": 0.0696,
"step": 946
},
{
"epoch": 1.456923076923077,
"grad_norm": 0.4077733821140671,
"learning_rate": 1.4148154067665305e-05,
"loss": 0.0699,
"step": 947
},
{
"epoch": 1.4584615384615385,
"grad_norm": 0.40284125050375413,
"learning_rate": 1.4137156608689469e-05,
"loss": 0.0742,
"step": 948
},
{
"epoch": 1.46,
"grad_norm": 0.42566126627046824,
"learning_rate": 1.4126153109463025e-05,
"loss": 0.0642,
"step": 949
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.32133228937421,
"learning_rate": 1.411514358605109e-05,
"loss": 0.068,
"step": 950
},
{
"epoch": 1.463076923076923,
"grad_norm": 0.36895430822571307,
"learning_rate": 1.410412805452757e-05,
"loss": 0.0667,
"step": 951
},
{
"epoch": 1.4646153846153847,
"grad_norm": 0.3166854845654635,
"learning_rate": 1.4093106530975146e-05,
"loss": 0.061,
"step": 952
},
{
"epoch": 1.4661538461538461,
"grad_norm": 0.5003043211716166,
"learning_rate": 1.4082079031485253e-05,
"loss": 0.0656,
"step": 953
},
{
"epoch": 1.4676923076923076,
"grad_norm": 0.39875601093346025,
"learning_rate": 1.4071045572158038e-05,
"loss": 0.0681,
"step": 954
},
{
"epoch": 1.4692307692307693,
"grad_norm": 0.3993223750486499,
"learning_rate": 1.4060006169102363e-05,
"loss": 0.0692,
"step": 955
},
{
"epoch": 1.4707692307692308,
"grad_norm": 0.3477397845990775,
"learning_rate": 1.4048960838435755e-05,
"loss": 0.0702,
"step": 956
},
{
"epoch": 1.4723076923076923,
"grad_norm": 0.4126048706500608,
"learning_rate": 1.403790959628441e-05,
"loss": 0.0703,
"step": 957
},
{
"epoch": 1.4738461538461538,
"grad_norm": 0.3012393164802371,
"learning_rate": 1.4026852458783141e-05,
"loss": 0.0631,
"step": 958
},
{
"epoch": 1.4753846153846153,
"grad_norm": 0.41825312268062614,
"learning_rate": 1.4015789442075376e-05,
"loss": 0.0639,
"step": 959
},
{
"epoch": 1.476923076923077,
"grad_norm": 0.35965621239571177,
"learning_rate": 1.4004720562313125e-05,
"loss": 0.0631,
"step": 960
},
{
"epoch": 1.4784615384615385,
"grad_norm": 0.329531529153342,
"learning_rate": 1.3993645835656955e-05,
"loss": 0.0794,
"step": 961
},
{
"epoch": 1.48,
"grad_norm": 0.34849163011002293,
"learning_rate": 1.3982565278275976e-05,
"loss": 0.067,
"step": 962
},
{
"epoch": 1.4815384615384615,
"grad_norm": 0.3857507607852618,
"learning_rate": 1.3971478906347806e-05,
"loss": 0.0786,
"step": 963
},
{
"epoch": 1.483076923076923,
"grad_norm": 0.3354179087450714,
"learning_rate": 1.3960386736058552e-05,
"loss": 0.0653,
"step": 964
},
{
"epoch": 1.4846153846153847,
"grad_norm": 0.4517425643119976,
"learning_rate": 1.394928878360279e-05,
"loss": 0.0696,
"step": 965
},
{
"epoch": 1.4861538461538462,
"grad_norm": 0.3726999175701167,
"learning_rate": 1.3938185065183534e-05,
"loss": 0.0703,
"step": 966
},
{
"epoch": 1.4876923076923076,
"grad_norm": 0.32753752193184116,
"learning_rate": 1.3927075597012215e-05,
"loss": 0.0688,
"step": 967
},
{
"epoch": 1.4892307692307694,
"grad_norm": 0.3625776900702585,
"learning_rate": 1.391596039530867e-05,
"loss": 0.0699,
"step": 968
},
{
"epoch": 1.4907692307692308,
"grad_norm": 0.4221103719027582,
"learning_rate": 1.3904839476301091e-05,
"loss": 0.0745,
"step": 969
},
{
"epoch": 1.4923076923076923,
"grad_norm": 0.3300247767704618,
"learning_rate": 1.3893712856226028e-05,
"loss": 0.063,
"step": 970
},
{
"epoch": 1.4938461538461538,
"grad_norm": 0.35465084506366884,
"learning_rate": 1.388258055132835e-05,
"loss": 0.0628,
"step": 971
},
{
"epoch": 1.4953846153846153,
"grad_norm": 0.3956402484871412,
"learning_rate": 1.3871442577861234e-05,
"loss": 0.0718,
"step": 972
},
{
"epoch": 1.496923076923077,
"grad_norm": 0.32259127869403026,
"learning_rate": 1.3860298952086118e-05,
"loss": 0.0673,
"step": 973
},
{
"epoch": 1.4984615384615385,
"grad_norm": 0.30587983895963383,
"learning_rate": 1.3849149690272704e-05,
"loss": 0.0642,
"step": 974
},
{
"epoch": 1.5,
"grad_norm": 0.3268162425630392,
"learning_rate": 1.383799480869892e-05,
"loss": 0.0655,
"step": 975
},
{
"epoch": 1.5015384615384615,
"grad_norm": 0.35344360561508237,
"learning_rate": 1.3826834323650899e-05,
"loss": 0.0717,
"step": 976
},
{
"epoch": 1.503076923076923,
"grad_norm": 0.4240517174731372,
"learning_rate": 1.3815668251422953e-05,
"loss": 0.0767,
"step": 977
},
{
"epoch": 1.5046153846153847,
"grad_norm": 0.3890020659351386,
"learning_rate": 1.3804496608317557e-05,
"loss": 0.0707,
"step": 978
},
{
"epoch": 1.5061538461538462,
"grad_norm": 0.34522079737200434,
"learning_rate": 1.3793319410645307e-05,
"loss": 0.0704,
"step": 979
},
{
"epoch": 1.5076923076923077,
"grad_norm": 0.35470349339672885,
"learning_rate": 1.3782136674724924e-05,
"loss": 0.0681,
"step": 980
},
{
"epoch": 1.5092307692307694,
"grad_norm": 0.37327575501099824,
"learning_rate": 1.3770948416883205e-05,
"loss": 0.0609,
"step": 981
},
{
"epoch": 1.5107692307692306,
"grad_norm": 0.3727506860019033,
"learning_rate": 1.3759754653455013e-05,
"loss": 0.0624,
"step": 982
},
{
"epoch": 1.5123076923076924,
"grad_norm": 0.3169662482724166,
"learning_rate": 1.3748555400783245e-05,
"loss": 0.0679,
"step": 983
},
{
"epoch": 1.5138461538461538,
"grad_norm": 0.318203785112587,
"learning_rate": 1.3737350675218819e-05,
"loss": 0.0721,
"step": 984
},
{
"epoch": 1.5153846153846153,
"grad_norm": 0.3459767757333529,
"learning_rate": 1.3726140493120639e-05,
"loss": 0.0692,
"step": 985
},
{
"epoch": 1.516923076923077,
"grad_norm": 0.35213169704344977,
"learning_rate": 1.3714924870855573e-05,
"loss": 0.059,
"step": 986
},
{
"epoch": 1.5184615384615383,
"grad_norm": 0.444276142733395,
"learning_rate": 1.3703703824798438e-05,
"loss": 0.0684,
"step": 987
},
{
"epoch": 1.52,
"grad_norm": 0.35207361668294573,
"learning_rate": 1.3692477371331965e-05,
"loss": 0.0744,
"step": 988
},
{
"epoch": 1.5215384615384615,
"grad_norm": 0.35999767828776713,
"learning_rate": 1.3681245526846782e-05,
"loss": 0.0713,
"step": 989
},
{
"epoch": 1.523076923076923,
"grad_norm": 0.39013944310312315,
"learning_rate": 1.3670008307741388e-05,
"loss": 0.0725,
"step": 990
},
{
"epoch": 1.5246153846153847,
"grad_norm": 0.38236299835311816,
"learning_rate": 1.3658765730422126e-05,
"loss": 0.0679,
"step": 991
},
{
"epoch": 1.5261538461538462,
"grad_norm": 0.3391508447409303,
"learning_rate": 1.3647517811303164e-05,
"loss": 0.0674,
"step": 992
},
{
"epoch": 1.5276923076923077,
"grad_norm": 0.6412950373687982,
"learning_rate": 1.3636264566806473e-05,
"loss": 0.0741,
"step": 993
},
{
"epoch": 1.5292307692307694,
"grad_norm": 0.3350577163257067,
"learning_rate": 1.362500601336179e-05,
"loss": 0.0614,
"step": 994
},
{
"epoch": 1.5307692307692307,
"grad_norm": 0.3626403986182635,
"learning_rate": 1.3613742167406614e-05,
"loss": 0.0691,
"step": 995
},
{
"epoch": 1.5323076923076924,
"grad_norm": 0.3470732304493993,
"learning_rate": 1.3602473045386165e-05,
"loss": 0.0742,
"step": 996
},
{
"epoch": 1.5338461538461539,
"grad_norm": 0.33260189078997204,
"learning_rate": 1.3591198663753358e-05,
"loss": 0.0686,
"step": 997
},
{
"epoch": 1.5353846153846153,
"grad_norm": 0.28916096300010014,
"learning_rate": 1.3579919038968805e-05,
"loss": 0.0638,
"step": 998
},
{
"epoch": 1.536923076923077,
"grad_norm": 0.43372086950966615,
"learning_rate": 1.3568634187500762e-05,
"loss": 0.07,
"step": 999
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.40435446045850537,
"learning_rate": 1.3557344125825113e-05,
"loss": 0.0686,
"step": 1000
},
{
"epoch": 1.54,
"grad_norm": 0.3600272169695901,
"learning_rate": 1.3546048870425356e-05,
"loss": 0.075,
"step": 1001
},
{
"epoch": 1.5415384615384615,
"grad_norm": 0.31816083037803755,
"learning_rate": 1.3534748437792573e-05,
"loss": 0.0737,
"step": 1002
},
{
"epoch": 1.543076923076923,
"grad_norm": 0.33273148854011647,
"learning_rate": 1.3523442844425393e-05,
"loss": 0.0665,
"step": 1003
},
{
"epoch": 1.5446153846153847,
"grad_norm": 0.37169937500283534,
"learning_rate": 1.3512132106829996e-05,
"loss": 0.0654,
"step": 1004
},
{
"epoch": 1.546153846153846,
"grad_norm": 0.329535717539556,
"learning_rate": 1.3500816241520059e-05,
"loss": 0.0648,
"step": 1005
},
{
"epoch": 1.5476923076923077,
"grad_norm": 0.25023308868192906,
"learning_rate": 1.3489495265016753e-05,
"loss": 0.0558,
"step": 1006
},
{
"epoch": 1.5492307692307692,
"grad_norm": 0.3555280320936877,
"learning_rate": 1.3478169193848705e-05,
"loss": 0.0695,
"step": 1007
},
{
"epoch": 1.5507692307692307,
"grad_norm": 0.3533955661877233,
"learning_rate": 1.346683804455199e-05,
"loss": 0.0651,
"step": 1008
},
{
"epoch": 1.5523076923076924,
"grad_norm": 0.3802507308411157,
"learning_rate": 1.3455501833670089e-05,
"loss": 0.0737,
"step": 1009
},
{
"epoch": 1.5538461538461539,
"grad_norm": 0.3088088179326214,
"learning_rate": 1.3444160577753872e-05,
"loss": 0.0681,
"step": 1010
},
{
"epoch": 1.5553846153846154,
"grad_norm": 0.2929009087213979,
"learning_rate": 1.3432814293361585e-05,
"loss": 0.062,
"step": 1011
},
{
"epoch": 1.556923076923077,
"grad_norm": 0.37467092752746656,
"learning_rate": 1.34214629970588e-05,
"loss": 0.0709,
"step": 1012
},
{
"epoch": 1.5584615384615383,
"grad_norm": 0.33157137643925366,
"learning_rate": 1.3410106705418424e-05,
"loss": 0.0655,
"step": 1013
},
{
"epoch": 1.56,
"grad_norm": 0.36425079800863963,
"learning_rate": 1.3398745435020642e-05,
"loss": 0.0717,
"step": 1014
},
{
"epoch": 1.5615384615384615,
"grad_norm": 0.3794215971913442,
"learning_rate": 1.3387379202452917e-05,
"loss": 0.07,
"step": 1015
},
{
"epoch": 1.563076923076923,
"grad_norm": 0.35292364976240903,
"learning_rate": 1.337600802430995e-05,
"loss": 0.0737,
"step": 1016
},
{
"epoch": 1.5646153846153847,
"grad_norm": 0.36309299670186634,
"learning_rate": 1.3364631917193671e-05,
"loss": 0.0678,
"step": 1017
},
{
"epoch": 1.566153846153846,
"grad_norm": 0.43174462874417424,
"learning_rate": 1.33532508977132e-05,
"loss": 0.0664,
"step": 1018
},
{
"epoch": 1.5676923076923077,
"grad_norm": 0.2980059236096669,
"learning_rate": 1.3341864982484828e-05,
"loss": 0.0592,
"step": 1019
},
{
"epoch": 1.5692307692307692,
"grad_norm": 0.3505262612950089,
"learning_rate": 1.3330474188132004e-05,
"loss": 0.0735,
"step": 1020
},
{
"epoch": 1.5707692307692307,
"grad_norm": 0.3640243702743533,
"learning_rate": 1.3319078531285286e-05,
"loss": 0.0684,
"step": 1021
},
{
"epoch": 1.5723076923076924,
"grad_norm": 0.37347682099255064,
"learning_rate": 1.3307678028582342e-05,
"loss": 0.0631,
"step": 1022
},
{
"epoch": 1.573846153846154,
"grad_norm": 0.30683658621574744,
"learning_rate": 1.329627269666791e-05,
"loss": 0.064,
"step": 1023
},
{
"epoch": 1.5753846153846154,
"grad_norm": 0.38325694612265304,
"learning_rate": 1.328486255219378e-05,
"loss": 0.0748,
"step": 1024
},
{
"epoch": 1.5769230769230769,
"grad_norm": 0.34834422809752846,
"learning_rate": 1.3273447611818768e-05,
"loss": 0.0706,
"step": 1025
},
{
"epoch": 1.5784615384615384,
"grad_norm": 0.33431857255788205,
"learning_rate": 1.3262027892208696e-05,
"loss": 0.0607,
"step": 1026
},
{
"epoch": 1.58,
"grad_norm": 0.36935448940068555,
"learning_rate": 1.3250603410036356e-05,
"loss": 0.0657,
"step": 1027
},
{
"epoch": 1.5815384615384616,
"grad_norm": 0.5029604354345404,
"learning_rate": 1.3239174181981496e-05,
"loss": 0.0709,
"step": 1028
},
{
"epoch": 1.583076923076923,
"grad_norm": 0.34010742851810977,
"learning_rate": 1.3227740224730799e-05,
"loss": 0.0639,
"step": 1029
},
{
"epoch": 1.5846153846153848,
"grad_norm": 0.35326235063290296,
"learning_rate": 1.3216301554977844e-05,
"loss": 0.07,
"step": 1030
},
{
"epoch": 1.586153846153846,
"grad_norm": 0.371088178596748,
"learning_rate": 1.3204858189423097e-05,
"loss": 0.0689,
"step": 1031
},
{
"epoch": 1.5876923076923077,
"grad_norm": 0.3525413313256349,
"learning_rate": 1.3193410144773876e-05,
"loss": 0.0646,
"step": 1032
},
{
"epoch": 1.5892307692307692,
"grad_norm": 0.47439467360396914,
"learning_rate": 1.3181957437744333e-05,
"loss": 0.0665,
"step": 1033
},
{
"epoch": 1.5907692307692307,
"grad_norm": 0.3389914679073154,
"learning_rate": 1.3170500085055424e-05,
"loss": 0.0663,
"step": 1034
},
{
"epoch": 1.5923076923076924,
"grad_norm": 0.34343403261475547,
"learning_rate": 1.3159038103434889e-05,
"loss": 0.0693,
"step": 1035
},
{
"epoch": 1.5938461538461537,
"grad_norm": 0.30611366359410186,
"learning_rate": 1.314757150961723e-05,
"loss": 0.0676,
"step": 1036
},
{
"epoch": 1.5953846153846154,
"grad_norm": 0.44257355506663903,
"learning_rate": 1.3136100320343674e-05,
"loss": 0.0701,
"step": 1037
},
{
"epoch": 1.596923076923077,
"grad_norm": 0.32909363170626144,
"learning_rate": 1.3124624552362166e-05,
"loss": 0.0627,
"step": 1038
},
{
"epoch": 1.5984615384615384,
"grad_norm": 0.38341925530098436,
"learning_rate": 1.3113144222427334e-05,
"loss": 0.0718,
"step": 1039
},
{
"epoch": 1.6,
"grad_norm": 0.3390749420830194,
"learning_rate": 1.3101659347300462e-05,
"loss": 0.0613,
"step": 1040
},
{
"epoch": 1.6015384615384616,
"grad_norm": 0.3462273945212573,
"learning_rate": 1.3090169943749475e-05,
"loss": 0.0735,
"step": 1041
},
{
"epoch": 1.603076923076923,
"grad_norm": 0.3695043310551725,
"learning_rate": 1.3078676028548908e-05,
"loss": 0.0694,
"step": 1042
},
{
"epoch": 1.6046153846153848,
"grad_norm": 0.3530125489309211,
"learning_rate": 1.3067177618479883e-05,
"loss": 0.0631,
"step": 1043
},
{
"epoch": 1.606153846153846,
"grad_norm": 0.32964686006837474,
"learning_rate": 1.305567473033008e-05,
"loss": 0.0679,
"step": 1044
},
{
"epoch": 1.6076923076923078,
"grad_norm": 0.3713682256507068,
"learning_rate": 1.3044167380893726e-05,
"loss": 0.0701,
"step": 1045
},
{
"epoch": 1.6092307692307692,
"grad_norm": 0.3762742954199106,
"learning_rate": 1.3032655586971552e-05,
"loss": 0.0628,
"step": 1046
},
{
"epoch": 1.6107692307692307,
"grad_norm": 0.3265101319762403,
"learning_rate": 1.3021139365370787e-05,
"loss": 0.0656,
"step": 1047
},
{
"epoch": 1.6123076923076924,
"grad_norm": 0.28711481275165773,
"learning_rate": 1.300961873290512e-05,
"loss": 0.068,
"step": 1048
},
{
"epoch": 1.6138461538461537,
"grad_norm": 0.4010476086338245,
"learning_rate": 1.2998093706394676e-05,
"loss": 0.0717,
"step": 1049
},
{
"epoch": 1.6153846153846154,
"grad_norm": 0.3145419470926991,
"learning_rate": 1.2986564302666e-05,
"loss": 0.0694,
"step": 1050
},
{
"epoch": 1.616923076923077,
"grad_norm": 0.3926012570647249,
"learning_rate": 1.297503053855203e-05,
"loss": 0.0713,
"step": 1051
},
{
"epoch": 1.6184615384615384,
"grad_norm": 0.3304899518184055,
"learning_rate": 1.2963492430892066e-05,
"loss": 0.0665,
"step": 1052
},
{
"epoch": 1.62,
"grad_norm": 0.3249058555050507,
"learning_rate": 1.295194999653175e-05,
"loss": 0.0716,
"step": 1053
},
{
"epoch": 1.6215384615384614,
"grad_norm": 0.34563682161792825,
"learning_rate": 1.294040325232304e-05,
"loss": 0.0689,
"step": 1054
},
{
"epoch": 1.623076923076923,
"grad_norm": 0.3623157841595741,
"learning_rate": 1.292885221512419e-05,
"loss": 0.0686,
"step": 1055
},
{
"epoch": 1.6246153846153846,
"grad_norm": 0.4032343951986735,
"learning_rate": 1.291729690179972e-05,
"loss": 0.0709,
"step": 1056
},
{
"epoch": 1.626153846153846,
"grad_norm": 0.42548798300770907,
"learning_rate": 1.2905737329220394e-05,
"loss": 0.0769,
"step": 1057
},
{
"epoch": 1.6276923076923078,
"grad_norm": 0.38063725968362255,
"learning_rate": 1.2894173514263191e-05,
"loss": 0.0718,
"step": 1058
},
{
"epoch": 1.6292307692307693,
"grad_norm": 0.4260166630050922,
"learning_rate": 1.2882605473811282e-05,
"loss": 0.0755,
"step": 1059
},
{
"epoch": 1.6307692307692307,
"grad_norm": 0.3183113830256884,
"learning_rate": 1.2871033224754022e-05,
"loss": 0.0659,
"step": 1060
},
{
"epoch": 1.6323076923076925,
"grad_norm": 0.33423253446578904,
"learning_rate": 1.2859456783986892e-05,
"loss": 0.0675,
"step": 1061
},
{
"epoch": 1.6338461538461537,
"grad_norm": 0.38425643174539403,
"learning_rate": 1.2847876168411506e-05,
"loss": 0.0668,
"step": 1062
},
{
"epoch": 1.6353846153846154,
"grad_norm": 0.39834995039582705,
"learning_rate": 1.2836291394935568e-05,
"loss": 0.0702,
"step": 1063
},
{
"epoch": 1.636923076923077,
"grad_norm": 0.3592062233324518,
"learning_rate": 1.2824702480472846e-05,
"loss": 0.0698,
"step": 1064
},
{
"epoch": 1.6384615384615384,
"grad_norm": 0.30141364654191644,
"learning_rate": 1.2813109441943166e-05,
"loss": 0.0608,
"step": 1065
},
{
"epoch": 1.6400000000000001,
"grad_norm": 0.31650470078418513,
"learning_rate": 1.280151229627237e-05,
"loss": 0.0727,
"step": 1066
},
{
"epoch": 1.6415384615384614,
"grad_norm": 0.3073328562892288,
"learning_rate": 1.2789911060392295e-05,
"loss": 0.0597,
"step": 1067
},
{
"epoch": 1.643076923076923,
"grad_norm": 0.352934869488259,
"learning_rate": 1.2778305751240749e-05,
"loss": 0.0646,
"step": 1068
},
{
"epoch": 1.6446153846153846,
"grad_norm": 0.3477210583426811,
"learning_rate": 1.2766696385761494e-05,
"loss": 0.0694,
"step": 1069
},
{
"epoch": 1.646153846153846,
"grad_norm": 0.31714834966112554,
"learning_rate": 1.2755082980904206e-05,
"loss": 0.0693,
"step": 1070
},
{
"epoch": 1.6476923076923078,
"grad_norm": 0.3328830298996729,
"learning_rate": 1.274346555362446e-05,
"loss": 0.0651,
"step": 1071
},
{
"epoch": 1.6492307692307693,
"grad_norm": 0.4081142662510038,
"learning_rate": 1.2731844120883705e-05,
"loss": 0.0677,
"step": 1072
},
{
"epoch": 1.6507692307692308,
"grad_norm": 0.3634578478803434,
"learning_rate": 1.2720218699649243e-05,
"loss": 0.0674,
"step": 1073
},
{
"epoch": 1.6523076923076923,
"grad_norm": 0.38192663091543777,
"learning_rate": 1.270858930689419e-05,
"loss": 0.0701,
"step": 1074
},
{
"epoch": 1.6538461538461537,
"grad_norm": 0.3172214274621381,
"learning_rate": 1.269695595959747e-05,
"loss": 0.0704,
"step": 1075
},
{
"epoch": 1.6553846153846155,
"grad_norm": 0.38108705084014816,
"learning_rate": 1.2685318674743769e-05,
"loss": 0.0773,
"step": 1076
},
{
"epoch": 1.656923076923077,
"grad_norm": 0.32308540109864475,
"learning_rate": 1.2673677469323532e-05,
"loss": 0.0648,
"step": 1077
},
{
"epoch": 1.6584615384615384,
"grad_norm": 0.397581930467522,
"learning_rate": 1.2662032360332926e-05,
"loss": 0.0639,
"step": 1078
},
{
"epoch": 1.6600000000000001,
"grad_norm": 0.428597873391097,
"learning_rate": 1.2650383364773812e-05,
"loss": 0.0649,
"step": 1079
},
{
"epoch": 1.6615384615384614,
"grad_norm": 0.3627584918043698,
"learning_rate": 1.2638730499653731e-05,
"loss": 0.0685,
"step": 1080
},
{
"epoch": 1.6630769230769231,
"grad_norm": 0.39434710297076264,
"learning_rate": 1.262707378198587e-05,
"loss": 0.0623,
"step": 1081
},
{
"epoch": 1.6646153846153846,
"grad_norm": 0.4398747205110735,
"learning_rate": 1.2615413228789044e-05,
"loss": 0.0731,
"step": 1082
},
{
"epoch": 1.666153846153846,
"grad_norm": 0.42581951735488976,
"learning_rate": 1.2603748857087668e-05,
"loss": 0.0722,
"step": 1083
},
{
"epoch": 1.6676923076923078,
"grad_norm": 0.33276187980893557,
"learning_rate": 1.2592080683911726e-05,
"loss": 0.0696,
"step": 1084
},
{
"epoch": 1.669230769230769,
"grad_norm": 0.41489629589539934,
"learning_rate": 1.258040872629676e-05,
"loss": 0.0719,
"step": 1085
},
{
"epoch": 1.6707692307692308,
"grad_norm": 0.3446331527576127,
"learning_rate": 1.2568733001283828e-05,
"loss": 0.0559,
"step": 1086
},
{
"epoch": 1.6723076923076923,
"grad_norm": 0.35834532855166323,
"learning_rate": 1.2557053525919503e-05,
"loss": 0.0746,
"step": 1087
},
{
"epoch": 1.6738461538461538,
"grad_norm": 0.4154468285899175,
"learning_rate": 1.2545370317255817e-05,
"loss": 0.0685,
"step": 1088
},
{
"epoch": 1.6753846153846155,
"grad_norm": 0.3870928793180616,
"learning_rate": 1.2533683392350264e-05,
"loss": 0.081,
"step": 1089
},
{
"epoch": 1.676923076923077,
"grad_norm": 0.32370828199921,
"learning_rate": 1.252199276826576e-05,
"loss": 0.0535,
"step": 1090
},
{
"epoch": 1.6784615384615384,
"grad_norm": 0.36882979778488023,
"learning_rate": 1.2510298462070619e-05,
"loss": 0.0671,
"step": 1091
},
{
"epoch": 1.6800000000000002,
"grad_norm": 0.32092949735914467,
"learning_rate": 1.2498600490838535e-05,
"loss": 0.0626,
"step": 1092
},
{
"epoch": 1.6815384615384614,
"grad_norm": 0.5727815255595513,
"learning_rate": 1.2486898871648552e-05,
"loss": 0.0734,
"step": 1093
},
{
"epoch": 1.6830769230769231,
"grad_norm": 0.3458495040870996,
"learning_rate": 1.2475193621585036e-05,
"loss": 0.0653,
"step": 1094
},
{
"epoch": 1.6846153846153846,
"grad_norm": 0.35967546735381495,
"learning_rate": 1.2463484757737663e-05,
"loss": 0.0705,
"step": 1095
},
{
"epoch": 1.6861538461538461,
"grad_norm": 0.38237562063590214,
"learning_rate": 1.2451772297201376e-05,
"loss": 0.0661,
"step": 1096
},
{
"epoch": 1.6876923076923078,
"grad_norm": 0.47686706186257616,
"learning_rate": 1.2440056257076376e-05,
"loss": 0.0613,
"step": 1097
},
{
"epoch": 1.689230769230769,
"grad_norm": 0.46719251208554496,
"learning_rate": 1.2428336654468085e-05,
"loss": 0.0662,
"step": 1098
},
{
"epoch": 1.6907692307692308,
"grad_norm": 0.3686979455788657,
"learning_rate": 1.241661350648713e-05,
"loss": 0.0735,
"step": 1099
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.33821254997323613,
"learning_rate": 1.240488683024931e-05,
"loss": 0.0671,
"step": 1100
},
{
"epoch": 1.6938461538461538,
"grad_norm": 0.314525906392349,
"learning_rate": 1.2393156642875579e-05,
"loss": 0.0593,
"step": 1101
},
{
"epoch": 1.6953846153846155,
"grad_norm": 0.37318354708691176,
"learning_rate": 1.2381422961492018e-05,
"loss": 0.0612,
"step": 1102
},
{
"epoch": 1.696923076923077,
"grad_norm": 0.38841429558968976,
"learning_rate": 1.2369685803229802e-05,
"loss": 0.0618,
"step": 1103
},
{
"epoch": 1.6984615384615385,
"grad_norm": 0.4236775137197504,
"learning_rate": 1.2357945185225194e-05,
"loss": 0.0657,
"step": 1104
},
{
"epoch": 1.7,
"grad_norm": 0.33764294486901314,
"learning_rate": 1.2346201124619502e-05,
"loss": 0.0649,
"step": 1105
},
{
"epoch": 1.7015384615384614,
"grad_norm": 0.4418410618698052,
"learning_rate": 1.2334453638559057e-05,
"loss": 0.0687,
"step": 1106
},
{
"epoch": 1.7030769230769232,
"grad_norm": 0.3121204113437058,
"learning_rate": 1.2322702744195192e-05,
"loss": 0.0611,
"step": 1107
},
{
"epoch": 1.7046153846153846,
"grad_norm": 0.3445242526989668,
"learning_rate": 1.231094845868422e-05,
"loss": 0.0702,
"step": 1108
},
{
"epoch": 1.7061538461538461,
"grad_norm": 0.3181762559945284,
"learning_rate": 1.2299190799187405e-05,
"loss": 0.0617,
"step": 1109
},
{
"epoch": 1.7076923076923078,
"grad_norm": 0.35189126228058387,
"learning_rate": 1.2287429782870936e-05,
"loss": 0.0654,
"step": 1110
},
{
"epoch": 1.709230769230769,
"grad_norm": 0.3893933792288036,
"learning_rate": 1.22756654269059e-05,
"loss": 0.0731,
"step": 1111
},
{
"epoch": 1.7107692307692308,
"grad_norm": 0.36384662807264734,
"learning_rate": 1.2263897748468265e-05,
"loss": 0.0748,
"step": 1112
},
{
"epoch": 1.7123076923076923,
"grad_norm": 0.35150416540561524,
"learning_rate": 1.2252126764738845e-05,
"loss": 0.063,
"step": 1113
},
{
"epoch": 1.7138461538461538,
"grad_norm": 0.37272242307664083,
"learning_rate": 1.2240352492903282e-05,
"loss": 0.0658,
"step": 1114
},
{
"epoch": 1.7153846153846155,
"grad_norm": 0.34210913074854066,
"learning_rate": 1.222857495015202e-05,
"loss": 0.0631,
"step": 1115
},
{
"epoch": 1.7169230769230768,
"grad_norm": 0.3261822986248889,
"learning_rate": 1.2216794153680274e-05,
"loss": 0.0682,
"step": 1116
},
{
"epoch": 1.7184615384615385,
"grad_norm": 0.38673754458534254,
"learning_rate": 1.2205010120688012e-05,
"loss": 0.0744,
"step": 1117
},
{
"epoch": 1.72,
"grad_norm": 0.322372777644759,
"learning_rate": 1.2193222868379933e-05,
"loss": 0.0695,
"step": 1118
},
{
"epoch": 1.7215384615384615,
"grad_norm": 0.3627083236569345,
"learning_rate": 1.2181432413965428e-05,
"loss": 0.0742,
"step": 1119
},
{
"epoch": 1.7230769230769232,
"grad_norm": 0.3603766452102839,
"learning_rate": 1.2169638774658566e-05,
"loss": 0.0701,
"step": 1120
},
{
"epoch": 1.7246153846153847,
"grad_norm": 0.400801215452547,
"learning_rate": 1.2157841967678064e-05,
"loss": 0.073,
"step": 1121
},
{
"epoch": 1.7261538461538461,
"grad_norm": 0.35897450361326094,
"learning_rate": 1.2146042010247268e-05,
"loss": 0.0713,
"step": 1122
},
{
"epoch": 1.7276923076923076,
"grad_norm": 0.32908439200162665,
"learning_rate": 1.2134238919594122e-05,
"loss": 0.073,
"step": 1123
},
{
"epoch": 1.7292307692307691,
"grad_norm": 0.3195562766482599,
"learning_rate": 1.2122432712951142e-05,
"loss": 0.0639,
"step": 1124
},
{
"epoch": 1.7307692307692308,
"grad_norm": 0.3845258420163186,
"learning_rate": 1.2110623407555398e-05,
"loss": 0.061,
"step": 1125
},
{
"epoch": 1.7323076923076923,
"grad_norm": 0.3101783646667082,
"learning_rate": 1.2098811020648475e-05,
"loss": 0.0612,
"step": 1126
},
{
"epoch": 1.7338461538461538,
"grad_norm": 0.40958171970575463,
"learning_rate": 1.2086995569476474e-05,
"loss": 0.0602,
"step": 1127
},
{
"epoch": 1.7353846153846155,
"grad_norm": 0.36131350860502126,
"learning_rate": 1.2075177071289952e-05,
"loss": 0.0716,
"step": 1128
},
{
"epoch": 1.7369230769230768,
"grad_norm": 0.3245396638728491,
"learning_rate": 1.2063355543343925e-05,
"loss": 0.0664,
"step": 1129
},
{
"epoch": 1.7384615384615385,
"grad_norm": 0.34948238997533076,
"learning_rate": 1.2051531002897823e-05,
"loss": 0.0743,
"step": 1130
},
{
"epoch": 1.74,
"grad_norm": 0.33482788469371083,
"learning_rate": 1.2039703467215489e-05,
"loss": 0.07,
"step": 1131
},
{
"epoch": 1.7415384615384615,
"grad_norm": 0.3106255436550464,
"learning_rate": 1.2027872953565125e-05,
"loss": 0.0656,
"step": 1132
},
{
"epoch": 1.7430769230769232,
"grad_norm": 0.3230656932594039,
"learning_rate": 1.2016039479219293e-05,
"loss": 0.0673,
"step": 1133
},
{
"epoch": 1.7446153846153845,
"grad_norm": 0.41382601752001624,
"learning_rate": 1.2004203061454864e-05,
"loss": 0.0665,
"step": 1134
},
{
"epoch": 1.7461538461538462,
"grad_norm": 0.3742034275244198,
"learning_rate": 1.1992363717553015e-05,
"loss": 0.0752,
"step": 1135
},
{
"epoch": 1.7476923076923077,
"grad_norm": 0.36413647568781654,
"learning_rate": 1.1980521464799197e-05,
"loss": 0.071,
"step": 1136
},
{
"epoch": 1.7492307692307691,
"grad_norm": 0.373128673273757,
"learning_rate": 1.1968676320483103e-05,
"loss": 0.0702,
"step": 1137
},
{
"epoch": 1.7507692307692309,
"grad_norm": 0.3372236802921592,
"learning_rate": 1.1956828301898648e-05,
"loss": 0.0678,
"step": 1138
},
{
"epoch": 1.7523076923076923,
"grad_norm": 0.40138229132959674,
"learning_rate": 1.194497742634395e-05,
"loss": 0.0717,
"step": 1139
},
{
"epoch": 1.7538461538461538,
"grad_norm": 0.33981746405629404,
"learning_rate": 1.1933123711121284e-05,
"loss": 0.0707,
"step": 1140
},
{
"epoch": 1.7553846153846155,
"grad_norm": 0.31343455180354113,
"learning_rate": 1.1921267173537085e-05,
"loss": 0.0649,
"step": 1141
},
{
"epoch": 1.7569230769230768,
"grad_norm": 0.40773047383140626,
"learning_rate": 1.1909407830901905e-05,
"loss": 0.0643,
"step": 1142
},
{
"epoch": 1.7584615384615385,
"grad_norm": 0.3528139994471679,
"learning_rate": 1.1897545700530387e-05,
"loss": 0.0611,
"step": 1143
},
{
"epoch": 1.76,
"grad_norm": 0.4034945403026721,
"learning_rate": 1.1885680799741249e-05,
"loss": 0.0743,
"step": 1144
},
{
"epoch": 1.7615384615384615,
"grad_norm": 0.41781100877516925,
"learning_rate": 1.187381314585725e-05,
"loss": 0.0678,
"step": 1145
},
{
"epoch": 1.7630769230769232,
"grad_norm": 0.36789269013488135,
"learning_rate": 1.186194275620517e-05,
"loss": 0.0662,
"step": 1146
},
{
"epoch": 1.7646153846153845,
"grad_norm": 0.4280148153142728,
"learning_rate": 1.1850069648115785e-05,
"loss": 0.0703,
"step": 1147
},
{
"epoch": 1.7661538461538462,
"grad_norm": 0.34082568199061963,
"learning_rate": 1.1838193838923835e-05,
"loss": 0.0664,
"step": 1148
},
{
"epoch": 1.7676923076923077,
"grad_norm": 0.40801522956984687,
"learning_rate": 1.1826315345968014e-05,
"loss": 0.0684,
"step": 1149
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.5130512133374202,
"learning_rate": 1.1814434186590922e-05,
"loss": 0.0673,
"step": 1150
},
{
"epoch": 1.7707692307692309,
"grad_norm": 0.41911565805686907,
"learning_rate": 1.180255037813906e-05,
"loss": 0.0721,
"step": 1151
},
{
"epoch": 1.7723076923076924,
"grad_norm": 0.4585867938819892,
"learning_rate": 1.1790663937962789e-05,
"loss": 0.07,
"step": 1152
},
{
"epoch": 1.7738461538461539,
"grad_norm": 0.4179899966673112,
"learning_rate": 1.1778774883416325e-05,
"loss": 0.0742,
"step": 1153
},
{
"epoch": 1.7753846153846153,
"grad_norm": 0.38592221008164235,
"learning_rate": 1.1766883231857686e-05,
"loss": 0.0744,
"step": 1154
},
{
"epoch": 1.7769230769230768,
"grad_norm": 0.35229801433260116,
"learning_rate": 1.1754989000648693e-05,
"loss": 0.0726,
"step": 1155
},
{
"epoch": 1.7784615384615385,
"grad_norm": 0.340673871886494,
"learning_rate": 1.1743092207154929e-05,
"loss": 0.0636,
"step": 1156
},
{
"epoch": 1.78,
"grad_norm": 0.3495561290874491,
"learning_rate": 1.1731192868745716e-05,
"loss": 0.056,
"step": 1157
},
{
"epoch": 1.7815384615384615,
"grad_norm": 0.3734216760946435,
"learning_rate": 1.1719291002794096e-05,
"loss": 0.0681,
"step": 1158
},
{
"epoch": 1.7830769230769232,
"grad_norm": 0.3417807928208906,
"learning_rate": 1.1707386626676798e-05,
"loss": 0.0729,
"step": 1159
},
{
"epoch": 1.7846153846153845,
"grad_norm": 0.3437047768651414,
"learning_rate": 1.1695479757774217e-05,
"loss": 0.0694,
"step": 1160
},
{
"epoch": 1.7861538461538462,
"grad_norm": 0.47420240578398315,
"learning_rate": 1.1683570413470384e-05,
"loss": 0.0676,
"step": 1161
},
{
"epoch": 1.7876923076923077,
"grad_norm": 0.3250761492618577,
"learning_rate": 1.1671658611152954e-05,
"loss": 0.0733,
"step": 1162
},
{
"epoch": 1.7892307692307692,
"grad_norm": 0.38754162428676897,
"learning_rate": 1.1659744368213159e-05,
"loss": 0.0621,
"step": 1163
},
{
"epoch": 1.790769230769231,
"grad_norm": 0.33675515675870277,
"learning_rate": 1.1647827702045802e-05,
"loss": 0.0672,
"step": 1164
},
{
"epoch": 1.7923076923076922,
"grad_norm": 0.44334916416488984,
"learning_rate": 1.163590863004922e-05,
"loss": 0.0767,
"step": 1165
},
{
"epoch": 1.7938461538461539,
"grad_norm": 0.36810606943536545,
"learning_rate": 1.1623987169625261e-05,
"loss": 0.0757,
"step": 1166
},
{
"epoch": 1.7953846153846154,
"grad_norm": 0.3878325616113022,
"learning_rate": 1.1612063338179269e-05,
"loss": 0.0674,
"step": 1167
},
{
"epoch": 1.7969230769230768,
"grad_norm": 0.36763716766560983,
"learning_rate": 1.1600137153120039e-05,
"loss": 0.0694,
"step": 1168
},
{
"epoch": 1.7984615384615386,
"grad_norm": 0.294038432326372,
"learning_rate": 1.1588208631859808e-05,
"loss": 0.0624,
"step": 1169
},
{
"epoch": 1.8,
"grad_norm": 0.3066013994606816,
"learning_rate": 1.1576277791814219e-05,
"loss": 0.064,
"step": 1170
},
{
"epoch": 1.8015384615384615,
"grad_norm": 0.3791728156018938,
"learning_rate": 1.156434465040231e-05,
"loss": 0.0685,
"step": 1171
},
{
"epoch": 1.803076923076923,
"grad_norm": 0.34991093402366064,
"learning_rate": 1.1552409225046472e-05,
"loss": 0.0702,
"step": 1172
},
{
"epoch": 1.8046153846153845,
"grad_norm": 0.35156981872263665,
"learning_rate": 1.154047153317243e-05,
"loss": 0.0648,
"step": 1173
},
{
"epoch": 1.8061538461538462,
"grad_norm": 0.47105508744307845,
"learning_rate": 1.152853159220922e-05,
"loss": 0.0662,
"step": 1174
},
{
"epoch": 1.8076923076923077,
"grad_norm": 0.34193953754457573,
"learning_rate": 1.1516589419589159e-05,
"loss": 0.0662,
"step": 1175
},
{
"epoch": 1.8092307692307692,
"grad_norm": 0.3080144424855269,
"learning_rate": 1.1504645032747832e-05,
"loss": 0.0718,
"step": 1176
},
{
"epoch": 1.810769230769231,
"grad_norm": 0.3166715432520699,
"learning_rate": 1.1492698449124042e-05,
"loss": 0.0632,
"step": 1177
},
{
"epoch": 1.8123076923076922,
"grad_norm": 0.3116243301245059,
"learning_rate": 1.148074968615981e-05,
"loss": 0.0668,
"step": 1178
},
{
"epoch": 1.8138461538461539,
"grad_norm": 0.3923187036355073,
"learning_rate": 1.1468798761300335e-05,
"loss": 0.0721,
"step": 1179
},
{
"epoch": 1.8153846153846154,
"grad_norm": 0.3293677766442383,
"learning_rate": 1.1456845691993975e-05,
"loss": 0.0636,
"step": 1180
},
{
"epoch": 1.8169230769230769,
"grad_norm": 0.33075818730038914,
"learning_rate": 1.1444890495692214e-05,
"loss": 0.0688,
"step": 1181
},
{
"epoch": 1.8184615384615386,
"grad_norm": 0.3270293740539091,
"learning_rate": 1.1432933189849647e-05,
"loss": 0.0675,
"step": 1182
},
{
"epoch": 1.8199999999999998,
"grad_norm": 0.2923335748923596,
"learning_rate": 1.1420973791923941e-05,
"loss": 0.0701,
"step": 1183
},
{
"epoch": 1.8215384615384616,
"grad_norm": 0.3093982567613493,
"learning_rate": 1.1409012319375828e-05,
"loss": 0.0659,
"step": 1184
},
{
"epoch": 1.823076923076923,
"grad_norm": 0.3799881014599721,
"learning_rate": 1.1397048789669061e-05,
"loss": 0.0714,
"step": 1185
},
{
"epoch": 1.8246153846153845,
"grad_norm": 0.36261575151573644,
"learning_rate": 1.13850832202704e-05,
"loss": 0.0693,
"step": 1186
},
{
"epoch": 1.8261538461538462,
"grad_norm": 0.3132085768992873,
"learning_rate": 1.1373115628649582e-05,
"loss": 0.0682,
"step": 1187
},
{
"epoch": 1.8276923076923077,
"grad_norm": 0.3461207920147732,
"learning_rate": 1.1361146032279295e-05,
"loss": 0.0711,
"step": 1188
},
{
"epoch": 1.8292307692307692,
"grad_norm": 0.3150515227081572,
"learning_rate": 1.1349174448635158e-05,
"loss": 0.0592,
"step": 1189
},
{
"epoch": 1.830769230769231,
"grad_norm": 0.33512808523002696,
"learning_rate": 1.1337200895195688e-05,
"loss": 0.0697,
"step": 1190
},
{
"epoch": 1.8323076923076922,
"grad_norm": 0.35007766877677027,
"learning_rate": 1.1325225389442278e-05,
"loss": 0.0667,
"step": 1191
},
{
"epoch": 1.833846153846154,
"grad_norm": 0.34216015472816874,
"learning_rate": 1.1313247948859168e-05,
"loss": 0.0769,
"step": 1192
},
{
"epoch": 1.8353846153846154,
"grad_norm": 0.34775267785403957,
"learning_rate": 1.1301268590933434e-05,
"loss": 0.0663,
"step": 1193
},
{
"epoch": 1.8369230769230769,
"grad_norm": 0.33480692927199907,
"learning_rate": 1.1289287333154941e-05,
"loss": 0.0762,
"step": 1194
},
{
"epoch": 1.8384615384615386,
"grad_norm": 0.32637056156167593,
"learning_rate": 1.1277304193016332e-05,
"loss": 0.0713,
"step": 1195
},
{
"epoch": 1.8399999999999999,
"grad_norm": 0.36829791104284954,
"learning_rate": 1.1265319188012995e-05,
"loss": 0.0737,
"step": 1196
},
{
"epoch": 1.8415384615384616,
"grad_norm": 0.36062035876290466,
"learning_rate": 1.1253332335643043e-05,
"loss": 0.0689,
"step": 1197
},
{
"epoch": 1.843076923076923,
"grad_norm": 0.3711995102972711,
"learning_rate": 1.124134365340729e-05,
"loss": 0.0642,
"step": 1198
},
{
"epoch": 1.8446153846153845,
"grad_norm": 0.30208317420183395,
"learning_rate": 1.1229353158809216e-05,
"loss": 0.0643,
"step": 1199
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.33621334713281376,
"learning_rate": 1.1217360869354948e-05,
"loss": 0.0716,
"step": 1200
},
{
"epoch": 1.8476923076923077,
"grad_norm": 0.4109500586936904,
"learning_rate": 1.1205366802553231e-05,
"loss": 0.0627,
"step": 1201
},
{
"epoch": 1.8492307692307692,
"grad_norm": 0.3355086005987176,
"learning_rate": 1.1193370975915414e-05,
"loss": 0.0651,
"step": 1202
},
{
"epoch": 1.8507692307692307,
"grad_norm": 0.33311821347747733,
"learning_rate": 1.118137340695541e-05,
"loss": 0.0707,
"step": 1203
},
{
"epoch": 1.8523076923076922,
"grad_norm": 0.3783217276792658,
"learning_rate": 1.1169374113189669e-05,
"loss": 0.0688,
"step": 1204
},
{
"epoch": 1.853846153846154,
"grad_norm": 0.43132614741663744,
"learning_rate": 1.1157373112137171e-05,
"loss": 0.0799,
"step": 1205
},
{
"epoch": 1.8553846153846154,
"grad_norm": 0.2922605690189237,
"learning_rate": 1.1145370421319377e-05,
"loss": 0.0592,
"step": 1206
},
{
"epoch": 1.856923076923077,
"grad_norm": 0.3223022612523878,
"learning_rate": 1.1133366058260232e-05,
"loss": 0.0647,
"step": 1207
},
{
"epoch": 1.8584615384615386,
"grad_norm": 0.3018670306666712,
"learning_rate": 1.11213600404861e-05,
"loss": 0.0685,
"step": 1208
},
{
"epoch": 1.8599999999999999,
"grad_norm": 0.3889446684181748,
"learning_rate": 1.1109352385525782e-05,
"loss": 0.0766,
"step": 1209
},
{
"epoch": 1.8615384615384616,
"grad_norm": 0.4477268822990713,
"learning_rate": 1.1097343110910452e-05,
"loss": 0.0645,
"step": 1210
},
{
"epoch": 1.863076923076923,
"grad_norm": 0.33752078106966593,
"learning_rate": 1.1085332234173664e-05,
"loss": 0.0724,
"step": 1211
},
{
"epoch": 1.8646153846153846,
"grad_norm": 0.33949138455564243,
"learning_rate": 1.1073319772851299e-05,
"loss": 0.0635,
"step": 1212
},
{
"epoch": 1.8661538461538463,
"grad_norm": 0.339715631924927,
"learning_rate": 1.106130574448156e-05,
"loss": 0.0708,
"step": 1213
},
{
"epoch": 1.8676923076923075,
"grad_norm": 0.38187248950026703,
"learning_rate": 1.1049290166604928e-05,
"loss": 0.0648,
"step": 1214
},
{
"epoch": 1.8692307692307693,
"grad_norm": 0.4755273814617505,
"learning_rate": 1.1037273056764157e-05,
"loss": 0.0676,
"step": 1215
},
{
"epoch": 1.8707692307692307,
"grad_norm": 0.35341205325615144,
"learning_rate": 1.1025254432504234e-05,
"loss": 0.0742,
"step": 1216
},
{
"epoch": 1.8723076923076922,
"grad_norm": 0.3254062177799031,
"learning_rate": 1.1013234311372353e-05,
"loss": 0.0649,
"step": 1217
},
{
"epoch": 1.873846153846154,
"grad_norm": 0.31246579893201815,
"learning_rate": 1.1001212710917897e-05,
"loss": 0.0649,
"step": 1218
},
{
"epoch": 1.8753846153846154,
"grad_norm": 0.33928603890752973,
"learning_rate": 1.0989189648692408e-05,
"loss": 0.07,
"step": 1219
},
{
"epoch": 1.876923076923077,
"grad_norm": 0.3450403297222314,
"learning_rate": 1.0977165142249566e-05,
"loss": 0.067,
"step": 1220
},
{
"epoch": 1.8784615384615386,
"grad_norm": 0.3423853230449218,
"learning_rate": 1.0965139209145153e-05,
"loss": 0.0702,
"step": 1221
},
{
"epoch": 1.88,
"grad_norm": 0.32174442781609414,
"learning_rate": 1.0953111866937038e-05,
"loss": 0.0701,
"step": 1222
},
{
"epoch": 1.8815384615384616,
"grad_norm": 0.3676532027177272,
"learning_rate": 1.0941083133185146e-05,
"loss": 0.0705,
"step": 1223
},
{
"epoch": 1.883076923076923,
"grad_norm": 0.41997872234324835,
"learning_rate": 1.0929053025451432e-05,
"loss": 0.0659,
"step": 1224
},
{
"epoch": 1.8846153846153846,
"grad_norm": 0.4110637997348029,
"learning_rate": 1.0917021561299864e-05,
"loss": 0.0729,
"step": 1225
},
{
"epoch": 1.8861538461538463,
"grad_norm": 0.38396879751089963,
"learning_rate": 1.090498875829638e-05,
"loss": 0.0786,
"step": 1226
},
{
"epoch": 1.8876923076923076,
"grad_norm": 0.3530760565665137,
"learning_rate": 1.089295463400888e-05,
"loss": 0.0736,
"step": 1227
},
{
"epoch": 1.8892307692307693,
"grad_norm": 0.3828535267508954,
"learning_rate": 1.0880919206007193e-05,
"loss": 0.0756,
"step": 1228
},
{
"epoch": 1.8907692307692308,
"grad_norm": 0.4620093651236917,
"learning_rate": 1.0868882491863048e-05,
"loss": 0.0672,
"step": 1229
},
{
"epoch": 1.8923076923076922,
"grad_norm": 0.34394347660813923,
"learning_rate": 1.0856844509150056e-05,
"loss": 0.0628,
"step": 1230
},
{
"epoch": 1.893846153846154,
"grad_norm": 0.34240923368033127,
"learning_rate": 1.0844805275443673e-05,
"loss": 0.0663,
"step": 1231
},
{
"epoch": 1.8953846153846152,
"grad_norm": 0.4216830435498585,
"learning_rate": 1.0832764808321186e-05,
"loss": 0.0731,
"step": 1232
},
{
"epoch": 1.896923076923077,
"grad_norm": 0.34817819208217915,
"learning_rate": 1.0820723125361685e-05,
"loss": 0.0696,
"step": 1233
},
{
"epoch": 1.8984615384615384,
"grad_norm": 0.3361061093793128,
"learning_rate": 1.0808680244146035e-05,
"loss": 0.0608,
"step": 1234
},
{
"epoch": 1.9,
"grad_norm": 0.3127731515392646,
"learning_rate": 1.0796636182256846e-05,
"loss": 0.0654,
"step": 1235
},
{
"epoch": 1.9015384615384616,
"grad_norm": 0.34974984864646785,
"learning_rate": 1.0784590957278452e-05,
"loss": 0.0696,
"step": 1236
},
{
"epoch": 1.903076923076923,
"grad_norm": 0.3241892888461775,
"learning_rate": 1.077254458679689e-05,
"loss": 0.0615,
"step": 1237
},
{
"epoch": 1.9046153846153846,
"grad_norm": 0.37471636624302423,
"learning_rate": 1.0760497088399863e-05,
"loss": 0.0657,
"step": 1238
},
{
"epoch": 1.9061538461538463,
"grad_norm": 0.4476088794896166,
"learning_rate": 1.074844847967673e-05,
"loss": 0.0696,
"step": 1239
},
{
"epoch": 1.9076923076923076,
"grad_norm": 0.35752628031586536,
"learning_rate": 1.0736398778218458e-05,
"loss": 0.0662,
"step": 1240
},
{
"epoch": 1.9092307692307693,
"grad_norm": 0.40745054566528704,
"learning_rate": 1.0724348001617626e-05,
"loss": 0.0688,
"step": 1241
},
{
"epoch": 1.9107692307692308,
"grad_norm": 0.39673775592008276,
"learning_rate": 1.0712296167468366e-05,
"loss": 0.0636,
"step": 1242
},
{
"epoch": 1.9123076923076923,
"grad_norm": 0.36255910467507374,
"learning_rate": 1.0700243293366365e-05,
"loss": 0.0606,
"step": 1243
},
{
"epoch": 1.913846153846154,
"grad_norm": 0.5136376481460118,
"learning_rate": 1.0688189396908826e-05,
"loss": 0.0645,
"step": 1244
},
{
"epoch": 1.9153846153846152,
"grad_norm": 0.3099567930694205,
"learning_rate": 1.0676134495694439e-05,
"loss": 0.0644,
"step": 1245
},
{
"epoch": 1.916923076923077,
"grad_norm": 0.37000890844695966,
"learning_rate": 1.0664078607323367e-05,
"loss": 0.0687,
"step": 1246
},
{
"epoch": 1.9184615384615384,
"grad_norm": 0.34210135022540517,
"learning_rate": 1.0652021749397216e-05,
"loss": 0.0699,
"step": 1247
},
{
"epoch": 1.92,
"grad_norm": 0.3609562310788875,
"learning_rate": 1.0639963939519005e-05,
"loss": 0.0738,
"step": 1248
},
{
"epoch": 1.9215384615384616,
"grad_norm": 0.3700975852964543,
"learning_rate": 1.0627905195293135e-05,
"loss": 0.0664,
"step": 1249
},
{
"epoch": 1.9230769230769231,
"grad_norm": 0.3859166868285145,
"learning_rate": 1.0615845534325384e-05,
"loss": 0.0713,
"step": 1250
},
{
"epoch": 1.9246153846153846,
"grad_norm": 0.4115264368388931,
"learning_rate": 1.0603784974222862e-05,
"loss": 0.0705,
"step": 1251
},
{
"epoch": 1.926153846153846,
"grad_norm": 0.3145083279396702,
"learning_rate": 1.0591723532593992e-05,
"loss": 0.0604,
"step": 1252
},
{
"epoch": 1.9276923076923076,
"grad_norm": 0.3197633328292509,
"learning_rate": 1.0579661227048484e-05,
"loss": 0.0604,
"step": 1253
},
{
"epoch": 1.9292307692307693,
"grad_norm": 0.3622577046638927,
"learning_rate": 1.056759807519731e-05,
"loss": 0.0716,
"step": 1254
},
{
"epoch": 1.9307692307692308,
"grad_norm": 0.3497451203917846,
"learning_rate": 1.0555534094652675e-05,
"loss": 0.0673,
"step": 1255
},
{
"epoch": 1.9323076923076923,
"grad_norm": 0.3320313053856611,
"learning_rate": 1.0543469303028002e-05,
"loss": 0.0635,
"step": 1256
},
{
"epoch": 1.933846153846154,
"grad_norm": 0.3233116362666333,
"learning_rate": 1.0531403717937888e-05,
"loss": 0.062,
"step": 1257
},
{
"epoch": 1.9353846153846153,
"grad_norm": 0.3685802421725291,
"learning_rate": 1.0519337356998094e-05,
"loss": 0.0692,
"step": 1258
},
{
"epoch": 1.936923076923077,
"grad_norm": 0.336210753592891,
"learning_rate": 1.0507270237825513e-05,
"loss": 0.0675,
"step": 1259
},
{
"epoch": 1.9384615384615385,
"grad_norm": 0.3289919747166384,
"learning_rate": 1.0495202378038144e-05,
"loss": 0.0657,
"step": 1260
},
{
"epoch": 1.94,
"grad_norm": 0.33651073591370484,
"learning_rate": 1.0483133795255072e-05,
"loss": 0.0717,
"step": 1261
},
{
"epoch": 1.9415384615384617,
"grad_norm": 0.3648322054454426,
"learning_rate": 1.0471064507096427e-05,
"loss": 0.0793,
"step": 1262
},
{
"epoch": 1.943076923076923,
"grad_norm": 0.33381701009336806,
"learning_rate": 1.045899453118338e-05,
"loss": 0.0617,
"step": 1263
},
{
"epoch": 1.9446153846153846,
"grad_norm": 0.3641612409279995,
"learning_rate": 1.0446923885138101e-05,
"loss": 0.0597,
"step": 1264
},
{
"epoch": 1.9461538461538461,
"grad_norm": 0.34585249190313433,
"learning_rate": 1.0434852586583737e-05,
"loss": 0.0674,
"step": 1265
},
{
"epoch": 1.9476923076923076,
"grad_norm": 0.3289388167113336,
"learning_rate": 1.0422780653144392e-05,
"loss": 0.0735,
"step": 1266
},
{
"epoch": 1.9492307692307693,
"grad_norm": 0.3887566915602357,
"learning_rate": 1.0410708102445091e-05,
"loss": 0.0766,
"step": 1267
},
{
"epoch": 1.9507692307692308,
"grad_norm": 0.30555149229418166,
"learning_rate": 1.0398634952111766e-05,
"loss": 0.0718,
"step": 1268
},
{
"epoch": 1.9523076923076923,
"grad_norm": 0.32317250467343567,
"learning_rate": 1.0386561219771222e-05,
"loss": 0.0652,
"step": 1269
},
{
"epoch": 1.953846153846154,
"grad_norm": 0.3006439298211834,
"learning_rate": 1.0374486923051117e-05,
"loss": 0.0623,
"step": 1270
},
{
"epoch": 1.9553846153846153,
"grad_norm": 0.33386303419709507,
"learning_rate": 1.0362412079579925e-05,
"loss": 0.0701,
"step": 1271
},
{
"epoch": 1.956923076923077,
"grad_norm": 0.33099092128522206,
"learning_rate": 1.0350336706986925e-05,
"loss": 0.068,
"step": 1272
},
{
"epoch": 1.9584615384615385,
"grad_norm": 0.3262986897907234,
"learning_rate": 1.0338260822902166e-05,
"loss": 0.0614,
"step": 1273
},
{
"epoch": 1.96,
"grad_norm": 0.34485599175028897,
"learning_rate": 1.0326184444956449e-05,
"loss": 0.0658,
"step": 1274
},
{
"epoch": 1.9615384615384617,
"grad_norm": 0.31792277168038097,
"learning_rate": 1.0314107590781284e-05,
"loss": 0.0574,
"step": 1275
},
{
"epoch": 1.963076923076923,
"grad_norm": 0.3208911821070251,
"learning_rate": 1.030203027800889e-05,
"loss": 0.0633,
"step": 1276
},
{
"epoch": 1.9646153846153847,
"grad_norm": 0.33055045597617766,
"learning_rate": 1.0289952524272147e-05,
"loss": 0.0706,
"step": 1277
},
{
"epoch": 1.9661538461538461,
"grad_norm": 0.34088210301412697,
"learning_rate": 1.027787434720458e-05,
"loss": 0.0739,
"step": 1278
},
{
"epoch": 1.9676923076923076,
"grad_norm": 0.32073888049730687,
"learning_rate": 1.0265795764440335e-05,
"loss": 0.0651,
"step": 1279
},
{
"epoch": 1.9692307692307693,
"grad_norm": 0.3076473967656385,
"learning_rate": 1.025371679361415e-05,
"loss": 0.0683,
"step": 1280
},
{
"epoch": 1.9707692307692306,
"grad_norm": 0.3621459788069735,
"learning_rate": 1.0241637452361323e-05,
"loss": 0.0677,
"step": 1281
},
{
"epoch": 1.9723076923076923,
"grad_norm": 0.3573542495717642,
"learning_rate": 1.0229557758317703e-05,
"loss": 0.0679,
"step": 1282
},
{
"epoch": 1.9738461538461538,
"grad_norm": 0.3483850527238064,
"learning_rate": 1.0217477729119648e-05,
"loss": 0.074,
"step": 1283
},
{
"epoch": 1.9753846153846153,
"grad_norm": 0.3063790927358412,
"learning_rate": 1.0205397382404006e-05,
"loss": 0.0652,
"step": 1284
},
{
"epoch": 1.976923076923077,
"grad_norm": 0.34619566294720616,
"learning_rate": 1.0193316735808085e-05,
"loss": 0.0676,
"step": 1285
},
{
"epoch": 1.9784615384615385,
"grad_norm": 0.46164644987132686,
"learning_rate": 1.018123580696964e-05,
"loss": 0.0711,
"step": 1286
},
{
"epoch": 1.98,
"grad_norm": 0.33352482622772156,
"learning_rate": 1.0169154613526831e-05,
"loss": 0.0696,
"step": 1287
},
{
"epoch": 1.9815384615384617,
"grad_norm": 0.3101921292214142,
"learning_rate": 1.0157073173118207e-05,
"loss": 0.0647,
"step": 1288
},
{
"epoch": 1.983076923076923,
"grad_norm": 0.31015657112545686,
"learning_rate": 1.0144991503382676e-05,
"loss": 0.0614,
"step": 1289
},
{
"epoch": 1.9846153846153847,
"grad_norm": 0.3503219066829283,
"learning_rate": 1.0132909621959482e-05,
"loss": 0.077,
"step": 1290
},
{
"epoch": 1.9861538461538462,
"grad_norm": 0.2992678727460164,
"learning_rate": 1.0120827546488175e-05,
"loss": 0.065,
"step": 1291
},
{
"epoch": 1.9876923076923076,
"grad_norm": 0.3232588558361219,
"learning_rate": 1.0108745294608595e-05,
"loss": 0.0596,
"step": 1292
},
{
"epoch": 1.9892307692307694,
"grad_norm": 0.3219005317029137,
"learning_rate": 1.0096662883960833e-05,
"loss": 0.0693,
"step": 1293
},
{
"epoch": 1.9907692307692306,
"grad_norm": 0.36479544578271617,
"learning_rate": 1.0084580332185214e-05,
"loss": 0.071,
"step": 1294
},
{
"epoch": 1.9923076923076923,
"grad_norm": 0.36208947868270336,
"learning_rate": 1.0072497656922266e-05,
"loss": 0.0697,
"step": 1295
},
{
"epoch": 1.9938461538461538,
"grad_norm": 0.374662600377983,
"learning_rate": 1.0060414875812709e-05,
"loss": 0.0724,
"step": 1296
},
{
"epoch": 1.9953846153846153,
"grad_norm": 0.4368778771016829,
"learning_rate": 1.0048332006497406e-05,
"loss": 0.0777,
"step": 1297
},
{
"epoch": 1.996923076923077,
"grad_norm": 0.40482597277564725,
"learning_rate": 1.003624906661735e-05,
"loss": 0.0672,
"step": 1298
},
{
"epoch": 1.9984615384615385,
"grad_norm": 0.3944268530980909,
"learning_rate": 1.0024166073813634e-05,
"loss": 0.0619,
"step": 1299
},
{
"epoch": 2.0,
"grad_norm": 0.3312157049217859,
"learning_rate": 1.0012083045727445e-05,
"loss": 0.0633,
"step": 1300
}
],
"logging_steps": 1,
"max_steps": 2600,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 263636178075648.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}