sedrickkeh's picture
End of training
83df9e1 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9904153354632586,
"eval_steps": 500,
"global_step": 468,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006389776357827476,
"grad_norm": 5.939312845608103,
"learning_rate": 4.2553191489361704e-07,
"loss": 0.8883,
"step": 1
},
{
"epoch": 0.012779552715654952,
"grad_norm": 5.6459943274568385,
"learning_rate": 8.510638297872341e-07,
"loss": 0.8365,
"step": 2
},
{
"epoch": 0.019169329073482427,
"grad_norm": 6.290489665408837,
"learning_rate": 1.276595744680851e-06,
"loss": 0.8864,
"step": 3
},
{
"epoch": 0.025559105431309903,
"grad_norm": 5.8544636560570735,
"learning_rate": 1.7021276595744682e-06,
"loss": 0.879,
"step": 4
},
{
"epoch": 0.03194888178913738,
"grad_norm": 5.478170251266386,
"learning_rate": 2.1276595744680853e-06,
"loss": 0.8245,
"step": 5
},
{
"epoch": 0.038338658146964855,
"grad_norm": 4.688336713722404,
"learning_rate": 2.553191489361702e-06,
"loss": 0.849,
"step": 6
},
{
"epoch": 0.04472843450479233,
"grad_norm": 4.278233500376253,
"learning_rate": 2.978723404255319e-06,
"loss": 0.8151,
"step": 7
},
{
"epoch": 0.051118210862619806,
"grad_norm": 2.367029075074496,
"learning_rate": 3.4042553191489363e-06,
"loss": 0.7566,
"step": 8
},
{
"epoch": 0.05750798722044728,
"grad_norm": 2.0707237905380484,
"learning_rate": 3.8297872340425535e-06,
"loss": 0.7551,
"step": 9
},
{
"epoch": 0.06389776357827476,
"grad_norm": 1.7866725538131587,
"learning_rate": 4.255319148936171e-06,
"loss": 0.7457,
"step": 10
},
{
"epoch": 0.07028753993610223,
"grad_norm": 3.404477977029493,
"learning_rate": 4.680851063829788e-06,
"loss": 0.699,
"step": 11
},
{
"epoch": 0.07667731629392971,
"grad_norm": 3.8759087344215093,
"learning_rate": 5.106382978723404e-06,
"loss": 0.8106,
"step": 12
},
{
"epoch": 0.08306709265175719,
"grad_norm": 4.206484510948959,
"learning_rate": 5.531914893617022e-06,
"loss": 0.7639,
"step": 13
},
{
"epoch": 0.08945686900958466,
"grad_norm": 3.597855342218176,
"learning_rate": 5.957446808510638e-06,
"loss": 0.7143,
"step": 14
},
{
"epoch": 0.09584664536741214,
"grad_norm": 3.493223809169224,
"learning_rate": 6.382978723404256e-06,
"loss": 0.7495,
"step": 15
},
{
"epoch": 0.10223642172523961,
"grad_norm": 2.7730036124678588,
"learning_rate": 6.808510638297873e-06,
"loss": 0.7079,
"step": 16
},
{
"epoch": 0.10862619808306709,
"grad_norm": 1.99917862715565,
"learning_rate": 7.234042553191491e-06,
"loss": 0.6637,
"step": 17
},
{
"epoch": 0.11501597444089456,
"grad_norm": 1.7234540912576959,
"learning_rate": 7.659574468085107e-06,
"loss": 0.6563,
"step": 18
},
{
"epoch": 0.12140575079872204,
"grad_norm": 1.9632219810386513,
"learning_rate": 8.085106382978723e-06,
"loss": 0.6753,
"step": 19
},
{
"epoch": 0.12779552715654952,
"grad_norm": 2.1614567368781157,
"learning_rate": 8.510638297872341e-06,
"loss": 0.6063,
"step": 20
},
{
"epoch": 0.134185303514377,
"grad_norm": 1.932507047332365,
"learning_rate": 8.936170212765958e-06,
"loss": 0.629,
"step": 21
},
{
"epoch": 0.14057507987220447,
"grad_norm": 1.5481507427216847,
"learning_rate": 9.361702127659576e-06,
"loss": 0.6444,
"step": 22
},
{
"epoch": 0.14696485623003194,
"grad_norm": 1.2177329752055572,
"learning_rate": 9.787234042553192e-06,
"loss": 0.641,
"step": 23
},
{
"epoch": 0.15335463258785942,
"grad_norm": 1.4206725631658237,
"learning_rate": 1.0212765957446808e-05,
"loss": 0.6496,
"step": 24
},
{
"epoch": 0.1597444089456869,
"grad_norm": 1.4667519530084758,
"learning_rate": 1.0638297872340426e-05,
"loss": 0.6304,
"step": 25
},
{
"epoch": 0.16613418530351437,
"grad_norm": 1.2452708214307595,
"learning_rate": 1.1063829787234044e-05,
"loss": 0.6413,
"step": 26
},
{
"epoch": 0.17252396166134185,
"grad_norm": 0.7759167145812482,
"learning_rate": 1.1489361702127662e-05,
"loss": 0.5927,
"step": 27
},
{
"epoch": 0.17891373801916932,
"grad_norm": 1.0321850071310563,
"learning_rate": 1.1914893617021277e-05,
"loss": 0.5898,
"step": 28
},
{
"epoch": 0.1853035143769968,
"grad_norm": 1.0266703283162124,
"learning_rate": 1.2340425531914895e-05,
"loss": 0.5895,
"step": 29
},
{
"epoch": 0.19169329073482427,
"grad_norm": 0.9674925840183656,
"learning_rate": 1.2765957446808513e-05,
"loss": 0.5853,
"step": 30
},
{
"epoch": 0.19808306709265175,
"grad_norm": 0.9805374530460067,
"learning_rate": 1.3191489361702127e-05,
"loss": 0.5756,
"step": 31
},
{
"epoch": 0.20447284345047922,
"grad_norm": 0.9258462587522315,
"learning_rate": 1.3617021276595745e-05,
"loss": 0.585,
"step": 32
},
{
"epoch": 0.2108626198083067,
"grad_norm": 0.8925757909283792,
"learning_rate": 1.4042553191489363e-05,
"loss": 0.584,
"step": 33
},
{
"epoch": 0.21725239616613418,
"grad_norm": 0.7738950177471798,
"learning_rate": 1.4468085106382981e-05,
"loss": 0.553,
"step": 34
},
{
"epoch": 0.22364217252396165,
"grad_norm": 0.823063478968403,
"learning_rate": 1.4893617021276596e-05,
"loss": 0.605,
"step": 35
},
{
"epoch": 0.23003194888178913,
"grad_norm": 0.7769591687125452,
"learning_rate": 1.5319148936170214e-05,
"loss": 0.5398,
"step": 36
},
{
"epoch": 0.2364217252396166,
"grad_norm": 0.724085749329525,
"learning_rate": 1.5744680851063832e-05,
"loss": 0.5541,
"step": 37
},
{
"epoch": 0.24281150159744408,
"grad_norm": 0.7937699708520545,
"learning_rate": 1.6170212765957446e-05,
"loss": 0.5516,
"step": 38
},
{
"epoch": 0.24920127795527156,
"grad_norm": 0.838915436660047,
"learning_rate": 1.6595744680851064e-05,
"loss": 0.5335,
"step": 39
},
{
"epoch": 0.25559105431309903,
"grad_norm": 0.7447841608522099,
"learning_rate": 1.7021276595744682e-05,
"loss": 0.5414,
"step": 40
},
{
"epoch": 0.26198083067092653,
"grad_norm": 0.8714110715774589,
"learning_rate": 1.74468085106383e-05,
"loss": 0.556,
"step": 41
},
{
"epoch": 0.268370607028754,
"grad_norm": 0.7411083091257032,
"learning_rate": 1.7872340425531915e-05,
"loss": 0.5113,
"step": 42
},
{
"epoch": 0.2747603833865815,
"grad_norm": 0.7680242229988357,
"learning_rate": 1.8297872340425533e-05,
"loss": 0.5684,
"step": 43
},
{
"epoch": 0.28115015974440893,
"grad_norm": 0.7374046719973165,
"learning_rate": 1.872340425531915e-05,
"loss": 0.577,
"step": 44
},
{
"epoch": 0.28753993610223644,
"grad_norm": 0.7709945486266545,
"learning_rate": 1.914893617021277e-05,
"loss": 0.5251,
"step": 45
},
{
"epoch": 0.2939297124600639,
"grad_norm": 0.7420014515431359,
"learning_rate": 1.9574468085106384e-05,
"loss": 0.5358,
"step": 46
},
{
"epoch": 0.3003194888178914,
"grad_norm": 0.7321704500165183,
"learning_rate": 2e-05,
"loss": 0.56,
"step": 47
},
{
"epoch": 0.30670926517571884,
"grad_norm": 0.7855274459068936,
"learning_rate": 1.9999721578003894e-05,
"loss": 0.5427,
"step": 48
},
{
"epoch": 0.31309904153354634,
"grad_norm": 0.7731901448013743,
"learning_rate": 1.9998886327519337e-05,
"loss": 0.5517,
"step": 49
},
{
"epoch": 0.3194888178913738,
"grad_norm": 0.7327331638939495,
"learning_rate": 1.999749429505675e-05,
"loss": 0.5468,
"step": 50
},
{
"epoch": 0.3258785942492013,
"grad_norm": 0.7419176272941636,
"learning_rate": 1.9995545558130624e-05,
"loss": 0.548,
"step": 51
},
{
"epoch": 0.33226837060702874,
"grad_norm": 0.7873113572745489,
"learning_rate": 1.9993040225255205e-05,
"loss": 0.5403,
"step": 52
},
{
"epoch": 0.33865814696485624,
"grad_norm": 0.6658693818077654,
"learning_rate": 1.998997843593845e-05,
"loss": 0.5054,
"step": 53
},
{
"epoch": 0.3450479233226837,
"grad_norm": 0.7409323699660156,
"learning_rate": 1.9986360360674252e-05,
"loss": 0.5555,
"step": 54
},
{
"epoch": 0.3514376996805112,
"grad_norm": 0.7094024540386773,
"learning_rate": 1.9982186200932964e-05,
"loss": 0.535,
"step": 55
},
{
"epoch": 0.35782747603833864,
"grad_norm": 0.9538171828161622,
"learning_rate": 1.9977456189150164e-05,
"loss": 0.5788,
"step": 56
},
{
"epoch": 0.36421725239616615,
"grad_norm": 0.7395956073936416,
"learning_rate": 1.9972170588713715e-05,
"loss": 0.5836,
"step": 57
},
{
"epoch": 0.3706070287539936,
"grad_norm": 0.8649724098491106,
"learning_rate": 1.9966329693949098e-05,
"loss": 0.4918,
"step": 58
},
{
"epoch": 0.3769968051118211,
"grad_norm": 0.8838321465136332,
"learning_rate": 1.9959933830103034e-05,
"loss": 0.514,
"step": 59
},
{
"epoch": 0.38338658146964855,
"grad_norm": 0.7993190296372936,
"learning_rate": 1.9952983353325358e-05,
"loss": 0.5191,
"step": 60
},
{
"epoch": 0.38977635782747605,
"grad_norm": 1.0003320683454586,
"learning_rate": 1.9945478650649192e-05,
"loss": 0.5395,
"step": 61
},
{
"epoch": 0.3961661341853035,
"grad_norm": 0.8174378430731664,
"learning_rate": 1.9937420139969397e-05,
"loss": 0.5428,
"step": 62
},
{
"epoch": 0.402555910543131,
"grad_norm": 0.7497946237416347,
"learning_rate": 1.9928808270019297e-05,
"loss": 0.5826,
"step": 63
},
{
"epoch": 0.40894568690095845,
"grad_norm": 0.8640015891418493,
"learning_rate": 1.9919643520345698e-05,
"loss": 0.5343,
"step": 64
},
{
"epoch": 0.41533546325878595,
"grad_norm": 1.0033449675940591,
"learning_rate": 1.990992640128218e-05,
"loss": 0.5623,
"step": 65
},
{
"epoch": 0.4217252396166134,
"grad_norm": 0.7213392023439573,
"learning_rate": 1.989965745392068e-05,
"loss": 0.5314,
"step": 66
},
{
"epoch": 0.4281150159744409,
"grad_norm": 0.8863830620435368,
"learning_rate": 1.988883725008136e-05,
"loss": 0.5458,
"step": 67
},
{
"epoch": 0.43450479233226835,
"grad_norm": 0.7362416113857194,
"learning_rate": 1.9877466392280773e-05,
"loss": 0.5428,
"step": 68
},
{
"epoch": 0.44089456869009586,
"grad_norm": 0.9788737478764125,
"learning_rate": 1.9865545513698304e-05,
"loss": 0.5412,
"step": 69
},
{
"epoch": 0.4472843450479233,
"grad_norm": 0.8021931238319318,
"learning_rate": 1.9853075278140913e-05,
"loss": 0.5522,
"step": 70
},
{
"epoch": 0.4536741214057508,
"grad_norm": 0.85887825826516,
"learning_rate": 1.984005638000618e-05,
"loss": 0.5702,
"step": 71
},
{
"epoch": 0.46006389776357826,
"grad_norm": 0.7815409304110528,
"learning_rate": 1.9826489544243623e-05,
"loss": 0.5482,
"step": 72
},
{
"epoch": 0.46645367412140576,
"grad_norm": 0.7632274039193099,
"learning_rate": 1.981237552631434e-05,
"loss": 0.5387,
"step": 73
},
{
"epoch": 0.4728434504792332,
"grad_norm": 0.7141397681245097,
"learning_rate": 1.9797715112148937e-05,
"loss": 0.5546,
"step": 74
},
{
"epoch": 0.4792332268370607,
"grad_norm": 0.9775621975797985,
"learning_rate": 1.9782509118103773e-05,
"loss": 0.5308,
"step": 75
},
{
"epoch": 0.48562300319488816,
"grad_norm": 0.7330287655613912,
"learning_rate": 1.9766758390915494e-05,
"loss": 0.5431,
"step": 76
},
{
"epoch": 0.49201277955271566,
"grad_norm": 0.8510881793812443,
"learning_rate": 1.9750463807653873e-05,
"loss": 0.5187,
"step": 77
},
{
"epoch": 0.4984025559105431,
"grad_norm": 0.8486211359598791,
"learning_rate": 1.9733626275673e-05,
"loss": 0.55,
"step": 78
},
{
"epoch": 0.5047923322683706,
"grad_norm": 0.7806116343240348,
"learning_rate": 1.9716246732560715e-05,
"loss": 0.4813,
"step": 79
},
{
"epoch": 0.5111821086261981,
"grad_norm": 0.6353268171693698,
"learning_rate": 1.9698326146086446e-05,
"loss": 0.5078,
"step": 80
},
{
"epoch": 0.5175718849840255,
"grad_norm": 0.8766273238149929,
"learning_rate": 1.967986551414728e-05,
"loss": 0.544,
"step": 81
},
{
"epoch": 0.5239616613418531,
"grad_norm": 0.7124929308186037,
"learning_rate": 1.9660865864712413e-05,
"loss": 0.5547,
"step": 82
},
{
"epoch": 0.5303514376996805,
"grad_norm": 0.7985052392042088,
"learning_rate": 1.9641328255765916e-05,
"loss": 0.5215,
"step": 83
},
{
"epoch": 0.536741214057508,
"grad_norm": 0.6406165534662566,
"learning_rate": 1.96212537752478e-05,
"loss": 0.4978,
"step": 84
},
{
"epoch": 0.5431309904153354,
"grad_norm": 0.7351154516097609,
"learning_rate": 1.9600643540993453e-05,
"loss": 0.5498,
"step": 85
},
{
"epoch": 0.549520766773163,
"grad_norm": 0.7023736701907295,
"learning_rate": 1.9579498700671386e-05,
"loss": 0.5271,
"step": 86
},
{
"epoch": 0.5559105431309904,
"grad_norm": 0.7585882232459317,
"learning_rate": 1.9557820431719333e-05,
"loss": 0.5419,
"step": 87
},
{
"epoch": 0.5623003194888179,
"grad_norm": 0.7609703576418915,
"learning_rate": 1.9535609941278676e-05,
"loss": 0.5115,
"step": 88
},
{
"epoch": 0.5686900958466453,
"grad_norm": 0.8355505315864735,
"learning_rate": 1.9512868466127232e-05,
"loss": 0.5214,
"step": 89
},
{
"epoch": 0.5750798722044729,
"grad_norm": 0.8985531652526637,
"learning_rate": 1.9489597272610377e-05,
"loss": 0.5217,
"step": 90
},
{
"epoch": 0.5814696485623003,
"grad_norm": 0.8296453033353365,
"learning_rate": 1.9465797656570546e-05,
"loss": 0.5235,
"step": 91
},
{
"epoch": 0.5878594249201278,
"grad_norm": 0.8400305646818227,
"learning_rate": 1.944147094327506e-05,
"loss": 0.5028,
"step": 92
},
{
"epoch": 0.5942492012779552,
"grad_norm": 0.8027178786911765,
"learning_rate": 1.9416618487342333e-05,
"loss": 0.4932,
"step": 93
},
{
"epoch": 0.6006389776357828,
"grad_norm": 0.831774187088593,
"learning_rate": 1.9391241672666438e-05,
"loss": 0.5015,
"step": 94
},
{
"epoch": 0.6070287539936102,
"grad_norm": 0.6880013029770987,
"learning_rate": 1.936534191234006e-05,
"loss": 0.4851,
"step": 95
},
{
"epoch": 0.6134185303514377,
"grad_norm": 0.7874317799707009,
"learning_rate": 1.9338920648575798e-05,
"loss": 0.4993,
"step": 96
},
{
"epoch": 0.6198083067092651,
"grad_norm": 0.8086603460274078,
"learning_rate": 1.9311979352625837e-05,
"loss": 0.5472,
"step": 97
},
{
"epoch": 0.6261980830670927,
"grad_norm": 0.7330826956678399,
"learning_rate": 1.928451952470007e-05,
"loss": 0.5267,
"step": 98
},
{
"epoch": 0.6325878594249201,
"grad_norm": 0.6774581525725395,
"learning_rate": 1.9256542693882505e-05,
"loss": 0.4964,
"step": 99
},
{
"epoch": 0.6389776357827476,
"grad_norm": 0.699096603267482,
"learning_rate": 1.922805041804617e-05,
"loss": 0.5051,
"step": 100
},
{
"epoch": 0.645367412140575,
"grad_norm": 0.7625712048870313,
"learning_rate": 1.919904428376632e-05,
"loss": 0.5301,
"step": 101
},
{
"epoch": 0.6517571884984026,
"grad_norm": 0.7442243954348883,
"learning_rate": 1.916952590623212e-05,
"loss": 0.5114,
"step": 102
},
{
"epoch": 0.65814696485623,
"grad_norm": 0.6319971486530133,
"learning_rate": 1.9139496929156685e-05,
"loss": 0.5106,
"step": 103
},
{
"epoch": 0.6645367412140575,
"grad_norm": 0.6634321632909197,
"learning_rate": 1.910895902468557e-05,
"loss": 0.5211,
"step": 104
},
{
"epoch": 0.670926517571885,
"grad_norm": 0.5959863886093508,
"learning_rate": 1.907791389330363e-05,
"loss": 0.5124,
"step": 105
},
{
"epoch": 0.6773162939297125,
"grad_norm": 0.7995435481771501,
"learning_rate": 1.904636326374036e-05,
"loss": 0.52,
"step": 106
},
{
"epoch": 0.6837060702875399,
"grad_norm": 0.7383352863257084,
"learning_rate": 1.9014308892873612e-05,
"loss": 0.5516,
"step": 107
},
{
"epoch": 0.6900958466453674,
"grad_norm": 0.8315686682034763,
"learning_rate": 1.8981752565631767e-05,
"loss": 0.4891,
"step": 108
},
{
"epoch": 0.6964856230031949,
"grad_norm": 0.6508951908046772,
"learning_rate": 1.8948696094894354e-05,
"loss": 0.5591,
"step": 109
},
{
"epoch": 0.7028753993610224,
"grad_norm": 0.8009701429333234,
"learning_rate": 1.8915141321391083e-05,
"loss": 0.5305,
"step": 110
},
{
"epoch": 0.7092651757188498,
"grad_norm": 0.7308231155353043,
"learning_rate": 1.8881090113599353e-05,
"loss": 0.5191,
"step": 111
},
{
"epoch": 0.7156549520766773,
"grad_norm": 0.7808253591246626,
"learning_rate": 1.8846544367640218e-05,
"loss": 0.4906,
"step": 112
},
{
"epoch": 0.7220447284345048,
"grad_norm": 0.8760149164716509,
"learning_rate": 1.881150600717279e-05,
"loss": 0.532,
"step": 113
},
{
"epoch": 0.7284345047923323,
"grad_norm": 0.6776840820965584,
"learning_rate": 1.8775976983287117e-05,
"loss": 0.521,
"step": 114
},
{
"epoch": 0.7348242811501597,
"grad_norm": 0.8898200759371685,
"learning_rate": 1.873995927439555e-05,
"loss": 0.5251,
"step": 115
},
{
"epoch": 0.7412140575079872,
"grad_norm": 0.7499195664294598,
"learning_rate": 1.8703454886122568e-05,
"loss": 0.5348,
"step": 116
},
{
"epoch": 0.7476038338658147,
"grad_norm": 0.8505865320259214,
"learning_rate": 1.86664658511931e-05,
"loss": 0.5054,
"step": 117
},
{
"epoch": 0.7539936102236422,
"grad_norm": 0.6522705814194051,
"learning_rate": 1.862899422931934e-05,
"loss": 0.4996,
"step": 118
},
{
"epoch": 0.7603833865814696,
"grad_norm": 0.6577321085300217,
"learning_rate": 1.859104210708604e-05,
"loss": 0.5106,
"step": 119
},
{
"epoch": 0.7667731629392971,
"grad_norm": 0.6237691862811225,
"learning_rate": 1.855261159783432e-05,
"loss": 0.4952,
"step": 120
},
{
"epoch": 0.7731629392971247,
"grad_norm": 0.7454463301186427,
"learning_rate": 1.8513704841543997e-05,
"loss": 0.5026,
"step": 121
},
{
"epoch": 0.7795527156549521,
"grad_norm": 0.7934145741008819,
"learning_rate": 1.847432400471443e-05,
"loss": 0.549,
"step": 122
},
{
"epoch": 0.7859424920127795,
"grad_norm": 0.6802649460778195,
"learning_rate": 1.8434471280243854e-05,
"loss": 0.506,
"step": 123
},
{
"epoch": 0.792332268370607,
"grad_norm": 0.7304620814404005,
"learning_rate": 1.8394148887307286e-05,
"loss": 0.5383,
"step": 124
},
{
"epoch": 0.7987220447284346,
"grad_norm": 0.6957782218041328,
"learning_rate": 1.8353359071232954e-05,
"loss": 0.5019,
"step": 125
},
{
"epoch": 0.805111821086262,
"grad_norm": 0.6826820301754135,
"learning_rate": 1.8312104103377266e-05,
"loss": 0.4937,
"step": 126
},
{
"epoch": 0.8115015974440895,
"grad_norm": 0.598906813674708,
"learning_rate": 1.827038628099831e-05,
"loss": 0.488,
"step": 127
},
{
"epoch": 0.8178913738019169,
"grad_norm": 0.6633449549765212,
"learning_rate": 1.822820792712797e-05,
"loss": 0.4971,
"step": 128
},
{
"epoch": 0.8242811501597445,
"grad_norm": 0.6710913601976871,
"learning_rate": 1.8185571390442542e-05,
"loss": 0.4836,
"step": 129
},
{
"epoch": 0.8306709265175719,
"grad_norm": 0.5762592167913351,
"learning_rate": 1.8142479045131956e-05,
"loss": 0.5192,
"step": 130
},
{
"epoch": 0.8370607028753994,
"grad_norm": 0.6738150719804582,
"learning_rate": 1.809893329076757e-05,
"loss": 0.5228,
"step": 131
},
{
"epoch": 0.8434504792332268,
"grad_norm": 0.6500141385276138,
"learning_rate": 1.8054936552168548e-05,
"loss": 0.5332,
"step": 132
},
{
"epoch": 0.8498402555910544,
"grad_norm": 0.5886713445337487,
"learning_rate": 1.801049127926686e-05,
"loss": 0.4787,
"step": 133
},
{
"epoch": 0.8562300319488818,
"grad_norm": 0.7649082888316344,
"learning_rate": 1.7965599946970814e-05,
"loss": 0.4914,
"step": 134
},
{
"epoch": 0.8626198083067093,
"grad_norm": 0.8014082101971098,
"learning_rate": 1.7920265055027285e-05,
"loss": 0.5171,
"step": 135
},
{
"epoch": 0.8690095846645367,
"grad_norm": 0.7706425976997076,
"learning_rate": 1.7874489127882496e-05,
"loss": 0.5267,
"step": 136
},
{
"epoch": 0.8753993610223643,
"grad_norm": 0.9916919676402002,
"learning_rate": 1.7828274714541445e-05,
"loss": 0.5225,
"step": 137
},
{
"epoch": 0.8817891373801917,
"grad_norm": 0.7510268200588113,
"learning_rate": 1.7781624388425974e-05,
"loss": 0.5056,
"step": 138
},
{
"epoch": 0.8881789137380192,
"grad_norm": 1.0622441121039243,
"learning_rate": 1.773454074723147e-05,
"loss": 0.4773,
"step": 139
},
{
"epoch": 0.8945686900958466,
"grad_norm": 0.8430739649427611,
"learning_rate": 1.76870264127822e-05,
"loss": 0.5384,
"step": 140
},
{
"epoch": 0.9009584664536742,
"grad_norm": 0.7217579935144164,
"learning_rate": 1.763908403088534e-05,
"loss": 0.5108,
"step": 141
},
{
"epoch": 0.9073482428115016,
"grad_norm": 1.0377962316213807,
"learning_rate": 1.759071627118362e-05,
"loss": 0.4961,
"step": 142
},
{
"epoch": 0.9137380191693291,
"grad_norm": 0.6423592379266523,
"learning_rate": 1.754192582700668e-05,
"loss": 0.5061,
"step": 143
},
{
"epoch": 0.9201277955271565,
"grad_norm": 0.7302473302785821,
"learning_rate": 1.7492715415221087e-05,
"loss": 0.5074,
"step": 144
},
{
"epoch": 0.9265175718849841,
"grad_norm": 0.709358335589859,
"learning_rate": 1.7443087776079068e-05,
"loss": 0.4867,
"step": 145
},
{
"epoch": 0.9329073482428115,
"grad_norm": 0.6708691396098521,
"learning_rate": 1.739304567306588e-05,
"loss": 0.4955,
"step": 146
},
{
"epoch": 0.939297124600639,
"grad_norm": 0.7224681216999947,
"learning_rate": 1.7342591892745978e-05,
"loss": 0.5063,
"step": 147
},
{
"epoch": 0.9456869009584664,
"grad_norm": 0.692060343253805,
"learning_rate": 1.72917292446078e-05,
"loss": 0.5395,
"step": 148
},
{
"epoch": 0.952076677316294,
"grad_norm": 0.6503466417528021,
"learning_rate": 1.7240460560907345e-05,
"loss": 0.502,
"step": 149
},
{
"epoch": 0.9584664536741214,
"grad_norm": 0.7852299377720712,
"learning_rate": 1.7188788696510477e-05,
"loss": 0.4982,
"step": 150
},
{
"epoch": 0.9648562300319489,
"grad_norm": 0.6487201669982385,
"learning_rate": 1.7136716528733916e-05,
"loss": 0.491,
"step": 151
},
{
"epoch": 0.9712460063897763,
"grad_norm": 0.7162643954228396,
"learning_rate": 1.7084246957185036e-05,
"loss": 0.4715,
"step": 152
},
{
"epoch": 0.9776357827476039,
"grad_norm": 0.6882308824929978,
"learning_rate": 1.703138290360041e-05,
"loss": 0.4884,
"step": 153
},
{
"epoch": 0.9840255591054313,
"grad_norm": 0.6558983320574183,
"learning_rate": 1.6978127311683103e-05,
"loss": 0.5053,
"step": 154
},
{
"epoch": 0.9904153354632588,
"grad_norm": 0.7793396179424056,
"learning_rate": 1.6924483146938756e-05,
"loss": 0.4891,
"step": 155
},
{
"epoch": 0.9968051118210862,
"grad_norm": 0.6408866926804856,
"learning_rate": 1.6870453396510456e-05,
"loss": 0.5253,
"step": 156
},
{
"epoch": 1.0031948881789137,
"grad_norm": 1.219315523607184,
"learning_rate": 1.681604106901239e-05,
"loss": 0.7727,
"step": 157
},
{
"epoch": 1.0095846645367412,
"grad_norm": 0.7363761093026954,
"learning_rate": 1.676124919436233e-05,
"loss": 0.4369,
"step": 158
},
{
"epoch": 1.0159744408945688,
"grad_norm": 0.6104584236502094,
"learning_rate": 1.6706080823612897e-05,
"loss": 0.4074,
"step": 159
},
{
"epoch": 1.0223642172523961,
"grad_norm": 0.6688279966321403,
"learning_rate": 1.665053902878167e-05,
"loss": 0.3875,
"step": 160
},
{
"epoch": 1.0287539936102237,
"grad_norm": 0.7714442502018793,
"learning_rate": 1.659462690268013e-05,
"loss": 0.4693,
"step": 161
},
{
"epoch": 1.035143769968051,
"grad_norm": 0.6581337601062037,
"learning_rate": 1.6538347558741424e-05,
"loss": 0.4088,
"step": 162
},
{
"epoch": 1.0415335463258786,
"grad_norm": 0.8907132608840105,
"learning_rate": 1.6481704130847013e-05,
"loss": 0.4421,
"step": 163
},
{
"epoch": 1.0479233226837061,
"grad_norm": 0.7281123005456749,
"learning_rate": 1.642469977315214e-05,
"loss": 0.4466,
"step": 164
},
{
"epoch": 1.0543130990415335,
"grad_norm": 0.8713456380104283,
"learning_rate": 1.6367337659910223e-05,
"loss": 0.4205,
"step": 165
},
{
"epoch": 1.060702875399361,
"grad_norm": 0.8221934081375099,
"learning_rate": 1.6309620985296075e-05,
"loss": 0.4811,
"step": 166
},
{
"epoch": 1.0670926517571886,
"grad_norm": 0.6414871893647255,
"learning_rate": 1.625155296322805e-05,
"loss": 0.3994,
"step": 167
},
{
"epoch": 1.073482428115016,
"grad_norm": 0.6020141646779518,
"learning_rate": 1.6193136827189067e-05,
"loss": 0.3544,
"step": 168
},
{
"epoch": 1.0798722044728435,
"grad_norm": 0.6620341470772692,
"learning_rate": 1.6134375830046566e-05,
"loss": 0.3865,
"step": 169
},
{
"epoch": 1.0862619808306708,
"grad_norm": 0.7529504403287007,
"learning_rate": 1.607527324387137e-05,
"loss": 0.5138,
"step": 170
},
{
"epoch": 1.0926517571884984,
"grad_norm": 0.5746068027715405,
"learning_rate": 1.6015832359755483e-05,
"loss": 0.4053,
"step": 171
},
{
"epoch": 1.099041533546326,
"grad_norm": 0.6423681462846987,
"learning_rate": 1.5956056487628832e-05,
"loss": 0.4878,
"step": 172
},
{
"epoch": 1.1054313099041533,
"grad_norm": 0.6192258722203541,
"learning_rate": 1.5895948956074937e-05,
"loss": 0.4188,
"step": 173
},
{
"epoch": 1.1118210862619808,
"grad_norm": 0.6163929815962566,
"learning_rate": 1.5835513112145583e-05,
"loss": 0.3854,
"step": 174
},
{
"epoch": 1.1182108626198084,
"grad_norm": 0.5697582130826278,
"learning_rate": 1.5774752321174428e-05,
"loss": 0.4073,
"step": 175
},
{
"epoch": 1.1246006389776357,
"grad_norm": 0.6223679982213549,
"learning_rate": 1.571366996658962e-05,
"loss": 0.4224,
"step": 176
},
{
"epoch": 1.1309904153354633,
"grad_norm": 0.6320866995001497,
"learning_rate": 1.5652269449725375e-05,
"loss": 0.4248,
"step": 177
},
{
"epoch": 1.1373801916932909,
"grad_norm": 0.5297458143137096,
"learning_rate": 1.5590554189632585e-05,
"loss": 0.3748,
"step": 178
},
{
"epoch": 1.1437699680511182,
"grad_norm": 0.6358623803264093,
"learning_rate": 1.552852762288843e-05,
"loss": 0.4747,
"step": 179
},
{
"epoch": 1.1501597444089458,
"grad_norm": 0.6278673693951984,
"learning_rate": 1.5466193203405017e-05,
"loss": 0.4036,
"step": 180
},
{
"epoch": 1.156549520766773,
"grad_norm": 0.6151559404690766,
"learning_rate": 1.540355440223704e-05,
"loss": 0.4124,
"step": 181
},
{
"epoch": 1.1629392971246006,
"grad_norm": 0.5891168526046718,
"learning_rate": 1.534061470738852e-05,
"loss": 0.4637,
"step": 182
},
{
"epoch": 1.1693290734824282,
"grad_norm": 0.6184989510741438,
"learning_rate": 1.527737762361855e-05,
"loss": 0.4161,
"step": 183
},
{
"epoch": 1.1757188498402555,
"grad_norm": 0.6419175983144096,
"learning_rate": 1.5213846672246139e-05,
"loss": 0.4533,
"step": 184
},
{
"epoch": 1.182108626198083,
"grad_norm": 0.5120065057835093,
"learning_rate": 1.5150025390954153e-05,
"loss": 0.375,
"step": 185
},
{
"epoch": 1.1884984025559104,
"grad_norm": 0.698299656514935,
"learning_rate": 1.5085917333592299e-05,
"loss": 0.4193,
"step": 186
},
{
"epoch": 1.194888178913738,
"grad_norm": 0.6532253196374253,
"learning_rate": 1.5021526069979232e-05,
"loss": 0.4596,
"step": 187
},
{
"epoch": 1.2012779552715656,
"grad_norm": 0.6028992706390495,
"learning_rate": 1.4956855185703787e-05,
"loss": 0.4207,
"step": 188
},
{
"epoch": 1.207667731629393,
"grad_norm": 0.6185875205322425,
"learning_rate": 1.48919082819253e-05,
"loss": 0.4243,
"step": 189
},
{
"epoch": 1.2140575079872205,
"grad_norm": 0.6043057260440954,
"learning_rate": 1.4826688975173085e-05,
"loss": 0.437,
"step": 190
},
{
"epoch": 1.220447284345048,
"grad_norm": 0.6821203725825057,
"learning_rate": 1.4761200897145063e-05,
"loss": 0.5037,
"step": 191
},
{
"epoch": 1.2268370607028753,
"grad_norm": 0.6008298374460067,
"learning_rate": 1.4695447694505512e-05,
"loss": 0.3661,
"step": 192
},
{
"epoch": 1.233226837060703,
"grad_norm": 0.7199503911011816,
"learning_rate": 1.4629433028682014e-05,
"loss": 0.5043,
"step": 193
},
{
"epoch": 1.2396166134185305,
"grad_norm": 0.5180806806501965,
"learning_rate": 1.456316057566158e-05,
"loss": 0.3819,
"step": 194
},
{
"epoch": 1.2460063897763578,
"grad_norm": 0.6928136290488852,
"learning_rate": 1.4496634025785938e-05,
"loss": 0.4263,
"step": 195
},
{
"epoch": 1.2523961661341854,
"grad_norm": 0.5665537937081578,
"learning_rate": 1.4429857083546054e-05,
"loss": 0.4844,
"step": 196
},
{
"epoch": 1.2587859424920127,
"grad_norm": 0.5563376996661965,
"learning_rate": 1.4362833467375839e-05,
"loss": 0.4156,
"step": 197
},
{
"epoch": 1.2651757188498403,
"grad_norm": 0.5377783142222045,
"learning_rate": 1.429556690944509e-05,
"loss": 0.4025,
"step": 198
},
{
"epoch": 1.2715654952076676,
"grad_norm": 0.5413290657480134,
"learning_rate": 1.4228061155451671e-05,
"loss": 0.4066,
"step": 199
},
{
"epoch": 1.2779552715654952,
"grad_norm": 0.5854137012981868,
"learning_rate": 1.4160319964412943e-05,
"loss": 0.4631,
"step": 200
},
{
"epoch": 1.2843450479233227,
"grad_norm": 0.5549643478979397,
"learning_rate": 1.4092347108456425e-05,
"loss": 0.4203,
"step": 201
},
{
"epoch": 1.29073482428115,
"grad_norm": 0.6090412261860937,
"learning_rate": 1.402414637260977e-05,
"loss": 0.4818,
"step": 202
},
{
"epoch": 1.2971246006389776,
"grad_norm": 0.5907663028777652,
"learning_rate": 1.3955721554589979e-05,
"loss": 0.4253,
"step": 203
},
{
"epoch": 1.3035143769968052,
"grad_norm": 0.6715293474231077,
"learning_rate": 1.388707646459193e-05,
"loss": 0.418,
"step": 204
},
{
"epoch": 1.3099041533546325,
"grad_norm": 0.6297284099203906,
"learning_rate": 1.3818214925076226e-05,
"loss": 0.4147,
"step": 205
},
{
"epoch": 1.31629392971246,
"grad_norm": 0.5667392659435537,
"learning_rate": 1.3749140770556322e-05,
"loss": 0.4311,
"step": 206
},
{
"epoch": 1.3226837060702876,
"grad_norm": 0.6322500298227558,
"learning_rate": 1.367985784738501e-05,
"loss": 0.4203,
"step": 207
},
{
"epoch": 1.329073482428115,
"grad_norm": 0.6001231519749656,
"learning_rate": 1.361037001354025e-05,
"loss": 0.4476,
"step": 208
},
{
"epoch": 1.3354632587859425,
"grad_norm": 0.5824816964719087,
"learning_rate": 1.3540681138410317e-05,
"loss": 0.3966,
"step": 209
},
{
"epoch": 1.34185303514377,
"grad_norm": 0.6931574895039062,
"learning_rate": 1.3470795102578358e-05,
"loss": 0.5033,
"step": 210
},
{
"epoch": 1.3482428115015974,
"grad_norm": 0.5416798455387101,
"learning_rate": 1.3400715797606293e-05,
"loss": 0.4008,
"step": 211
},
{
"epoch": 1.354632587859425,
"grad_norm": 0.6524420733651487,
"learning_rate": 1.3330447125818115e-05,
"loss": 0.4436,
"step": 212
},
{
"epoch": 1.3610223642172525,
"grad_norm": 0.6109732592249346,
"learning_rate": 1.3259993000082599e-05,
"loss": 0.4139,
"step": 213
},
{
"epoch": 1.3674121405750799,
"grad_norm": 0.6089156914272807,
"learning_rate": 1.3189357343595405e-05,
"loss": 0.4287,
"step": 214
},
{
"epoch": 1.3738019169329074,
"grad_norm": 0.5100478253771064,
"learning_rate": 1.3118544089660635e-05,
"loss": 0.3773,
"step": 215
},
{
"epoch": 1.3801916932907348,
"grad_norm": 0.665636694875197,
"learning_rate": 1.3047557181471784e-05,
"loss": 0.3922,
"step": 216
},
{
"epoch": 1.3865814696485623,
"grad_norm": 0.6421315985345839,
"learning_rate": 1.2976400571892189e-05,
"loss": 0.4432,
"step": 217
},
{
"epoch": 1.3929712460063897,
"grad_norm": 0.5630313538184247,
"learning_rate": 1.2905078223234907e-05,
"loss": 0.4191,
"step": 218
},
{
"epoch": 1.3993610223642172,
"grad_norm": 0.6745179299688657,
"learning_rate": 1.2833594107042078e-05,
"loss": 0.4333,
"step": 219
},
{
"epoch": 1.4057507987220448,
"grad_norm": 0.5454374317298349,
"learning_rate": 1.2761952203863759e-05,
"loss": 0.4277,
"step": 220
},
{
"epoch": 1.4121405750798721,
"grad_norm": 0.5789898470817271,
"learning_rate": 1.2690156503036288e-05,
"loss": 0.4036,
"step": 221
},
{
"epoch": 1.4185303514376997,
"grad_norm": 0.5949332555607629,
"learning_rate": 1.2618211002460135e-05,
"loss": 0.4785,
"step": 222
},
{
"epoch": 1.4249201277955272,
"grad_norm": 0.5587473061257787,
"learning_rate": 1.2546119708377273e-05,
"loss": 0.4109,
"step": 223
},
{
"epoch": 1.4313099041533546,
"grad_norm": 0.5092986860891325,
"learning_rate": 1.2473886635148109e-05,
"loss": 0.4375,
"step": 224
},
{
"epoch": 1.4376996805111821,
"grad_norm": 0.49324268916930714,
"learning_rate": 1.2401515805027924e-05,
"loss": 0.4339,
"step": 225
},
{
"epoch": 1.4440894568690097,
"grad_norm": 0.521973180886398,
"learning_rate": 1.2329011247942915e-05,
"loss": 0.3929,
"step": 226
},
{
"epoch": 1.450479233226837,
"grad_norm": 0.49449177011483986,
"learning_rate": 1.2256377001265785e-05,
"loss": 0.3688,
"step": 227
},
{
"epoch": 1.4568690095846646,
"grad_norm": 0.688803254610615,
"learning_rate": 1.2183617109590925e-05,
"loss": 0.5054,
"step": 228
},
{
"epoch": 1.4632587859424921,
"grad_norm": 0.5222053229962778,
"learning_rate": 1.2110735624509184e-05,
"loss": 0.3946,
"step": 229
},
{
"epoch": 1.4696485623003195,
"grad_norm": 0.5032950143354372,
"learning_rate": 1.2037736604382279e-05,
"loss": 0.4158,
"step": 230
},
{
"epoch": 1.476038338658147,
"grad_norm": 0.5739740428574489,
"learning_rate": 1.1964624114116784e-05,
"loss": 0.3794,
"step": 231
},
{
"epoch": 1.4824281150159744,
"grad_norm": 0.6710884805647371,
"learning_rate": 1.1891402224937805e-05,
"loss": 0.466,
"step": 232
},
{
"epoch": 1.488817891373802,
"grad_norm": 0.4582879331386117,
"learning_rate": 1.1818075014162243e-05,
"loss": 0.3585,
"step": 233
},
{
"epoch": 1.4952076677316293,
"grad_norm": 0.5448073298830566,
"learning_rate": 1.1744646564971777e-05,
"loss": 0.4238,
"step": 234
},
{
"epoch": 1.5015974440894568,
"grad_norm": 0.5760693930458392,
"learning_rate": 1.1671120966185486e-05,
"loss": 0.4324,
"step": 235
},
{
"epoch": 1.5079872204472844,
"grad_norm": 0.49993740420542576,
"learning_rate": 1.159750231203217e-05,
"loss": 0.4078,
"step": 236
},
{
"epoch": 1.5143769968051117,
"grad_norm": 0.563168928495261,
"learning_rate": 1.1523794701922351e-05,
"loss": 0.4351,
"step": 237
},
{
"epoch": 1.5207667731629393,
"grad_norm": 0.5313983639628916,
"learning_rate": 1.145000224022002e-05,
"loss": 0.3571,
"step": 238
},
{
"epoch": 1.5271565495207668,
"grad_norm": 0.501524364831029,
"learning_rate": 1.1376129036014073e-05,
"loss": 0.4304,
"step": 239
},
{
"epoch": 1.5335463258785942,
"grad_norm": 0.44709921144798975,
"learning_rate": 1.1302179202889505e-05,
"loss": 0.3948,
"step": 240
},
{
"epoch": 1.5399361022364217,
"grad_norm": 0.6024736966395187,
"learning_rate": 1.1228156858698344e-05,
"loss": 0.5055,
"step": 241
},
{
"epoch": 1.5463258785942493,
"grad_norm": 0.5039190418971187,
"learning_rate": 1.1154066125330358e-05,
"loss": 0.3853,
"step": 242
},
{
"epoch": 1.5527156549520766,
"grad_norm": 0.5082416278400913,
"learning_rate": 1.107991112848352e-05,
"loss": 0.365,
"step": 243
},
{
"epoch": 1.5591054313099042,
"grad_norm": 0.571424017024242,
"learning_rate": 1.100569599743428e-05,
"loss": 0.4157,
"step": 244
},
{
"epoch": 1.5654952076677318,
"grad_norm": 0.5185521716853748,
"learning_rate": 1.0931424864807624e-05,
"loss": 0.4278,
"step": 245
},
{
"epoch": 1.571884984025559,
"grad_norm": 0.48800280720450917,
"learning_rate": 1.0857101866346953e-05,
"loss": 0.3713,
"step": 246
},
{
"epoch": 1.5782747603833864,
"grad_norm": 0.6095490026665179,
"learning_rate": 1.0782731140683786e-05,
"loss": 0.4746,
"step": 247
},
{
"epoch": 1.5846645367412142,
"grad_norm": 0.44928616025284457,
"learning_rate": 1.0708316829107295e-05,
"loss": 0.3991,
"step": 248
},
{
"epoch": 1.5910543130990416,
"grad_norm": 0.527002704565493,
"learning_rate": 1.0633863075333713e-05,
"loss": 0.4499,
"step": 249
},
{
"epoch": 1.5974440894568689,
"grad_norm": 0.5073694564198865,
"learning_rate": 1.0559374025275597e-05,
"loss": 0.4451,
"step": 250
},
{
"epoch": 1.6038338658146964,
"grad_norm": 0.43991601364268823,
"learning_rate": 1.0484853826810942e-05,
"loss": 0.3585,
"step": 251
},
{
"epoch": 1.610223642172524,
"grad_norm": 0.5793726292991741,
"learning_rate": 1.0410306629552231e-05,
"loss": 0.4504,
"step": 252
},
{
"epoch": 1.6166134185303513,
"grad_norm": 0.5109980782065752,
"learning_rate": 1.0335736584615357e-05,
"loss": 0.4323,
"step": 253
},
{
"epoch": 1.623003194888179,
"grad_norm": 0.5541438519921948,
"learning_rate": 1.0261147844388472e-05,
"loss": 0.4108,
"step": 254
},
{
"epoch": 1.6293929712460065,
"grad_norm": 0.49521914364363934,
"learning_rate": 1.0186544562300766e-05,
"loss": 0.3646,
"step": 255
},
{
"epoch": 1.6357827476038338,
"grad_norm": 0.5712895311131073,
"learning_rate": 1.011193089259118e-05,
"loss": 0.4436,
"step": 256
},
{
"epoch": 1.6421725239616614,
"grad_norm": 0.4759908770222787,
"learning_rate": 1.0037310990077083e-05,
"loss": 0.4114,
"step": 257
},
{
"epoch": 1.648562300319489,
"grad_norm": 0.5916396180237442,
"learning_rate": 9.962689009922918e-06,
"loss": 0.4351,
"step": 258
},
{
"epoch": 1.6549520766773163,
"grad_norm": 0.5121135747386429,
"learning_rate": 9.888069107408824e-06,
"loss": 0.4054,
"step": 259
},
{
"epoch": 1.6613418530351438,
"grad_norm": 0.5138320946770819,
"learning_rate": 9.813455437699238e-06,
"loss": 0.3825,
"step": 260
},
{
"epoch": 1.6677316293929714,
"grad_norm": 0.6049983944001385,
"learning_rate": 9.738852155611531e-06,
"loss": 0.4499,
"step": 261
},
{
"epoch": 1.6741214057507987,
"grad_norm": 0.5783031442949135,
"learning_rate": 9.664263415384644e-06,
"loss": 0.4481,
"step": 262
},
{
"epoch": 1.680511182108626,
"grad_norm": 0.5079212196836366,
"learning_rate": 9.589693370447769e-06,
"loss": 0.4215,
"step": 263
},
{
"epoch": 1.6869009584664538,
"grad_norm": 0.5362072676088528,
"learning_rate": 9.515146173189058e-06,
"loss": 0.382,
"step": 264
},
{
"epoch": 1.6932907348242812,
"grad_norm": 0.5939862572487508,
"learning_rate": 9.440625974724408e-06,
"loss": 0.4359,
"step": 265
},
{
"epoch": 1.6996805111821085,
"grad_norm": 0.509251006459368,
"learning_rate": 9.366136924666289e-06,
"loss": 0.4187,
"step": 266
},
{
"epoch": 1.706070287539936,
"grad_norm": 0.5360988896841928,
"learning_rate": 9.291683170892712e-06,
"loss": 0.4397,
"step": 267
},
{
"epoch": 1.7124600638977636,
"grad_norm": 0.4969550304271485,
"learning_rate": 9.217268859316219e-06,
"loss": 0.4203,
"step": 268
},
{
"epoch": 1.718849840255591,
"grad_norm": 0.5191197891519107,
"learning_rate": 9.142898133653049e-06,
"loss": 0.3609,
"step": 269
},
{
"epoch": 1.7252396166134185,
"grad_norm": 0.4714057411588144,
"learning_rate": 9.068575135192377e-06,
"loss": 0.4465,
"step": 270
},
{
"epoch": 1.731629392971246,
"grad_norm": 0.4652015886422851,
"learning_rate": 8.994304002565723e-06,
"loss": 0.4304,
"step": 271
},
{
"epoch": 1.7380191693290734,
"grad_norm": 0.5542538380015932,
"learning_rate": 8.920088871516482e-06,
"loss": 0.4401,
"step": 272
},
{
"epoch": 1.744408945686901,
"grad_norm": 0.5308379584880095,
"learning_rate": 8.845933874669645e-06,
"loss": 0.4408,
"step": 273
},
{
"epoch": 1.7507987220447285,
"grad_norm": 0.5930183810947137,
"learning_rate": 8.771843141301659e-06,
"loss": 0.4715,
"step": 274
},
{
"epoch": 1.7571884984025559,
"grad_norm": 0.5544444716982725,
"learning_rate": 8.697820797110499e-06,
"loss": 0.4054,
"step": 275
},
{
"epoch": 1.7635782747603834,
"grad_norm": 0.5332926586542652,
"learning_rate": 8.62387096398593e-06,
"loss": 0.4075,
"step": 276
},
{
"epoch": 1.769968051118211,
"grad_norm": 0.46783125273604637,
"learning_rate": 8.549997759779981e-06,
"loss": 0.383,
"step": 277
},
{
"epoch": 1.7763578274760383,
"grad_norm": 0.5521530233312871,
"learning_rate": 8.47620529807765e-06,
"loss": 0.4533,
"step": 278
},
{
"epoch": 1.7827476038338657,
"grad_norm": 0.5736693711263415,
"learning_rate": 8.402497687967837e-06,
"loss": 0.4041,
"step": 279
},
{
"epoch": 1.7891373801916934,
"grad_norm": 0.5178503013621817,
"learning_rate": 8.328879033814516e-06,
"loss": 0.397,
"step": 280
},
{
"epoch": 1.7955271565495208,
"grad_norm": 0.5005614658444079,
"learning_rate": 8.255353435028228e-06,
"loss": 0.432,
"step": 281
},
{
"epoch": 1.8019169329073481,
"grad_norm": 0.466589550480172,
"learning_rate": 8.181924985837762e-06,
"loss": 0.3921,
"step": 282
},
{
"epoch": 1.8083067092651757,
"grad_norm": 0.492210299786068,
"learning_rate": 8.108597775062198e-06,
"loss": 0.3952,
"step": 283
},
{
"epoch": 1.8146964856230032,
"grad_norm": 0.5488180113006943,
"learning_rate": 8.035375885883217e-06,
"loss": 0.4721,
"step": 284
},
{
"epoch": 1.8210862619808306,
"grad_norm": 0.46872330736771156,
"learning_rate": 7.962263395617724e-06,
"loss": 0.4027,
"step": 285
},
{
"epoch": 1.8274760383386581,
"grad_norm": 0.5921155853332334,
"learning_rate": 7.88926437549082e-06,
"loss": 0.4784,
"step": 286
},
{
"epoch": 1.8338658146964857,
"grad_norm": 0.44822680013304916,
"learning_rate": 7.81638289040908e-06,
"loss": 0.3745,
"step": 287
},
{
"epoch": 1.840255591054313,
"grad_norm": 0.5070334055575236,
"learning_rate": 7.743622998734217e-06,
"loss": 0.4493,
"step": 288
},
{
"epoch": 1.8466453674121406,
"grad_norm": 0.45829357905039364,
"learning_rate": 7.670988752057088e-06,
"loss": 0.3906,
"step": 289
},
{
"epoch": 1.8530351437699681,
"grad_norm": 0.4401957393457272,
"learning_rate": 7.598484194972076e-06,
"loss": 0.4351,
"step": 290
},
{
"epoch": 1.8594249201277955,
"grad_norm": 0.45173410773764394,
"learning_rate": 7.526113364851891e-06,
"loss": 0.404,
"step": 291
},
{
"epoch": 1.865814696485623,
"grad_norm": 0.48332838338770834,
"learning_rate": 7.453880291622726e-06,
"loss": 0.39,
"step": 292
},
{
"epoch": 1.8722044728434506,
"grad_norm": 0.49775533365807295,
"learning_rate": 7.381788997539869e-06,
"loss": 0.4263,
"step": 293
},
{
"epoch": 1.878594249201278,
"grad_norm": 0.45957391883705206,
"learning_rate": 7.309843496963715e-06,
"loss": 0.362,
"step": 294
},
{
"epoch": 1.8849840255591053,
"grad_norm": 0.5317384668186387,
"learning_rate": 7.238047796136247e-06,
"loss": 0.4593,
"step": 295
},
{
"epoch": 1.891373801916933,
"grad_norm": 0.4766410704227857,
"learning_rate": 7.166405892957926e-06,
"loss": 0.3991,
"step": 296
},
{
"epoch": 1.8977635782747604,
"grad_norm": 0.4468475803781737,
"learning_rate": 7.094921776765095e-06,
"loss": 0.4128,
"step": 297
},
{
"epoch": 1.9041533546325877,
"grad_norm": 0.5191666964746032,
"learning_rate": 7.023599428107815e-06,
"loss": 0.457,
"step": 298
},
{
"epoch": 1.9105431309904153,
"grad_norm": 0.44496722949584133,
"learning_rate": 6.95244281852822e-06,
"loss": 0.3908,
"step": 299
},
{
"epoch": 1.9169329073482428,
"grad_norm": 0.4827518065178973,
"learning_rate": 6.881455910339369e-06,
"loss": 0.4515,
"step": 300
},
{
"epoch": 1.9233226837060702,
"grad_norm": 0.5066972710057507,
"learning_rate": 6.8106426564045965e-06,
"loss": 0.4084,
"step": 301
},
{
"epoch": 1.9297124600638977,
"grad_norm": 0.468517851202735,
"learning_rate": 6.740006999917406e-06,
"loss": 0.4292,
"step": 302
},
{
"epoch": 1.9361022364217253,
"grad_norm": 0.5620319742163961,
"learning_rate": 6.669552874181888e-06,
"loss": 0.4285,
"step": 303
},
{
"epoch": 1.9424920127795526,
"grad_norm": 0.46212229948639855,
"learning_rate": 6.599284202393709e-06,
"loss": 0.4009,
"step": 304
},
{
"epoch": 1.9488817891373802,
"grad_norm": 0.4618831752703247,
"learning_rate": 6.529204897421644e-06,
"loss": 0.471,
"step": 305
},
{
"epoch": 1.9552715654952078,
"grad_norm": 0.4453523115078921,
"learning_rate": 6.4593188615896855e-06,
"loss": 0.4149,
"step": 306
},
{
"epoch": 1.961661341853035,
"grad_norm": 0.5565537423113835,
"learning_rate": 6.389629986459756e-06,
"loss": 0.4595,
"step": 307
},
{
"epoch": 1.9680511182108626,
"grad_norm": 0.4817525299898893,
"learning_rate": 6.3201421526149945e-06,
"loss": 0.351,
"step": 308
},
{
"epoch": 1.9744408945686902,
"grad_norm": 0.49529548437523874,
"learning_rate": 6.250859229443684e-06,
"loss": 0.4219,
"step": 309
},
{
"epoch": 1.9808306709265175,
"grad_norm": 0.4677244453368462,
"learning_rate": 6.181785074923778e-06,
"loss": 0.4204,
"step": 310
},
{
"epoch": 1.9872204472843449,
"grad_norm": 0.5533811435739016,
"learning_rate": 6.112923535408074e-06,
"loss": 0.4005,
"step": 311
},
{
"epoch": 1.9936102236421727,
"grad_norm": 0.6829408529085472,
"learning_rate": 6.044278445410025e-06,
"loss": 0.457,
"step": 312
},
{
"epoch": 2.0,
"grad_norm": 0.7425708478665342,
"learning_rate": 5.975853627390233e-06,
"loss": 0.5428,
"step": 313
},
{
"epoch": 2.0063897763578273,
"grad_norm": 0.6185014849673437,
"learning_rate": 5.907652891543576e-06,
"loss": 0.3446,
"step": 314
},
{
"epoch": 2.012779552715655,
"grad_norm": 0.550249298005834,
"learning_rate": 5.839680035587061e-06,
"loss": 0.3567,
"step": 315
},
{
"epoch": 2.0191693290734825,
"grad_norm": 0.6406389855268264,
"learning_rate": 5.771938844548331e-06,
"loss": 0.3439,
"step": 316
},
{
"epoch": 2.02555910543131,
"grad_norm": 0.7498793828860592,
"learning_rate": 5.704433090554912e-06,
"loss": 0.3525,
"step": 317
},
{
"epoch": 2.0319488817891376,
"grad_norm": 0.6058522485872244,
"learning_rate": 5.6371665326241635e-06,
"loss": 0.3493,
"step": 318
},
{
"epoch": 2.038338658146965,
"grad_norm": 0.4684273692024481,
"learning_rate": 5.570142916453944e-06,
"loss": 0.3529,
"step": 319
},
{
"epoch": 2.0447284345047922,
"grad_norm": 0.5736798457088337,
"learning_rate": 5.503365974214059e-06,
"loss": 0.3601,
"step": 320
},
{
"epoch": 2.0511182108626196,
"grad_norm": 0.6933501821794038,
"learning_rate": 5.436839424338426e-06,
"loss": 0.3495,
"step": 321
},
{
"epoch": 2.0575079872204474,
"grad_norm": 0.5220030865474038,
"learning_rate": 5.37056697131799e-06,
"loss": 0.3313,
"step": 322
},
{
"epoch": 2.0638977635782747,
"grad_norm": 0.48932445764444477,
"learning_rate": 5.304552305494493e-06,
"loss": 0.3429,
"step": 323
},
{
"epoch": 2.070287539936102,
"grad_norm": 0.5367328609042773,
"learning_rate": 5.238799102854941e-06,
"loss": 0.3631,
"step": 324
},
{
"epoch": 2.07667731629393,
"grad_norm": 0.4753482664474989,
"learning_rate": 5.173311024826916e-06,
"loss": 0.3164,
"step": 325
},
{
"epoch": 2.083067092651757,
"grad_norm": 0.5857488280714297,
"learning_rate": 5.108091718074706e-06,
"loss": 0.3684,
"step": 326
},
{
"epoch": 2.0894568690095845,
"grad_norm": 0.5577030340771072,
"learning_rate": 5.043144814296214e-06,
"loss": 0.3393,
"step": 327
},
{
"epoch": 2.0958466453674123,
"grad_norm": 0.5001282088257198,
"learning_rate": 4.9784739300207675e-06,
"loss": 0.3289,
"step": 328
},
{
"epoch": 2.1022364217252396,
"grad_norm": 0.49155760192395415,
"learning_rate": 4.914082666407705e-06,
"loss": 0.3126,
"step": 329
},
{
"epoch": 2.108626198083067,
"grad_norm": 0.4721435423436379,
"learning_rate": 4.849974609045849e-06,
"loss": 0.3412,
"step": 330
},
{
"epoch": 2.1150159744408947,
"grad_norm": 0.4887160051279591,
"learning_rate": 4.786153327753865e-06,
"loss": 0.344,
"step": 331
},
{
"epoch": 2.121405750798722,
"grad_norm": 0.4325584192349116,
"learning_rate": 4.722622376381455e-06,
"loss": 0.3316,
"step": 332
},
{
"epoch": 2.1277955271565494,
"grad_norm": 0.4294996718579918,
"learning_rate": 4.659385292611479e-06,
"loss": 0.3339,
"step": 333
},
{
"epoch": 2.134185303514377,
"grad_norm": 0.41116978589443015,
"learning_rate": 4.59644559776296e-06,
"loss": 0.3325,
"step": 334
},
{
"epoch": 2.1405750798722045,
"grad_norm": 0.4411913227153377,
"learning_rate": 4.533806796594989e-06,
"loss": 0.342,
"step": 335
},
{
"epoch": 2.146964856230032,
"grad_norm": 0.44530187962960605,
"learning_rate": 4.471472377111574e-06,
"loss": 0.364,
"step": 336
},
{
"epoch": 2.1533546325878596,
"grad_norm": 0.4681980949823063,
"learning_rate": 4.409445810367421e-06,
"loss": 0.3749,
"step": 337
},
{
"epoch": 2.159744408945687,
"grad_norm": 0.4082359456108807,
"learning_rate": 4.347730550274628e-06,
"loss": 0.3586,
"step": 338
},
{
"epoch": 2.1661341853035143,
"grad_norm": 0.41583681737409806,
"learning_rate": 4.286330033410384e-06,
"loss": 0.3047,
"step": 339
},
{
"epoch": 2.1725239616613417,
"grad_norm": 0.4540111795301057,
"learning_rate": 4.2252476788255735e-06,
"loss": 0.3254,
"step": 340
},
{
"epoch": 2.1789137380191694,
"grad_norm": 0.45433063555043707,
"learning_rate": 4.164486887854424e-06,
"loss": 0.3459,
"step": 341
},
{
"epoch": 2.1853035143769968,
"grad_norm": 0.43892759194821435,
"learning_rate": 4.104051043925068e-06,
"loss": 0.354,
"step": 342
},
{
"epoch": 2.191693290734824,
"grad_norm": 0.43626351327017776,
"learning_rate": 4.043943512371171e-06,
"loss": 0.3484,
"step": 343
},
{
"epoch": 2.198083067092652,
"grad_norm": 0.41866320314755173,
"learning_rate": 3.984167640244518e-06,
"loss": 0.3595,
"step": 344
},
{
"epoch": 2.2044728434504792,
"grad_norm": 0.43171890711936267,
"learning_rate": 3.924726756128632e-06,
"loss": 0.3683,
"step": 345
},
{
"epoch": 2.2108626198083066,
"grad_norm": 0.39657734240201464,
"learning_rate": 3.8656241699534396e-06,
"loss": 0.3319,
"step": 346
},
{
"epoch": 2.2172523961661343,
"grad_norm": 0.4390716578058822,
"learning_rate": 3.8068631728109364e-06,
"loss": 0.3718,
"step": 347
},
{
"epoch": 2.2236421725239617,
"grad_norm": 0.43526087538819297,
"learning_rate": 3.7484470367719493e-06,
"loss": 0.3586,
"step": 348
},
{
"epoch": 2.230031948881789,
"grad_norm": 0.41150940940323066,
"learning_rate": 3.6903790147039286e-06,
"loss": 0.3219,
"step": 349
},
{
"epoch": 2.236421725239617,
"grad_norm": 0.4434399201886998,
"learning_rate": 3.6326623400897797e-06,
"loss": 0.3496,
"step": 350
},
{
"epoch": 2.242811501597444,
"grad_norm": 0.4226836887457401,
"learning_rate": 3.575300226847863e-06,
"loss": 0.3369,
"step": 351
},
{
"epoch": 2.2492012779552715,
"grad_norm": 0.449011167578069,
"learning_rate": 3.5182958691529945e-06,
"loss": 0.3456,
"step": 352
},
{
"epoch": 2.255591054313099,
"grad_norm": 0.40337825402760535,
"learning_rate": 3.4616524412585797e-06,
"loss": 0.3569,
"step": 353
},
{
"epoch": 2.2619808306709266,
"grad_norm": 0.4417946658839781,
"learning_rate": 3.405373097319875e-06,
"loss": 0.2965,
"step": 354
},
{
"epoch": 2.268370607028754,
"grad_norm": 0.4226400522886019,
"learning_rate": 3.3494609712183323e-06,
"loss": 0.3143,
"step": 355
},
{
"epoch": 2.2747603833865817,
"grad_norm": 0.42990286288068097,
"learning_rate": 3.2939191763871047e-06,
"loss": 0.325,
"step": 356
},
{
"epoch": 2.281150159744409,
"grad_norm": 0.48668783932063814,
"learning_rate": 3.2387508056376726e-06,
"loss": 0.3393,
"step": 357
},
{
"epoch": 2.2875399361022364,
"grad_norm": 0.45993768274586794,
"learning_rate": 3.183958930987612e-06,
"loss": 0.3271,
"step": 358
},
{
"epoch": 2.2939297124600637,
"grad_norm": 0.3696554039297162,
"learning_rate": 3.1295466034895482e-06,
"loss": 0.356,
"step": 359
},
{
"epoch": 2.3003194888178915,
"grad_norm": 0.41320713175521795,
"learning_rate": 3.0755168530612444e-06,
"loss": 0.321,
"step": 360
},
{
"epoch": 2.306709265175719,
"grad_norm": 0.42742117582586886,
"learning_rate": 3.021872688316896e-06,
"loss": 0.3161,
"step": 361
},
{
"epoch": 2.313099041533546,
"grad_norm": 0.41256761998572333,
"learning_rate": 2.968617096399592e-06,
"loss": 0.3095,
"step": 362
},
{
"epoch": 2.319488817891374,
"grad_norm": 0.418828546634445,
"learning_rate": 2.9157530428149683e-06,
"loss": 0.3434,
"step": 363
},
{
"epoch": 2.3258785942492013,
"grad_norm": 0.40161025260603606,
"learning_rate": 2.8632834712660882e-06,
"loss": 0.3203,
"step": 364
},
{
"epoch": 2.3322683706070286,
"grad_norm": 0.4182135730630846,
"learning_rate": 2.8112113034895273e-06,
"loss": 0.3135,
"step": 365
},
{
"epoch": 2.3386581469648564,
"grad_norm": 0.40613861239190874,
"learning_rate": 2.7595394390926557e-06,
"loss": 0.3643,
"step": 366
},
{
"epoch": 2.3450479233226837,
"grad_norm": 0.41302826534326115,
"learning_rate": 2.708270755392207e-06,
"loss": 0.3424,
"step": 367
},
{
"epoch": 2.351437699680511,
"grad_norm": 0.3930172226452873,
"learning_rate": 2.657408107254027e-06,
"loss": 0.3429,
"step": 368
},
{
"epoch": 2.357827476038339,
"grad_norm": 0.41867946397201244,
"learning_rate": 2.60695432693412e-06,
"loss": 0.3657,
"step": 369
},
{
"epoch": 2.364217252396166,
"grad_norm": 0.3952956485859315,
"learning_rate": 2.5569122239209366e-06,
"loss": 0.2999,
"step": 370
},
{
"epoch": 2.3706070287539935,
"grad_norm": 0.38547686501477857,
"learning_rate": 2.507284584778913e-06,
"loss": 0.3229,
"step": 371
},
{
"epoch": 2.376996805111821,
"grad_norm": 0.4071024941451391,
"learning_rate": 2.4580741729933246e-06,
"loss": 0.3398,
"step": 372
},
{
"epoch": 2.3833865814696487,
"grad_norm": 0.3975504941590724,
"learning_rate": 2.4092837288163807e-06,
"loss": 0.3246,
"step": 373
},
{
"epoch": 2.389776357827476,
"grad_norm": 0.4007814100724813,
"learning_rate": 2.3609159691146577e-06,
"loss": 0.3834,
"step": 374
},
{
"epoch": 2.3961661341853033,
"grad_norm": 0.4132465009460835,
"learning_rate": 2.312973587217798e-06,
"loss": 0.3552,
"step": 375
},
{
"epoch": 2.402555910543131,
"grad_norm": 0.38136466401080477,
"learning_rate": 2.2654592527685305e-06,
"loss": 0.3394,
"step": 376
},
{
"epoch": 2.4089456869009584,
"grad_norm": 0.38378352088164336,
"learning_rate": 2.2183756115740274e-06,
"loss": 0.3417,
"step": 377
},
{
"epoch": 2.415335463258786,
"grad_norm": 0.42705803790306746,
"learning_rate": 2.171725285458559e-06,
"loss": 0.3657,
"step": 378
},
{
"epoch": 2.4217252396166136,
"grad_norm": 0.3893534985584009,
"learning_rate": 2.1255108721175066e-06,
"loss": 0.3439,
"step": 379
},
{
"epoch": 2.428115015974441,
"grad_norm": 0.443399810426828,
"learning_rate": 2.079734944972717e-06,
"loss": 0.3216,
"step": 380
},
{
"epoch": 2.4345047923322682,
"grad_norm": 0.39874922783339656,
"learning_rate": 2.0344000530291875e-06,
"loss": 0.3434,
"step": 381
},
{
"epoch": 2.440894568690096,
"grad_norm": 0.3771297738742276,
"learning_rate": 1.9895087207331422e-06,
"loss": 0.3409,
"step": 382
},
{
"epoch": 2.4472843450479234,
"grad_norm": 0.40027082321568275,
"learning_rate": 1.945063447831452e-06,
"loss": 0.3265,
"step": 383
},
{
"epoch": 2.4536741214057507,
"grad_norm": 0.3737581626959035,
"learning_rate": 1.9010667092324342e-06,
"loss": 0.34,
"step": 384
},
{
"epoch": 2.460063897763578,
"grad_norm": 0.3841031529537794,
"learning_rate": 1.8575209548680472e-06,
"loss": 0.3503,
"step": 385
},
{
"epoch": 2.466453674121406,
"grad_norm": 0.4038855857442773,
"learning_rate": 1.814428609557458e-06,
"loss": 0.3201,
"step": 386
},
{
"epoch": 2.472843450479233,
"grad_norm": 0.40823279792933154,
"learning_rate": 1.7717920728720284e-06,
"loss": 0.3219,
"step": 387
},
{
"epoch": 2.479233226837061,
"grad_norm": 0.45571229187170187,
"learning_rate": 1.7296137190016916e-06,
"loss": 0.3646,
"step": 388
},
{
"epoch": 2.4856230031948883,
"grad_norm": 0.3992759518176683,
"learning_rate": 1.6878958966227366e-06,
"loss": 0.3514,
"step": 389
},
{
"epoch": 2.4920127795527156,
"grad_norm": 0.44997582744977993,
"learning_rate": 1.646640928767047e-06,
"loss": 0.3676,
"step": 390
},
{
"epoch": 2.498402555910543,
"grad_norm": 0.4036812441443117,
"learning_rate": 1.6058511126927178e-06,
"loss": 0.3572,
"step": 391
},
{
"epoch": 2.5047923322683707,
"grad_norm": 0.38861999088805715,
"learning_rate": 1.5655287197561497e-06,
"loss": 0.3176,
"step": 392
},
{
"epoch": 2.511182108626198,
"grad_norm": 0.3958987299112533,
"learning_rate": 1.5256759952855737e-06,
"loss": 0.3411,
"step": 393
},
{
"epoch": 2.5175718849840254,
"grad_norm": 0.3958900572864119,
"learning_rate": 1.4862951584560037e-06,
"loss": 0.3391,
"step": 394
},
{
"epoch": 2.523961661341853,
"grad_norm": 0.3918198479937304,
"learning_rate": 1.447388402165686e-06,
"loss": 0.3178,
"step": 395
},
{
"epoch": 2.5303514376996805,
"grad_norm": 0.40756583772176336,
"learning_rate": 1.4089578929139635e-06,
"loss": 0.3264,
"step": 396
},
{
"epoch": 2.536741214057508,
"grad_norm": 0.40960058637131896,
"learning_rate": 1.371005770680659e-06,
"loss": 0.3385,
"step": 397
},
{
"epoch": 2.543130990415335,
"grad_norm": 0.40903897126869604,
"learning_rate": 1.3335341488068997e-06,
"loss": 0.3805,
"step": 398
},
{
"epoch": 2.549520766773163,
"grad_norm": 0.39992597114439177,
"learning_rate": 1.2965451138774343e-06,
"loss": 0.3196,
"step": 399
},
{
"epoch": 2.5559105431309903,
"grad_norm": 0.3894861786588525,
"learning_rate": 1.2600407256044544e-06,
"loss": 0.354,
"step": 400
},
{
"epoch": 2.562300319488818,
"grad_norm": 0.3773260960595135,
"learning_rate": 1.2240230167128863e-06,
"loss": 0.3408,
"step": 401
},
{
"epoch": 2.5686900958466454,
"grad_norm": 0.3916021706010166,
"learning_rate": 1.188493992827211e-06,
"loss": 0.3141,
"step": 402
},
{
"epoch": 2.5750798722044728,
"grad_norm": 0.3906836314398566,
"learning_rate": 1.1534556323597824e-06,
"loss": 0.2924,
"step": 403
},
{
"epoch": 2.5814696485623,
"grad_norm": 0.4038749879309955,
"learning_rate": 1.1189098864006488e-06,
"loss": 0.3365,
"step": 404
},
{
"epoch": 2.587859424920128,
"grad_norm": 0.40279549296487077,
"learning_rate": 1.084858678608922e-06,
"loss": 0.3239,
"step": 405
},
{
"epoch": 2.594249201277955,
"grad_norm": 0.36365947820301786,
"learning_rate": 1.0513039051056507e-06,
"loss": 0.3326,
"step": 406
},
{
"epoch": 2.600638977635783,
"grad_norm": 0.3696286967249599,
"learning_rate": 1.0182474343682346e-06,
"loss": 0.3292,
"step": 407
},
{
"epoch": 2.6070287539936103,
"grad_norm": 0.4175231953214467,
"learning_rate": 9.85691107126392e-07,
"loss": 0.3744,
"step": 408
},
{
"epoch": 2.6134185303514377,
"grad_norm": 0.39195370451842787,
"learning_rate": 9.536367362596422e-07,
"loss": 0.3695,
"step": 409
},
{
"epoch": 2.619808306709265,
"grad_norm": 0.37605332595306007,
"learning_rate": 9.220861066963715e-07,
"loss": 0.3418,
"step": 410
},
{
"epoch": 2.626198083067093,
"grad_norm": 0.3995977702745196,
"learning_rate": 8.910409753144344e-07,
"loss": 0.3549,
"step": 411
},
{
"epoch": 2.63258785942492,
"grad_norm": 0.3834229012458761,
"learning_rate": 8.605030708433149e-07,
"loss": 0.3445,
"step": 412
},
{
"epoch": 2.6389776357827475,
"grad_norm": 0.37644074285723944,
"learning_rate": 8.304740937678835e-07,
"loss": 0.3668,
"step": 413
},
{
"epoch": 2.6453674121405752,
"grad_norm": 0.38149522660404983,
"learning_rate": 8.009557162336823e-07,
"loss": 0.3286,
"step": 414
},
{
"epoch": 2.6517571884984026,
"grad_norm": 0.3965348662763524,
"learning_rate": 7.719495819538325e-07,
"loss": 0.3562,
"step": 415
},
{
"epoch": 2.65814696485623,
"grad_norm": 0.3740820985387306,
"learning_rate": 7.434573061174966e-07,
"loss": 0.3171,
"step": 416
},
{
"epoch": 2.6645367412140573,
"grad_norm": 0.37288133480679564,
"learning_rate": 7.154804752999344e-07,
"loss": 0.3263,
"step": 417
},
{
"epoch": 2.670926517571885,
"grad_norm": 0.37023486110236786,
"learning_rate": 6.880206473741646e-07,
"loss": 0.368,
"step": 418
},
{
"epoch": 2.6773162939297124,
"grad_norm": 0.3924590139873951,
"learning_rate": 6.610793514242075e-07,
"loss": 0.3348,
"step": 419
},
{
"epoch": 2.68370607028754,
"grad_norm": 0.4026773365921225,
"learning_rate": 6.346580876599395e-07,
"loss": 0.3405,
"step": 420
},
{
"epoch": 2.6900958466453675,
"grad_norm": 0.3855444929570631,
"learning_rate": 6.08758327333564e-07,
"loss": 0.3588,
"step": 421
},
{
"epoch": 2.696485623003195,
"grad_norm": 0.4018291966488895,
"learning_rate": 5.833815126576714e-07,
"loss": 0.3259,
"step": 422
},
{
"epoch": 2.702875399361022,
"grad_norm": 0.36389835567175366,
"learning_rate": 5.585290567249424e-07,
"loss": 0.3514,
"step": 423
},
{
"epoch": 2.70926517571885,
"grad_norm": 0.40458156951380986,
"learning_rate": 5.342023434294552e-07,
"loss": 0.337,
"step": 424
},
{
"epoch": 2.7156549520766773,
"grad_norm": 0.36807317654930927,
"learning_rate": 5.104027273896239e-07,
"loss": 0.3148,
"step": 425
},
{
"epoch": 2.722044728434505,
"grad_norm": 0.3689483940462552,
"learning_rate": 4.871315338727711e-07,
"loss": 0.3181,
"step": 426
},
{
"epoch": 2.7284345047923324,
"grad_norm": 0.3833457229816993,
"learning_rate": 4.6439005872132457e-07,
"loss": 0.3368,
"step": 427
},
{
"epoch": 2.7348242811501597,
"grad_norm": 0.36604943775735366,
"learning_rate": 4.421795682806662e-07,
"loss": 0.3364,
"step": 428
},
{
"epoch": 2.741214057507987,
"grad_norm": 0.3773811652537217,
"learning_rate": 4.2050129932861394e-07,
"loss": 0.3229,
"step": 429
},
{
"epoch": 2.747603833865815,
"grad_norm": 0.3796116399398149,
"learning_rate": 3.9935645900654906e-07,
"loss": 0.3633,
"step": 430
},
{
"epoch": 2.753993610223642,
"grad_norm": 0.37803460451421467,
"learning_rate": 3.7874622475220336e-07,
"loss": 0.3524,
"step": 431
},
{
"epoch": 2.7603833865814695,
"grad_norm": 0.36675174735385174,
"learning_rate": 3.58671744234087e-07,
"loss": 0.365,
"step": 432
},
{
"epoch": 2.7667731629392973,
"grad_norm": 0.3898919999844523,
"learning_rate": 3.3913413528758877e-07,
"loss": 0.3323,
"step": 433
},
{
"epoch": 2.7731629392971247,
"grad_norm": 0.3936657666166714,
"learning_rate": 3.2013448585272333e-07,
"loss": 0.3337,
"step": 434
},
{
"epoch": 2.779552715654952,
"grad_norm": 0.38046637386341803,
"learning_rate": 3.016738539135566e-07,
"loss": 0.344,
"step": 435
},
{
"epoch": 2.7859424920127793,
"grad_norm": 0.3770525251793597,
"learning_rate": 2.837532674392862e-07,
"loss": 0.374,
"step": 436
},
{
"epoch": 2.792332268370607,
"grad_norm": 0.3916319385244186,
"learning_rate": 2.6637372432700483e-07,
"loss": 0.3492,
"step": 437
},
{
"epoch": 2.7987220447284344,
"grad_norm": 0.4083221537063404,
"learning_rate": 2.4953619234612816e-07,
"loss": 0.3056,
"step": 438
},
{
"epoch": 2.8051118210862622,
"grad_norm": 0.36900391336674243,
"learning_rate": 2.332416090845102e-07,
"loss": 0.3071,
"step": 439
},
{
"epoch": 2.8115015974440896,
"grad_norm": 0.4181996987699531,
"learning_rate": 2.1749088189622848e-07,
"loss": 0.3351,
"step": 440
},
{
"epoch": 2.817891373801917,
"grad_norm": 0.3547619175700847,
"learning_rate": 2.0228488785106636e-07,
"loss": 0.3406,
"step": 441
},
{
"epoch": 2.8242811501597442,
"grad_norm": 0.3977218103469981,
"learning_rate": 1.8762447368566582e-07,
"loss": 0.3505,
"step": 442
},
{
"epoch": 2.830670926517572,
"grad_norm": 0.39003543295538634,
"learning_rate": 1.7351045575638047e-07,
"loss": 0.3753,
"step": 443
},
{
"epoch": 2.8370607028753994,
"grad_norm": 0.37845868272626343,
"learning_rate": 1.5994361999382202e-07,
"loss": 0.3278,
"step": 444
},
{
"epoch": 2.8434504792332267,
"grad_norm": 0.39842394374774287,
"learning_rate": 1.4692472185908635e-07,
"loss": 0.3682,
"step": 445
},
{
"epoch": 2.8498402555910545,
"grad_norm": 0.3457049068963664,
"learning_rate": 1.344544863016961e-07,
"loss": 0.3438,
"step": 446
},
{
"epoch": 2.856230031948882,
"grad_norm": 0.40463187861502226,
"learning_rate": 1.225336077192274e-07,
"loss": 0.3356,
"step": 447
},
{
"epoch": 2.862619808306709,
"grad_norm": 0.5473639905587789,
"learning_rate": 1.1116274991864073e-07,
"loss": 0.3612,
"step": 448
},
{
"epoch": 2.8690095846645365,
"grad_norm": 0.36738048359514414,
"learning_rate": 1.003425460793217e-07,
"loss": 0.3618,
"step": 449
},
{
"epoch": 2.8753993610223643,
"grad_norm": 0.39397443170606444,
"learning_rate": 9.00735987178214e-08,
"loss": 0.3325,
"step": 450
},
{
"epoch": 2.8817891373801916,
"grad_norm": 0.3641877860232395,
"learning_rate": 8.035647965430215e-08,
"loss": 0.3705,
"step": 451
},
{
"epoch": 2.8881789137380194,
"grad_norm": 0.3766602850517281,
"learning_rate": 7.119172998070412e-08,
"loss": 0.3602,
"step": 452
},
{
"epoch": 2.8945686900958467,
"grad_norm": 0.3504672529434601,
"learning_rate": 6.25798600306049e-08,
"loss": 0.345,
"step": 453
},
{
"epoch": 2.900958466453674,
"grad_norm": 0.3689087024207003,
"learning_rate": 5.4521349350808996e-08,
"loss": 0.3711,
"step": 454
},
{
"epoch": 2.9073482428115014,
"grad_norm": 0.3816320750179523,
"learning_rate": 4.701664667464245e-08,
"loss": 0.3646,
"step": 455
},
{
"epoch": 2.913738019169329,
"grad_norm": 0.3679437431040902,
"learning_rate": 4.006616989696621e-08,
"loss": 0.3615,
"step": 456
},
{
"epoch": 2.9201277955271565,
"grad_norm": 0.38571927082544183,
"learning_rate": 3.367030605090249e-08,
"loss": 0.3222,
"step": 457
},
{
"epoch": 2.9265175718849843,
"grad_norm": 0.3971237004311223,
"learning_rate": 2.7829411286287577e-08,
"loss": 0.3485,
"step": 458
},
{
"epoch": 2.9329073482428116,
"grad_norm": 0.3630064489830963,
"learning_rate": 2.254381084983659e-08,
"loss": 0.3065,
"step": 459
},
{
"epoch": 2.939297124600639,
"grad_norm": 0.3754819920105193,
"learning_rate": 1.781379906703573e-08,
"loss": 0.36,
"step": 460
},
{
"epoch": 2.9456869009584663,
"grad_norm": 0.37643335933636624,
"learning_rate": 1.3639639325748744e-08,
"loss": 0.3725,
"step": 461
},
{
"epoch": 2.952076677316294,
"grad_norm": 0.397354621784988,
"learning_rate": 1.0021564061554189e-08,
"loss": 0.3378,
"step": 462
},
{
"epoch": 2.9584664536741214,
"grad_norm": 0.37424179888551773,
"learning_rate": 6.959774744796921e-09,
"loss": 0.3399,
"step": 463
},
{
"epoch": 2.9648562300319488,
"grad_norm": 0.37190360281608004,
"learning_rate": 4.454441869377047e-09,
"loss": 0.3219,
"step": 464
},
{
"epoch": 2.9712460063897765,
"grad_norm": 0.3657736795626134,
"learning_rate": 2.5057049432519744e-09,
"loss": 0.3705,
"step": 465
},
{
"epoch": 2.977635782747604,
"grad_norm": 0.35303669900268125,
"learning_rate": 1.1136724806637411e-09,
"loss": 0.3318,
"step": 466
},
{
"epoch": 2.984025559105431,
"grad_norm": 0.36854160936154184,
"learning_rate": 2.784219961060597e-10,
"loss": 0.3518,
"step": 467
},
{
"epoch": 2.9904153354632586,
"grad_norm": 0.3989983000303115,
"learning_rate": 0.0,
"loss": 0.3546,
"step": 468
},
{
"epoch": 2.9904153354632586,
"step": 468,
"total_flos": 3.310734730054861e+17,
"train_loss": 0.4434524042229367,
"train_runtime": 10205.5973,
"train_samples_per_second": 2.939,
"train_steps_per_second": 0.046
}
],
"logging_steps": 1.0,
"max_steps": 468,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.310734730054861e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}