{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9904153354632586, "eval_steps": 500, "global_step": 468, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006389776357827476, "grad_norm": 5.939312845608103, "learning_rate": 4.2553191489361704e-07, "loss": 0.8883, "step": 1 }, { "epoch": 0.012779552715654952, "grad_norm": 5.6459943274568385, "learning_rate": 8.510638297872341e-07, "loss": 0.8365, "step": 2 }, { "epoch": 0.019169329073482427, "grad_norm": 6.290489665408837, "learning_rate": 1.276595744680851e-06, "loss": 0.8864, "step": 3 }, { "epoch": 0.025559105431309903, "grad_norm": 5.8544636560570735, "learning_rate": 1.7021276595744682e-06, "loss": 0.879, "step": 4 }, { "epoch": 0.03194888178913738, "grad_norm": 5.478170251266386, "learning_rate": 2.1276595744680853e-06, "loss": 0.8245, "step": 5 }, { "epoch": 0.038338658146964855, "grad_norm": 4.688336713722404, "learning_rate": 2.553191489361702e-06, "loss": 0.849, "step": 6 }, { "epoch": 0.04472843450479233, "grad_norm": 4.278233500376253, "learning_rate": 2.978723404255319e-06, "loss": 0.8151, "step": 7 }, { "epoch": 0.051118210862619806, "grad_norm": 2.367029075074496, "learning_rate": 3.4042553191489363e-06, "loss": 0.7566, "step": 8 }, { "epoch": 0.05750798722044728, "grad_norm": 2.0707237905380484, "learning_rate": 3.8297872340425535e-06, "loss": 0.7551, "step": 9 }, { "epoch": 0.06389776357827476, "grad_norm": 1.7866725538131587, "learning_rate": 4.255319148936171e-06, "loss": 0.7457, "step": 10 }, { "epoch": 0.07028753993610223, "grad_norm": 3.404477977029493, "learning_rate": 4.680851063829788e-06, "loss": 0.699, "step": 11 }, { "epoch": 0.07667731629392971, "grad_norm": 3.8759087344215093, "learning_rate": 5.106382978723404e-06, "loss": 0.8106, "step": 12 }, { "epoch": 0.08306709265175719, "grad_norm": 4.206484510948959, "learning_rate": 5.531914893617022e-06, "loss": 0.7639, "step": 13 }, { "epoch": 0.08945686900958466, "grad_norm": 3.597855342218176, "learning_rate": 5.957446808510638e-06, "loss": 0.7143, "step": 14 }, { "epoch": 0.09584664536741214, "grad_norm": 3.493223809169224, "learning_rate": 6.382978723404256e-06, "loss": 0.7495, "step": 15 }, { "epoch": 0.10223642172523961, "grad_norm": 2.7730036124678588, "learning_rate": 6.808510638297873e-06, "loss": 0.7079, "step": 16 }, { "epoch": 0.10862619808306709, "grad_norm": 1.99917862715565, "learning_rate": 7.234042553191491e-06, "loss": 0.6637, "step": 17 }, { "epoch": 0.11501597444089456, "grad_norm": 1.7234540912576959, "learning_rate": 7.659574468085107e-06, "loss": 0.6563, "step": 18 }, { "epoch": 0.12140575079872204, "grad_norm": 1.9632219810386513, "learning_rate": 8.085106382978723e-06, "loss": 0.6753, "step": 19 }, { "epoch": 0.12779552715654952, "grad_norm": 2.1614567368781157, "learning_rate": 8.510638297872341e-06, "loss": 0.6063, "step": 20 }, { "epoch": 0.134185303514377, "grad_norm": 1.932507047332365, "learning_rate": 8.936170212765958e-06, "loss": 0.629, "step": 21 }, { "epoch": 0.14057507987220447, "grad_norm": 1.5481507427216847, "learning_rate": 9.361702127659576e-06, "loss": 0.6444, "step": 22 }, { "epoch": 0.14696485623003194, "grad_norm": 1.2177329752055572, "learning_rate": 9.787234042553192e-06, "loss": 0.641, "step": 23 }, { "epoch": 0.15335463258785942, "grad_norm": 1.4206725631658237, "learning_rate": 1.0212765957446808e-05, "loss": 0.6496, "step": 24 }, { "epoch": 0.1597444089456869, "grad_norm": 1.4667519530084758, "learning_rate": 1.0638297872340426e-05, "loss": 0.6304, "step": 25 }, { "epoch": 0.16613418530351437, "grad_norm": 1.2452708214307595, "learning_rate": 1.1063829787234044e-05, "loss": 0.6413, "step": 26 }, { "epoch": 0.17252396166134185, "grad_norm": 0.7759167145812482, "learning_rate": 1.1489361702127662e-05, "loss": 0.5927, "step": 27 }, { "epoch": 0.17891373801916932, "grad_norm": 1.0321850071310563, "learning_rate": 1.1914893617021277e-05, "loss": 0.5898, "step": 28 }, { "epoch": 0.1853035143769968, "grad_norm": 1.0266703283162124, "learning_rate": 1.2340425531914895e-05, "loss": 0.5895, "step": 29 }, { "epoch": 0.19169329073482427, "grad_norm": 0.9674925840183656, "learning_rate": 1.2765957446808513e-05, "loss": 0.5853, "step": 30 }, { "epoch": 0.19808306709265175, "grad_norm": 0.9805374530460067, "learning_rate": 1.3191489361702127e-05, "loss": 0.5756, "step": 31 }, { "epoch": 0.20447284345047922, "grad_norm": 0.9258462587522315, "learning_rate": 1.3617021276595745e-05, "loss": 0.585, "step": 32 }, { "epoch": 0.2108626198083067, "grad_norm": 0.8925757909283792, "learning_rate": 1.4042553191489363e-05, "loss": 0.584, "step": 33 }, { "epoch": 0.21725239616613418, "grad_norm": 0.7738950177471798, "learning_rate": 1.4468085106382981e-05, "loss": 0.553, "step": 34 }, { "epoch": 0.22364217252396165, "grad_norm": 0.823063478968403, "learning_rate": 1.4893617021276596e-05, "loss": 0.605, "step": 35 }, { "epoch": 0.23003194888178913, "grad_norm": 0.7769591687125452, "learning_rate": 1.5319148936170214e-05, "loss": 0.5398, "step": 36 }, { "epoch": 0.2364217252396166, "grad_norm": 0.724085749329525, "learning_rate": 1.5744680851063832e-05, "loss": 0.5541, "step": 37 }, { "epoch": 0.24281150159744408, "grad_norm": 0.7937699708520545, "learning_rate": 1.6170212765957446e-05, "loss": 0.5516, "step": 38 }, { "epoch": 0.24920127795527156, "grad_norm": 0.838915436660047, "learning_rate": 1.6595744680851064e-05, "loss": 0.5335, "step": 39 }, { "epoch": 0.25559105431309903, "grad_norm": 0.7447841608522099, "learning_rate": 1.7021276595744682e-05, "loss": 0.5414, "step": 40 }, { "epoch": 0.26198083067092653, "grad_norm": 0.8714110715774589, "learning_rate": 1.74468085106383e-05, "loss": 0.556, "step": 41 }, { "epoch": 0.268370607028754, "grad_norm": 0.7411083091257032, "learning_rate": 1.7872340425531915e-05, "loss": 0.5113, "step": 42 }, { "epoch": 0.2747603833865815, "grad_norm": 0.7680242229988357, "learning_rate": 1.8297872340425533e-05, "loss": 0.5684, "step": 43 }, { "epoch": 0.28115015974440893, "grad_norm": 0.7374046719973165, "learning_rate": 1.872340425531915e-05, "loss": 0.577, "step": 44 }, { "epoch": 0.28753993610223644, "grad_norm": 0.7709945486266545, "learning_rate": 1.914893617021277e-05, "loss": 0.5251, "step": 45 }, { "epoch": 0.2939297124600639, "grad_norm": 0.7420014515431359, "learning_rate": 1.9574468085106384e-05, "loss": 0.5358, "step": 46 }, { "epoch": 0.3003194888178914, "grad_norm": 0.7321704500165183, "learning_rate": 2e-05, "loss": 0.56, "step": 47 }, { "epoch": 0.30670926517571884, "grad_norm": 0.7855274459068936, "learning_rate": 1.9999721578003894e-05, "loss": 0.5427, "step": 48 }, { "epoch": 0.31309904153354634, "grad_norm": 0.7731901448013743, "learning_rate": 1.9998886327519337e-05, "loss": 0.5517, "step": 49 }, { "epoch": 0.3194888178913738, "grad_norm": 0.7327331638939495, "learning_rate": 1.999749429505675e-05, "loss": 0.5468, "step": 50 }, { "epoch": 0.3258785942492013, "grad_norm": 0.7419176272941636, "learning_rate": 1.9995545558130624e-05, "loss": 0.548, "step": 51 }, { "epoch": 0.33226837060702874, "grad_norm": 0.7873113572745489, "learning_rate": 1.9993040225255205e-05, "loss": 0.5403, "step": 52 }, { "epoch": 0.33865814696485624, "grad_norm": 0.6658693818077654, "learning_rate": 1.998997843593845e-05, "loss": 0.5054, "step": 53 }, { "epoch": 0.3450479233226837, "grad_norm": 0.7409323699660156, "learning_rate": 1.9986360360674252e-05, "loss": 0.5555, "step": 54 }, { "epoch": 0.3514376996805112, "grad_norm": 0.7094024540386773, "learning_rate": 1.9982186200932964e-05, "loss": 0.535, "step": 55 }, { "epoch": 0.35782747603833864, "grad_norm": 0.9538171828161622, "learning_rate": 1.9977456189150164e-05, "loss": 0.5788, "step": 56 }, { "epoch": 0.36421725239616615, "grad_norm": 0.7395956073936416, "learning_rate": 1.9972170588713715e-05, "loss": 0.5836, "step": 57 }, { "epoch": 0.3706070287539936, "grad_norm": 0.8649724098491106, "learning_rate": 1.9966329693949098e-05, "loss": 0.4918, "step": 58 }, { "epoch": 0.3769968051118211, "grad_norm": 0.8838321465136332, "learning_rate": 1.9959933830103034e-05, "loss": 0.514, "step": 59 }, { "epoch": 0.38338658146964855, "grad_norm": 0.7993190296372936, "learning_rate": 1.9952983353325358e-05, "loss": 0.5191, "step": 60 }, { "epoch": 0.38977635782747605, "grad_norm": 1.0003320683454586, "learning_rate": 1.9945478650649192e-05, "loss": 0.5395, "step": 61 }, { "epoch": 0.3961661341853035, "grad_norm": 0.8174378430731664, "learning_rate": 1.9937420139969397e-05, "loss": 0.5428, "step": 62 }, { "epoch": 0.402555910543131, "grad_norm": 0.7497946237416347, "learning_rate": 1.9928808270019297e-05, "loss": 0.5826, "step": 63 }, { "epoch": 0.40894568690095845, "grad_norm": 0.8640015891418493, "learning_rate": 1.9919643520345698e-05, "loss": 0.5343, "step": 64 }, { "epoch": 0.41533546325878595, "grad_norm": 1.0033449675940591, "learning_rate": 1.990992640128218e-05, "loss": 0.5623, "step": 65 }, { "epoch": 0.4217252396166134, "grad_norm": 0.7213392023439573, "learning_rate": 1.989965745392068e-05, "loss": 0.5314, "step": 66 }, { "epoch": 0.4281150159744409, "grad_norm": 0.8863830620435368, "learning_rate": 1.988883725008136e-05, "loss": 0.5458, "step": 67 }, { "epoch": 0.43450479233226835, "grad_norm": 0.7362416113857194, "learning_rate": 1.9877466392280773e-05, "loss": 0.5428, "step": 68 }, { "epoch": 0.44089456869009586, "grad_norm": 0.9788737478764125, "learning_rate": 1.9865545513698304e-05, "loss": 0.5412, "step": 69 }, { "epoch": 0.4472843450479233, "grad_norm": 0.8021931238319318, "learning_rate": 1.9853075278140913e-05, "loss": 0.5522, "step": 70 }, { "epoch": 0.4536741214057508, "grad_norm": 0.85887825826516, "learning_rate": 1.984005638000618e-05, "loss": 0.5702, "step": 71 }, { "epoch": 0.46006389776357826, "grad_norm": 0.7815409304110528, "learning_rate": 1.9826489544243623e-05, "loss": 0.5482, "step": 72 }, { "epoch": 0.46645367412140576, "grad_norm": 0.7632274039193099, "learning_rate": 1.981237552631434e-05, "loss": 0.5387, "step": 73 }, { "epoch": 0.4728434504792332, "grad_norm": 0.7141397681245097, "learning_rate": 1.9797715112148937e-05, "loss": 0.5546, "step": 74 }, { "epoch": 0.4792332268370607, "grad_norm": 0.9775621975797985, "learning_rate": 1.9782509118103773e-05, "loss": 0.5308, "step": 75 }, { "epoch": 0.48562300319488816, "grad_norm": 0.7330287655613912, "learning_rate": 1.9766758390915494e-05, "loss": 0.5431, "step": 76 }, { "epoch": 0.49201277955271566, "grad_norm": 0.8510881793812443, "learning_rate": 1.9750463807653873e-05, "loss": 0.5187, "step": 77 }, { "epoch": 0.4984025559105431, "grad_norm": 0.8486211359598791, "learning_rate": 1.9733626275673e-05, "loss": 0.55, "step": 78 }, { "epoch": 0.5047923322683706, "grad_norm": 0.7806116343240348, "learning_rate": 1.9716246732560715e-05, "loss": 0.4813, "step": 79 }, { "epoch": 0.5111821086261981, "grad_norm": 0.6353268171693698, "learning_rate": 1.9698326146086446e-05, "loss": 0.5078, "step": 80 }, { "epoch": 0.5175718849840255, "grad_norm": 0.8766273238149929, "learning_rate": 1.967986551414728e-05, "loss": 0.544, "step": 81 }, { "epoch": 0.5239616613418531, "grad_norm": 0.7124929308186037, "learning_rate": 1.9660865864712413e-05, "loss": 0.5547, "step": 82 }, { "epoch": 0.5303514376996805, "grad_norm": 0.7985052392042088, "learning_rate": 1.9641328255765916e-05, "loss": 0.5215, "step": 83 }, { "epoch": 0.536741214057508, "grad_norm": 0.6406165534662566, "learning_rate": 1.96212537752478e-05, "loss": 0.4978, "step": 84 }, { "epoch": 0.5431309904153354, "grad_norm": 0.7351154516097609, "learning_rate": 1.9600643540993453e-05, "loss": 0.5498, "step": 85 }, { "epoch": 0.549520766773163, "grad_norm": 0.7023736701907295, "learning_rate": 1.9579498700671386e-05, "loss": 0.5271, "step": 86 }, { "epoch": 0.5559105431309904, "grad_norm": 0.7585882232459317, "learning_rate": 1.9557820431719333e-05, "loss": 0.5419, "step": 87 }, { "epoch": 0.5623003194888179, "grad_norm": 0.7609703576418915, "learning_rate": 1.9535609941278676e-05, "loss": 0.5115, "step": 88 }, { "epoch": 0.5686900958466453, "grad_norm": 0.8355505315864735, "learning_rate": 1.9512868466127232e-05, "loss": 0.5214, "step": 89 }, { "epoch": 0.5750798722044729, "grad_norm": 0.8985531652526637, "learning_rate": 1.9489597272610377e-05, "loss": 0.5217, "step": 90 }, { "epoch": 0.5814696485623003, "grad_norm": 0.8296453033353365, "learning_rate": 1.9465797656570546e-05, "loss": 0.5235, "step": 91 }, { "epoch": 0.5878594249201278, "grad_norm": 0.8400305646818227, "learning_rate": 1.944147094327506e-05, "loss": 0.5028, "step": 92 }, { "epoch": 0.5942492012779552, "grad_norm": 0.8027178786911765, "learning_rate": 1.9416618487342333e-05, "loss": 0.4932, "step": 93 }, { "epoch": 0.6006389776357828, "grad_norm": 0.831774187088593, "learning_rate": 1.9391241672666438e-05, "loss": 0.5015, "step": 94 }, { "epoch": 0.6070287539936102, "grad_norm": 0.6880013029770987, "learning_rate": 1.936534191234006e-05, "loss": 0.4851, "step": 95 }, { "epoch": 0.6134185303514377, "grad_norm": 0.7874317799707009, "learning_rate": 1.9338920648575798e-05, "loss": 0.4993, "step": 96 }, { "epoch": 0.6198083067092651, "grad_norm": 0.8086603460274078, "learning_rate": 1.9311979352625837e-05, "loss": 0.5472, "step": 97 }, { "epoch": 0.6261980830670927, "grad_norm": 0.7330826956678399, "learning_rate": 1.928451952470007e-05, "loss": 0.5267, "step": 98 }, { "epoch": 0.6325878594249201, "grad_norm": 0.6774581525725395, "learning_rate": 1.9256542693882505e-05, "loss": 0.4964, "step": 99 }, { "epoch": 0.6389776357827476, "grad_norm": 0.699096603267482, "learning_rate": 1.922805041804617e-05, "loss": 0.5051, "step": 100 }, { "epoch": 0.645367412140575, "grad_norm": 0.7625712048870313, "learning_rate": 1.919904428376632e-05, "loss": 0.5301, "step": 101 }, { "epoch": 0.6517571884984026, "grad_norm": 0.7442243954348883, "learning_rate": 1.916952590623212e-05, "loss": 0.5114, "step": 102 }, { "epoch": 0.65814696485623, "grad_norm": 0.6319971486530133, "learning_rate": 1.9139496929156685e-05, "loss": 0.5106, "step": 103 }, { "epoch": 0.6645367412140575, "grad_norm": 0.6634321632909197, "learning_rate": 1.910895902468557e-05, "loss": 0.5211, "step": 104 }, { "epoch": 0.670926517571885, "grad_norm": 0.5959863886093508, "learning_rate": 1.907791389330363e-05, "loss": 0.5124, "step": 105 }, { "epoch": 0.6773162939297125, "grad_norm": 0.7995435481771501, "learning_rate": 1.904636326374036e-05, "loss": 0.52, "step": 106 }, { "epoch": 0.6837060702875399, "grad_norm": 0.7383352863257084, "learning_rate": 1.9014308892873612e-05, "loss": 0.5516, "step": 107 }, { "epoch": 0.6900958466453674, "grad_norm": 0.8315686682034763, "learning_rate": 1.8981752565631767e-05, "loss": 0.4891, "step": 108 }, { "epoch": 0.6964856230031949, "grad_norm": 0.6508951908046772, "learning_rate": 1.8948696094894354e-05, "loss": 0.5591, "step": 109 }, { "epoch": 0.7028753993610224, "grad_norm": 0.8009701429333234, "learning_rate": 1.8915141321391083e-05, "loss": 0.5305, "step": 110 }, { "epoch": 0.7092651757188498, "grad_norm": 0.7308231155353043, "learning_rate": 1.8881090113599353e-05, "loss": 0.5191, "step": 111 }, { "epoch": 0.7156549520766773, "grad_norm": 0.7808253591246626, "learning_rate": 1.8846544367640218e-05, "loss": 0.4906, "step": 112 }, { "epoch": 0.7220447284345048, "grad_norm": 0.8760149164716509, "learning_rate": 1.881150600717279e-05, "loss": 0.532, "step": 113 }, { "epoch": 0.7284345047923323, "grad_norm": 0.6776840820965584, "learning_rate": 1.8775976983287117e-05, "loss": 0.521, "step": 114 }, { "epoch": 0.7348242811501597, "grad_norm": 0.8898200759371685, "learning_rate": 1.873995927439555e-05, "loss": 0.5251, "step": 115 }, { "epoch": 0.7412140575079872, "grad_norm": 0.7499195664294598, "learning_rate": 1.8703454886122568e-05, "loss": 0.5348, "step": 116 }, { "epoch": 0.7476038338658147, "grad_norm": 0.8505865320259214, "learning_rate": 1.86664658511931e-05, "loss": 0.5054, "step": 117 }, { "epoch": 0.7539936102236422, "grad_norm": 0.6522705814194051, "learning_rate": 1.862899422931934e-05, "loss": 0.4996, "step": 118 }, { "epoch": 0.7603833865814696, "grad_norm": 0.6577321085300217, "learning_rate": 1.859104210708604e-05, "loss": 0.5106, "step": 119 }, { "epoch": 0.7667731629392971, "grad_norm": 0.6237691862811225, "learning_rate": 1.855261159783432e-05, "loss": 0.4952, "step": 120 }, { "epoch": 0.7731629392971247, "grad_norm": 0.7454463301186427, "learning_rate": 1.8513704841543997e-05, "loss": 0.5026, "step": 121 }, { "epoch": 0.7795527156549521, "grad_norm": 0.7934145741008819, "learning_rate": 1.847432400471443e-05, "loss": 0.549, "step": 122 }, { "epoch": 0.7859424920127795, "grad_norm": 0.6802649460778195, "learning_rate": 1.8434471280243854e-05, "loss": 0.506, "step": 123 }, { "epoch": 0.792332268370607, "grad_norm": 0.7304620814404005, "learning_rate": 1.8394148887307286e-05, "loss": 0.5383, "step": 124 }, { "epoch": 0.7987220447284346, "grad_norm": 0.6957782218041328, "learning_rate": 1.8353359071232954e-05, "loss": 0.5019, "step": 125 }, { "epoch": 0.805111821086262, "grad_norm": 0.6826820301754135, "learning_rate": 1.8312104103377266e-05, "loss": 0.4937, "step": 126 }, { "epoch": 0.8115015974440895, "grad_norm": 0.598906813674708, "learning_rate": 1.827038628099831e-05, "loss": 0.488, "step": 127 }, { "epoch": 0.8178913738019169, "grad_norm": 0.6633449549765212, "learning_rate": 1.822820792712797e-05, "loss": 0.4971, "step": 128 }, { "epoch": 0.8242811501597445, "grad_norm": 0.6710913601976871, "learning_rate": 1.8185571390442542e-05, "loss": 0.4836, "step": 129 }, { "epoch": 0.8306709265175719, "grad_norm": 0.5762592167913351, "learning_rate": 1.8142479045131956e-05, "loss": 0.5192, "step": 130 }, { "epoch": 0.8370607028753994, "grad_norm": 0.6738150719804582, "learning_rate": 1.809893329076757e-05, "loss": 0.5228, "step": 131 }, { "epoch": 0.8434504792332268, "grad_norm": 0.6500141385276138, "learning_rate": 1.8054936552168548e-05, "loss": 0.5332, "step": 132 }, { "epoch": 0.8498402555910544, "grad_norm": 0.5886713445337487, "learning_rate": 1.801049127926686e-05, "loss": 0.4787, "step": 133 }, { "epoch": 0.8562300319488818, "grad_norm": 0.7649082888316344, "learning_rate": 1.7965599946970814e-05, "loss": 0.4914, "step": 134 }, { "epoch": 0.8626198083067093, "grad_norm": 0.8014082101971098, "learning_rate": 1.7920265055027285e-05, "loss": 0.5171, "step": 135 }, { "epoch": 0.8690095846645367, "grad_norm": 0.7706425976997076, "learning_rate": 1.7874489127882496e-05, "loss": 0.5267, "step": 136 }, { "epoch": 0.8753993610223643, "grad_norm": 0.9916919676402002, "learning_rate": 1.7828274714541445e-05, "loss": 0.5225, "step": 137 }, { "epoch": 0.8817891373801917, "grad_norm": 0.7510268200588113, "learning_rate": 1.7781624388425974e-05, "loss": 0.5056, "step": 138 }, { "epoch": 0.8881789137380192, "grad_norm": 1.0622441121039243, "learning_rate": 1.773454074723147e-05, "loss": 0.4773, "step": 139 }, { "epoch": 0.8945686900958466, "grad_norm": 0.8430739649427611, "learning_rate": 1.76870264127822e-05, "loss": 0.5384, "step": 140 }, { "epoch": 0.9009584664536742, "grad_norm": 0.7217579935144164, "learning_rate": 1.763908403088534e-05, "loss": 0.5108, "step": 141 }, { "epoch": 0.9073482428115016, "grad_norm": 1.0377962316213807, "learning_rate": 1.759071627118362e-05, "loss": 0.4961, "step": 142 }, { "epoch": 0.9137380191693291, "grad_norm": 0.6423592379266523, "learning_rate": 1.754192582700668e-05, "loss": 0.5061, "step": 143 }, { "epoch": 0.9201277955271565, "grad_norm": 0.7302473302785821, "learning_rate": 1.7492715415221087e-05, "loss": 0.5074, "step": 144 }, { "epoch": 0.9265175718849841, "grad_norm": 0.709358335589859, "learning_rate": 1.7443087776079068e-05, "loss": 0.4867, "step": 145 }, { "epoch": 0.9329073482428115, "grad_norm": 0.6708691396098521, "learning_rate": 1.739304567306588e-05, "loss": 0.4955, "step": 146 }, { "epoch": 0.939297124600639, "grad_norm": 0.7224681216999947, "learning_rate": 1.7342591892745978e-05, "loss": 0.5063, "step": 147 }, { "epoch": 0.9456869009584664, "grad_norm": 0.692060343253805, "learning_rate": 1.72917292446078e-05, "loss": 0.5395, "step": 148 }, { "epoch": 0.952076677316294, "grad_norm": 0.6503466417528021, "learning_rate": 1.7240460560907345e-05, "loss": 0.502, "step": 149 }, { "epoch": 0.9584664536741214, "grad_norm": 0.7852299377720712, "learning_rate": 1.7188788696510477e-05, "loss": 0.4982, "step": 150 }, { "epoch": 0.9648562300319489, "grad_norm": 0.6487201669982385, "learning_rate": 1.7136716528733916e-05, "loss": 0.491, "step": 151 }, { "epoch": 0.9712460063897763, "grad_norm": 0.7162643954228396, "learning_rate": 1.7084246957185036e-05, "loss": 0.4715, "step": 152 }, { "epoch": 0.9776357827476039, "grad_norm": 0.6882308824929978, "learning_rate": 1.703138290360041e-05, "loss": 0.4884, "step": 153 }, { "epoch": 0.9840255591054313, "grad_norm": 0.6558983320574183, "learning_rate": 1.6978127311683103e-05, "loss": 0.5053, "step": 154 }, { "epoch": 0.9904153354632588, "grad_norm": 0.7793396179424056, "learning_rate": 1.6924483146938756e-05, "loss": 0.4891, "step": 155 }, { "epoch": 0.9968051118210862, "grad_norm": 0.6408866926804856, "learning_rate": 1.6870453396510456e-05, "loss": 0.5253, "step": 156 }, { "epoch": 1.0031948881789137, "grad_norm": 1.219315523607184, "learning_rate": 1.681604106901239e-05, "loss": 0.7727, "step": 157 }, { "epoch": 1.0095846645367412, "grad_norm": 0.7363761093026954, "learning_rate": 1.676124919436233e-05, "loss": 0.4369, "step": 158 }, { "epoch": 1.0159744408945688, "grad_norm": 0.6104584236502094, "learning_rate": 1.6706080823612897e-05, "loss": 0.4074, "step": 159 }, { "epoch": 1.0223642172523961, "grad_norm": 0.6688279966321403, "learning_rate": 1.665053902878167e-05, "loss": 0.3875, "step": 160 }, { "epoch": 1.0287539936102237, "grad_norm": 0.7714442502018793, "learning_rate": 1.659462690268013e-05, "loss": 0.4693, "step": 161 }, { "epoch": 1.035143769968051, "grad_norm": 0.6581337601062037, "learning_rate": 1.6538347558741424e-05, "loss": 0.4088, "step": 162 }, { "epoch": 1.0415335463258786, "grad_norm": 0.8907132608840105, "learning_rate": 1.6481704130847013e-05, "loss": 0.4421, "step": 163 }, { "epoch": 1.0479233226837061, "grad_norm": 0.7281123005456749, "learning_rate": 1.642469977315214e-05, "loss": 0.4466, "step": 164 }, { "epoch": 1.0543130990415335, "grad_norm": 0.8713456380104283, "learning_rate": 1.6367337659910223e-05, "loss": 0.4205, "step": 165 }, { "epoch": 1.060702875399361, "grad_norm": 0.8221934081375099, "learning_rate": 1.6309620985296075e-05, "loss": 0.4811, "step": 166 }, { "epoch": 1.0670926517571886, "grad_norm": 0.6414871893647255, "learning_rate": 1.625155296322805e-05, "loss": 0.3994, "step": 167 }, { "epoch": 1.073482428115016, "grad_norm": 0.6020141646779518, "learning_rate": 1.6193136827189067e-05, "loss": 0.3544, "step": 168 }, { "epoch": 1.0798722044728435, "grad_norm": 0.6620341470772692, "learning_rate": 1.6134375830046566e-05, "loss": 0.3865, "step": 169 }, { "epoch": 1.0862619808306708, "grad_norm": 0.7529504403287007, "learning_rate": 1.607527324387137e-05, "loss": 0.5138, "step": 170 }, { "epoch": 1.0926517571884984, "grad_norm": 0.5746068027715405, "learning_rate": 1.6015832359755483e-05, "loss": 0.4053, "step": 171 }, { "epoch": 1.099041533546326, "grad_norm": 0.6423681462846987, "learning_rate": 1.5956056487628832e-05, "loss": 0.4878, "step": 172 }, { "epoch": 1.1054313099041533, "grad_norm": 0.6192258722203541, "learning_rate": 1.5895948956074937e-05, "loss": 0.4188, "step": 173 }, { "epoch": 1.1118210862619808, "grad_norm": 0.6163929815962566, "learning_rate": 1.5835513112145583e-05, "loss": 0.3854, "step": 174 }, { "epoch": 1.1182108626198084, "grad_norm": 0.5697582130826278, "learning_rate": 1.5774752321174428e-05, "loss": 0.4073, "step": 175 }, { "epoch": 1.1246006389776357, "grad_norm": 0.6223679982213549, "learning_rate": 1.571366996658962e-05, "loss": 0.4224, "step": 176 }, { "epoch": 1.1309904153354633, "grad_norm": 0.6320866995001497, "learning_rate": 1.5652269449725375e-05, "loss": 0.4248, "step": 177 }, { "epoch": 1.1373801916932909, "grad_norm": 0.5297458143137096, "learning_rate": 1.5590554189632585e-05, "loss": 0.3748, "step": 178 }, { "epoch": 1.1437699680511182, "grad_norm": 0.6358623803264093, "learning_rate": 1.552852762288843e-05, "loss": 0.4747, "step": 179 }, { "epoch": 1.1501597444089458, "grad_norm": 0.6278673693951984, "learning_rate": 1.5466193203405017e-05, "loss": 0.4036, "step": 180 }, { "epoch": 1.156549520766773, "grad_norm": 0.6151559404690766, "learning_rate": 1.540355440223704e-05, "loss": 0.4124, "step": 181 }, { "epoch": 1.1629392971246006, "grad_norm": 0.5891168526046718, "learning_rate": 1.534061470738852e-05, "loss": 0.4637, "step": 182 }, { "epoch": 1.1693290734824282, "grad_norm": 0.6184989510741438, "learning_rate": 1.527737762361855e-05, "loss": 0.4161, "step": 183 }, { "epoch": 1.1757188498402555, "grad_norm": 0.6419175983144096, "learning_rate": 1.5213846672246139e-05, "loss": 0.4533, "step": 184 }, { "epoch": 1.182108626198083, "grad_norm": 0.5120065057835093, "learning_rate": 1.5150025390954153e-05, "loss": 0.375, "step": 185 }, { "epoch": 1.1884984025559104, "grad_norm": 0.698299656514935, "learning_rate": 1.5085917333592299e-05, "loss": 0.4193, "step": 186 }, { "epoch": 1.194888178913738, "grad_norm": 0.6532253196374253, "learning_rate": 1.5021526069979232e-05, "loss": 0.4596, "step": 187 }, { "epoch": 1.2012779552715656, "grad_norm": 0.6028992706390495, "learning_rate": 1.4956855185703787e-05, "loss": 0.4207, "step": 188 }, { "epoch": 1.207667731629393, "grad_norm": 0.6185875205322425, "learning_rate": 1.48919082819253e-05, "loss": 0.4243, "step": 189 }, { "epoch": 1.2140575079872205, "grad_norm": 0.6043057260440954, "learning_rate": 1.4826688975173085e-05, "loss": 0.437, "step": 190 }, { "epoch": 1.220447284345048, "grad_norm": 0.6821203725825057, "learning_rate": 1.4761200897145063e-05, "loss": 0.5037, "step": 191 }, { "epoch": 1.2268370607028753, "grad_norm": 0.6008298374460067, "learning_rate": 1.4695447694505512e-05, "loss": 0.3661, "step": 192 }, { "epoch": 1.233226837060703, "grad_norm": 0.7199503911011816, "learning_rate": 1.4629433028682014e-05, "loss": 0.5043, "step": 193 }, { "epoch": 1.2396166134185305, "grad_norm": 0.5180806806501965, "learning_rate": 1.456316057566158e-05, "loss": 0.3819, "step": 194 }, { "epoch": 1.2460063897763578, "grad_norm": 0.6928136290488852, "learning_rate": 1.4496634025785938e-05, "loss": 0.4263, "step": 195 }, { "epoch": 1.2523961661341854, "grad_norm": 0.5665537937081578, "learning_rate": 1.4429857083546054e-05, "loss": 0.4844, "step": 196 }, { "epoch": 1.2587859424920127, "grad_norm": 0.5563376996661965, "learning_rate": 1.4362833467375839e-05, "loss": 0.4156, "step": 197 }, { "epoch": 1.2651757188498403, "grad_norm": 0.5377783142222045, "learning_rate": 1.429556690944509e-05, "loss": 0.4025, "step": 198 }, { "epoch": 1.2715654952076676, "grad_norm": 0.5413290657480134, "learning_rate": 1.4228061155451671e-05, "loss": 0.4066, "step": 199 }, { "epoch": 1.2779552715654952, "grad_norm": 0.5854137012981868, "learning_rate": 1.4160319964412943e-05, "loss": 0.4631, "step": 200 }, { "epoch": 1.2843450479233227, "grad_norm": 0.5549643478979397, "learning_rate": 1.4092347108456425e-05, "loss": 0.4203, "step": 201 }, { "epoch": 1.29073482428115, "grad_norm": 0.6090412261860937, "learning_rate": 1.402414637260977e-05, "loss": 0.4818, "step": 202 }, { "epoch": 1.2971246006389776, "grad_norm": 0.5907663028777652, "learning_rate": 1.3955721554589979e-05, "loss": 0.4253, "step": 203 }, { "epoch": 1.3035143769968052, "grad_norm": 0.6715293474231077, "learning_rate": 1.388707646459193e-05, "loss": 0.418, "step": 204 }, { "epoch": 1.3099041533546325, "grad_norm": 0.6297284099203906, "learning_rate": 1.3818214925076226e-05, "loss": 0.4147, "step": 205 }, { "epoch": 1.31629392971246, "grad_norm": 0.5667392659435537, "learning_rate": 1.3749140770556322e-05, "loss": 0.4311, "step": 206 }, { "epoch": 1.3226837060702876, "grad_norm": 0.6322500298227558, "learning_rate": 1.367985784738501e-05, "loss": 0.4203, "step": 207 }, { "epoch": 1.329073482428115, "grad_norm": 0.6001231519749656, "learning_rate": 1.361037001354025e-05, "loss": 0.4476, "step": 208 }, { "epoch": 1.3354632587859425, "grad_norm": 0.5824816964719087, "learning_rate": 1.3540681138410317e-05, "loss": 0.3966, "step": 209 }, { "epoch": 1.34185303514377, "grad_norm": 0.6931574895039062, "learning_rate": 1.3470795102578358e-05, "loss": 0.5033, "step": 210 }, { "epoch": 1.3482428115015974, "grad_norm": 0.5416798455387101, "learning_rate": 1.3400715797606293e-05, "loss": 0.4008, "step": 211 }, { "epoch": 1.354632587859425, "grad_norm": 0.6524420733651487, "learning_rate": 1.3330447125818115e-05, "loss": 0.4436, "step": 212 }, { "epoch": 1.3610223642172525, "grad_norm": 0.6109732592249346, "learning_rate": 1.3259993000082599e-05, "loss": 0.4139, "step": 213 }, { "epoch": 1.3674121405750799, "grad_norm": 0.6089156914272807, "learning_rate": 1.3189357343595405e-05, "loss": 0.4287, "step": 214 }, { "epoch": 1.3738019169329074, "grad_norm": 0.5100478253771064, "learning_rate": 1.3118544089660635e-05, "loss": 0.3773, "step": 215 }, { "epoch": 1.3801916932907348, "grad_norm": 0.665636694875197, "learning_rate": 1.3047557181471784e-05, "loss": 0.3922, "step": 216 }, { "epoch": 1.3865814696485623, "grad_norm": 0.6421315985345839, "learning_rate": 1.2976400571892189e-05, "loss": 0.4432, "step": 217 }, { "epoch": 1.3929712460063897, "grad_norm": 0.5630313538184247, "learning_rate": 1.2905078223234907e-05, "loss": 0.4191, "step": 218 }, { "epoch": 1.3993610223642172, "grad_norm": 0.6745179299688657, "learning_rate": 1.2833594107042078e-05, "loss": 0.4333, "step": 219 }, { "epoch": 1.4057507987220448, "grad_norm": 0.5454374317298349, "learning_rate": 1.2761952203863759e-05, "loss": 0.4277, "step": 220 }, { "epoch": 1.4121405750798721, "grad_norm": 0.5789898470817271, "learning_rate": 1.2690156503036288e-05, "loss": 0.4036, "step": 221 }, { "epoch": 1.4185303514376997, "grad_norm": 0.5949332555607629, "learning_rate": 1.2618211002460135e-05, "loss": 0.4785, "step": 222 }, { "epoch": 1.4249201277955272, "grad_norm": 0.5587473061257787, "learning_rate": 1.2546119708377273e-05, "loss": 0.4109, "step": 223 }, { "epoch": 1.4313099041533546, "grad_norm": 0.5092986860891325, "learning_rate": 1.2473886635148109e-05, "loss": 0.4375, "step": 224 }, { "epoch": 1.4376996805111821, "grad_norm": 0.49324268916930714, "learning_rate": 1.2401515805027924e-05, "loss": 0.4339, "step": 225 }, { "epoch": 1.4440894568690097, "grad_norm": 0.521973180886398, "learning_rate": 1.2329011247942915e-05, "loss": 0.3929, "step": 226 }, { "epoch": 1.450479233226837, "grad_norm": 0.49449177011483986, "learning_rate": 1.2256377001265785e-05, "loss": 0.3688, "step": 227 }, { "epoch": 1.4568690095846646, "grad_norm": 0.688803254610615, "learning_rate": 1.2183617109590925e-05, "loss": 0.5054, "step": 228 }, { "epoch": 1.4632587859424921, "grad_norm": 0.5222053229962778, "learning_rate": 1.2110735624509184e-05, "loss": 0.3946, "step": 229 }, { "epoch": 1.4696485623003195, "grad_norm": 0.5032950143354372, "learning_rate": 1.2037736604382279e-05, "loss": 0.4158, "step": 230 }, { "epoch": 1.476038338658147, "grad_norm": 0.5739740428574489, "learning_rate": 1.1964624114116784e-05, "loss": 0.3794, "step": 231 }, { "epoch": 1.4824281150159744, "grad_norm": 0.6710884805647371, "learning_rate": 1.1891402224937805e-05, "loss": 0.466, "step": 232 }, { "epoch": 1.488817891373802, "grad_norm": 0.4582879331386117, "learning_rate": 1.1818075014162243e-05, "loss": 0.3585, "step": 233 }, { "epoch": 1.4952076677316293, "grad_norm": 0.5448073298830566, "learning_rate": 1.1744646564971777e-05, "loss": 0.4238, "step": 234 }, { "epoch": 1.5015974440894568, "grad_norm": 0.5760693930458392, "learning_rate": 1.1671120966185486e-05, "loss": 0.4324, "step": 235 }, { "epoch": 1.5079872204472844, "grad_norm": 0.49993740420542576, "learning_rate": 1.159750231203217e-05, "loss": 0.4078, "step": 236 }, { "epoch": 1.5143769968051117, "grad_norm": 0.563168928495261, "learning_rate": 1.1523794701922351e-05, "loss": 0.4351, "step": 237 }, { "epoch": 1.5207667731629393, "grad_norm": 0.5313983639628916, "learning_rate": 1.145000224022002e-05, "loss": 0.3571, "step": 238 }, { "epoch": 1.5271565495207668, "grad_norm": 0.501524364831029, "learning_rate": 1.1376129036014073e-05, "loss": 0.4304, "step": 239 }, { "epoch": 1.5335463258785942, "grad_norm": 0.44709921144798975, "learning_rate": 1.1302179202889505e-05, "loss": 0.3948, "step": 240 }, { "epoch": 1.5399361022364217, "grad_norm": 0.6024736966395187, "learning_rate": 1.1228156858698344e-05, "loss": 0.5055, "step": 241 }, { "epoch": 1.5463258785942493, "grad_norm": 0.5039190418971187, "learning_rate": 1.1154066125330358e-05, "loss": 0.3853, "step": 242 }, { "epoch": 1.5527156549520766, "grad_norm": 0.5082416278400913, "learning_rate": 1.107991112848352e-05, "loss": 0.365, "step": 243 }, { "epoch": 1.5591054313099042, "grad_norm": 0.571424017024242, "learning_rate": 1.100569599743428e-05, "loss": 0.4157, "step": 244 }, { "epoch": 1.5654952076677318, "grad_norm": 0.5185521716853748, "learning_rate": 1.0931424864807624e-05, "loss": 0.4278, "step": 245 }, { "epoch": 1.571884984025559, "grad_norm": 0.48800280720450917, "learning_rate": 1.0857101866346953e-05, "loss": 0.3713, "step": 246 }, { "epoch": 1.5782747603833864, "grad_norm": 0.6095490026665179, "learning_rate": 1.0782731140683786e-05, "loss": 0.4746, "step": 247 }, { "epoch": 1.5846645367412142, "grad_norm": 0.44928616025284457, "learning_rate": 1.0708316829107295e-05, "loss": 0.3991, "step": 248 }, { "epoch": 1.5910543130990416, "grad_norm": 0.527002704565493, "learning_rate": 1.0633863075333713e-05, "loss": 0.4499, "step": 249 }, { "epoch": 1.5974440894568689, "grad_norm": 0.5073694564198865, "learning_rate": 1.0559374025275597e-05, "loss": 0.4451, "step": 250 }, { "epoch": 1.6038338658146964, "grad_norm": 0.43991601364268823, "learning_rate": 1.0484853826810942e-05, "loss": 0.3585, "step": 251 }, { "epoch": 1.610223642172524, "grad_norm": 0.5793726292991741, "learning_rate": 1.0410306629552231e-05, "loss": 0.4504, "step": 252 }, { "epoch": 1.6166134185303513, "grad_norm": 0.5109980782065752, "learning_rate": 1.0335736584615357e-05, "loss": 0.4323, "step": 253 }, { "epoch": 1.623003194888179, "grad_norm": 0.5541438519921948, "learning_rate": 1.0261147844388472e-05, "loss": 0.4108, "step": 254 }, { "epoch": 1.6293929712460065, "grad_norm": 0.49521914364363934, "learning_rate": 1.0186544562300766e-05, "loss": 0.3646, "step": 255 }, { "epoch": 1.6357827476038338, "grad_norm": 0.5712895311131073, "learning_rate": 1.011193089259118e-05, "loss": 0.4436, "step": 256 }, { "epoch": 1.6421725239616614, "grad_norm": 0.4759908770222787, "learning_rate": 1.0037310990077083e-05, "loss": 0.4114, "step": 257 }, { "epoch": 1.648562300319489, "grad_norm": 0.5916396180237442, "learning_rate": 9.962689009922918e-06, "loss": 0.4351, "step": 258 }, { "epoch": 1.6549520766773163, "grad_norm": 0.5121135747386429, "learning_rate": 9.888069107408824e-06, "loss": 0.4054, "step": 259 }, { "epoch": 1.6613418530351438, "grad_norm": 0.5138320946770819, "learning_rate": 9.813455437699238e-06, "loss": 0.3825, "step": 260 }, { "epoch": 1.6677316293929714, "grad_norm": 0.6049983944001385, "learning_rate": 9.738852155611531e-06, "loss": 0.4499, "step": 261 }, { "epoch": 1.6741214057507987, "grad_norm": 0.5783031442949135, "learning_rate": 9.664263415384644e-06, "loss": 0.4481, "step": 262 }, { "epoch": 1.680511182108626, "grad_norm": 0.5079212196836366, "learning_rate": 9.589693370447769e-06, "loss": 0.4215, "step": 263 }, { "epoch": 1.6869009584664538, "grad_norm": 0.5362072676088528, "learning_rate": 9.515146173189058e-06, "loss": 0.382, "step": 264 }, { "epoch": 1.6932907348242812, "grad_norm": 0.5939862572487508, "learning_rate": 9.440625974724408e-06, "loss": 0.4359, "step": 265 }, { "epoch": 1.6996805111821085, "grad_norm": 0.509251006459368, "learning_rate": 9.366136924666289e-06, "loss": 0.4187, "step": 266 }, { "epoch": 1.706070287539936, "grad_norm": 0.5360988896841928, "learning_rate": 9.291683170892712e-06, "loss": 0.4397, "step": 267 }, { "epoch": 1.7124600638977636, "grad_norm": 0.4969550304271485, "learning_rate": 9.217268859316219e-06, "loss": 0.4203, "step": 268 }, { "epoch": 1.718849840255591, "grad_norm": 0.5191197891519107, "learning_rate": 9.142898133653049e-06, "loss": 0.3609, "step": 269 }, { "epoch": 1.7252396166134185, "grad_norm": 0.4714057411588144, "learning_rate": 9.068575135192377e-06, "loss": 0.4465, "step": 270 }, { "epoch": 1.731629392971246, "grad_norm": 0.4652015886422851, "learning_rate": 8.994304002565723e-06, "loss": 0.4304, "step": 271 }, { "epoch": 1.7380191693290734, "grad_norm": 0.5542538380015932, "learning_rate": 8.920088871516482e-06, "loss": 0.4401, "step": 272 }, { "epoch": 1.744408945686901, "grad_norm": 0.5308379584880095, "learning_rate": 8.845933874669645e-06, "loss": 0.4408, "step": 273 }, { "epoch": 1.7507987220447285, "grad_norm": 0.5930183810947137, "learning_rate": 8.771843141301659e-06, "loss": 0.4715, "step": 274 }, { "epoch": 1.7571884984025559, "grad_norm": 0.5544444716982725, "learning_rate": 8.697820797110499e-06, "loss": 0.4054, "step": 275 }, { "epoch": 1.7635782747603834, "grad_norm": 0.5332926586542652, "learning_rate": 8.62387096398593e-06, "loss": 0.4075, "step": 276 }, { "epoch": 1.769968051118211, "grad_norm": 0.46783125273604637, "learning_rate": 8.549997759779981e-06, "loss": 0.383, "step": 277 }, { "epoch": 1.7763578274760383, "grad_norm": 0.5521530233312871, "learning_rate": 8.47620529807765e-06, "loss": 0.4533, "step": 278 }, { "epoch": 1.7827476038338657, "grad_norm": 0.5736693711263415, "learning_rate": 8.402497687967837e-06, "loss": 0.4041, "step": 279 }, { "epoch": 1.7891373801916934, "grad_norm": 0.5178503013621817, "learning_rate": 8.328879033814516e-06, "loss": 0.397, "step": 280 }, { "epoch": 1.7955271565495208, "grad_norm": 0.5005614658444079, "learning_rate": 8.255353435028228e-06, "loss": 0.432, "step": 281 }, { "epoch": 1.8019169329073481, "grad_norm": 0.466589550480172, "learning_rate": 8.181924985837762e-06, "loss": 0.3921, "step": 282 }, { "epoch": 1.8083067092651757, "grad_norm": 0.492210299786068, "learning_rate": 8.108597775062198e-06, "loss": 0.3952, "step": 283 }, { "epoch": 1.8146964856230032, "grad_norm": 0.5488180113006943, "learning_rate": 8.035375885883217e-06, "loss": 0.4721, "step": 284 }, { "epoch": 1.8210862619808306, "grad_norm": 0.46872330736771156, "learning_rate": 7.962263395617724e-06, "loss": 0.4027, "step": 285 }, { "epoch": 1.8274760383386581, "grad_norm": 0.5921155853332334, "learning_rate": 7.88926437549082e-06, "loss": 0.4784, "step": 286 }, { "epoch": 1.8338658146964857, "grad_norm": 0.44822680013304916, "learning_rate": 7.81638289040908e-06, "loss": 0.3745, "step": 287 }, { "epoch": 1.840255591054313, "grad_norm": 0.5070334055575236, "learning_rate": 7.743622998734217e-06, "loss": 0.4493, "step": 288 }, { "epoch": 1.8466453674121406, "grad_norm": 0.45829357905039364, "learning_rate": 7.670988752057088e-06, "loss": 0.3906, "step": 289 }, { "epoch": 1.8530351437699681, "grad_norm": 0.4401957393457272, "learning_rate": 7.598484194972076e-06, "loss": 0.4351, "step": 290 }, { "epoch": 1.8594249201277955, "grad_norm": 0.45173410773764394, "learning_rate": 7.526113364851891e-06, "loss": 0.404, "step": 291 }, { "epoch": 1.865814696485623, "grad_norm": 0.48332838338770834, "learning_rate": 7.453880291622726e-06, "loss": 0.39, "step": 292 }, { "epoch": 1.8722044728434506, "grad_norm": 0.49775533365807295, "learning_rate": 7.381788997539869e-06, "loss": 0.4263, "step": 293 }, { "epoch": 1.878594249201278, "grad_norm": 0.45957391883705206, "learning_rate": 7.309843496963715e-06, "loss": 0.362, "step": 294 }, { "epoch": 1.8849840255591053, "grad_norm": 0.5317384668186387, "learning_rate": 7.238047796136247e-06, "loss": 0.4593, "step": 295 }, { "epoch": 1.891373801916933, "grad_norm": 0.4766410704227857, "learning_rate": 7.166405892957926e-06, "loss": 0.3991, "step": 296 }, { "epoch": 1.8977635782747604, "grad_norm": 0.4468475803781737, "learning_rate": 7.094921776765095e-06, "loss": 0.4128, "step": 297 }, { "epoch": 1.9041533546325877, "grad_norm": 0.5191666964746032, "learning_rate": 7.023599428107815e-06, "loss": 0.457, "step": 298 }, { "epoch": 1.9105431309904153, "grad_norm": 0.44496722949584133, "learning_rate": 6.95244281852822e-06, "loss": 0.3908, "step": 299 }, { "epoch": 1.9169329073482428, "grad_norm": 0.4827518065178973, "learning_rate": 6.881455910339369e-06, "loss": 0.4515, "step": 300 }, { "epoch": 1.9233226837060702, "grad_norm": 0.5066972710057507, "learning_rate": 6.8106426564045965e-06, "loss": 0.4084, "step": 301 }, { "epoch": 1.9297124600638977, "grad_norm": 0.468517851202735, "learning_rate": 6.740006999917406e-06, "loss": 0.4292, "step": 302 }, { "epoch": 1.9361022364217253, "grad_norm": 0.5620319742163961, "learning_rate": 6.669552874181888e-06, "loss": 0.4285, "step": 303 }, { "epoch": 1.9424920127795526, "grad_norm": 0.46212229948639855, "learning_rate": 6.599284202393709e-06, "loss": 0.4009, "step": 304 }, { "epoch": 1.9488817891373802, "grad_norm": 0.4618831752703247, "learning_rate": 6.529204897421644e-06, "loss": 0.471, "step": 305 }, { "epoch": 1.9552715654952078, "grad_norm": 0.4453523115078921, "learning_rate": 6.4593188615896855e-06, "loss": 0.4149, "step": 306 }, { "epoch": 1.961661341853035, "grad_norm": 0.5565537423113835, "learning_rate": 6.389629986459756e-06, "loss": 0.4595, "step": 307 }, { "epoch": 1.9680511182108626, "grad_norm": 0.4817525299898893, "learning_rate": 6.3201421526149945e-06, "loss": 0.351, "step": 308 }, { "epoch": 1.9744408945686902, "grad_norm": 0.49529548437523874, "learning_rate": 6.250859229443684e-06, "loss": 0.4219, "step": 309 }, { "epoch": 1.9808306709265175, "grad_norm": 0.4677244453368462, "learning_rate": 6.181785074923778e-06, "loss": 0.4204, "step": 310 }, { "epoch": 1.9872204472843449, "grad_norm": 0.5533811435739016, "learning_rate": 6.112923535408074e-06, "loss": 0.4005, "step": 311 }, { "epoch": 1.9936102236421727, "grad_norm": 0.6829408529085472, "learning_rate": 6.044278445410025e-06, "loss": 0.457, "step": 312 }, { "epoch": 2.0, "grad_norm": 0.7425708478665342, "learning_rate": 5.975853627390233e-06, "loss": 0.5428, "step": 313 }, { "epoch": 2.0063897763578273, "grad_norm": 0.6185014849673437, "learning_rate": 5.907652891543576e-06, "loss": 0.3446, "step": 314 }, { "epoch": 2.012779552715655, "grad_norm": 0.550249298005834, "learning_rate": 5.839680035587061e-06, "loss": 0.3567, "step": 315 }, { "epoch": 2.0191693290734825, "grad_norm": 0.6406389855268264, "learning_rate": 5.771938844548331e-06, "loss": 0.3439, "step": 316 }, { "epoch": 2.02555910543131, "grad_norm": 0.7498793828860592, "learning_rate": 5.704433090554912e-06, "loss": 0.3525, "step": 317 }, { "epoch": 2.0319488817891376, "grad_norm": 0.6058522485872244, "learning_rate": 5.6371665326241635e-06, "loss": 0.3493, "step": 318 }, { "epoch": 2.038338658146965, "grad_norm": 0.4684273692024481, "learning_rate": 5.570142916453944e-06, "loss": 0.3529, "step": 319 }, { "epoch": 2.0447284345047922, "grad_norm": 0.5736798457088337, "learning_rate": 5.503365974214059e-06, "loss": 0.3601, "step": 320 }, { "epoch": 2.0511182108626196, "grad_norm": 0.6933501821794038, "learning_rate": 5.436839424338426e-06, "loss": 0.3495, "step": 321 }, { "epoch": 2.0575079872204474, "grad_norm": 0.5220030865474038, "learning_rate": 5.37056697131799e-06, "loss": 0.3313, "step": 322 }, { "epoch": 2.0638977635782747, "grad_norm": 0.48932445764444477, "learning_rate": 5.304552305494493e-06, "loss": 0.3429, "step": 323 }, { "epoch": 2.070287539936102, "grad_norm": 0.5367328609042773, "learning_rate": 5.238799102854941e-06, "loss": 0.3631, "step": 324 }, { "epoch": 2.07667731629393, "grad_norm": 0.4753482664474989, "learning_rate": 5.173311024826916e-06, "loss": 0.3164, "step": 325 }, { "epoch": 2.083067092651757, "grad_norm": 0.5857488280714297, "learning_rate": 5.108091718074706e-06, "loss": 0.3684, "step": 326 }, { "epoch": 2.0894568690095845, "grad_norm": 0.5577030340771072, "learning_rate": 5.043144814296214e-06, "loss": 0.3393, "step": 327 }, { "epoch": 2.0958466453674123, "grad_norm": 0.5001282088257198, "learning_rate": 4.9784739300207675e-06, "loss": 0.3289, "step": 328 }, { "epoch": 2.1022364217252396, "grad_norm": 0.49155760192395415, "learning_rate": 4.914082666407705e-06, "loss": 0.3126, "step": 329 }, { "epoch": 2.108626198083067, "grad_norm": 0.4721435423436379, "learning_rate": 4.849974609045849e-06, "loss": 0.3412, "step": 330 }, { "epoch": 2.1150159744408947, "grad_norm": 0.4887160051279591, "learning_rate": 4.786153327753865e-06, "loss": 0.344, "step": 331 }, { "epoch": 2.121405750798722, "grad_norm": 0.4325584192349116, "learning_rate": 4.722622376381455e-06, "loss": 0.3316, "step": 332 }, { "epoch": 2.1277955271565494, "grad_norm": 0.4294996718579918, "learning_rate": 4.659385292611479e-06, "loss": 0.3339, "step": 333 }, { "epoch": 2.134185303514377, "grad_norm": 0.41116978589443015, "learning_rate": 4.59644559776296e-06, "loss": 0.3325, "step": 334 }, { "epoch": 2.1405750798722045, "grad_norm": 0.4411913227153377, "learning_rate": 4.533806796594989e-06, "loss": 0.342, "step": 335 }, { "epoch": 2.146964856230032, "grad_norm": 0.44530187962960605, "learning_rate": 4.471472377111574e-06, "loss": 0.364, "step": 336 }, { "epoch": 2.1533546325878596, "grad_norm": 0.4681980949823063, "learning_rate": 4.409445810367421e-06, "loss": 0.3749, "step": 337 }, { "epoch": 2.159744408945687, "grad_norm": 0.4082359456108807, "learning_rate": 4.347730550274628e-06, "loss": 0.3586, "step": 338 }, { "epoch": 2.1661341853035143, "grad_norm": 0.41583681737409806, "learning_rate": 4.286330033410384e-06, "loss": 0.3047, "step": 339 }, { "epoch": 2.1725239616613417, "grad_norm": 0.4540111795301057, "learning_rate": 4.2252476788255735e-06, "loss": 0.3254, "step": 340 }, { "epoch": 2.1789137380191694, "grad_norm": 0.45433063555043707, "learning_rate": 4.164486887854424e-06, "loss": 0.3459, "step": 341 }, { "epoch": 2.1853035143769968, "grad_norm": 0.43892759194821435, "learning_rate": 4.104051043925068e-06, "loss": 0.354, "step": 342 }, { "epoch": 2.191693290734824, "grad_norm": 0.43626351327017776, "learning_rate": 4.043943512371171e-06, "loss": 0.3484, "step": 343 }, { "epoch": 2.198083067092652, "grad_norm": 0.41866320314755173, "learning_rate": 3.984167640244518e-06, "loss": 0.3595, "step": 344 }, { "epoch": 2.2044728434504792, "grad_norm": 0.43171890711936267, "learning_rate": 3.924726756128632e-06, "loss": 0.3683, "step": 345 }, { "epoch": 2.2108626198083066, "grad_norm": 0.39657734240201464, "learning_rate": 3.8656241699534396e-06, "loss": 0.3319, "step": 346 }, { "epoch": 2.2172523961661343, "grad_norm": 0.4390716578058822, "learning_rate": 3.8068631728109364e-06, "loss": 0.3718, "step": 347 }, { "epoch": 2.2236421725239617, "grad_norm": 0.43526087538819297, "learning_rate": 3.7484470367719493e-06, "loss": 0.3586, "step": 348 }, { "epoch": 2.230031948881789, "grad_norm": 0.41150940940323066, "learning_rate": 3.6903790147039286e-06, "loss": 0.3219, "step": 349 }, { "epoch": 2.236421725239617, "grad_norm": 0.4434399201886998, "learning_rate": 3.6326623400897797e-06, "loss": 0.3496, "step": 350 }, { "epoch": 2.242811501597444, "grad_norm": 0.4226836887457401, "learning_rate": 3.575300226847863e-06, "loss": 0.3369, "step": 351 }, { "epoch": 2.2492012779552715, "grad_norm": 0.449011167578069, "learning_rate": 3.5182958691529945e-06, "loss": 0.3456, "step": 352 }, { "epoch": 2.255591054313099, "grad_norm": 0.40337825402760535, "learning_rate": 3.4616524412585797e-06, "loss": 0.3569, "step": 353 }, { "epoch": 2.2619808306709266, "grad_norm": 0.4417946658839781, "learning_rate": 3.405373097319875e-06, "loss": 0.2965, "step": 354 }, { "epoch": 2.268370607028754, "grad_norm": 0.4226400522886019, "learning_rate": 3.3494609712183323e-06, "loss": 0.3143, "step": 355 }, { "epoch": 2.2747603833865817, "grad_norm": 0.42990286288068097, "learning_rate": 3.2939191763871047e-06, "loss": 0.325, "step": 356 }, { "epoch": 2.281150159744409, "grad_norm": 0.48668783932063814, "learning_rate": 3.2387508056376726e-06, "loss": 0.3393, "step": 357 }, { "epoch": 2.2875399361022364, "grad_norm": 0.45993768274586794, "learning_rate": 3.183958930987612e-06, "loss": 0.3271, "step": 358 }, { "epoch": 2.2939297124600637, "grad_norm": 0.3696554039297162, "learning_rate": 3.1295466034895482e-06, "loss": 0.356, "step": 359 }, { "epoch": 2.3003194888178915, "grad_norm": 0.41320713175521795, "learning_rate": 3.0755168530612444e-06, "loss": 0.321, "step": 360 }, { "epoch": 2.306709265175719, "grad_norm": 0.42742117582586886, "learning_rate": 3.021872688316896e-06, "loss": 0.3161, "step": 361 }, { "epoch": 2.313099041533546, "grad_norm": 0.41256761998572333, "learning_rate": 2.968617096399592e-06, "loss": 0.3095, "step": 362 }, { "epoch": 2.319488817891374, "grad_norm": 0.418828546634445, "learning_rate": 2.9157530428149683e-06, "loss": 0.3434, "step": 363 }, { "epoch": 2.3258785942492013, "grad_norm": 0.40161025260603606, "learning_rate": 2.8632834712660882e-06, "loss": 0.3203, "step": 364 }, { "epoch": 2.3322683706070286, "grad_norm": 0.4182135730630846, "learning_rate": 2.8112113034895273e-06, "loss": 0.3135, "step": 365 }, { "epoch": 2.3386581469648564, "grad_norm": 0.40613861239190874, "learning_rate": 2.7595394390926557e-06, "loss": 0.3643, "step": 366 }, { "epoch": 2.3450479233226837, "grad_norm": 0.41302826534326115, "learning_rate": 2.708270755392207e-06, "loss": 0.3424, "step": 367 }, { "epoch": 2.351437699680511, "grad_norm": 0.3930172226452873, "learning_rate": 2.657408107254027e-06, "loss": 0.3429, "step": 368 }, { "epoch": 2.357827476038339, "grad_norm": 0.41867946397201244, "learning_rate": 2.60695432693412e-06, "loss": 0.3657, "step": 369 }, { "epoch": 2.364217252396166, "grad_norm": 0.3952956485859315, "learning_rate": 2.5569122239209366e-06, "loss": 0.2999, "step": 370 }, { "epoch": 2.3706070287539935, "grad_norm": 0.38547686501477857, "learning_rate": 2.507284584778913e-06, "loss": 0.3229, "step": 371 }, { "epoch": 2.376996805111821, "grad_norm": 0.4071024941451391, "learning_rate": 2.4580741729933246e-06, "loss": 0.3398, "step": 372 }, { "epoch": 2.3833865814696487, "grad_norm": 0.3975504941590724, "learning_rate": 2.4092837288163807e-06, "loss": 0.3246, "step": 373 }, { "epoch": 2.389776357827476, "grad_norm": 0.4007814100724813, "learning_rate": 2.3609159691146577e-06, "loss": 0.3834, "step": 374 }, { "epoch": 2.3961661341853033, "grad_norm": 0.4132465009460835, "learning_rate": 2.312973587217798e-06, "loss": 0.3552, "step": 375 }, { "epoch": 2.402555910543131, "grad_norm": 0.38136466401080477, "learning_rate": 2.2654592527685305e-06, "loss": 0.3394, "step": 376 }, { "epoch": 2.4089456869009584, "grad_norm": 0.38378352088164336, "learning_rate": 2.2183756115740274e-06, "loss": 0.3417, "step": 377 }, { "epoch": 2.415335463258786, "grad_norm": 0.42705803790306746, "learning_rate": 2.171725285458559e-06, "loss": 0.3657, "step": 378 }, { "epoch": 2.4217252396166136, "grad_norm": 0.3893534985584009, "learning_rate": 2.1255108721175066e-06, "loss": 0.3439, "step": 379 }, { "epoch": 2.428115015974441, "grad_norm": 0.443399810426828, "learning_rate": 2.079734944972717e-06, "loss": 0.3216, "step": 380 }, { "epoch": 2.4345047923322682, "grad_norm": 0.39874922783339656, "learning_rate": 2.0344000530291875e-06, "loss": 0.3434, "step": 381 }, { "epoch": 2.440894568690096, "grad_norm": 0.3771297738742276, "learning_rate": 1.9895087207331422e-06, "loss": 0.3409, "step": 382 }, { "epoch": 2.4472843450479234, "grad_norm": 0.40027082321568275, "learning_rate": 1.945063447831452e-06, "loss": 0.3265, "step": 383 }, { "epoch": 2.4536741214057507, "grad_norm": 0.3737581626959035, "learning_rate": 1.9010667092324342e-06, "loss": 0.34, "step": 384 }, { "epoch": 2.460063897763578, "grad_norm": 0.3841031529537794, "learning_rate": 1.8575209548680472e-06, "loss": 0.3503, "step": 385 }, { "epoch": 2.466453674121406, "grad_norm": 0.4038855857442773, "learning_rate": 1.814428609557458e-06, "loss": 0.3201, "step": 386 }, { "epoch": 2.472843450479233, "grad_norm": 0.40823279792933154, "learning_rate": 1.7717920728720284e-06, "loss": 0.3219, "step": 387 }, { "epoch": 2.479233226837061, "grad_norm": 0.45571229187170187, "learning_rate": 1.7296137190016916e-06, "loss": 0.3646, "step": 388 }, { "epoch": 2.4856230031948883, "grad_norm": 0.3992759518176683, "learning_rate": 1.6878958966227366e-06, "loss": 0.3514, "step": 389 }, { "epoch": 2.4920127795527156, "grad_norm": 0.44997582744977993, "learning_rate": 1.646640928767047e-06, "loss": 0.3676, "step": 390 }, { "epoch": 2.498402555910543, "grad_norm": 0.4036812441443117, "learning_rate": 1.6058511126927178e-06, "loss": 0.3572, "step": 391 }, { "epoch": 2.5047923322683707, "grad_norm": 0.38861999088805715, "learning_rate": 1.5655287197561497e-06, "loss": 0.3176, "step": 392 }, { "epoch": 2.511182108626198, "grad_norm": 0.3958987299112533, "learning_rate": 1.5256759952855737e-06, "loss": 0.3411, "step": 393 }, { "epoch": 2.5175718849840254, "grad_norm": 0.3958900572864119, "learning_rate": 1.4862951584560037e-06, "loss": 0.3391, "step": 394 }, { "epoch": 2.523961661341853, "grad_norm": 0.3918198479937304, "learning_rate": 1.447388402165686e-06, "loss": 0.3178, "step": 395 }, { "epoch": 2.5303514376996805, "grad_norm": 0.40756583772176336, "learning_rate": 1.4089578929139635e-06, "loss": 0.3264, "step": 396 }, { "epoch": 2.536741214057508, "grad_norm": 0.40960058637131896, "learning_rate": 1.371005770680659e-06, "loss": 0.3385, "step": 397 }, { "epoch": 2.543130990415335, "grad_norm": 0.40903897126869604, "learning_rate": 1.3335341488068997e-06, "loss": 0.3805, "step": 398 }, { "epoch": 2.549520766773163, "grad_norm": 0.39992597114439177, "learning_rate": 1.2965451138774343e-06, "loss": 0.3196, "step": 399 }, { "epoch": 2.5559105431309903, "grad_norm": 0.3894861786588525, "learning_rate": 1.2600407256044544e-06, "loss": 0.354, "step": 400 }, { "epoch": 2.562300319488818, "grad_norm": 0.3773260960595135, "learning_rate": 1.2240230167128863e-06, "loss": 0.3408, "step": 401 }, { "epoch": 2.5686900958466454, "grad_norm": 0.3916021706010166, "learning_rate": 1.188493992827211e-06, "loss": 0.3141, "step": 402 }, { "epoch": 2.5750798722044728, "grad_norm": 0.3906836314398566, "learning_rate": 1.1534556323597824e-06, "loss": 0.2924, "step": 403 }, { "epoch": 2.5814696485623, "grad_norm": 0.4038749879309955, "learning_rate": 1.1189098864006488e-06, "loss": 0.3365, "step": 404 }, { "epoch": 2.587859424920128, "grad_norm": 0.40279549296487077, "learning_rate": 1.084858678608922e-06, "loss": 0.3239, "step": 405 }, { "epoch": 2.594249201277955, "grad_norm": 0.36365947820301786, "learning_rate": 1.0513039051056507e-06, "loss": 0.3326, "step": 406 }, { "epoch": 2.600638977635783, "grad_norm": 0.3696286967249599, "learning_rate": 1.0182474343682346e-06, "loss": 0.3292, "step": 407 }, { "epoch": 2.6070287539936103, "grad_norm": 0.4175231953214467, "learning_rate": 9.85691107126392e-07, "loss": 0.3744, "step": 408 }, { "epoch": 2.6134185303514377, "grad_norm": 0.39195370451842787, "learning_rate": 9.536367362596422e-07, "loss": 0.3695, "step": 409 }, { "epoch": 2.619808306709265, "grad_norm": 0.37605332595306007, "learning_rate": 9.220861066963715e-07, "loss": 0.3418, "step": 410 }, { "epoch": 2.626198083067093, "grad_norm": 0.3995977702745196, "learning_rate": 8.910409753144344e-07, "loss": 0.3549, "step": 411 }, { "epoch": 2.63258785942492, "grad_norm": 0.3834229012458761, "learning_rate": 8.605030708433149e-07, "loss": 0.3445, "step": 412 }, { "epoch": 2.6389776357827475, "grad_norm": 0.37644074285723944, "learning_rate": 8.304740937678835e-07, "loss": 0.3668, "step": 413 }, { "epoch": 2.6453674121405752, "grad_norm": 0.38149522660404983, "learning_rate": 8.009557162336823e-07, "loss": 0.3286, "step": 414 }, { "epoch": 2.6517571884984026, "grad_norm": 0.3965348662763524, "learning_rate": 7.719495819538325e-07, "loss": 0.3562, "step": 415 }, { "epoch": 2.65814696485623, "grad_norm": 0.3740820985387306, "learning_rate": 7.434573061174966e-07, "loss": 0.3171, "step": 416 }, { "epoch": 2.6645367412140573, "grad_norm": 0.37288133480679564, "learning_rate": 7.154804752999344e-07, "loss": 0.3263, "step": 417 }, { "epoch": 2.670926517571885, "grad_norm": 0.37023486110236786, "learning_rate": 6.880206473741646e-07, "loss": 0.368, "step": 418 }, { "epoch": 2.6773162939297124, "grad_norm": 0.3924590139873951, "learning_rate": 6.610793514242075e-07, "loss": 0.3348, "step": 419 }, { "epoch": 2.68370607028754, "grad_norm": 0.4026773365921225, "learning_rate": 6.346580876599395e-07, "loss": 0.3405, "step": 420 }, { "epoch": 2.6900958466453675, "grad_norm": 0.3855444929570631, "learning_rate": 6.08758327333564e-07, "loss": 0.3588, "step": 421 }, { "epoch": 2.696485623003195, "grad_norm": 0.4018291966488895, "learning_rate": 5.833815126576714e-07, "loss": 0.3259, "step": 422 }, { "epoch": 2.702875399361022, "grad_norm": 0.36389835567175366, "learning_rate": 5.585290567249424e-07, "loss": 0.3514, "step": 423 }, { "epoch": 2.70926517571885, "grad_norm": 0.40458156951380986, "learning_rate": 5.342023434294552e-07, "loss": 0.337, "step": 424 }, { "epoch": 2.7156549520766773, "grad_norm": 0.36807317654930927, "learning_rate": 5.104027273896239e-07, "loss": 0.3148, "step": 425 }, { "epoch": 2.722044728434505, "grad_norm": 0.3689483940462552, "learning_rate": 4.871315338727711e-07, "loss": 0.3181, "step": 426 }, { "epoch": 2.7284345047923324, "grad_norm": 0.3833457229816993, "learning_rate": 4.6439005872132457e-07, "loss": 0.3368, "step": 427 }, { "epoch": 2.7348242811501597, "grad_norm": 0.36604943775735366, "learning_rate": 4.421795682806662e-07, "loss": 0.3364, "step": 428 }, { "epoch": 2.741214057507987, "grad_norm": 0.3773811652537217, "learning_rate": 4.2050129932861394e-07, "loss": 0.3229, "step": 429 }, { "epoch": 2.747603833865815, "grad_norm": 0.3796116399398149, "learning_rate": 3.9935645900654906e-07, "loss": 0.3633, "step": 430 }, { "epoch": 2.753993610223642, "grad_norm": 0.37803460451421467, "learning_rate": 3.7874622475220336e-07, "loss": 0.3524, "step": 431 }, { "epoch": 2.7603833865814695, "grad_norm": 0.36675174735385174, "learning_rate": 3.58671744234087e-07, "loss": 0.365, "step": 432 }, { "epoch": 2.7667731629392973, "grad_norm": 0.3898919999844523, "learning_rate": 3.3913413528758877e-07, "loss": 0.3323, "step": 433 }, { "epoch": 2.7731629392971247, "grad_norm": 0.3936657666166714, "learning_rate": 3.2013448585272333e-07, "loss": 0.3337, "step": 434 }, { "epoch": 2.779552715654952, "grad_norm": 0.38046637386341803, "learning_rate": 3.016738539135566e-07, "loss": 0.344, "step": 435 }, { "epoch": 2.7859424920127793, "grad_norm": 0.3770525251793597, "learning_rate": 2.837532674392862e-07, "loss": 0.374, "step": 436 }, { "epoch": 2.792332268370607, "grad_norm": 0.3916319385244186, "learning_rate": 2.6637372432700483e-07, "loss": 0.3492, "step": 437 }, { "epoch": 2.7987220447284344, "grad_norm": 0.4083221537063404, "learning_rate": 2.4953619234612816e-07, "loss": 0.3056, "step": 438 }, { "epoch": 2.8051118210862622, "grad_norm": 0.36900391336674243, "learning_rate": 2.332416090845102e-07, "loss": 0.3071, "step": 439 }, { "epoch": 2.8115015974440896, "grad_norm": 0.4181996987699531, "learning_rate": 2.1749088189622848e-07, "loss": 0.3351, "step": 440 }, { "epoch": 2.817891373801917, "grad_norm": 0.3547619175700847, "learning_rate": 2.0228488785106636e-07, "loss": 0.3406, "step": 441 }, { "epoch": 2.8242811501597442, "grad_norm": 0.3977218103469981, "learning_rate": 1.8762447368566582e-07, "loss": 0.3505, "step": 442 }, { "epoch": 2.830670926517572, "grad_norm": 0.39003543295538634, "learning_rate": 1.7351045575638047e-07, "loss": 0.3753, "step": 443 }, { "epoch": 2.8370607028753994, "grad_norm": 0.37845868272626343, "learning_rate": 1.5994361999382202e-07, "loss": 0.3278, "step": 444 }, { "epoch": 2.8434504792332267, "grad_norm": 0.39842394374774287, "learning_rate": 1.4692472185908635e-07, "loss": 0.3682, "step": 445 }, { "epoch": 2.8498402555910545, "grad_norm": 0.3457049068963664, "learning_rate": 1.344544863016961e-07, "loss": 0.3438, "step": 446 }, { "epoch": 2.856230031948882, "grad_norm": 0.40463187861502226, "learning_rate": 1.225336077192274e-07, "loss": 0.3356, "step": 447 }, { "epoch": 2.862619808306709, "grad_norm": 0.5473639905587789, "learning_rate": 1.1116274991864073e-07, "loss": 0.3612, "step": 448 }, { "epoch": 2.8690095846645365, "grad_norm": 0.36738048359514414, "learning_rate": 1.003425460793217e-07, "loss": 0.3618, "step": 449 }, { "epoch": 2.8753993610223643, "grad_norm": 0.39397443170606444, "learning_rate": 9.00735987178214e-08, "loss": 0.3325, "step": 450 }, { "epoch": 2.8817891373801916, "grad_norm": 0.3641877860232395, "learning_rate": 8.035647965430215e-08, "loss": 0.3705, "step": 451 }, { "epoch": 2.8881789137380194, "grad_norm": 0.3766602850517281, "learning_rate": 7.119172998070412e-08, "loss": 0.3602, "step": 452 }, { "epoch": 2.8945686900958467, "grad_norm": 0.3504672529434601, "learning_rate": 6.25798600306049e-08, "loss": 0.345, "step": 453 }, { "epoch": 2.900958466453674, "grad_norm": 0.3689087024207003, "learning_rate": 5.4521349350808996e-08, "loss": 0.3711, "step": 454 }, { "epoch": 2.9073482428115014, "grad_norm": 0.3816320750179523, "learning_rate": 4.701664667464245e-08, "loss": 0.3646, "step": 455 }, { "epoch": 2.913738019169329, "grad_norm": 0.3679437431040902, "learning_rate": 4.006616989696621e-08, "loss": 0.3615, "step": 456 }, { "epoch": 2.9201277955271565, "grad_norm": 0.38571927082544183, "learning_rate": 3.367030605090249e-08, "loss": 0.3222, "step": 457 }, { "epoch": 2.9265175718849843, "grad_norm": 0.3971237004311223, "learning_rate": 2.7829411286287577e-08, "loss": 0.3485, "step": 458 }, { "epoch": 2.9329073482428116, "grad_norm": 0.3630064489830963, "learning_rate": 2.254381084983659e-08, "loss": 0.3065, "step": 459 }, { "epoch": 2.939297124600639, "grad_norm": 0.3754819920105193, "learning_rate": 1.781379906703573e-08, "loss": 0.36, "step": 460 }, { "epoch": 2.9456869009584663, "grad_norm": 0.37643335933636624, "learning_rate": 1.3639639325748744e-08, "loss": 0.3725, "step": 461 }, { "epoch": 2.952076677316294, "grad_norm": 0.397354621784988, "learning_rate": 1.0021564061554189e-08, "loss": 0.3378, "step": 462 }, { "epoch": 2.9584664536741214, "grad_norm": 0.37424179888551773, "learning_rate": 6.959774744796921e-09, "loss": 0.3399, "step": 463 }, { "epoch": 2.9648562300319488, "grad_norm": 0.37190360281608004, "learning_rate": 4.454441869377047e-09, "loss": 0.3219, "step": 464 }, { "epoch": 2.9712460063897765, "grad_norm": 0.3657736795626134, "learning_rate": 2.5057049432519744e-09, "loss": 0.3705, "step": 465 }, { "epoch": 2.977635782747604, "grad_norm": 0.35303669900268125, "learning_rate": 1.1136724806637411e-09, "loss": 0.3318, "step": 466 }, { "epoch": 2.984025559105431, "grad_norm": 0.36854160936154184, "learning_rate": 2.784219961060597e-10, "loss": 0.3518, "step": 467 }, { "epoch": 2.9904153354632586, "grad_norm": 0.3989983000303115, "learning_rate": 0.0, "loss": 0.3546, "step": 468 }, { "epoch": 2.9904153354632586, "step": 468, "total_flos": 3.310734730054861e+17, "train_loss": 0.4434524042229367, "train_runtime": 10205.5973, "train_samples_per_second": 2.939, "train_steps_per_second": 0.046 } ], "logging_steps": 1.0, "max_steps": 468, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.310734730054861e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }