diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5099 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 3160, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.007911392405063292, + "grad_norm": 2.3738328422587087, + "learning_rate": 1.5822784810126583e-06, + "loss": 1.1261, + "num_tokens": 5242880.0, + "step": 5 + }, + { + "epoch": 0.015822784810126583, + "grad_norm": 1.7722440301570659, + "learning_rate": 3.1645569620253167e-06, + "loss": 1.1058, + "num_tokens": 10470029.0, + "step": 10 + }, + { + "epoch": 0.023734177215189875, + "grad_norm": 1.083276838182905, + "learning_rate": 4.746835443037975e-06, + "loss": 1.0806, + "num_tokens": 15712909.0, + "step": 15 + }, + { + "epoch": 0.03164556962025317, + "grad_norm": 0.8611689905530553, + "learning_rate": 6.329113924050633e-06, + "loss": 1.02, + "num_tokens": 20951872.0, + "step": 20 + }, + { + "epoch": 0.03955696202531646, + "grad_norm": 0.7250421722694049, + "learning_rate": 7.911392405063292e-06, + "loss": 0.9766, + "num_tokens": 26185341.0, + "step": 25 + }, + { + "epoch": 0.04746835443037975, + "grad_norm": 0.6382913905764741, + "learning_rate": 9.49367088607595e-06, + "loss": 0.9456, + "num_tokens": 31428221.0, + "step": 30 + }, + { + "epoch": 0.055379746835443035, + "grad_norm": 0.5436432996737252, + "learning_rate": 1.1075949367088608e-05, + "loss": 0.9187, + "num_tokens": 36669330.0, + "step": 35 + }, + { + "epoch": 0.06329113924050633, + "grad_norm": 0.45175884987072906, + "learning_rate": 1.2658227848101267e-05, + "loss": 0.8956, + "num_tokens": 41910897.0, + "step": 40 + }, + { + "epoch": 0.07120253164556962, + "grad_norm": 0.3906460936286765, + "learning_rate": 1.4240506329113925e-05, + "loss": 0.8787, + "num_tokens": 47153777.0, + "step": 45 + }, + { + "epoch": 0.07911392405063292, + "grad_norm": 0.38060702422443193, + "learning_rate": 1.5822784810126583e-05, + "loss": 0.8674, + "num_tokens": 52391193.0, + "step": 50 + }, + { + "epoch": 0.08702531645569621, + "grad_norm": 0.3976893756810291, + "learning_rate": 1.7405063291139243e-05, + "loss": 0.8589, + "num_tokens": 57634073.0, + "step": 55 + }, + { + "epoch": 0.0949367088607595, + "grad_norm": 0.39610453911616206, + "learning_rate": 1.89873417721519e-05, + "loss": 0.8577, + "num_tokens": 62876953.0, + "step": 60 + }, + { + "epoch": 0.10284810126582279, + "grad_norm": 0.3232899679188804, + "learning_rate": 2.056962025316456e-05, + "loss": 0.8456, + "num_tokens": 68119833.0, + "step": 65 + }, + { + "epoch": 0.11075949367088607, + "grad_norm": 0.3785345815272384, + "learning_rate": 2.2151898734177217e-05, + "loss": 0.8354, + "num_tokens": 73362713.0, + "step": 70 + }, + { + "epoch": 0.11867088607594936, + "grad_norm": 0.42958423811773794, + "learning_rate": 2.3734177215189873e-05, + "loss": 0.8277, + "num_tokens": 78605593.0, + "step": 75 + }, + { + "epoch": 0.12658227848101267, + "grad_norm": 0.3759983838903421, + "learning_rate": 2.5316455696202533e-05, + "loss": 0.8257, + "num_tokens": 83848473.0, + "step": 80 + }, + { + "epoch": 0.13449367088607594, + "grad_norm": 0.4694604862178893, + "learning_rate": 2.689873417721519e-05, + "loss": 0.8212, + "num_tokens": 89091353.0, + "step": 85 + }, + { + "epoch": 0.14240506329113925, + "grad_norm": 0.41926661815869387, + "learning_rate": 2.848101265822785e-05, + "loss": 0.8132, + "num_tokens": 94334233.0, + "step": 90 + }, + { + "epoch": 0.15031645569620253, + "grad_norm": 0.8644371279587081, + "learning_rate": 3.0063291139240506e-05, + "loss": 0.8356, + "num_tokens": 99577113.0, + "step": 95 + }, + { + "epoch": 0.15822784810126583, + "grad_norm": 0.7239620838593109, + "learning_rate": 3.1645569620253167e-05, + "loss": 0.8115, + "num_tokens": 104819993.0, + "step": 100 + }, + { + "epoch": 0.1661392405063291, + "grad_norm": 0.6759442627573765, + "learning_rate": 3.322784810126582e-05, + "loss": 0.8131, + "num_tokens": 110062873.0, + "step": 105 + }, + { + "epoch": 0.17405063291139242, + "grad_norm": 0.5445796635610345, + "learning_rate": 3.4810126582278487e-05, + "loss": 0.8051, + "num_tokens": 115305753.0, + "step": 110 + }, + { + "epoch": 0.1819620253164557, + "grad_norm": 0.6030674874140043, + "learning_rate": 3.639240506329114e-05, + "loss": 0.8069, + "num_tokens": 120539230.0, + "step": 115 + }, + { + "epoch": 0.189873417721519, + "grad_norm": 0.5775646724321816, + "learning_rate": 3.79746835443038e-05, + "loss": 0.8001, + "num_tokens": 125782110.0, + "step": 120 + }, + { + "epoch": 0.19778481012658228, + "grad_norm": 1.1999946695600228, + "learning_rate": 3.9556962025316456e-05, + "loss": 0.7923, + "num_tokens": 131014137.0, + "step": 125 + }, + { + "epoch": 0.20569620253164558, + "grad_norm": 0.9396643672179282, + "learning_rate": 4.113924050632912e-05, + "loss": 0.8032, + "num_tokens": 136257017.0, + "step": 130 + }, + { + "epoch": 0.21360759493670886, + "grad_norm": 0.989562272615721, + "learning_rate": 4.2721518987341776e-05, + "loss": 0.7945, + "num_tokens": 141499897.0, + "step": 135 + }, + { + "epoch": 0.22151898734177214, + "grad_norm": 0.7938020654922184, + "learning_rate": 4.430379746835443e-05, + "loss": 0.7911, + "num_tokens": 146712990.0, + "step": 140 + }, + { + "epoch": 0.22943037974683544, + "grad_norm": 0.772547411213572, + "learning_rate": 4.588607594936709e-05, + "loss": 0.788, + "num_tokens": 151927597.0, + "step": 145 + }, + { + "epoch": 0.23734177215189872, + "grad_norm": 0.5427524570272351, + "learning_rate": 4.7468354430379746e-05, + "loss": 0.7931, + "num_tokens": 157170477.0, + "step": 150 + }, + { + "epoch": 0.24525316455696203, + "grad_norm": 0.520616997801594, + "learning_rate": 4.905063291139241e-05, + "loss": 0.7761, + "num_tokens": 162397909.0, + "step": 155 + }, + { + "epoch": 0.25316455696202533, + "grad_norm": 1.321171420646125, + "learning_rate": 4.9999950717727614e-05, + "loss": 0.8103, + "num_tokens": 167639712.0, + "step": 160 + }, + { + "epoch": 0.2610759493670886, + "grad_norm": 0.8179961728401967, + "learning_rate": 4.999939629464255e-05, + "loss": 0.7829, + "num_tokens": 172882592.0, + "step": 165 + }, + { + "epoch": 0.2689873417721519, + "grad_norm": 1.07950233433435, + "learning_rate": 4.999822586086208e-05, + "loss": 0.7815, + "num_tokens": 178125472.0, + "step": 170 + }, + { + "epoch": 0.27689873417721517, + "grad_norm": 0.764049579486957, + "learning_rate": 4.9996439448431495e-05, + "loss": 0.7936, + "num_tokens": 183355239.0, + "step": 175 + }, + { + "epoch": 0.2848101265822785, + "grad_norm": 0.6643682550526433, + "learning_rate": 4.9994037106260965e-05, + "loss": 0.777, + "num_tokens": 188598119.0, + "step": 180 + }, + { + "epoch": 0.2927215189873418, + "grad_norm": 0.4988533797702287, + "learning_rate": 4.999101890012418e-05, + "loss": 0.7755, + "num_tokens": 193840999.0, + "step": 185 + }, + { + "epoch": 0.30063291139240506, + "grad_norm": 0.49381791616490034, + "learning_rate": 4.998738491265655e-05, + "loss": 0.7646, + "num_tokens": 199082063.0, + "step": 190 + }, + { + "epoch": 0.30854430379746833, + "grad_norm": 0.5570710586204755, + "learning_rate": 4.998313524335302e-05, + "loss": 0.7698, + "num_tokens": 204324943.0, + "step": 195 + }, + { + "epoch": 0.31645569620253167, + "grad_norm": 0.4911267149039519, + "learning_rate": 4.9978270008565184e-05, + "loss": 0.7641, + "num_tokens": 209561370.0, + "step": 200 + }, + { + "epoch": 0.32436708860759494, + "grad_norm": 0.6409945115911554, + "learning_rate": 4.997278934149828e-05, + "loss": 0.7677, + "num_tokens": 214804250.0, + "step": 205 + }, + { + "epoch": 0.3322784810126582, + "grad_norm": 0.5401344460770882, + "learning_rate": 4.996669339220741e-05, + "loss": 0.7657, + "num_tokens": 220047130.0, + "step": 210 + }, + { + "epoch": 0.3401898734177215, + "grad_norm": 0.5361201681791574, + "learning_rate": 4.995998232759349e-05, + "loss": 0.7663, + "num_tokens": 225290010.0, + "step": 215 + }, + { + "epoch": 0.34810126582278483, + "grad_norm": 0.5803989669904568, + "learning_rate": 4.995265633139869e-05, + "loss": 0.7636, + "num_tokens": 230532890.0, + "step": 220 + }, + { + "epoch": 0.3560126582278481, + "grad_norm": 0.726211587240982, + "learning_rate": 4.994471560420137e-05, + "loss": 0.7631, + "num_tokens": 235775770.0, + "step": 225 + }, + { + "epoch": 0.3639240506329114, + "grad_norm": 0.698163213622165, + "learning_rate": 4.9936160363410584e-05, + "loss": 0.7545, + "num_tokens": 241018650.0, + "step": 230 + }, + { + "epoch": 0.37183544303797467, + "grad_norm": 0.5153045338635585, + "learning_rate": 4.992699084326018e-05, + "loss": 0.7541, + "num_tokens": 246261530.0, + "step": 235 + }, + { + "epoch": 0.379746835443038, + "grad_norm": 0.6959186613242923, + "learning_rate": 4.9917207294802346e-05, + "loss": 0.7635, + "num_tokens": 251504410.0, + "step": 240 + }, + { + "epoch": 0.3876582278481013, + "grad_norm": 0.475983385570461, + "learning_rate": 4.990680998590071e-05, + "loss": 0.755, + "num_tokens": 256747290.0, + "step": 245 + }, + { + "epoch": 0.39556962025316456, + "grad_norm": 0.6721509783638105, + "learning_rate": 4.989579920122307e-05, + "loss": 0.7532, + "num_tokens": 261990170.0, + "step": 250 + }, + { + "epoch": 0.40348101265822783, + "grad_norm": 0.4813755853658302, + "learning_rate": 4.9884175242233585e-05, + "loss": 0.7552, + "num_tokens": 267233050.0, + "step": 255 + }, + { + "epoch": 0.41139240506329117, + "grad_norm": 0.46711611820253635, + "learning_rate": 4.987193842718448e-05, + "loss": 0.7513, + "num_tokens": 272475930.0, + "step": 260 + }, + { + "epoch": 0.41930379746835444, + "grad_norm": 0.5972032077244416, + "learning_rate": 4.985908909110735e-05, + "loss": 0.7443, + "num_tokens": 277718810.0, + "step": 265 + }, + { + "epoch": 0.4272151898734177, + "grad_norm": 0.5059990764921701, + "learning_rate": 4.9845627585804025e-05, + "loss": 0.7458, + "num_tokens": 282942349.0, + "step": 270 + }, + { + "epoch": 0.435126582278481, + "grad_norm": 0.5582209701112856, + "learning_rate": 4.9831554279836884e-05, + "loss": 0.7521, + "num_tokens": 288185229.0, + "step": 275 + }, + { + "epoch": 0.4430379746835443, + "grad_norm": 0.4595435461264603, + "learning_rate": 4.9816869558518796e-05, + "loss": 0.7517, + "num_tokens": 293428109.0, + "step": 280 + }, + { + "epoch": 0.4509493670886076, + "grad_norm": 0.4675296642906588, + "learning_rate": 4.980157382390258e-05, + "loss": 0.7427, + "num_tokens": 298670989.0, + "step": 285 + }, + { + "epoch": 0.4588607594936709, + "grad_norm": 0.4659606979020265, + "learning_rate": 4.978566749476995e-05, + "loss": 0.7397, + "num_tokens": 303913869.0, + "step": 290 + }, + { + "epoch": 0.46677215189873417, + "grad_norm": 0.5143550840646993, + "learning_rate": 4.9769151006620115e-05, + "loss": 0.7488, + "num_tokens": 309151588.0, + "step": 295 + }, + { + "epoch": 0.47468354430379744, + "grad_norm": 0.5527231062409742, + "learning_rate": 4.975202481165778e-05, + "loss": 0.7528, + "num_tokens": 314394468.0, + "step": 300 + }, + { + "epoch": 0.4825949367088608, + "grad_norm": 0.6582518727489046, + "learning_rate": 4.973428937878085e-05, + "loss": 0.7388, + "num_tokens": 319622567.0, + "step": 305 + }, + { + "epoch": 0.49050632911392406, + "grad_norm": 0.5202900734758532, + "learning_rate": 4.9715945193567506e-05, + "loss": 0.7332, + "num_tokens": 324865447.0, + "step": 310 + }, + { + "epoch": 0.49841772151898733, + "grad_norm": 0.38453517395197584, + "learning_rate": 4.969699275826298e-05, + "loss": 0.7364, + "num_tokens": 330108327.0, + "step": 315 + }, + { + "epoch": 0.5063291139240507, + "grad_norm": 0.5632153181654556, + "learning_rate": 4.967743259176575e-05, + "loss": 0.7387, + "num_tokens": 335351207.0, + "step": 320 + }, + { + "epoch": 0.5142405063291139, + "grad_norm": 0.44887504740519496, + "learning_rate": 4.96572652296134e-05, + "loss": 0.7354, + "num_tokens": 340594087.0, + "step": 325 + }, + { + "epoch": 0.5221518987341772, + "grad_norm": 0.48735207873608405, + "learning_rate": 4.963649122396788e-05, + "loss": 0.7376, + "num_tokens": 345825159.0, + "step": 330 + }, + { + "epoch": 0.5300632911392406, + "grad_norm": 0.4193108797718534, + "learning_rate": 4.961511114360043e-05, + "loss": 0.7361, + "num_tokens": 351068039.0, + "step": 335 + }, + { + "epoch": 0.5379746835443038, + "grad_norm": 0.5662379674502074, + "learning_rate": 4.959312557387601e-05, + "loss": 0.739, + "num_tokens": 356304704.0, + "step": 340 + }, + { + "epoch": 0.5458860759493671, + "grad_norm": 0.5775103716382664, + "learning_rate": 4.957053511673728e-05, + "loss": 0.739, + "num_tokens": 361547584.0, + "step": 345 + }, + { + "epoch": 0.5537974683544303, + "grad_norm": 0.4784805413121098, + "learning_rate": 4.954734039068806e-05, + "loss": 0.7369, + "num_tokens": 366775047.0, + "step": 350 + }, + { + "epoch": 0.5617088607594937, + "grad_norm": 0.3826615225742145, + "learning_rate": 4.95235420307765e-05, + "loss": 0.7371, + "num_tokens": 372017927.0, + "step": 355 + }, + { + "epoch": 0.569620253164557, + "grad_norm": 0.5250012256720773, + "learning_rate": 4.949914068857758e-05, + "loss": 0.7401, + "num_tokens": 377260807.0, + "step": 360 + }, + { + "epoch": 0.5775316455696202, + "grad_norm": 0.5294573268848846, + "learning_rate": 4.947413703217537e-05, + "loss": 0.7303, + "num_tokens": 382503687.0, + "step": 365 + }, + { + "epoch": 0.5854430379746836, + "grad_norm": 0.4657426470359354, + "learning_rate": 4.944853174614465e-05, + "loss": 0.7276, + "num_tokens": 387746567.0, + "step": 370 + }, + { + "epoch": 0.5933544303797469, + "grad_norm": 0.6110352375166948, + "learning_rate": 4.9422325531532254e-05, + "loss": 0.7332, + "num_tokens": 392989447.0, + "step": 375 + }, + { + "epoch": 0.6012658227848101, + "grad_norm": 0.5173863425147587, + "learning_rate": 4.9395519105837794e-05, + "loss": 0.7345, + "num_tokens": 398220435.0, + "step": 380 + }, + { + "epoch": 0.6091772151898734, + "grad_norm": 0.44630851057828674, + "learning_rate": 4.936811320299409e-05, + "loss": 0.7344, + "num_tokens": 403454091.0, + "step": 385 + }, + { + "epoch": 0.6170886075949367, + "grad_norm": 0.5511830323483669, + "learning_rate": 4.9340108573346994e-05, + "loss": 0.7277, + "num_tokens": 408695566.0, + "step": 390 + }, + { + "epoch": 0.625, + "grad_norm": 0.488590082197144, + "learning_rate": 4.931150598363494e-05, + "loss": 0.7262, + "num_tokens": 413938446.0, + "step": 395 + }, + { + "epoch": 0.6329113924050633, + "grad_norm": 0.5980267797290539, + "learning_rate": 4.9282306216967866e-05, + "loss": 0.7367, + "num_tokens": 419181326.0, + "step": 400 + }, + { + "epoch": 0.6408227848101266, + "grad_norm": 0.4042639446511123, + "learning_rate": 4.925251007280584e-05, + "loss": 0.7277, + "num_tokens": 424410675.0, + "step": 405 + }, + { + "epoch": 0.6487341772151899, + "grad_norm": 0.5873987854936606, + "learning_rate": 4.922211836693711e-05, + "loss": 0.7216, + "num_tokens": 429653555.0, + "step": 410 + }, + { + "epoch": 0.6566455696202531, + "grad_norm": 0.49135762238140734, + "learning_rate": 4.9191131931455836e-05, + "loss": 0.7249, + "num_tokens": 434896435.0, + "step": 415 + }, + { + "epoch": 0.6645569620253164, + "grad_norm": 0.5977070057438132, + "learning_rate": 4.915955161473925e-05, + "loss": 0.7213, + "num_tokens": 440139315.0, + "step": 420 + }, + { + "epoch": 0.6724683544303798, + "grad_norm": 0.5865545939218848, + "learning_rate": 4.9127378281424454e-05, + "loss": 0.7291, + "num_tokens": 445382195.0, + "step": 425 + }, + { + "epoch": 0.680379746835443, + "grad_norm": 0.5141049741136753, + "learning_rate": 4.909461281238475e-05, + "loss": 0.7275, + "num_tokens": 450625075.0, + "step": 430 + }, + { + "epoch": 0.6882911392405063, + "grad_norm": 0.586173282775543, + "learning_rate": 4.9061256104705514e-05, + "loss": 0.7147, + "num_tokens": 455867955.0, + "step": 435 + }, + { + "epoch": 0.6962025316455697, + "grad_norm": 0.4884462198074093, + "learning_rate": 4.902730907165962e-05, + "loss": 0.7262, + "num_tokens": 461110835.0, + "step": 440 + }, + { + "epoch": 0.7041139240506329, + "grad_norm": 0.5155782485627577, + "learning_rate": 4.899277264268249e-05, + "loss": 0.726, + "num_tokens": 466329951.0, + "step": 445 + }, + { + "epoch": 0.7120253164556962, + "grad_norm": 0.5894348115226837, + "learning_rate": 4.8957647763346575e-05, + "loss": 0.7276, + "num_tokens": 471572831.0, + "step": 450 + }, + { + "epoch": 0.7199367088607594, + "grad_norm": 0.6178813430658108, + "learning_rate": 4.892193539533553e-05, + "loss": 0.7234, + "num_tokens": 476815711.0, + "step": 455 + }, + { + "epoch": 0.7278481012658228, + "grad_norm": 0.7787566109612348, + "learning_rate": 4.888563651641784e-05, + "loss": 0.7297, + "num_tokens": 482044487.0, + "step": 460 + }, + { + "epoch": 0.7357594936708861, + "grad_norm": 0.5460153094045689, + "learning_rate": 4.884875212042005e-05, + "loss": 0.7257, + "num_tokens": 487287367.0, + "step": 465 + }, + { + "epoch": 0.7436708860759493, + "grad_norm": 0.5019931926593302, + "learning_rate": 4.88112832171996e-05, + "loss": 0.7219, + "num_tokens": 492530247.0, + "step": 470 + }, + { + "epoch": 0.7515822784810127, + "grad_norm": 0.46922376523806203, + "learning_rate": 4.877323083261713e-05, + "loss": 0.7207, + "num_tokens": 497773127.0, + "step": 475 + }, + { + "epoch": 0.759493670886076, + "grad_norm": 0.4491131631584376, + "learning_rate": 4.873459600850841e-05, + "loss": 0.718, + "num_tokens": 503013415.0, + "step": 480 + }, + { + "epoch": 0.7674050632911392, + "grad_norm": 0.4576902411009356, + "learning_rate": 4.869537980265581e-05, + "loss": 0.7119, + "num_tokens": 508255611.0, + "step": 485 + }, + { + "epoch": 0.7753164556962026, + "grad_norm": 0.6179813478813936, + "learning_rate": 4.865558328875937e-05, + "loss": 0.7086, + "num_tokens": 513498491.0, + "step": 490 + }, + { + "epoch": 0.7832278481012658, + "grad_norm": 0.5471403707624125, + "learning_rate": 4.8615207556407324e-05, + "loss": 0.7242, + "num_tokens": 518723580.0, + "step": 495 + }, + { + "epoch": 0.7911392405063291, + "grad_norm": 0.4589250795263245, + "learning_rate": 4.857425371104638e-05, + "loss": 0.7159, + "num_tokens": 523966460.0, + "step": 500 + }, + { + "epoch": 0.7990506329113924, + "grad_norm": 0.4660626316391112, + "learning_rate": 4.853272287395135e-05, + "loss": 0.7251, + "num_tokens": 529209340.0, + "step": 505 + }, + { + "epoch": 0.8069620253164557, + "grad_norm": 0.5382100486924232, + "learning_rate": 4.8490616182194534e-05, + "loss": 0.7187, + "num_tokens": 534429844.0, + "step": 510 + }, + { + "epoch": 0.814873417721519, + "grad_norm": 0.5238715230138492, + "learning_rate": 4.8447934788614515e-05, + "loss": 0.7187, + "num_tokens": 539672724.0, + "step": 515 + }, + { + "epoch": 0.8227848101265823, + "grad_norm": 0.5194663159593585, + "learning_rate": 4.840467986178464e-05, + "loss": 0.7168, + "num_tokens": 544915604.0, + "step": 520 + }, + { + "epoch": 0.8306962025316456, + "grad_norm": 0.4135374652648048, + "learning_rate": 4.8360852585981034e-05, + "loss": 0.7236, + "num_tokens": 550158484.0, + "step": 525 + }, + { + "epoch": 0.8386075949367089, + "grad_norm": 0.47687000377153205, + "learning_rate": 4.831645416115014e-05, + "loss": 0.7091, + "num_tokens": 555401364.0, + "step": 530 + }, + { + "epoch": 0.8465189873417721, + "grad_norm": 0.4659031703902228, + "learning_rate": 4.827148580287588e-05, + "loss": 0.7042, + "num_tokens": 560644244.0, + "step": 535 + }, + { + "epoch": 0.8544303797468354, + "grad_norm": 0.432869438663762, + "learning_rate": 4.82259487423464e-05, + "loss": 0.7119, + "num_tokens": 565887124.0, + "step": 540 + }, + { + "epoch": 0.8623417721518988, + "grad_norm": 0.3989063394441302, + "learning_rate": 4.81798442263203e-05, + "loss": 0.7164, + "num_tokens": 571114334.0, + "step": 545 + }, + { + "epoch": 0.870253164556962, + "grad_norm": 0.5839119244408789, + "learning_rate": 4.8133173517092575e-05, + "loss": 0.7147, + "num_tokens": 576351352.0, + "step": 550 + }, + { + "epoch": 0.8781645569620253, + "grad_norm": 0.4989762196407589, + "learning_rate": 4.808593789246e-05, + "loss": 0.7116, + "num_tokens": 581594232.0, + "step": 555 + }, + { + "epoch": 0.8860759493670886, + "grad_norm": 0.6013182449359468, + "learning_rate": 4.803813864568616e-05, + "loss": 0.7101, + "num_tokens": 586837112.0, + "step": 560 + }, + { + "epoch": 0.8939873417721519, + "grad_norm": 0.47611561215232195, + "learning_rate": 4.7989777085466054e-05, + "loss": 0.7089, + "num_tokens": 592079992.0, + "step": 565 + }, + { + "epoch": 0.9018987341772152, + "grad_norm": 0.5649210203680972, + "learning_rate": 4.794085453589022e-05, + "loss": 0.7074, + "num_tokens": 597322872.0, + "step": 570 + }, + { + "epoch": 0.9098101265822784, + "grad_norm": 0.4954711968927705, + "learning_rate": 4.789137233640858e-05, + "loss": 0.7057, + "num_tokens": 602545993.0, + "step": 575 + }, + { + "epoch": 0.9177215189873418, + "grad_norm": 0.45694982466683615, + "learning_rate": 4.7841331841793646e-05, + "loss": 0.7099, + "num_tokens": 607788873.0, + "step": 580 + }, + { + "epoch": 0.9256329113924051, + "grad_norm": 0.4331078260141868, + "learning_rate": 4.7790734422103503e-05, + "loss": 0.7074, + "num_tokens": 613031753.0, + "step": 585 + }, + { + "epoch": 0.9335443037974683, + "grad_norm": 0.6029099335069901, + "learning_rate": 4.77395814626443e-05, + "loss": 0.7132, + "num_tokens": 618274069.0, + "step": 590 + }, + { + "epoch": 0.9414556962025317, + "grad_norm": 0.3964222914078214, + "learning_rate": 4.7687874363932277e-05, + "loss": 0.7107, + "num_tokens": 623516949.0, + "step": 595 + }, + { + "epoch": 0.9493670886075949, + "grad_norm": 0.40557860921547184, + "learning_rate": 4.763561454165546e-05, + "loss": 0.7114, + "num_tokens": 628759829.0, + "step": 600 + }, + { + "epoch": 0.9572784810126582, + "grad_norm": 0.3911037173148, + "learning_rate": 4.758280342663488e-05, + "loss": 0.7011, + "num_tokens": 634002709.0, + "step": 605 + }, + { + "epoch": 0.9651898734177216, + "grad_norm": 0.5205226454815663, + "learning_rate": 4.7529442464785404e-05, + "loss": 0.7035, + "num_tokens": 639245589.0, + "step": 610 + }, + { + "epoch": 0.9731012658227848, + "grad_norm": 0.6081952511481521, + "learning_rate": 4.747553311707616e-05, + "loss": 0.7109, + "num_tokens": 644488469.0, + "step": 615 + }, + { + "epoch": 0.9810126582278481, + "grad_norm": 0.5230701333389108, + "learning_rate": 4.742107685949052e-05, + "loss": 0.7054, + "num_tokens": 649731349.0, + "step": 620 + }, + { + "epoch": 0.9889240506329114, + "grad_norm": 0.45553268701488436, + "learning_rate": 4.736607518298568e-05, + "loss": 0.7084, + "num_tokens": 654971226.0, + "step": 625 + }, + { + "epoch": 0.9968354430379747, + "grad_norm": 0.4252255621189171, + "learning_rate": 4.731052959345188e-05, + "loss": 0.698, + "num_tokens": 660214106.0, + "step": 630 + }, + { + "epoch": 1.004746835443038, + "grad_norm": 0.5213909939853125, + "learning_rate": 4.725444161167113e-05, + "loss": 0.6984, + "num_tokens": 665456986.0, + "step": 635 + }, + { + "epoch": 1.0126582278481013, + "grad_norm": 0.6092766382306241, + "learning_rate": 4.71978127732756e-05, + "loss": 0.6834, + "num_tokens": 670699866.0, + "step": 640 + }, + { + "epoch": 1.0205696202531644, + "grad_norm": 0.6016879209190829, + "learning_rate": 4.714064462870556e-05, + "loss": 0.6799, + "num_tokens": 675941341.0, + "step": 645 + }, + { + "epoch": 1.0284810126582278, + "grad_norm": 0.46770946063736324, + "learning_rate": 4.708293874316693e-05, + "loss": 0.6718, + "num_tokens": 681184221.0, + "step": 650 + }, + { + "epoch": 1.0363924050632911, + "grad_norm": 0.5193625042396812, + "learning_rate": 4.702469669658845e-05, + "loss": 0.6824, + "num_tokens": 686427101.0, + "step": 655 + }, + { + "epoch": 1.0443037974683544, + "grad_norm": 0.4310707796867463, + "learning_rate": 4.6965920083578406e-05, + "loss": 0.6814, + "num_tokens": 691643199.0, + "step": 660 + }, + { + "epoch": 1.0522151898734178, + "grad_norm": 0.46509456309972086, + "learning_rate": 4.690661051338096e-05, + "loss": 0.6783, + "num_tokens": 696885395.0, + "step": 665 + }, + { + "epoch": 1.0601265822784811, + "grad_norm": 0.4290977849490662, + "learning_rate": 4.6846769609832106e-05, + "loss": 0.6745, + "num_tokens": 702099431.0, + "step": 670 + }, + { + "epoch": 1.0680379746835442, + "grad_norm": 0.44562761654453537, + "learning_rate": 4.6786399011315215e-05, + "loss": 0.6708, + "num_tokens": 707335241.0, + "step": 675 + }, + { + "epoch": 1.0759493670886076, + "grad_norm": 0.4027263571632704, + "learning_rate": 4.672550037071616e-05, + "loss": 0.6742, + "num_tokens": 712578121.0, + "step": 680 + }, + { + "epoch": 1.0838607594936709, + "grad_norm": 0.4509630202842952, + "learning_rate": 4.6664075355378064e-05, + "loss": 0.6787, + "num_tokens": 717813308.0, + "step": 685 + }, + { + "epoch": 1.0917721518987342, + "grad_norm": 0.579102128781878, + "learning_rate": 4.660212564705569e-05, + "loss": 0.6827, + "num_tokens": 723056188.0, + "step": 690 + }, + { + "epoch": 1.0996835443037976, + "grad_norm": 0.4750686074161037, + "learning_rate": 4.653965294186933e-05, + "loss": 0.6846, + "num_tokens": 728299068.0, + "step": 695 + }, + { + "epoch": 1.1075949367088607, + "grad_norm": 0.3720445468877555, + "learning_rate": 4.647665895025842e-05, + "loss": 0.68, + "num_tokens": 733541948.0, + "step": 700 + }, + { + "epoch": 1.115506329113924, + "grad_norm": 0.41896058368158273, + "learning_rate": 4.6413145396934677e-05, + "loss": 0.6727, + "num_tokens": 738784828.0, + "step": 705 + }, + { + "epoch": 1.1234177215189873, + "grad_norm": 0.41709918206048263, + "learning_rate": 4.634911402083491e-05, + "loss": 0.683, + "num_tokens": 744027708.0, + "step": 710 + }, + { + "epoch": 1.1313291139240507, + "grad_norm": 0.44679889502370607, + "learning_rate": 4.628456657507336e-05, + "loss": 0.6745, + "num_tokens": 749270588.0, + "step": 715 + }, + { + "epoch": 1.139240506329114, + "grad_norm": 0.4049444543836899, + "learning_rate": 4.6219504826893774e-05, + "loss": 0.6754, + "num_tokens": 754513468.0, + "step": 720 + }, + { + "epoch": 1.1471518987341773, + "grad_norm": 0.46788618407855487, + "learning_rate": 4.615393055762095e-05, + "loss": 0.6705, + "num_tokens": 759755275.0, + "step": 725 + }, + { + "epoch": 1.1550632911392404, + "grad_norm": 0.45889187811298193, + "learning_rate": 4.6087845562612e-05, + "loss": 0.6775, + "num_tokens": 764993238.0, + "step": 730 + }, + { + "epoch": 1.1629746835443038, + "grad_norm": 0.5053059074244526, + "learning_rate": 4.602125165120721e-05, + "loss": 0.6739, + "num_tokens": 770236118.0, + "step": 735 + }, + { + "epoch": 1.1708860759493671, + "grad_norm": 0.5163271050928327, + "learning_rate": 4.595415064668044e-05, + "loss": 0.6768, + "num_tokens": 775478998.0, + "step": 740 + }, + { + "epoch": 1.1787974683544304, + "grad_norm": 0.4282870421961532, + "learning_rate": 4.5886544386189286e-05, + "loss": 0.6681, + "num_tokens": 780721878.0, + "step": 745 + }, + { + "epoch": 1.1867088607594938, + "grad_norm": 0.46026364834735695, + "learning_rate": 4.581843472072472e-05, + "loss": 0.677, + "num_tokens": 785964758.0, + "step": 750 + }, + { + "epoch": 1.1946202531645569, + "grad_norm": 0.3647491363050659, + "learning_rate": 4.574982351506044e-05, + "loss": 0.673, + "num_tokens": 791207638.0, + "step": 755 + }, + { + "epoch": 1.2025316455696202, + "grad_norm": 0.40334584668377854, + "learning_rate": 4.568071264770179e-05, + "loss": 0.6733, + "num_tokens": 796450518.0, + "step": 760 + }, + { + "epoch": 1.2104430379746836, + "grad_norm": 0.6172425933341248, + "learning_rate": 4.5611104010834384e-05, + "loss": 0.6714, + "num_tokens": 801691627.0, + "step": 765 + }, + { + "epoch": 1.2183544303797469, + "grad_norm": 0.5660318449559009, + "learning_rate": 4.554099951027223e-05, + "loss": 0.6695, + "num_tokens": 806934507.0, + "step": 770 + }, + { + "epoch": 1.2262658227848102, + "grad_norm": 0.5013160792670336, + "learning_rate": 4.5470401065405574e-05, + "loss": 0.6733, + "num_tokens": 812177387.0, + "step": 775 + }, + { + "epoch": 1.2341772151898733, + "grad_norm": 0.48633536466547755, + "learning_rate": 4.539931060914841e-05, + "loss": 0.6775, + "num_tokens": 817420267.0, + "step": 780 + }, + { + "epoch": 1.2420886075949367, + "grad_norm": 0.5345960534136472, + "learning_rate": 4.532773008788542e-05, + "loss": 0.6742, + "num_tokens": 822663147.0, + "step": 785 + }, + { + "epoch": 1.25, + "grad_norm": 0.46751637317248607, + "learning_rate": 4.5255661461418854e-05, + "loss": 0.6683, + "num_tokens": 827902185.0, + "step": 790 + }, + { + "epoch": 1.2579113924050633, + "grad_norm": 0.40210236140036404, + "learning_rate": 4.5183106702914744e-05, + "loss": 0.6684, + "num_tokens": 833145065.0, + "step": 795 + }, + { + "epoch": 1.2658227848101267, + "grad_norm": 0.40800636198331647, + "learning_rate": 4.511006779884894e-05, + "loss": 0.6717, + "num_tokens": 838387945.0, + "step": 800 + }, + { + "epoch": 1.2737341772151898, + "grad_norm": 0.39143626003055415, + "learning_rate": 4.503654674895268e-05, + "loss": 0.6729, + "num_tokens": 843630825.0, + "step": 805 + }, + { + "epoch": 1.2816455696202531, + "grad_norm": 0.4114368812178738, + "learning_rate": 4.49625455661579e-05, + "loss": 0.6662, + "num_tokens": 848873705.0, + "step": 810 + }, + { + "epoch": 1.2895569620253164, + "grad_norm": 0.3850979319103424, + "learning_rate": 4.4888066276542076e-05, + "loss": 0.6676, + "num_tokens": 854116585.0, + "step": 815 + }, + { + "epoch": 1.2974683544303798, + "grad_norm": 0.3942547463552602, + "learning_rate": 4.481311091927278e-05, + "loss": 0.6697, + "num_tokens": 859340010.0, + "step": 820 + }, + { + "epoch": 1.3053797468354431, + "grad_norm": 0.3827683736050918, + "learning_rate": 4.47376815465518e-05, + "loss": 0.6648, + "num_tokens": 864560605.0, + "step": 825 + }, + { + "epoch": 1.3132911392405062, + "grad_norm": 0.4613241812192372, + "learning_rate": 4.466178022355902e-05, + "loss": 0.6757, + "num_tokens": 869803485.0, + "step": 830 + }, + { + "epoch": 1.3212025316455696, + "grad_norm": 0.3407656518079187, + "learning_rate": 4.458540902839582e-05, + "loss": 0.6643, + "num_tokens": 875046365.0, + "step": 835 + }, + { + "epoch": 1.3291139240506329, + "grad_norm": 0.3433476064598642, + "learning_rate": 4.450857005202823e-05, + "loss": 0.6684, + "num_tokens": 880272119.0, + "step": 840 + }, + { + "epoch": 1.3370253164556962, + "grad_norm": 0.4072649317968132, + "learning_rate": 4.443126539822962e-05, + "loss": 0.6758, + "num_tokens": 885514999.0, + "step": 845 + }, + { + "epoch": 1.3449367088607596, + "grad_norm": 0.38388206068514286, + "learning_rate": 4.435349718352319e-05, + "loss": 0.6651, + "num_tokens": 890751664.0, + "step": 850 + }, + { + "epoch": 1.3528481012658227, + "grad_norm": 0.4906281471488268, + "learning_rate": 4.427526753712392e-05, + "loss": 0.6758, + "num_tokens": 895978998.0, + "step": 855 + }, + { + "epoch": 1.360759493670886, + "grad_norm": 0.5690161031589217, + "learning_rate": 4.419657860088033e-05, + "loss": 0.6661, + "num_tokens": 901206430.0, + "step": 860 + }, + { + "epoch": 1.3686708860759493, + "grad_norm": 0.5239364843573313, + "learning_rate": 4.411743252921587e-05, + "loss": 0.6711, + "num_tokens": 906449310.0, + "step": 865 + }, + { + "epoch": 1.3765822784810127, + "grad_norm": 0.4590626844046868, + "learning_rate": 4.403783148906984e-05, + "loss": 0.6742, + "num_tokens": 911692190.0, + "step": 870 + }, + { + "epoch": 1.384493670886076, + "grad_norm": 0.6369825573287001, + "learning_rate": 4.3957777659838156e-05, + "loss": 0.6691, + "num_tokens": 916923938.0, + "step": 875 + }, + { + "epoch": 1.3924050632911391, + "grad_norm": 0.5309325547971587, + "learning_rate": 4.387727323331362e-05, + "loss": 0.6671, + "num_tokens": 922166818.0, + "step": 880 + }, + { + "epoch": 1.4003164556962027, + "grad_norm": 0.4182221238855379, + "learning_rate": 4.379632041362594e-05, + "loss": 0.6673, + "num_tokens": 927409698.0, + "step": 885 + }, + { + "epoch": 1.4082278481012658, + "grad_norm": 0.37453984190952516, + "learning_rate": 4.371492141718138e-05, + "loss": 0.6669, + "num_tokens": 932652578.0, + "step": 890 + }, + { + "epoch": 1.4161392405063291, + "grad_norm": 0.4472314832066788, + "learning_rate": 4.363307847260206e-05, + "loss": 0.6609, + "num_tokens": 937883810.0, + "step": 895 + }, + { + "epoch": 1.4240506329113924, + "grad_norm": 0.40759290291866934, + "learning_rate": 4.3550793820664965e-05, + "loss": 0.6657, + "num_tokens": 943113159.0, + "step": 900 + }, + { + "epoch": 1.4319620253164558, + "grad_norm": 0.38119165043031206, + "learning_rate": 4.3468069714240556e-05, + "loss": 0.6674, + "num_tokens": 948346815.0, + "step": 905 + }, + { + "epoch": 1.439873417721519, + "grad_norm": 0.3408121507908347, + "learning_rate": 4.3384908418231144e-05, + "loss": 0.6649, + "num_tokens": 953589695.0, + "step": 910 + }, + { + "epoch": 1.4477848101265822, + "grad_norm": 0.3824255603451359, + "learning_rate": 4.330131220950883e-05, + "loss": 0.664, + "num_tokens": 958831498.0, + "step": 915 + }, + { + "epoch": 1.4556962025316456, + "grad_norm": 0.42035687659685655, + "learning_rate": 4.321728337685318e-05, + "loss": 0.667, + "num_tokens": 964074378.0, + "step": 920 + }, + { + "epoch": 1.4636075949367089, + "grad_norm": 0.45598858937412523, + "learning_rate": 4.313282422088859e-05, + "loss": 0.6685, + "num_tokens": 969317258.0, + "step": 925 + }, + { + "epoch": 1.4715189873417722, + "grad_norm": 0.44573187825703264, + "learning_rate": 4.304793705402124e-05, + "loss": 0.6613, + "num_tokens": 974546034.0, + "step": 930 + }, + { + "epoch": 1.4794303797468356, + "grad_norm": 0.3946235668573835, + "learning_rate": 4.2962624200375835e-05, + "loss": 0.6671, + "num_tokens": 979788914.0, + "step": 935 + }, + { + "epoch": 1.4873417721518987, + "grad_norm": 0.3228711597439979, + "learning_rate": 4.287688799573195e-05, + "loss": 0.6686, + "num_tokens": 985031794.0, + "step": 940 + }, + { + "epoch": 1.495253164556962, + "grad_norm": 0.3636228559548819, + "learning_rate": 4.2790730787460096e-05, + "loss": 0.6657, + "num_tokens": 990255694.0, + "step": 945 + }, + { + "epoch": 1.5031645569620253, + "grad_norm": 0.36277524743915257, + "learning_rate": 4.270415493445739e-05, + "loss": 0.6618, + "num_tokens": 995498574.0, + "step": 950 + }, + { + "epoch": 1.5110759493670884, + "grad_norm": 0.449377328776007, + "learning_rate": 4.2617162807083084e-05, + "loss": 0.664, + "num_tokens": 1000741454.0, + "step": 955 + }, + { + "epoch": 1.518987341772152, + "grad_norm": 0.5226848678404952, + "learning_rate": 4.252975678709354e-05, + "loss": 0.664, + "num_tokens": 1005981331.0, + "step": 960 + }, + { + "epoch": 1.5268987341772151, + "grad_norm": 0.426961497295756, + "learning_rate": 4.244193926757713e-05, + "loss": 0.6672, + "num_tokens": 1011224211.0, + "step": 965 + }, + { + "epoch": 1.5348101265822784, + "grad_norm": 0.3336409027569924, + "learning_rate": 4.235371265288864e-05, + "loss": 0.6681, + "num_tokens": 1016467091.0, + "step": 970 + }, + { + "epoch": 1.5427215189873418, + "grad_norm": 0.4036839066490488, + "learning_rate": 4.2265079358583455e-05, + "loss": 0.6643, + "num_tokens": 1021695784.0, + "step": 975 + }, + { + "epoch": 1.5506329113924051, + "grad_norm": 0.7140710792273346, + "learning_rate": 4.217604181135147e-05, + "loss": 0.6697, + "num_tokens": 1026938664.0, + "step": 980 + }, + { + "epoch": 1.5585443037974684, + "grad_norm": 0.5962932136354121, + "learning_rate": 4.20866024489506e-05, + "loss": 0.6636, + "num_tokens": 1032181544.0, + "step": 985 + }, + { + "epoch": 1.5664556962025316, + "grad_norm": 0.5399117523136845, + "learning_rate": 4.199676372014004e-05, + "loss": 0.6642, + "num_tokens": 1037424424.0, + "step": 990 + }, + { + "epoch": 1.5743670886075949, + "grad_norm": 0.5727471985317703, + "learning_rate": 4.190652808461326e-05, + "loss": 0.6673, + "num_tokens": 1042667304.0, + "step": 995 + }, + { + "epoch": 1.5822784810126582, + "grad_norm": 0.4218917245287635, + "learning_rate": 4.1815898012930603e-05, + "loss": 0.6597, + "num_tokens": 1047910184.0, + "step": 1000 + }, + { + "epoch": 1.5901898734177216, + "grad_norm": 0.3919982695655947, + "learning_rate": 4.172487598645171e-05, + "loss": 0.6587, + "num_tokens": 1053153064.0, + "step": 1005 + }, + { + "epoch": 1.5981012658227849, + "grad_norm": 0.35433285883968835, + "learning_rate": 4.163346449726752e-05, + "loss": 0.6658, + "num_tokens": 1058395944.0, + "step": 1010 + }, + { + "epoch": 1.606012658227848, + "grad_norm": 0.40624024878352494, + "learning_rate": 4.154166604813206e-05, + "loss": 0.661, + "num_tokens": 1063638824.0, + "step": 1015 + }, + { + "epoch": 1.6139240506329116, + "grad_norm": 0.45360696653600874, + "learning_rate": 4.144948315239397e-05, + "loss": 0.6636, + "num_tokens": 1068880391.0, + "step": 1020 + }, + { + "epoch": 1.6218354430379747, + "grad_norm": 0.528510428398564, + "learning_rate": 4.13569183339276e-05, + "loss": 0.6561, + "num_tokens": 1074123271.0, + "step": 1025 + }, + { + "epoch": 1.629746835443038, + "grad_norm": 0.435117205930865, + "learning_rate": 4.1263974127063994e-05, + "loss": 0.6572, + "num_tokens": 1079366151.0, + "step": 1030 + }, + { + "epoch": 1.6376582278481013, + "grad_norm": 0.35010239698525997, + "learning_rate": 4.117065307652145e-05, + "loss": 0.6675, + "num_tokens": 1084603567.0, + "step": 1035 + }, + { + "epoch": 1.6455696202531644, + "grad_norm": 0.34340761605514514, + "learning_rate": 4.107695773733586e-05, + "loss": 0.6634, + "num_tokens": 1089846447.0, + "step": 1040 + }, + { + "epoch": 1.653481012658228, + "grad_norm": 0.3800590366717773, + "learning_rate": 4.098289067479077e-05, + "loss": 0.6529, + "num_tokens": 1095089327.0, + "step": 1045 + }, + { + "epoch": 1.6613924050632911, + "grad_norm": 0.5301647765231198, + "learning_rate": 4.0888454464347156e-05, + "loss": 0.6608, + "num_tokens": 1100328290.0, + "step": 1050 + }, + { + "epoch": 1.6693037974683544, + "grad_norm": 0.40371596065925885, + "learning_rate": 4.079365169157283e-05, + "loss": 0.6596, + "num_tokens": 1105565945.0, + "step": 1055 + }, + { + "epoch": 1.6772151898734178, + "grad_norm": 0.44878584304658214, + "learning_rate": 4.069848495207176e-05, + "loss": 0.6584, + "num_tokens": 1110808825.0, + "step": 1060 + }, + { + "epoch": 1.685126582278481, + "grad_norm": 0.42760590310826146, + "learning_rate": 4.060295685141295e-05, + "loss": 0.6601, + "num_tokens": 1116051705.0, + "step": 1065 + }, + { + "epoch": 1.6930379746835444, + "grad_norm": 0.42094155594543997, + "learning_rate": 4.0507070005059086e-05, + "loss": 0.6636, + "num_tokens": 1121294585.0, + "step": 1070 + }, + { + "epoch": 1.7009493670886076, + "grad_norm": 0.366450602326517, + "learning_rate": 4.0410827038294966e-05, + "loss": 0.6565, + "num_tokens": 1126537465.0, + "step": 1075 + }, + { + "epoch": 1.7088607594936709, + "grad_norm": 0.3382264432193073, + "learning_rate": 4.031423058615559e-05, + "loss": 0.6637, + "num_tokens": 1131778529.0, + "step": 1080 + }, + { + "epoch": 1.7167721518987342, + "grad_norm": 0.4135053453091331, + "learning_rate": 4.0217283293354044e-05, + "loss": 0.6598, + "num_tokens": 1137021409.0, + "step": 1085 + }, + { + "epoch": 1.7246835443037973, + "grad_norm": 0.32789223717337174, + "learning_rate": 4.011998781420907e-05, + "loss": 0.66, + "num_tokens": 1142264289.0, + "step": 1090 + }, + { + "epoch": 1.7325949367088609, + "grad_norm": 0.37159495208450527, + "learning_rate": 4.002234681257239e-05, + "loss": 0.6634, + "num_tokens": 1147488495.0, + "step": 1095 + }, + { + "epoch": 1.740506329113924, + "grad_norm": 0.41706053572545077, + "learning_rate": 3.992436296175581e-05, + "loss": 0.6554, + "num_tokens": 1152731375.0, + "step": 1100 + }, + { + "epoch": 1.7484177215189873, + "grad_norm": 0.3410031339365452, + "learning_rate": 3.982603894445796e-05, + "loss": 0.6574, + "num_tokens": 1157974255.0, + "step": 1105 + }, + { + "epoch": 1.7563291139240507, + "grad_norm": 0.40387170849869736, + "learning_rate": 3.97273774526909e-05, + "loss": 0.6518, + "num_tokens": 1163196277.0, + "step": 1110 + }, + { + "epoch": 1.7642405063291138, + "grad_norm": 0.41494353564073877, + "learning_rate": 3.962838118770643e-05, + "loss": 0.6521, + "num_tokens": 1168438593.0, + "step": 1115 + }, + { + "epoch": 1.7721518987341773, + "grad_norm": 0.33845287541155555, + "learning_rate": 3.952905285992206e-05, + "loss": 0.6636, + "num_tokens": 1173674462.0, + "step": 1120 + }, + { + "epoch": 1.7800632911392404, + "grad_norm": 0.3616932835037642, + "learning_rate": 3.942939518884686e-05, + "loss": 0.6566, + "num_tokens": 1178914750.0, + "step": 1125 + }, + { + "epoch": 1.7879746835443038, + "grad_norm": 0.38849401008078355, + "learning_rate": 3.932941090300699e-05, + "loss": 0.6546, + "num_tokens": 1184150094.0, + "step": 1130 + }, + { + "epoch": 1.7958860759493671, + "grad_norm": 0.42281182227916236, + "learning_rate": 3.922910273987098e-05, + "loss": 0.6548, + "num_tokens": 1189392974.0, + "step": 1135 + }, + { + "epoch": 1.8037974683544302, + "grad_norm": 0.4210942734836147, + "learning_rate": 3.912847344577481e-05, + "loss": 0.6555, + "num_tokens": 1194635854.0, + "step": 1140 + }, + { + "epoch": 1.8117088607594938, + "grad_norm": 0.5474931691520276, + "learning_rate": 3.9027525775846666e-05, + "loss": 0.6533, + "num_tokens": 1199878734.0, + "step": 1145 + }, + { + "epoch": 1.8196202531645569, + "grad_norm": 0.3732844050319297, + "learning_rate": 3.892626249393159e-05, + "loss": 0.6533, + "num_tokens": 1205121614.0, + "step": 1150 + }, + { + "epoch": 1.8275316455696202, + "grad_norm": 0.45982960737179485, + "learning_rate": 3.882468637251573e-05, + "loss": 0.658, + "num_tokens": 1210364494.0, + "step": 1155 + }, + { + "epoch": 1.8354430379746836, + "grad_norm": 0.3489396715842175, + "learning_rate": 3.872280019265046e-05, + "loss": 0.6583, + "num_tokens": 1215607374.0, + "step": 1160 + }, + { + "epoch": 1.8433544303797469, + "grad_norm": 0.35751783832753387, + "learning_rate": 3.8620606743876264e-05, + "loss": 0.6522, + "num_tokens": 1220843801.0, + "step": 1165 + }, + { + "epoch": 1.8512658227848102, + "grad_norm": 0.4188675688211132, + "learning_rate": 3.851810882414632e-05, + "loss": 0.6497, + "num_tokens": 1226086681.0, + "step": 1170 + }, + { + "epoch": 1.8591772151898733, + "grad_norm": 0.3704256158964576, + "learning_rate": 3.841530923974991e-05, + "loss": 0.6542, + "num_tokens": 1231329561.0, + "step": 1175 + }, + { + "epoch": 1.8670886075949367, + "grad_norm": 0.35739760458738973, + "learning_rate": 3.83122108052356e-05, + "loss": 0.6558, + "num_tokens": 1236572441.0, + "step": 1180 + }, + { + "epoch": 1.875, + "grad_norm": 0.4108797442075266, + "learning_rate": 3.8208816343334156e-05, + "loss": 0.6503, + "num_tokens": 1241815321.0, + "step": 1185 + }, + { + "epoch": 1.8829113924050633, + "grad_norm": 0.3989382930649431, + "learning_rate": 3.810512868488129e-05, + "loss": 0.6567, + "num_tokens": 1247058201.0, + "step": 1190 + }, + { + "epoch": 1.8908227848101267, + "grad_norm": 0.36123580918393156, + "learning_rate": 3.800115066874014e-05, + "loss": 0.6539, + "num_tokens": 1252301081.0, + "step": 1195 + }, + { + "epoch": 1.8987341772151898, + "grad_norm": 0.3372531720914223, + "learning_rate": 3.789688514172353e-05, + "loss": 0.6473, + "num_tokens": 1257543961.0, + "step": 1200 + }, + { + "epoch": 1.9066455696202531, + "grad_norm": 0.3512013159202198, + "learning_rate": 3.779233495851604e-05, + "loss": 0.6581, + "num_tokens": 1262786841.0, + "step": 1205 + }, + { + "epoch": 1.9145569620253164, + "grad_norm": 0.3542753269510345, + "learning_rate": 3.768750298159587e-05, + "loss": 0.6445, + "num_tokens": 1268029721.0, + "step": 1210 + }, + { + "epoch": 1.9224683544303798, + "grad_norm": 0.4031550450820841, + "learning_rate": 3.75823920811564e-05, + "loss": 0.6508, + "num_tokens": 1273258360.0, + "step": 1215 + }, + { + "epoch": 1.9303797468354431, + "grad_norm": 0.4403211137259473, + "learning_rate": 3.747700513502772e-05, + "loss": 0.6573, + "num_tokens": 1278501240.0, + "step": 1220 + }, + { + "epoch": 1.9382911392405062, + "grad_norm": 0.4812191269085378, + "learning_rate": 3.737134502859772e-05, + "loss": 0.6536, + "num_tokens": 1283744120.0, + "step": 1225 + }, + { + "epoch": 1.9462025316455698, + "grad_norm": 0.39299909706187625, + "learning_rate": 3.726541465473317e-05, + "loss": 0.649, + "num_tokens": 1288987000.0, + "step": 1230 + }, + { + "epoch": 1.9541139240506329, + "grad_norm": 0.39824695051848524, + "learning_rate": 3.7159216913700456e-05, + "loss": 0.6478, + "num_tokens": 1294229880.0, + "step": 1235 + }, + { + "epoch": 1.9620253164556962, + "grad_norm": 0.42039758141791633, + "learning_rate": 3.7052754713086246e-05, + "loss": 0.6534, + "num_tokens": 1299472760.0, + "step": 1240 + }, + { + "epoch": 1.9699367088607596, + "grad_norm": 0.3960672982967281, + "learning_rate": 3.694603096771781e-05, + "loss": 0.6568, + "num_tokens": 1304715640.0, + "step": 1245 + }, + { + "epoch": 1.9778481012658227, + "grad_norm": 0.3747599520232236, + "learning_rate": 3.683904859958329e-05, + "loss": 0.651, + "num_tokens": 1309942452.0, + "step": 1250 + }, + { + "epoch": 1.9857594936708862, + "grad_norm": 0.34963602368732627, + "learning_rate": 3.673181053775162e-05, + "loss": 0.6591, + "num_tokens": 1315185332.0, + "step": 1255 + }, + { + "epoch": 1.9936708860759493, + "grad_norm": 0.3336233742912806, + "learning_rate": 3.662431971829237e-05, + "loss": 0.6567, + "num_tokens": 1320428212.0, + "step": 1260 + }, + { + "epoch": 2.0015822784810124, + "grad_norm": 0.34626223476299406, + "learning_rate": 3.651657908419537e-05, + "loss": 0.6528, + "num_tokens": 1325656988.0, + "step": 1265 + }, + { + "epoch": 2.009493670886076, + "grad_norm": 0.3739591180690111, + "learning_rate": 3.640859158529014e-05, + "loss": 0.6228, + "num_tokens": 1330899868.0, + "step": 1270 + }, + { + "epoch": 2.017405063291139, + "grad_norm": 0.36163026913569957, + "learning_rate": 3.6300360178165065e-05, + "loss": 0.6198, + "num_tokens": 1336142748.0, + "step": 1275 + }, + { + "epoch": 2.0253164556962027, + "grad_norm": 0.3413456124109847, + "learning_rate": 3.619188782608653e-05, + "loss": 0.6249, + "num_tokens": 1341371387.0, + "step": 1280 + }, + { + "epoch": 2.0332278481012658, + "grad_norm": 0.41603416587113357, + "learning_rate": 3.6083177498917745e-05, + "loss": 0.6137, + "num_tokens": 1346614267.0, + "step": 1285 + }, + { + "epoch": 2.041139240506329, + "grad_norm": 0.39514773706497214, + "learning_rate": 3.5974232173037385e-05, + "loss": 0.6292, + "num_tokens": 1351857147.0, + "step": 1290 + }, + { + "epoch": 2.0490506329113924, + "grad_norm": 0.3949477431561823, + "learning_rate": 3.586505483125823e-05, + "loss": 0.6204, + "num_tokens": 1357100027.0, + "step": 1295 + }, + { + "epoch": 2.0569620253164556, + "grad_norm": 0.324048846717354, + "learning_rate": 3.5755648462745366e-05, + "loss": 0.6263, + "num_tokens": 1362342907.0, + "step": 1300 + }, + { + "epoch": 2.064873417721519, + "grad_norm": 0.3478785681280589, + "learning_rate": 3.5646016062934413e-05, + "loss": 0.6216, + "num_tokens": 1367576563.0, + "step": 1305 + }, + { + "epoch": 2.0727848101265822, + "grad_norm": 0.38436773833943666, + "learning_rate": 3.553616063344951e-05, + "loss": 0.6255, + "num_tokens": 1372800486.0, + "step": 1310 + }, + { + "epoch": 2.0806962025316458, + "grad_norm": 0.38048821813178435, + "learning_rate": 3.5426085182021114e-05, + "loss": 0.6208, + "num_tokens": 1378043366.0, + "step": 1315 + }, + { + "epoch": 2.088607594936709, + "grad_norm": 0.3868214542706034, + "learning_rate": 3.531579272240366e-05, + "loss": 0.6235, + "num_tokens": 1383286246.0, + "step": 1320 + }, + { + "epoch": 2.096518987341772, + "grad_norm": 0.29924119376240554, + "learning_rate": 3.520528627429304e-05, + "loss": 0.616, + "num_tokens": 1388529126.0, + "step": 1325 + }, + { + "epoch": 2.1044303797468356, + "grad_norm": 0.30143694906012813, + "learning_rate": 3.509456886324395e-05, + "loss": 0.6228, + "num_tokens": 1393772006.0, + "step": 1330 + }, + { + "epoch": 2.1123417721518987, + "grad_norm": 0.3430834549488217, + "learning_rate": 3.498364352058703e-05, + "loss": 0.6231, + "num_tokens": 1399014886.0, + "step": 1335 + }, + { + "epoch": 2.1202531645569622, + "grad_norm": 0.3918758750528519, + "learning_rate": 3.487251328334588e-05, + "loss": 0.6184, + "num_tokens": 1404254509.0, + "step": 1340 + }, + { + "epoch": 2.1281645569620253, + "grad_norm": 0.3860740822731293, + "learning_rate": 3.47611811941539e-05, + "loss": 0.6223, + "num_tokens": 1409497389.0, + "step": 1345 + }, + { + "epoch": 2.1360759493670884, + "grad_norm": 0.30507806139472327, + "learning_rate": 3.464965030117099e-05, + "loss": 0.6227, + "num_tokens": 1414740269.0, + "step": 1350 + }, + { + "epoch": 2.143987341772152, + "grad_norm": 0.31931935671102873, + "learning_rate": 3.453792365800011e-05, + "loss": 0.6213, + "num_tokens": 1419976138.0, + "step": 1355 + }, + { + "epoch": 2.151898734177215, + "grad_norm": 0.4296009660896887, + "learning_rate": 3.4426004323603655e-05, + "loss": 0.6195, + "num_tokens": 1425219018.0, + "step": 1360 + }, + { + "epoch": 2.1598101265822787, + "grad_norm": 0.36080348698318637, + "learning_rate": 3.4313895362219704e-05, + "loss": 0.6321, + "num_tokens": 1430461898.0, + "step": 1365 + }, + { + "epoch": 2.1677215189873418, + "grad_norm": 0.31266622534218946, + "learning_rate": 3.420159984327814e-05, + "loss": 0.6329, + "num_tokens": 1435704778.0, + "step": 1370 + }, + { + "epoch": 2.175632911392405, + "grad_norm": 0.29368200019966956, + "learning_rate": 3.40891208413166e-05, + "loss": 0.622, + "num_tokens": 1440941205.0, + "step": 1375 + }, + { + "epoch": 2.1835443037974684, + "grad_norm": 0.3318857805197203, + "learning_rate": 3.397646143589629e-05, + "loss": 0.6218, + "num_tokens": 1446182176.0, + "step": 1380 + }, + { + "epoch": 2.1914556962025316, + "grad_norm": 0.3469231890169258, + "learning_rate": 3.38636247115177e-05, + "loss": 0.6241, + "num_tokens": 1451425056.0, + "step": 1385 + }, + { + "epoch": 2.199367088607595, + "grad_norm": 0.3555477367643954, + "learning_rate": 3.3750613757536134e-05, + "loss": 0.6237, + "num_tokens": 1456667936.0, + "step": 1390 + }, + { + "epoch": 2.2072784810126582, + "grad_norm": 0.4308050328882218, + "learning_rate": 3.36374316680771e-05, + "loss": 0.6269, + "num_tokens": 1461909000.0, + "step": 1395 + }, + { + "epoch": 2.2151898734177213, + "grad_norm": 0.35870973114272886, + "learning_rate": 3.3524081541951644e-05, + "loss": 0.6222, + "num_tokens": 1467150803.0, + "step": 1400 + }, + { + "epoch": 2.223101265822785, + "grad_norm": 0.3606193086420081, + "learning_rate": 3.3410566482571484e-05, + "loss": 0.625, + "num_tokens": 1472393683.0, + "step": 1405 + }, + { + "epoch": 2.231012658227848, + "grad_norm": 0.3173225301332037, + "learning_rate": 3.329688959786403e-05, + "loss": 0.6225, + "num_tokens": 1477635158.0, + "step": 1410 + }, + { + "epoch": 2.2389240506329116, + "grad_norm": 0.3691990634713518, + "learning_rate": 3.318305400018732e-05, + "loss": 0.6231, + "num_tokens": 1482878038.0, + "step": 1415 + }, + { + "epoch": 2.2468354430379747, + "grad_norm": 0.4172329830772149, + "learning_rate": 3.306906280624479e-05, + "loss": 0.6286, + "num_tokens": 1488105187.0, + "step": 1420 + }, + { + "epoch": 2.254746835443038, + "grad_norm": 0.3279503525329187, + "learning_rate": 3.295491913699994e-05, + "loss": 0.6233, + "num_tokens": 1493346994.0, + "step": 1425 + }, + { + "epoch": 2.2626582278481013, + "grad_norm": 0.3632060308412657, + "learning_rate": 3.284062611759089e-05, + "loss": 0.6226, + "num_tokens": 1498589874.0, + "step": 1430 + }, + { + "epoch": 2.2705696202531644, + "grad_norm": 0.46252170164225265, + "learning_rate": 3.272618687724481e-05, + "loss": 0.6225, + "num_tokens": 1503832754.0, + "step": 1435 + }, + { + "epoch": 2.278481012658228, + "grad_norm": 0.3317711593176921, + "learning_rate": 3.261160454919228e-05, + "loss": 0.6236, + "num_tokens": 1509075634.0, + "step": 1440 + }, + { + "epoch": 2.286392405063291, + "grad_norm": 0.3470857725220561, + "learning_rate": 3.249688227058145e-05, + "loss": 0.6273, + "num_tokens": 1514318514.0, + "step": 1445 + }, + { + "epoch": 2.2943037974683547, + "grad_norm": 0.29280649749694565, + "learning_rate": 3.238202318239216e-05, + "loss": 0.6304, + "num_tokens": 1519551983.0, + "step": 1450 + }, + { + "epoch": 2.3022151898734178, + "grad_norm": 0.37241388988066654, + "learning_rate": 3.226703042935e-05, + "loss": 0.6184, + "num_tokens": 1524794863.0, + "step": 1455 + }, + { + "epoch": 2.310126582278481, + "grad_norm": 0.33976317790538424, + "learning_rate": 3.215190715984013e-05, + "loss": 0.622, + "num_tokens": 1530031528.0, + "step": 1460 + }, + { + "epoch": 2.3180379746835444, + "grad_norm": 0.29066911940116874, + "learning_rate": 3.2036656525821144e-05, + "loss": 0.6231, + "num_tokens": 1535274408.0, + "step": 1465 + }, + { + "epoch": 2.3259493670886076, + "grad_norm": 0.28162811956261474, + "learning_rate": 3.1921281682738734e-05, + "loss": 0.6182, + "num_tokens": 1540517288.0, + "step": 1470 + }, + { + "epoch": 2.333860759493671, + "grad_norm": 0.3524371457891882, + "learning_rate": 3.180578578943933e-05, + "loss": 0.6294, + "num_tokens": 1545748994.0, + "step": 1475 + }, + { + "epoch": 2.3417721518987342, + "grad_norm": 0.3468858213633566, + "learning_rate": 3.1690172008083554e-05, + "loss": 0.619, + "num_tokens": 1550991874.0, + "step": 1480 + }, + { + "epoch": 2.3496835443037973, + "grad_norm": 0.4224520445380759, + "learning_rate": 3.157444350405972e-05, + "loss": 0.6257, + "num_tokens": 1556223622.0, + "step": 1485 + }, + { + "epoch": 2.357594936708861, + "grad_norm": 0.3754485859450013, + "learning_rate": 3.1458603445897136e-05, + "loss": 0.6255, + "num_tokens": 1561439202.0, + "step": 1490 + }, + { + "epoch": 2.365506329113924, + "grad_norm": 0.3607328127982716, + "learning_rate": 3.134265500517932e-05, + "loss": 0.6184, + "num_tokens": 1566661187.0, + "step": 1495 + }, + { + "epoch": 2.3734177215189876, + "grad_norm": 0.3016702669665355, + "learning_rate": 3.122660135645721e-05, + "loss": 0.6226, + "num_tokens": 1571904067.0, + "step": 1500 + }, + { + "epoch": 2.3813291139240507, + "grad_norm": 0.3397158516715143, + "learning_rate": 3.111044567716223e-05, + "loss": 0.6237, + "num_tokens": 1577146947.0, + "step": 1505 + }, + { + "epoch": 2.3892405063291138, + "grad_norm": 0.34056042132518993, + "learning_rate": 3.0994191147519296e-05, + "loss": 0.6205, + "num_tokens": 1582389827.0, + "step": 1510 + }, + { + "epoch": 2.3971518987341773, + "grad_norm": 0.34821134798667835, + "learning_rate": 3.0877840950459763e-05, + "loss": 0.6217, + "num_tokens": 1587632707.0, + "step": 1515 + }, + { + "epoch": 2.4050632911392404, + "grad_norm": 0.29071305472990666, + "learning_rate": 3.076139827153425e-05, + "loss": 0.6212, + "num_tokens": 1592875587.0, + "step": 1520 + }, + { + "epoch": 2.412974683544304, + "grad_norm": 0.3213307109034431, + "learning_rate": 3.064486629882543e-05, + "loss": 0.6261, + "num_tokens": 1598118467.0, + "step": 1525 + }, + { + "epoch": 2.420886075949367, + "grad_norm": 0.317409439920391, + "learning_rate": 3.052824822286075e-05, + "loss": 0.6192, + "num_tokens": 1603361347.0, + "step": 1530 + }, + { + "epoch": 2.4287974683544302, + "grad_norm": 0.31888869908174583, + "learning_rate": 3.041154723652509e-05, + "loss": 0.6208, + "num_tokens": 1608604227.0, + "step": 1535 + }, + { + "epoch": 2.4367088607594938, + "grad_norm": 0.2946434998719226, + "learning_rate": 3.0294766534973335e-05, + "loss": 0.6291, + "num_tokens": 1613847107.0, + "step": 1540 + }, + { + "epoch": 2.444620253164557, + "grad_norm": 0.34443852085875215, + "learning_rate": 3.0177909315542862e-05, + "loss": 0.6256, + "num_tokens": 1619089987.0, + "step": 1545 + }, + { + "epoch": 2.4525316455696204, + "grad_norm": 0.35106351147093817, + "learning_rate": 3.0060978777666054e-05, + "loss": 0.6217, + "num_tokens": 1624332867.0, + "step": 1550 + }, + { + "epoch": 2.4604430379746836, + "grad_norm": 0.31920875041898267, + "learning_rate": 2.994397812278269e-05, + "loss": 0.6229, + "num_tokens": 1629575747.0, + "step": 1555 + }, + { + "epoch": 2.4683544303797467, + "grad_norm": 0.3877394978298516, + "learning_rate": 2.9826910554252253e-05, + "loss": 0.6272, + "num_tokens": 1634818627.0, + "step": 1560 + }, + { + "epoch": 2.4762658227848102, + "grad_norm": 0.31347418712210595, + "learning_rate": 2.9709779277266258e-05, + "loss": 0.6276, + "num_tokens": 1640061507.0, + "step": 1565 + }, + { + "epoch": 2.4841772151898733, + "grad_norm": 0.3663502650533136, + "learning_rate": 2.959258749876052e-05, + "loss": 0.6228, + "num_tokens": 1645292739.0, + "step": 1570 + }, + { + "epoch": 2.492088607594937, + "grad_norm": 0.4369820023540263, + "learning_rate": 2.9475338427327298e-05, + "loss": 0.622, + "num_tokens": 1650520838.0, + "step": 1575 + }, + { + "epoch": 2.5, + "grad_norm": 0.38676813237014673, + "learning_rate": 2.9358035273127483e-05, + "loss": 0.6269, + "num_tokens": 1655763718.0, + "step": 1580 + }, + { + "epoch": 2.507911392405063, + "grad_norm": 0.3206842281049778, + "learning_rate": 2.924068124780269e-05, + "loss": 0.625, + "num_tokens": 1661006598.0, + "step": 1585 + }, + { + "epoch": 2.5158227848101267, + "grad_norm": 0.3173236770874382, + "learning_rate": 2.9123279564387355e-05, + "loss": 0.6134, + "num_tokens": 1666249478.0, + "step": 1590 + }, + { + "epoch": 2.5237341772151898, + "grad_norm": 0.32043576387698464, + "learning_rate": 2.900583343722072e-05, + "loss": 0.6215, + "num_tokens": 1671492358.0, + "step": 1595 + }, + { + "epoch": 2.5316455696202533, + "grad_norm": 0.34279988720816984, + "learning_rate": 2.8888346081858858e-05, + "loss": 0.6176, + "num_tokens": 1676735238.0, + "step": 1600 + }, + { + "epoch": 2.5395569620253164, + "grad_norm": 0.3445942670476214, + "learning_rate": 2.8770820714986647e-05, + "loss": 0.6256, + "num_tokens": 1681960978.0, + "step": 1605 + }, + { + "epoch": 2.5474683544303796, + "grad_norm": 0.3249239294351048, + "learning_rate": 2.8653260554329643e-05, + "loss": 0.6229, + "num_tokens": 1687203858.0, + "step": 1610 + }, + { + "epoch": 2.555379746835443, + "grad_norm": 0.4195352553077635, + "learning_rate": 2.8535668818566065e-05, + "loss": 0.6209, + "num_tokens": 1692446738.0, + "step": 1615 + }, + { + "epoch": 2.5632911392405062, + "grad_norm": 0.3148180466101991, + "learning_rate": 2.8418048727238585e-05, + "loss": 0.6243, + "num_tokens": 1697685701.0, + "step": 1620 + }, + { + "epoch": 2.5712025316455698, + "grad_norm": 0.3145292653905644, + "learning_rate": 2.8300403500666256e-05, + "loss": 0.6118, + "num_tokens": 1702928581.0, + "step": 1625 + }, + { + "epoch": 2.579113924050633, + "grad_norm": 0.34661533020564694, + "learning_rate": 2.81827363598563e-05, + "loss": 0.6209, + "num_tokens": 1708171461.0, + "step": 1630 + }, + { + "epoch": 2.587025316455696, + "grad_norm": 0.32721458461346575, + "learning_rate": 2.8065050526415916e-05, + "loss": 0.6231, + "num_tokens": 1713414341.0, + "step": 1635 + }, + { + "epoch": 2.5949367088607596, + "grad_norm": 0.3293080910499085, + "learning_rate": 2.794734922246408e-05, + "loss": 0.6227, + "num_tokens": 1718657221.0, + "step": 1640 + }, + { + "epoch": 2.6028481012658227, + "grad_norm": 0.3006610821739955, + "learning_rate": 2.7829635670543375e-05, + "loss": 0.6198, + "num_tokens": 1723897509.0, + "step": 1645 + }, + { + "epoch": 2.6107594936708862, + "grad_norm": 0.3221265574540506, + "learning_rate": 2.7711913093531682e-05, + "loss": 0.6215, + "num_tokens": 1729134925.0, + "step": 1650 + }, + { + "epoch": 2.6186708860759493, + "grad_norm": 0.3629658305187627, + "learning_rate": 2.7594184714554e-05, + "loss": 0.6193, + "num_tokens": 1734377805.0, + "step": 1655 + }, + { + "epoch": 2.6265822784810124, + "grad_norm": 0.3305255086161955, + "learning_rate": 2.7476453756894173e-05, + "loss": 0.6163, + "num_tokens": 1739620685.0, + "step": 1660 + }, + { + "epoch": 2.634493670886076, + "grad_norm": 0.30385118713526643, + "learning_rate": 2.7358723443906653e-05, + "loss": 0.6168, + "num_tokens": 1744863565.0, + "step": 1665 + }, + { + "epoch": 2.642405063291139, + "grad_norm": 0.3729938271215438, + "learning_rate": 2.724099699892822e-05, + "loss": 0.6247, + "num_tokens": 1750106445.0, + "step": 1670 + }, + { + "epoch": 2.6503164556962027, + "grad_norm": 0.3363024823177763, + "learning_rate": 2.712327764518979e-05, + "loss": 0.6227, + "num_tokens": 1755349325.0, + "step": 1675 + }, + { + "epoch": 2.6582278481012658, + "grad_norm": 0.3055930853692746, + "learning_rate": 2.7005568605728087e-05, + "loss": 0.6229, + "num_tokens": 1760592205.0, + "step": 1680 + }, + { + "epoch": 2.666139240506329, + "grad_norm": 0.3298226038292838, + "learning_rate": 2.6887873103297474e-05, + "loss": 0.6214, + "num_tokens": 1765835085.0, + "step": 1685 + }, + { + "epoch": 2.6740506329113924, + "grad_norm": 0.2867154405096028, + "learning_rate": 2.6770194360281653e-05, + "loss": 0.6199, + "num_tokens": 1771077965.0, + "step": 1690 + }, + { + "epoch": 2.6819620253164556, + "grad_norm": 0.3195638173711784, + "learning_rate": 2.6652535598605504e-05, + "loss": 0.6191, + "num_tokens": 1776320845.0, + "step": 1695 + }, + { + "epoch": 2.689873417721519, + "grad_norm": 0.32434124204868575, + "learning_rate": 2.6534900039646816e-05, + "loss": 0.6181, + "num_tokens": 1781532860.0, + "step": 1700 + }, + { + "epoch": 2.6977848101265822, + "grad_norm": 0.2918992438055131, + "learning_rate": 2.6417290904148116e-05, + "loss": 0.6197, + "num_tokens": 1786775740.0, + "step": 1705 + }, + { + "epoch": 2.7056962025316453, + "grad_norm": 0.3055847287718812, + "learning_rate": 2.629971141212848e-05, + "loss": 0.6173, + "num_tokens": 1792018620.0, + "step": 1710 + }, + { + "epoch": 2.713607594936709, + "grad_norm": 0.2985953721458501, + "learning_rate": 2.618216478279538e-05, + "loss": 0.6165, + "num_tokens": 1797261500.0, + "step": 1715 + }, + { + "epoch": 2.721518987341772, + "grad_norm": 0.3212918053380019, + "learning_rate": 2.6064654234456554e-05, + "loss": 0.6214, + "num_tokens": 1802504380.0, + "step": 1720 + }, + { + "epoch": 2.7294303797468356, + "grad_norm": 0.2919099745172365, + "learning_rate": 2.5947182984431827e-05, + "loss": 0.6255, + "num_tokens": 1807747260.0, + "step": 1725 + }, + { + "epoch": 2.7373417721518987, + "grad_norm": 0.2860751140237954, + "learning_rate": 2.5829754248965154e-05, + "loss": 0.6189, + "num_tokens": 1812990140.0, + "step": 1730 + }, + { + "epoch": 2.745253164556962, + "grad_norm": 0.3019949199367483, + "learning_rate": 2.571237124313643e-05, + "loss": 0.6124, + "num_tokens": 1818233020.0, + "step": 1735 + }, + { + "epoch": 2.7531645569620253, + "grad_norm": 0.3187461502303022, + "learning_rate": 2.5595037180773524e-05, + "loss": 0.6196, + "num_tokens": 1823475900.0, + "step": 1740 + }, + { + "epoch": 2.7610759493670884, + "grad_norm": 0.31735001215836417, + "learning_rate": 2.5477755274364308e-05, + "loss": 0.623, + "num_tokens": 1828718780.0, + "step": 1745 + }, + { + "epoch": 2.768987341772152, + "grad_norm": 0.33876769709652277, + "learning_rate": 2.5360528734968647e-05, + "loss": 0.611, + "num_tokens": 1833961660.0, + "step": 1750 + }, + { + "epoch": 2.776898734177215, + "grad_norm": 0.3542233915904492, + "learning_rate": 2.5243360772130532e-05, + "loss": 0.625, + "num_tokens": 1839200698.0, + "step": 1755 + }, + { + "epoch": 2.7848101265822782, + "grad_norm": 0.3630368367495965, + "learning_rate": 2.5126254593790182e-05, + "loss": 0.6179, + "num_tokens": 1844443578.0, + "step": 1760 + }, + { + "epoch": 2.7927215189873418, + "grad_norm": 0.3826705315790488, + "learning_rate": 2.500921340619621e-05, + "loss": 0.6205, + "num_tokens": 1849680596.0, + "step": 1765 + }, + { + "epoch": 2.8006329113924053, + "grad_norm": 0.3644864833948713, + "learning_rate": 2.489224041381787e-05, + "loss": 0.6156, + "num_tokens": 1854907408.0, + "step": 1770 + }, + { + "epoch": 2.8085443037974684, + "grad_norm": 0.3386672412828899, + "learning_rate": 2.477533881925727e-05, + "loss": 0.6238, + "num_tokens": 1860134742.0, + "step": 1775 + }, + { + "epoch": 2.8164556962025316, + "grad_norm": 0.2921538507032039, + "learning_rate": 2.465851182316176e-05, + "loss": 0.6202, + "num_tokens": 1865377622.0, + "step": 1780 + }, + { + "epoch": 2.8243670886075947, + "grad_norm": 0.28494983081811437, + "learning_rate": 2.4541762624136212e-05, + "loss": 0.6176, + "num_tokens": 1870620502.0, + "step": 1785 + }, + { + "epoch": 2.8322784810126582, + "grad_norm": 0.29021546121305103, + "learning_rate": 2.4425094418655526e-05, + "loss": 0.6177, + "num_tokens": 1875863382.0, + "step": 1790 + }, + { + "epoch": 2.8401898734177218, + "grad_norm": 0.27953269634031763, + "learning_rate": 2.430851040097707e-05, + "loss": 0.6158, + "num_tokens": 1881101588.0, + "step": 1795 + }, + { + "epoch": 2.848101265822785, + "grad_norm": 0.28367003015841197, + "learning_rate": 2.4192013763053222e-05, + "loss": 0.6151, + "num_tokens": 1886333356.0, + "step": 1800 + }, + { + "epoch": 2.856012658227848, + "grad_norm": 0.30695047341331355, + "learning_rate": 2.4075607694444014e-05, + "loss": 0.615, + "num_tokens": 1891576236.0, + "step": 1805 + }, + { + "epoch": 2.8639240506329116, + "grad_norm": 0.3389129881606512, + "learning_rate": 2.395929538222978e-05, + "loss": 0.6194, + "num_tokens": 1896816113.0, + "step": 1810 + }, + { + "epoch": 2.8718354430379747, + "grad_norm": 0.2993951418009312, + "learning_rate": 2.3843080010923873e-05, + "loss": 0.6151, + "num_tokens": 1902045462.0, + "step": 1815 + }, + { + "epoch": 2.879746835443038, + "grad_norm": 0.2747247961261455, + "learning_rate": 2.3726964762385546e-05, + "loss": 0.6152, + "num_tokens": 1907275468.0, + "step": 1820 + }, + { + "epoch": 2.8876582278481013, + "grad_norm": 0.27656654436718103, + "learning_rate": 2.361095281573274e-05, + "loss": 0.6204, + "num_tokens": 1912517035.0, + "step": 1825 + }, + { + "epoch": 2.8955696202531644, + "grad_norm": 0.26450516299877097, + "learning_rate": 2.3495047347255122e-05, + "loss": 0.6141, + "num_tokens": 1917752379.0, + "step": 1830 + }, + { + "epoch": 2.903481012658228, + "grad_norm": 0.2680090640698674, + "learning_rate": 2.3379251530327112e-05, + "loss": 0.6098, + "num_tokens": 1922995259.0, + "step": 1835 + }, + { + "epoch": 2.911392405063291, + "grad_norm": 0.3020470058551071, + "learning_rate": 2.326356853532094e-05, + "loss": 0.6121, + "num_tokens": 1928238139.0, + "step": 1840 + }, + { + "epoch": 2.9193037974683547, + "grad_norm": 0.35207295628986174, + "learning_rate": 2.3148001529519918e-05, + "loss": 0.6176, + "num_tokens": 1933481019.0, + "step": 1845 + }, + { + "epoch": 2.9272151898734178, + "grad_norm": 0.31909804688235877, + "learning_rate": 2.303255367703169e-05, + "loss": 0.6206, + "num_tokens": 1938714322.0, + "step": 1850 + }, + { + "epoch": 2.935126582278481, + "grad_norm": 0.30462726559520964, + "learning_rate": 2.2917228138701596e-05, + "loss": 0.6154, + "num_tokens": 1943957202.0, + "step": 1855 + }, + { + "epoch": 2.9430379746835444, + "grad_norm": 0.32103060565165314, + "learning_rate": 2.2802028072026115e-05, + "loss": 0.6089, + "num_tokens": 1949191447.0, + "step": 1860 + }, + { + "epoch": 2.9509493670886076, + "grad_norm": 0.3062127051035162, + "learning_rate": 2.268695663106648e-05, + "loss": 0.6127, + "num_tokens": 1954434327.0, + "step": 1865 + }, + { + "epoch": 2.958860759493671, + "grad_norm": 0.3239852849229511, + "learning_rate": 2.257201696636225e-05, + "loss": 0.615, + "num_tokens": 1959677207.0, + "step": 1870 + }, + { + "epoch": 2.9667721518987342, + "grad_norm": 0.2893785533195971, + "learning_rate": 2.2457212224845115e-05, + "loss": 0.6182, + "num_tokens": 1964920087.0, + "step": 1875 + }, + { + "epoch": 2.9746835443037973, + "grad_norm": 0.3565583813030718, + "learning_rate": 2.2342545549752658e-05, + "loss": 0.6187, + "num_tokens": 1970162967.0, + "step": 1880 + }, + { + "epoch": 2.982594936708861, + "grad_norm": 0.2901982983640066, + "learning_rate": 2.2228020080542394e-05, + "loss": 0.6155, + "num_tokens": 1975405847.0, + "step": 1885 + }, + { + "epoch": 2.990506329113924, + "grad_norm": 0.2552747365991554, + "learning_rate": 2.211363895280573e-05, + "loss": 0.617, + "num_tokens": 1980648163.0, + "step": 1890 + }, + { + "epoch": 2.9984177215189876, + "grad_norm": 0.2925757603065985, + "learning_rate": 2.1999405298182176e-05, + "loss": 0.6114, + "num_tokens": 1985885198.0, + "step": 1895 + }, + { + "epoch": 3.0063291139240507, + "grad_norm": 0.3169901185514265, + "learning_rate": 2.1885322244273537e-05, + "loss": 0.6001, + "num_tokens": 1991128078.0, + "step": 1900 + }, + { + "epoch": 3.0142405063291138, + "grad_norm": 0.2733791677119331, + "learning_rate": 2.177139291455837e-05, + "loss": 0.5909, + "num_tokens": 1996370958.0, + "step": 1905 + }, + { + "epoch": 3.0221518987341773, + "grad_norm": 0.2969117337975924, + "learning_rate": 2.1657620428306386e-05, + "loss": 0.5979, + "num_tokens": 2001613838.0, + "step": 1910 + }, + { + "epoch": 3.0300632911392404, + "grad_norm": 0.259222996285033, + "learning_rate": 2.154400790049309e-05, + "loss": 0.5935, + "num_tokens": 2006856718.0, + "step": 1915 + }, + { + "epoch": 3.037974683544304, + "grad_norm": 0.29751625615948685, + "learning_rate": 2.1430558441714488e-05, + "loss": 0.5892, + "num_tokens": 2012099598.0, + "step": 1920 + }, + { + "epoch": 3.045886075949367, + "grad_norm": 0.2808339111842118, + "learning_rate": 2.131727515810193e-05, + "loss": 0.5881, + "num_tokens": 2017342478.0, + "step": 1925 + }, + { + "epoch": 3.0537974683544302, + "grad_norm": 0.2665237486778169, + "learning_rate": 2.1204161151237044e-05, + "loss": 0.5869, + "num_tokens": 2022585358.0, + "step": 1930 + }, + { + "epoch": 3.0617088607594938, + "grad_norm": 0.28147532434254685, + "learning_rate": 2.1091219518066846e-05, + "loss": 0.5857, + "num_tokens": 2027828238.0, + "step": 1935 + }, + { + "epoch": 3.069620253164557, + "grad_norm": 0.27827204435611375, + "learning_rate": 2.0978453350818923e-05, + "loss": 0.5892, + "num_tokens": 2033071118.0, + "step": 1940 + }, + { + "epoch": 3.0775316455696204, + "grad_norm": 0.27504620542469616, + "learning_rate": 2.0865865736916767e-05, + "loss": 0.5927, + "num_tokens": 2038313998.0, + "step": 1945 + }, + { + "epoch": 3.0854430379746836, + "grad_norm": 0.2625667359010651, + "learning_rate": 2.0753459758895305e-05, + "loss": 0.5878, + "num_tokens": 2043556878.0, + "step": 1950 + }, + { + "epoch": 3.0933544303797467, + "grad_norm": 0.31412193685640266, + "learning_rate": 2.0641238494316412e-05, + "loss": 0.5954, + "num_tokens": 2048795916.0, + "step": 1955 + }, + { + "epoch": 3.1012658227848102, + "grad_norm": 0.2551742932124429, + "learning_rate": 2.0529205015684733e-05, + "loss": 0.5896, + "num_tokens": 2054038796.0, + "step": 1960 + }, + { + "epoch": 3.1091772151898733, + "grad_norm": 0.27107418641432374, + "learning_rate": 2.0417362390363497e-05, + "loss": 0.5949, + "num_tokens": 2059281676.0, + "step": 1965 + }, + { + "epoch": 3.117088607594937, + "grad_norm": 0.2910292647945281, + "learning_rate": 2.030571368049058e-05, + "loss": 0.593, + "num_tokens": 2064524556.0, + "step": 1970 + }, + { + "epoch": 3.125, + "grad_norm": 0.2702497087199772, + "learning_rate": 2.0194261942894628e-05, + "loss": 0.5913, + "num_tokens": 2069763519.0, + "step": 1975 + }, + { + "epoch": 3.132911392405063, + "grad_norm": 0.27879096153789196, + "learning_rate": 2.0083010229011405e-05, + "loss": 0.5871, + "num_tokens": 2075006399.0, + "step": 1980 + }, + { + "epoch": 3.1408227848101267, + "grad_norm": 0.2760119851379016, + "learning_rate": 1.9971961584800218e-05, + "loss": 0.5889, + "num_tokens": 2080249279.0, + "step": 1985 + }, + { + "epoch": 3.1487341772151898, + "grad_norm": 0.2980189742569214, + "learning_rate": 1.986111905066055e-05, + "loss": 0.5936, + "num_tokens": 2085492159.0, + "step": 1990 + }, + { + "epoch": 3.1566455696202533, + "grad_norm": 0.27221033878781475, + "learning_rate": 1.9750485661348762e-05, + "loss": 0.5925, + "num_tokens": 2090735039.0, + "step": 1995 + }, + { + "epoch": 3.1645569620253164, + "grad_norm": 0.282708489024635, + "learning_rate": 1.964006444589509e-05, + "loss": 0.5906, + "num_tokens": 2095977919.0, + "step": 2000 + }, + { + "epoch": 3.1724683544303796, + "grad_norm": 0.2689336862752935, + "learning_rate": 1.952985842752062e-05, + "loss": 0.5908, + "num_tokens": 2101206695.0, + "step": 2005 + }, + { + "epoch": 3.180379746835443, + "grad_norm": 0.26260733949618775, + "learning_rate": 1.941987062355458e-05, + "loss": 0.5907, + "num_tokens": 2106449575.0, + "step": 2010 + }, + { + "epoch": 3.1882911392405062, + "grad_norm": 0.2829230017612488, + "learning_rate": 1.9310104045351724e-05, + "loss": 0.5869, + "num_tokens": 2111692455.0, + "step": 2015 + }, + { + "epoch": 3.1962025316455698, + "grad_norm": 0.2496685091023237, + "learning_rate": 1.9200561698209828e-05, + "loss": 0.5895, + "num_tokens": 2116935335.0, + "step": 2020 + }, + { + "epoch": 3.204113924050633, + "grad_norm": 0.30588754751321845, + "learning_rate": 1.9091246581287487e-05, + "loss": 0.5911, + "num_tokens": 2122178215.0, + "step": 2025 + }, + { + "epoch": 3.212025316455696, + "grad_norm": 0.3013756834853421, + "learning_rate": 1.8982161687521938e-05, + "loss": 0.5926, + "num_tokens": 2127416178.0, + "step": 2030 + }, + { + "epoch": 3.2199367088607596, + "grad_norm": 0.3260497257360817, + "learning_rate": 1.8873310003547152e-05, + "loss": 0.5957, + "num_tokens": 2132659058.0, + "step": 2035 + }, + { + "epoch": 3.2278481012658227, + "grad_norm": 18.37765842944507, + "learning_rate": 1.8764694509612057e-05, + "loss": 0.6045, + "num_tokens": 2137900029.0, + "step": 2040 + }, + { + "epoch": 3.2357594936708862, + "grad_norm": 0.2844360272669497, + "learning_rate": 1.865631817949893e-05, + "loss": 0.5961, + "num_tokens": 2143139652.0, + "step": 2045 + }, + { + "epoch": 3.2436708860759493, + "grad_norm": 0.25708276328077995, + "learning_rate": 1.8548183980441985e-05, + "loss": 0.5988, + "num_tokens": 2148382532.0, + "step": 2050 + }, + { + "epoch": 3.2515822784810124, + "grad_norm": 0.27434743791141486, + "learning_rate": 1.8440294873046126e-05, + "loss": 0.5904, + "num_tokens": 2153625412.0, + "step": 2055 + }, + { + "epoch": 3.259493670886076, + "grad_norm": 0.25975458422652054, + "learning_rate": 1.8332653811205906e-05, + "loss": 0.5928, + "num_tokens": 2158854105.0, + "step": 2060 + }, + { + "epoch": 3.267405063291139, + "grad_norm": 0.26925828595298595, + "learning_rate": 1.8225263742024638e-05, + "loss": 0.5833, + "num_tokens": 2164096985.0, + "step": 2065 + }, + { + "epoch": 3.2753164556962027, + "grad_norm": 0.282185762273433, + "learning_rate": 1.8118127605733697e-05, + "loss": 0.5981, + "num_tokens": 2169337273.0, + "step": 2070 + }, + { + "epoch": 3.2832278481012658, + "grad_norm": 0.25816071189825335, + "learning_rate": 1.801124833561208e-05, + "loss": 0.5961, + "num_tokens": 2174564736.0, + "step": 2075 + }, + { + "epoch": 3.291139240506329, + "grad_norm": 0.24619197376885332, + "learning_rate": 1.7904628857905982e-05, + "loss": 0.6011, + "num_tokens": 2179797828.0, + "step": 2080 + }, + { + "epoch": 3.2990506329113924, + "grad_norm": 0.24421414534208322, + "learning_rate": 1.7798272091748785e-05, + "loss": 0.5865, + "num_tokens": 2185040708.0, + "step": 2085 + }, + { + "epoch": 3.3069620253164556, + "grad_norm": 0.2524526157340073, + "learning_rate": 1.7692180949081093e-05, + "loss": 0.5959, + "num_tokens": 2190283588.0, + "step": 2090 + }, + { + "epoch": 3.314873417721519, + "grad_norm": 0.27130450376385096, + "learning_rate": 1.758635833457099e-05, + "loss": 0.5848, + "num_tokens": 2195513594.0, + "step": 2095 + }, + { + "epoch": 3.3227848101265822, + "grad_norm": 0.26771998653275103, + "learning_rate": 1.748080714553455e-05, + "loss": 0.5879, + "num_tokens": 2200745342.0, + "step": 2100 + }, + { + "epoch": 3.3306962025316453, + "grad_norm": 0.26138164229192307, + "learning_rate": 1.7375530271856487e-05, + "loss": 0.5922, + "num_tokens": 2205988222.0, + "step": 2105 + }, + { + "epoch": 3.338607594936709, + "grad_norm": 0.2715626154228517, + "learning_rate": 1.727053059591101e-05, + "loss": 0.5971, + "num_tokens": 2211231102.0, + "step": 2110 + }, + { + "epoch": 3.346518987341772, + "grad_norm": 0.2455227895466599, + "learning_rate": 1.7165810992482994e-05, + "loss": 0.5887, + "num_tokens": 2216460869.0, + "step": 2115 + }, + { + "epoch": 3.3544303797468356, + "grad_norm": 0.2777380961416323, + "learning_rate": 1.706137432868917e-05, + "loss": 0.5943, + "num_tokens": 2221703749.0, + "step": 2120 + }, + { + "epoch": 3.3623417721518987, + "grad_norm": 0.25563580174980555, + "learning_rate": 1.6957223463899658e-05, + "loss": 0.5897, + "num_tokens": 2226946629.0, + "step": 2125 + }, + { + "epoch": 3.370253164556962, + "grad_norm": 0.25164624412979186, + "learning_rate": 1.6853361249659722e-05, + "loss": 0.5929, + "num_tokens": 2232189509.0, + "step": 2130 + }, + { + "epoch": 3.3781645569620253, + "grad_norm": 0.2806108194410442, + "learning_rate": 1.674979052961166e-05, + "loss": 0.5975, + "num_tokens": 2237432389.0, + "step": 2135 + }, + { + "epoch": 3.3860759493670884, + "grad_norm": 0.2605837907585764, + "learning_rate": 1.664651413941693e-05, + "loss": 0.5907, + "num_tokens": 2242665858.0, + "step": 2140 + }, + { + "epoch": 3.393987341772152, + "grad_norm": 0.2598679967531081, + "learning_rate": 1.6543534906678552e-05, + "loss": 0.592, + "num_tokens": 2247908738.0, + "step": 2145 + }, + { + "epoch": 3.401898734177215, + "grad_norm": 0.25797499837806687, + "learning_rate": 1.6440855650863695e-05, + "loss": 0.5894, + "num_tokens": 2253151618.0, + "step": 2150 + }, + { + "epoch": 3.4098101265822787, + "grad_norm": 0.2427493258825347, + "learning_rate": 1.6338479183226436e-05, + "loss": 0.5924, + "num_tokens": 2258394498.0, + "step": 2155 + }, + { + "epoch": 3.4177215189873418, + "grad_norm": 0.2359671364270129, + "learning_rate": 1.6236408306730828e-05, + "loss": 0.5916, + "num_tokens": 2263637378.0, + "step": 2160 + }, + { + "epoch": 3.425632911392405, + "grad_norm": 0.23492984902262662, + "learning_rate": 1.6134645815974153e-05, + "loss": 0.588, + "num_tokens": 2268880258.0, + "step": 2165 + }, + { + "epoch": 3.4335443037974684, + "grad_norm": 0.2684129651198138, + "learning_rate": 1.6033194497110387e-05, + "loss": 0.5863, + "num_tokens": 2274107070.0, + "step": 2170 + }, + { + "epoch": 3.4414556962025316, + "grad_norm": 0.2551028754173729, + "learning_rate": 1.5932057127773956e-05, + "loss": 0.5912, + "num_tokens": 2279335709.0, + "step": 2175 + }, + { + "epoch": 3.449367088607595, + "grad_norm": 0.27552281820576113, + "learning_rate": 1.5831236477003657e-05, + "loss": 0.5901, + "num_tokens": 2284578589.0, + "step": 2180 + }, + { + "epoch": 3.4572784810126582, + "grad_norm": 0.245355717229245, + "learning_rate": 1.573073530516683e-05, + "loss": 0.5984, + "num_tokens": 2289805799.0, + "step": 2185 + }, + { + "epoch": 3.4651898734177213, + "grad_norm": 0.2711024543904534, + "learning_rate": 1.563055636388385e-05, + "loss": 0.5901, + "num_tokens": 2295041668.0, + "step": 2190 + }, + { + "epoch": 3.473101265822785, + "grad_norm": 0.25534456241862863, + "learning_rate": 1.5530702395952726e-05, + "loss": 0.5885, + "num_tokens": 2300284548.0, + "step": 2195 + }, + { + "epoch": 3.481012658227848, + "grad_norm": 0.28472473513191726, + "learning_rate": 1.5431176135274004e-05, + "loss": 0.5875, + "num_tokens": 2305527428.0, + "step": 2200 + }, + { + "epoch": 3.4889240506329116, + "grad_norm": 0.26401107803555085, + "learning_rate": 1.533198030677596e-05, + "loss": 0.587, + "num_tokens": 2310770308.0, + "step": 2205 + }, + { + "epoch": 3.4968354430379747, + "grad_norm": 0.23581382627455794, + "learning_rate": 1.5233117626339988e-05, + "loss": 0.5829, + "num_tokens": 2316013188.0, + "step": 2210 + }, + { + "epoch": 3.504746835443038, + "grad_norm": 0.2919647262861195, + "learning_rate": 1.513459080072617e-05, + "loss": 0.593, + "num_tokens": 2321256068.0, + "step": 2215 + }, + { + "epoch": 3.5126582278481013, + "grad_norm": 0.27795974121603223, + "learning_rate": 1.503640252749928e-05, + "loss": 0.5948, + "num_tokens": 2326498948.0, + "step": 2220 + }, + { + "epoch": 3.5205696202531644, + "grad_norm": 0.2513485774758495, + "learning_rate": 1.4938555494954838e-05, + "loss": 0.5928, + "num_tokens": 2331740515.0, + "step": 2225 + }, + { + "epoch": 3.528481012658228, + "grad_norm": 0.27468633702239875, + "learning_rate": 1.484105238204555e-05, + "loss": 0.592, + "num_tokens": 2336983395.0, + "step": 2230 + }, + { + "epoch": 3.536392405063291, + "grad_norm": 0.24973282593959645, + "learning_rate": 1.4743895858307951e-05, + "loss": 0.5858, + "num_tokens": 2342226275.0, + "step": 2235 + }, + { + "epoch": 3.5443037974683547, + "grad_norm": 0.236224838159416, + "learning_rate": 1.4647088583789325e-05, + "loss": 0.5943, + "num_tokens": 2347461462.0, + "step": 2240 + }, + { + "epoch": 3.5522151898734178, + "grad_norm": 0.25744574090772937, + "learning_rate": 1.4550633208974832e-05, + "loss": 0.5888, + "num_tokens": 2352704342.0, + "step": 2245 + }, + { + "epoch": 3.560126582278481, + "grad_norm": 0.24904139861794714, + "learning_rate": 1.445453237471501e-05, + "loss": 0.5937, + "num_tokens": 2357947222.0, + "step": 2250 + }, + { + "epoch": 3.5680379746835444, + "grad_norm": 0.26268941423266656, + "learning_rate": 1.435878871215342e-05, + "loss": 0.5903, + "num_tokens": 2363190102.0, + "step": 2255 + }, + { + "epoch": 3.5759493670886076, + "grad_norm": 0.24790895208789265, + "learning_rate": 1.4263404842654638e-05, + "loss": 0.5972, + "num_tokens": 2368432982.0, + "step": 2260 + }, + { + "epoch": 3.583860759493671, + "grad_norm": 0.2613242366100201, + "learning_rate": 1.416838337773243e-05, + "loss": 0.595, + "num_tokens": 2373661081.0, + "step": 2265 + }, + { + "epoch": 3.5917721518987342, + "grad_norm": 0.2922830460241977, + "learning_rate": 1.4073726918978347e-05, + "loss": 0.594, + "num_tokens": 2378903961.0, + "step": 2270 + }, + { + "epoch": 3.5996835443037973, + "grad_norm": 0.27101780547058585, + "learning_rate": 1.3979438057990399e-05, + "loss": 0.592, + "num_tokens": 2384146841.0, + "step": 2275 + }, + { + "epoch": 3.607594936708861, + "grad_norm": 0.26512432923141993, + "learning_rate": 1.3885519376302159e-05, + "loss": 0.5878, + "num_tokens": 2389389721.0, + "step": 2280 + }, + { + "epoch": 3.615506329113924, + "grad_norm": 0.23952238274180007, + "learning_rate": 1.3791973445312058e-05, + "loss": 0.589, + "num_tokens": 2394632601.0, + "step": 2285 + }, + { + "epoch": 3.6234177215189876, + "grad_norm": 0.25423971405929885, + "learning_rate": 1.3698802826213008e-05, + "loss": 0.5908, + "num_tokens": 2399860042.0, + "step": 2290 + }, + { + "epoch": 3.6313291139240507, + "grad_norm": 0.24740192193842994, + "learning_rate": 1.3606010069922232e-05, + "loss": 0.5882, + "num_tokens": 2405102922.0, + "step": 2295 + }, + { + "epoch": 3.6392405063291138, + "grad_norm": 0.5035309107226975, + "learning_rate": 1.351359771701147e-05, + "loss": 0.5966, + "num_tokens": 2410345802.0, + "step": 2300 + }, + { + "epoch": 3.6471518987341773, + "grad_norm": 0.231935472046041, + "learning_rate": 1.3421568297637404e-05, + "loss": 0.5873, + "num_tokens": 2415588682.0, + "step": 2305 + }, + { + "epoch": 3.6550632911392404, + "grad_norm": 0.2584802973832586, + "learning_rate": 1.3329924331472376e-05, + "loss": 0.5925, + "num_tokens": 2420806919.0, + "step": 2310 + }, + { + "epoch": 3.662974683544304, + "grad_norm": 0.31314246833170745, + "learning_rate": 1.3238668327635422e-05, + "loss": 0.5922, + "num_tokens": 2426049799.0, + "step": 2315 + }, + { + "epoch": 3.670886075949367, + "grad_norm": 0.2663616787106021, + "learning_rate": 1.3147802784623536e-05, + "loss": 0.5962, + "num_tokens": 2431292679.0, + "step": 2320 + }, + { + "epoch": 3.6787974683544302, + "grad_norm": 0.23711780957648615, + "learning_rate": 1.3057330190243317e-05, + "loss": 0.5894, + "num_tokens": 2436523336.0, + "step": 2325 + }, + { + "epoch": 3.6867088607594938, + "grad_norm": 0.2549540114874716, + "learning_rate": 1.2967253021542813e-05, + "loss": 0.5957, + "num_tokens": 2441766216.0, + "step": 2330 + }, + { + "epoch": 3.694620253164557, + "grad_norm": 0.24052219589182647, + "learning_rate": 1.2877573744743737e-05, + "loss": 0.5935, + "num_tokens": 2447009096.0, + "step": 2335 + }, + { + "epoch": 3.7025316455696204, + "grad_norm": 0.23451782912483585, + "learning_rate": 1.278829481517388e-05, + "loss": 0.5907, + "num_tokens": 2452248973.0, + "step": 2340 + }, + { + "epoch": 3.7104430379746836, + "grad_norm": 0.2572531172651155, + "learning_rate": 1.2699418677199992e-05, + "loss": 0.5871, + "num_tokens": 2457491853.0, + "step": 2345 + }, + { + "epoch": 3.7183544303797467, + "grad_norm": 0.22205564327904748, + "learning_rate": 1.2610947764160743e-05, + "loss": 0.5911, + "num_tokens": 2462732233.0, + "step": 2350 + }, + { + "epoch": 3.7262658227848102, + "grad_norm": 0.25061314597163403, + "learning_rate": 1.2522884498300185e-05, + "loss": 0.589, + "num_tokens": 2467975113.0, + "step": 2355 + }, + { + "epoch": 3.7341772151898733, + "grad_norm": 0.24341650567160555, + "learning_rate": 1.2435231290701398e-05, + "loss": 0.5839, + "num_tokens": 2473217993.0, + "step": 2360 + }, + { + "epoch": 3.742088607594937, + "grad_norm": 0.23957280785298024, + "learning_rate": 1.234799054122048e-05, + "loss": 0.5856, + "num_tokens": 2478460873.0, + "step": 2365 + }, + { + "epoch": 3.75, + "grad_norm": 0.23496985031896744, + "learning_rate": 1.2261164638420832e-05, + "loss": 0.5943, + "num_tokens": 2483695118.0, + "step": 2370 + }, + { + "epoch": 3.757911392405063, + "grad_norm": 0.2562958360759268, + "learning_rate": 1.2174755959507785e-05, + "loss": 0.5905, + "num_tokens": 2488915999.0, + "step": 2375 + }, + { + "epoch": 3.7658227848101267, + "grad_norm": 0.23818568022151865, + "learning_rate": 1.2088766870263504e-05, + "loss": 0.5833, + "num_tokens": 2494158879.0, + "step": 2380 + }, + { + "epoch": 3.7737341772151898, + "grad_norm": 0.271749751046968, + "learning_rate": 1.2003199724982206e-05, + "loss": 0.5898, + "num_tokens": 2499392356.0, + "step": 2385 + }, + { + "epoch": 3.7816455696202533, + "grad_norm": 0.2768085008184621, + "learning_rate": 1.1918056866405717e-05, + "loss": 0.5964, + "num_tokens": 2504635236.0, + "step": 2390 + }, + { + "epoch": 3.7895569620253164, + "grad_norm": 0.24209821208112645, + "learning_rate": 1.1833340625659325e-05, + "loss": 0.5907, + "num_tokens": 2509878116.0, + "step": 2395 + }, + { + "epoch": 3.7974683544303796, + "grad_norm": 0.2587091544484298, + "learning_rate": 1.1749053322187933e-05, + "loss": 0.5887, + "num_tokens": 2515098012.0, + "step": 2400 + }, + { + "epoch": 3.805379746835443, + "grad_norm": 0.241562313900529, + "learning_rate": 1.1665197263692593e-05, + "loss": 0.5914, + "num_tokens": 2520340892.0, + "step": 2405 + }, + { + "epoch": 3.8132911392405062, + "grad_norm": 0.21894193923034058, + "learning_rate": 1.1581774746067315e-05, + "loss": 0.5861, + "num_tokens": 2525582367.0, + "step": 2410 + }, + { + "epoch": 3.8212025316455698, + "grad_norm": 0.241635293617956, + "learning_rate": 1.149878805333616e-05, + "loss": 0.5875, + "num_tokens": 2530819032.0, + "step": 2415 + }, + { + "epoch": 3.829113924050633, + "grad_norm": 0.2377026625673294, + "learning_rate": 1.1416239457590797e-05, + "loss": 0.5895, + "num_tokens": 2536061912.0, + "step": 2420 + }, + { + "epoch": 3.837025316455696, + "grad_norm": 0.2500766928834093, + "learning_rate": 1.1334131218928215e-05, + "loss": 0.5955, + "num_tokens": 2541294406.0, + "step": 2425 + }, + { + "epoch": 3.8449367088607596, + "grad_norm": 0.24157809368048716, + "learning_rate": 1.1252465585388861e-05, + "loss": 0.593, + "num_tokens": 2546537286.0, + "step": 2430 + }, + { + "epoch": 3.8528481012658227, + "grad_norm": 0.23659644820965803, + "learning_rate": 1.1171244792895122e-05, + "loss": 0.5875, + "num_tokens": 2551780166.0, + "step": 2435 + }, + { + "epoch": 3.8607594936708862, + "grad_norm": 0.22665381320819925, + "learning_rate": 1.1090471065190087e-05, + "loss": 0.5895, + "num_tokens": 2557023046.0, + "step": 2440 + }, + { + "epoch": 3.8686708860759493, + "grad_norm": 0.2531032288030383, + "learning_rate": 1.1010146613776646e-05, + "loss": 0.5911, + "num_tokens": 2562249118.0, + "step": 2445 + }, + { + "epoch": 3.8765822784810124, + "grad_norm": 0.2375978750430171, + "learning_rate": 1.0930273637856969e-05, + "loss": 0.5913, + "num_tokens": 2567491998.0, + "step": 2450 + }, + { + "epoch": 3.884493670886076, + "grad_norm": 0.24286465990334033, + "learning_rate": 1.0850854324272289e-05, + "loss": 0.5877, + "num_tokens": 2572734878.0, + "step": 2455 + }, + { + "epoch": 3.892405063291139, + "grad_norm": 0.2575357182094295, + "learning_rate": 1.0771890847443022e-05, + "loss": 0.596, + "num_tokens": 2577975987.0, + "step": 2460 + }, + { + "epoch": 3.9003164556962027, + "grad_norm": 0.27638888789609667, + "learning_rate": 1.069338536930923e-05, + "loss": 0.5865, + "num_tokens": 2583218867.0, + "step": 2465 + }, + { + "epoch": 3.9082278481012658, + "grad_norm": 0.2344539051269508, + "learning_rate": 1.0615340039271443e-05, + "loss": 0.5892, + "num_tokens": 2588461747.0, + "step": 2470 + }, + { + "epoch": 3.916139240506329, + "grad_norm": 0.27080111151637076, + "learning_rate": 1.0537756994131783e-05, + "loss": 0.5895, + "num_tokens": 2593704627.0, + "step": 2475 + }, + { + "epoch": 3.9240506329113924, + "grad_norm": 0.2249920243150278, + "learning_rate": 1.0460638358035496e-05, + "loss": 0.5924, + "num_tokens": 2598947507.0, + "step": 2480 + }, + { + "epoch": 3.9319620253164556, + "grad_norm": 0.28780985267539305, + "learning_rate": 1.0383986242412785e-05, + "loss": 0.5871, + "num_tokens": 2604190387.0, + "step": 2485 + }, + { + "epoch": 3.939873417721519, + "grad_norm": 0.23427765653838825, + "learning_rate": 1.0307802745920964e-05, + "loss": 0.5839, + "num_tokens": 2609433267.0, + "step": 2490 + }, + { + "epoch": 3.9477848101265822, + "grad_norm": 0.2511572467034619, + "learning_rate": 1.0232089954387068e-05, + "loss": 0.591, + "num_tokens": 2614676147.0, + "step": 2495 + }, + { + "epoch": 3.9556962025316453, + "grad_norm": 0.24133671906166224, + "learning_rate": 1.0156849940750688e-05, + "loss": 0.5873, + "num_tokens": 2619919027.0, + "step": 2500 + }, + { + "epoch": 3.963607594936709, + "grad_norm": 0.24130871799729056, + "learning_rate": 1.0082084765007226e-05, + "loss": 0.5919, + "num_tokens": 2625150643.0, + "step": 2505 + }, + { + "epoch": 3.971518987341772, + "grad_norm": 0.23545792709918428, + "learning_rate": 1.0007796474151514e-05, + "loss": 0.5972, + "num_tokens": 2630393523.0, + "step": 2510 + }, + { + "epoch": 3.9794303797468356, + "grad_norm": 0.2506437876450052, + "learning_rate": 9.933987102121764e-06, + "loss": 0.5814, + "num_tokens": 2635636403.0, + "step": 2515 + }, + { + "epoch": 3.9873417721518987, + "grad_norm": 0.2284653998147006, + "learning_rate": 9.860658669743861e-06, + "loss": 0.5897, + "num_tokens": 2640879283.0, + "step": 2520 + }, + { + "epoch": 3.995253164556962, + "grad_norm": 0.22307394105006934, + "learning_rate": 9.787813184676056e-06, + "loss": 0.5861, + "num_tokens": 2646099304.0, + "step": 2525 + }, + { + "epoch": 4.003164556962025, + "grad_norm": 0.28089029948590827, + "learning_rate": 9.715452641353992e-06, + "loss": 0.5826, + "num_tokens": 2651342184.0, + "step": 2530 + }, + { + "epoch": 4.011075949367089, + "grad_norm": 0.23002963489023126, + "learning_rate": 9.643579020936106e-06, + "loss": 0.5662, + "num_tokens": 2656585064.0, + "step": 2535 + }, + { + "epoch": 4.018987341772152, + "grad_norm": 0.24076823558688779, + "learning_rate": 9.572194291249362e-06, + "loss": 0.5767, + "num_tokens": 2661827944.0, + "step": 2540 + }, + { + "epoch": 4.026898734177215, + "grad_norm": 0.22483920300780846, + "learning_rate": 9.501300406735406e-06, + "loss": 0.5743, + "num_tokens": 2667070260.0, + "step": 2545 + }, + { + "epoch": 4.034810126582278, + "grad_norm": 0.2425268841198098, + "learning_rate": 9.430899308397024e-06, + "loss": 0.5691, + "num_tokens": 2672313140.0, + "step": 2550 + }, + { + "epoch": 4.042721518987341, + "grad_norm": 0.22463870074019276, + "learning_rate": 9.360992923745032e-06, + "loss": 0.5727, + "num_tokens": 2677556020.0, + "step": 2555 + }, + { + "epoch": 4.050632911392405, + "grad_norm": 0.2257862359365115, + "learning_rate": 9.29158316674548e-06, + "loss": 0.5771, + "num_tokens": 2682798900.0, + "step": 2560 + }, + { + "epoch": 4.0585443037974684, + "grad_norm": 0.21855947078972188, + "learning_rate": 9.222671937767247e-06, + "loss": 0.5693, + "num_tokens": 2688041780.0, + "step": 2565 + }, + { + "epoch": 4.0664556962025316, + "grad_norm": 0.22655998729700472, + "learning_rate": 9.154261123530024e-06, + "loss": 0.5688, + "num_tokens": 2693280743.0, + "step": 2570 + }, + { + "epoch": 4.074367088607595, + "grad_norm": 0.2417803758293443, + "learning_rate": 9.086352597052674e-06, + "loss": 0.5699, + "num_tokens": 2698523623.0, + "step": 2575 + }, + { + "epoch": 4.082278481012658, + "grad_norm": 0.23491529993846316, + "learning_rate": 9.018948217601894e-06, + "loss": 0.5776, + "num_tokens": 2703760050.0, + "step": 2580 + }, + { + "epoch": 4.090189873417722, + "grad_norm": 0.21683310271118783, + "learning_rate": 8.952049830641368e-06, + "loss": 0.569, + "num_tokens": 2709001853.0, + "step": 2585 + }, + { + "epoch": 4.098101265822785, + "grad_norm": 0.22683758792329442, + "learning_rate": 8.885659267781218e-06, + "loss": 0.5723, + "num_tokens": 2714244733.0, + "step": 2590 + }, + { + "epoch": 4.106012658227848, + "grad_norm": 0.23440417433660363, + "learning_rate": 8.819778346727839e-06, + "loss": 0.571, + "num_tokens": 2719487613.0, + "step": 2595 + }, + { + "epoch": 4.113924050632911, + "grad_norm": 0.2192458553426772, + "learning_rate": 8.754408871234168e-06, + "loss": 0.5733, + "num_tokens": 2724715076.0, + "step": 2600 + }, + { + "epoch": 4.121835443037975, + "grad_norm": 0.2530876360164053, + "learning_rate": 8.689552631050274e-06, + "loss": 0.5714, + "num_tokens": 2729957956.0, + "step": 2605 + }, + { + "epoch": 4.129746835443038, + "grad_norm": 0.21492163566185005, + "learning_rate": 8.625211401874362e-06, + "loss": 0.5694, + "num_tokens": 2735200836.0, + "step": 2610 + }, + { + "epoch": 4.137658227848101, + "grad_norm": 0.22363359389581403, + "learning_rate": 8.561386945304155e-06, + "loss": 0.5727, + "num_tokens": 2740443716.0, + "step": 2615 + }, + { + "epoch": 4.1455696202531644, + "grad_norm": 0.2663268042449171, + "learning_rate": 8.498081008788677e-06, + "loss": 0.5689, + "num_tokens": 2745686596.0, + "step": 2620 + }, + { + "epoch": 4.1534810126582276, + "grad_norm": 0.2379773928525528, + "learning_rate": 8.435295325580373e-06, + "loss": 0.5693, + "num_tokens": 2750929476.0, + "step": 2625 + }, + { + "epoch": 4.1613924050632916, + "grad_norm": 0.2227866522568964, + "learning_rate": 8.373031614687699e-06, + "loss": 0.5698, + "num_tokens": 2756172356.0, + "step": 2630 + }, + { + "epoch": 4.169303797468355, + "grad_norm": 0.23722314775734296, + "learning_rate": 8.311291580828034e-06, + "loss": 0.5723, + "num_tokens": 2761415236.0, + "step": 2635 + }, + { + "epoch": 4.177215189873418, + "grad_norm": 0.2122410135791533, + "learning_rate": 8.250076914381017e-06, + "loss": 0.5661, + "num_tokens": 2766658116.0, + "step": 2640 + }, + { + "epoch": 4.185126582278481, + "grad_norm": 0.21872455435466664, + "learning_rate": 8.189389291342229e-06, + "loss": 0.5668, + "num_tokens": 2771900996.0, + "step": 2645 + }, + { + "epoch": 4.193037974683544, + "grad_norm": 0.22736670887055646, + "learning_rate": 8.12923037327738e-06, + "loss": 0.5723, + "num_tokens": 2777143876.0, + "step": 2650 + }, + { + "epoch": 4.200949367088608, + "grad_norm": 0.24119729706862408, + "learning_rate": 8.06960180727674e-06, + "loss": 0.5689, + "num_tokens": 2782386756.0, + "step": 2655 + }, + { + "epoch": 4.208860759493671, + "grad_norm": 0.22351050734609607, + "learning_rate": 8.010505225910083e-06, + "loss": 0.5677, + "num_tokens": 2787629636.0, + "step": 2660 + }, + { + "epoch": 4.216772151898734, + "grad_norm": 0.265494072724875, + "learning_rate": 7.951942247181992e-06, + "loss": 0.5718, + "num_tokens": 2792872516.0, + "step": 2665 + }, + { + "epoch": 4.224683544303797, + "grad_norm": 0.21820442695718212, + "learning_rate": 7.89391447448755e-06, + "loss": 0.5682, + "num_tokens": 2798109181.0, + "step": 2670 + }, + { + "epoch": 4.2325949367088604, + "grad_norm": 0.2420612432104863, + "learning_rate": 7.836423496568418e-06, + "loss": 0.5753, + "num_tokens": 2803337280.0, + "step": 2675 + }, + { + "epoch": 4.2405063291139244, + "grad_norm": 0.23517468870764166, + "learning_rate": 7.779470887469387e-06, + "loss": 0.5716, + "num_tokens": 2808580160.0, + "step": 2680 + }, + { + "epoch": 4.2484177215189876, + "grad_norm": 0.2267612123573776, + "learning_rate": 7.723058206495242e-06, + "loss": 0.5746, + "num_tokens": 2813823040.0, + "step": 2685 + }, + { + "epoch": 4.256329113924051, + "grad_norm": 0.22002454461146628, + "learning_rate": 7.667186998168082e-06, + "loss": 0.5684, + "num_tokens": 2819065920.0, + "step": 2690 + }, + { + "epoch": 4.264240506329114, + "grad_norm": 0.22912652065839756, + "learning_rate": 7.611858792185038e-06, + "loss": 0.5751, + "num_tokens": 2824308800.0, + "step": 2695 + }, + { + "epoch": 4.272151898734177, + "grad_norm": 0.22408651950567107, + "learning_rate": 7.557075103376383e-06, + "loss": 0.5724, + "num_tokens": 2829542456.0, + "step": 2700 + }, + { + "epoch": 4.280063291139241, + "grad_norm": 0.22342018834903565, + "learning_rate": 7.502837431664059e-06, + "loss": 0.5731, + "num_tokens": 2834785336.0, + "step": 2705 + }, + { + "epoch": 4.287974683544304, + "grad_norm": 0.22864174140723967, + "learning_rate": 7.449147262020616e-06, + "loss": 0.5679, + "num_tokens": 2840015342.0, + "step": 2710 + }, + { + "epoch": 4.295886075949367, + "grad_norm": 0.22865644771070281, + "learning_rate": 7.396006064428554e-06, + "loss": 0.5627, + "num_tokens": 2845248811.0, + "step": 2715 + }, + { + "epoch": 4.30379746835443, + "grad_norm": 0.2674648148165605, + "learning_rate": 7.34341529384006e-06, + "loss": 0.5765, + "num_tokens": 2850491691.0, + "step": 2720 + }, + { + "epoch": 4.311708860759493, + "grad_norm": 0.22345156469242552, + "learning_rate": 7.291376390137214e-06, + "loss": 0.5805, + "num_tokens": 2855734571.0, + "step": 2725 + }, + { + "epoch": 4.319620253164557, + "grad_norm": 0.2122765810320669, + "learning_rate": 7.239890778092509e-06, + "loss": 0.5712, + "num_tokens": 2860977451.0, + "step": 2730 + }, + { + "epoch": 4.3275316455696204, + "grad_norm": 0.2477019061431497, + "learning_rate": 7.188959867329893e-06, + "loss": 0.5814, + "num_tokens": 2866220331.0, + "step": 2735 + }, + { + "epoch": 4.3354430379746836, + "grad_norm": 0.21235713827113287, + "learning_rate": 7.138585052286142e-06, + "loss": 0.5732, + "num_tokens": 2871451563.0, + "step": 2740 + }, + { + "epoch": 4.343354430379747, + "grad_norm": 0.2146641271693942, + "learning_rate": 7.088767712172711e-06, + "loss": 0.5717, + "num_tokens": 2876694443.0, + "step": 2745 + }, + { + "epoch": 4.35126582278481, + "grad_norm": 0.23309761275535962, + "learning_rate": 7.039509210937932e-06, + "loss": 0.5732, + "num_tokens": 2881937323.0, + "step": 2750 + }, + { + "epoch": 4.359177215189874, + "grad_norm": 0.2234172047939556, + "learning_rate": 6.990810897229709e-06, + "loss": 0.57, + "num_tokens": 2887179519.0, + "step": 2755 + }, + { + "epoch": 4.367088607594937, + "grad_norm": 0.22237797838934822, + "learning_rate": 6.942674104358582e-06, + "loss": 0.5738, + "num_tokens": 2892422399.0, + "step": 2760 + }, + { + "epoch": 4.375, + "grad_norm": 0.21593708745753062, + "learning_rate": 6.8951001502612065e-06, + "loss": 0.5658, + "num_tokens": 2897636988.0, + "step": 2765 + }, + { + "epoch": 4.382911392405063, + "grad_norm": 0.22987570961197548, + "learning_rate": 6.848090337464294e-06, + "loss": 0.5682, + "num_tokens": 2902879868.0, + "step": 2770 + }, + { + "epoch": 4.390822784810126, + "grad_norm": 0.22029484227034318, + "learning_rate": 6.8016459530489335e-06, + "loss": 0.5746, + "num_tokens": 2908122748.0, + "step": 2775 + }, + { + "epoch": 4.39873417721519, + "grad_norm": 0.2171355743200665, + "learning_rate": 6.755768268615354e-06, + "loss": 0.567, + "num_tokens": 2913363036.0, + "step": 2780 + }, + { + "epoch": 4.406645569620253, + "grad_norm": 0.2213627866085991, + "learning_rate": 6.710458540248109e-06, + "loss": 0.5725, + "num_tokens": 2918605916.0, + "step": 2785 + }, + { + "epoch": 4.4145569620253164, + "grad_norm": 0.21863302278329563, + "learning_rate": 6.66571800848171e-06, + "loss": 0.5733, + "num_tokens": 2923835265.0, + "step": 2790 + }, + { + "epoch": 4.4224683544303796, + "grad_norm": 0.2114144591578047, + "learning_rate": 6.621547898266615e-06, + "loss": 0.5731, + "num_tokens": 2929076236.0, + "step": 2795 + }, + { + "epoch": 4.430379746835443, + "grad_norm": 0.22099052084655507, + "learning_rate": 6.577949418935732e-06, + "loss": 0.5741, + "num_tokens": 2934319116.0, + "step": 2800 + }, + { + "epoch": 4.438291139240507, + "grad_norm": 0.23002587290313806, + "learning_rate": 6.534923764171305e-06, + "loss": 0.572, + "num_tokens": 2939556532.0, + "step": 2805 + }, + { + "epoch": 4.44620253164557, + "grad_norm": 0.2525898694118869, + "learning_rate": 6.492472111972193e-06, + "loss": 0.5758, + "num_tokens": 2944799412.0, + "step": 2810 + }, + { + "epoch": 4.454113924050633, + "grad_norm": 0.22052197863760098, + "learning_rate": 6.450595624621672e-06, + "loss": 0.5687, + "num_tokens": 2950042292.0, + "step": 2815 + }, + { + "epoch": 4.462025316455696, + "grad_norm": 0.22012961127429426, + "learning_rate": 6.409295448655572e-06, + "loss": 0.5701, + "num_tokens": 2955275769.0, + "step": 2820 + }, + { + "epoch": 4.469936708860759, + "grad_norm": 0.22397593983188469, + "learning_rate": 6.3685727148309014e-06, + "loss": 0.5708, + "num_tokens": 2960518649.0, + "step": 2825 + }, + { + "epoch": 4.477848101265823, + "grad_norm": 0.21586061078083138, + "learning_rate": 6.328428538094895e-06, + "loss": 0.5763, + "num_tokens": 2965750136.0, + "step": 2830 + }, + { + "epoch": 4.485759493670886, + "grad_norm": 0.24077949998848439, + "learning_rate": 6.288864017554471e-06, + "loss": 0.5733, + "num_tokens": 2970993016.0, + "step": 2835 + }, + { + "epoch": 4.493670886075949, + "grad_norm": 0.225684038542863, + "learning_rate": 6.249880236446157e-06, + "loss": 0.5733, + "num_tokens": 2976232054.0, + "step": 2840 + }, + { + "epoch": 4.5015822784810124, + "grad_norm": 0.212864190216861, + "learning_rate": 6.211478262106416e-06, + "loss": 0.5757, + "num_tokens": 2981474934.0, + "step": 2845 + }, + { + "epoch": 4.509493670886076, + "grad_norm": 0.2240303672598969, + "learning_rate": 6.173659145942439e-06, + "loss": 0.5715, + "num_tokens": 2986716501.0, + "step": 2850 + }, + { + "epoch": 4.5174050632911396, + "grad_norm": 0.21964401447515855, + "learning_rate": 6.136423923403332e-06, + "loss": 0.5729, + "num_tokens": 2991948269.0, + "step": 2855 + }, + { + "epoch": 4.525316455696203, + "grad_norm": 0.21745396683364604, + "learning_rate": 6.099773613951805e-06, + "loss": 0.5691, + "num_tokens": 2997172161.0, + "step": 2860 + }, + { + "epoch": 4.533227848101266, + "grad_norm": 0.2051009997691557, + "learning_rate": 6.063709221036231e-06, + "loss": 0.5729, + "num_tokens": 3002415041.0, + "step": 2865 + }, + { + "epoch": 4.541139240506329, + "grad_norm": 0.2082354488706409, + "learning_rate": 6.028231732063179e-06, + "loss": 0.5683, + "num_tokens": 3007657921.0, + "step": 2870 + }, + { + "epoch": 4.549050632911392, + "grad_norm": 0.20333733144092334, + "learning_rate": 5.993342118370382e-06, + "loss": 0.5728, + "num_tokens": 3012900801.0, + "step": 2875 + }, + { + "epoch": 4.556962025316456, + "grad_norm": 0.22876886064683172, + "learning_rate": 5.959041335200154e-06, + "loss": 0.5713, + "num_tokens": 3018143681.0, + "step": 2880 + }, + { + "epoch": 4.564873417721519, + "grad_norm": 0.23107232316277548, + "learning_rate": 5.925330321673209e-06, + "loss": 0.5695, + "num_tokens": 3023386561.0, + "step": 2885 + }, + { + "epoch": 4.572784810126582, + "grad_norm": 0.2134849519354519, + "learning_rate": 5.892210000762978e-06, + "loss": 0.5669, + "num_tokens": 3028629441.0, + "step": 2890 + }, + { + "epoch": 4.580696202531645, + "grad_norm": 0.22796244574283725, + "learning_rate": 5.859681279270323e-06, + "loss": 0.5737, + "num_tokens": 3033867096.0, + "step": 2895 + }, + { + "epoch": 4.588607594936709, + "grad_norm": 0.2241995743982485, + "learning_rate": 5.827745047798706e-06, + "loss": 0.574, + "num_tokens": 3039109976.0, + "step": 2900 + }, + { + "epoch": 4.5965189873417724, + "grad_norm": 0.2432096355202257, + "learning_rate": 5.796402180729816e-06, + "loss": 0.5727, + "num_tokens": 3044352856.0, + "step": 2905 + }, + { + "epoch": 4.6044303797468356, + "grad_norm": 0.22578199424608314, + "learning_rate": 5.7656535361996335e-06, + "loss": 0.5732, + "num_tokens": 3049595736.0, + "step": 2910 + }, + { + "epoch": 4.612341772151899, + "grad_norm": 0.237878631640341, + "learning_rate": 5.735499956074914e-06, + "loss": 0.5694, + "num_tokens": 3054838616.0, + "step": 2915 + }, + { + "epoch": 4.620253164556962, + "grad_norm": 0.20655029863391322, + "learning_rate": 5.705942265930159e-06, + "loss": 0.568, + "num_tokens": 3060076579.0, + "step": 2920 + }, + { + "epoch": 4.628164556962025, + "grad_norm": 0.2102514850204724, + "learning_rate": 5.676981275025011e-06, + "loss": 0.5716, + "num_tokens": 3065319459.0, + "step": 2925 + }, + { + "epoch": 4.636075949367089, + "grad_norm": 0.21521201282981267, + "learning_rate": 5.648617776282077e-06, + "loss": 0.5727, + "num_tokens": 3070562339.0, + "step": 2930 + }, + { + "epoch": 4.643987341772152, + "grad_norm": 0.22220846495501412, + "learning_rate": 5.620852546265245e-06, + "loss": 0.5734, + "num_tokens": 3075805219.0, + "step": 2935 + }, + { + "epoch": 4.651898734177215, + "grad_norm": 0.20514030112007264, + "learning_rate": 5.5936863451584056e-06, + "loss": 0.5746, + "num_tokens": 3081039464.0, + "step": 2940 + }, + { + "epoch": 4.659810126582278, + "grad_norm": 0.1984930661211808, + "learning_rate": 5.56711991674465e-06, + "loss": 0.5775, + "num_tokens": 3086282344.0, + "step": 2945 + }, + { + "epoch": 4.667721518987342, + "grad_norm": 0.21848535912224024, + "learning_rate": 5.54115398838589e-06, + "loss": 0.5747, + "num_tokens": 3091525224.0, + "step": 2950 + }, + { + "epoch": 4.675632911392405, + "grad_norm": 0.21623565097615416, + "learning_rate": 5.515789271002967e-06, + "loss": 0.572, + "num_tokens": 3096751361.0, + "step": 2955 + }, + { + "epoch": 4.6835443037974684, + "grad_norm": 0.2067122434127961, + "learning_rate": 5.491026459056163e-06, + "loss": 0.5702, + "num_tokens": 3101978695.0, + "step": 2960 + }, + { + "epoch": 4.6914556962025316, + "grad_norm": 0.23059832785126186, + "learning_rate": 5.466866230526202e-06, + "loss": 0.5763, + "num_tokens": 3107221575.0, + "step": 2965 + }, + { + "epoch": 4.699367088607595, + "grad_norm": 0.21776141057424536, + "learning_rate": 5.443309246895688e-06, + "loss": 0.5724, + "num_tokens": 3112464455.0, + "step": 2970 + }, + { + "epoch": 4.707278481012658, + "grad_norm": 0.2086382988283168, + "learning_rate": 5.420356153130986e-06, + "loss": 0.5756, + "num_tokens": 3117704332.0, + "step": 2975 + }, + { + "epoch": 4.715189873417722, + "grad_norm": 0.22426323834508063, + "learning_rate": 5.398007577664566e-06, + "loss": 0.5766, + "num_tokens": 3122940201.0, + "step": 2980 + }, + { + "epoch": 4.723101265822785, + "grad_norm": 0.2134984091805231, + "learning_rate": 5.376264132377806e-06, + "loss": 0.5715, + "num_tokens": 3128183081.0, + "step": 2985 + }, + { + "epoch": 4.731012658227848, + "grad_norm": 0.2078865591432323, + "learning_rate": 5.355126412584226e-06, + "loss": 0.5796, + "num_tokens": 3133425961.0, + "step": 2990 + }, + { + "epoch": 4.738924050632911, + "grad_norm": 0.21033825281525734, + "learning_rate": 5.334594997013194e-06, + "loss": 0.5722, + "num_tokens": 3138668841.0, + "step": 2995 + }, + { + "epoch": 4.746835443037975, + "grad_norm": 0.20218242486649357, + "learning_rate": 5.314670447794094e-06, + "loss": 0.5697, + "num_tokens": 3143911721.0, + "step": 3000 + }, + { + "epoch": 4.754746835443038, + "grad_norm": 0.20528211055221243, + "learning_rate": 5.295353310440908e-06, + "loss": 0.5775, + "num_tokens": 3149154601.0, + "step": 3005 + }, + { + "epoch": 4.762658227848101, + "grad_norm": 0.20386433324381634, + "learning_rate": 5.276644113837305e-06, + "loss": 0.5703, + "num_tokens": 3154397481.0, + "step": 3010 + }, + { + "epoch": 4.7705696202531644, + "grad_norm": 0.21879527292522297, + "learning_rate": 5.258543370222146e-06, + "loss": 0.5739, + "num_tokens": 3159640361.0, + "step": 3015 + }, + { + "epoch": 4.7784810126582276, + "grad_norm": 0.238651659867707, + "learning_rate": 5.241051575175476e-06, + "loss": 0.567, + "num_tokens": 3164869054.0, + "step": 3020 + }, + { + "epoch": 4.786392405063291, + "grad_norm": 0.21930826191732458, + "learning_rate": 5.2241692076049294e-06, + "loss": 0.5713, + "num_tokens": 3170097693.0, + "step": 3025 + }, + { + "epoch": 4.794303797468355, + "grad_norm": 0.24549638164106027, + "learning_rate": 5.207896729732644e-06, + "loss": 0.5784, + "num_tokens": 3175338802.0, + "step": 3030 + }, + { + "epoch": 4.802215189873418, + "grad_norm": 0.20497121752386122, + "learning_rate": 5.1922345870825936e-06, + "loss": 0.5718, + "num_tokens": 3180581682.0, + "step": 3035 + }, + { + "epoch": 4.810126582278481, + "grad_norm": 0.21556589673067728, + "learning_rate": 5.1771832084683874e-06, + "loss": 0.5755, + "num_tokens": 3185824562.0, + "step": 3040 + }, + { + "epoch": 4.818037974683544, + "grad_norm": 0.2177509045329286, + "learning_rate": 5.162743005981538e-06, + "loss": 0.5693, + "num_tokens": 3191054329.0, + "step": 3045 + }, + { + "epoch": 4.825949367088608, + "grad_norm": 0.22258143026660726, + "learning_rate": 5.148914374980175e-06, + "loss": 0.5697, + "num_tokens": 3196297209.0, + "step": 3050 + }, + { + "epoch": 4.833860759493671, + "grad_norm": 0.21877472366506615, + "learning_rate": 5.135697694078209e-06, + "loss": 0.5732, + "num_tokens": 3201522963.0, + "step": 3055 + }, + { + "epoch": 4.841772151898734, + "grad_norm": 0.2194924983288656, + "learning_rate": 5.123093325134992e-06, + "loss": 0.577, + "num_tokens": 3206765843.0, + "step": 3060 + }, + { + "epoch": 4.849683544303797, + "grad_norm": 0.24236174132097704, + "learning_rate": 5.111101613245384e-06, + "loss": 0.5734, + "num_tokens": 3212008723.0, + "step": 3065 + }, + { + "epoch": 4.8575949367088604, + "grad_norm": 0.22136019033385568, + "learning_rate": 5.099722886730315e-06, + "loss": 0.5738, + "num_tokens": 3217244067.0, + "step": 3070 + }, + { + "epoch": 4.865506329113924, + "grad_norm": 0.21908628318619908, + "learning_rate": 5.088957457127804e-06, + "loss": 0.5657, + "num_tokens": 3222479254.0, + "step": 3075 + }, + { + "epoch": 4.8734177215189876, + "grad_norm": 0.23149848485508415, + "learning_rate": 5.078805619184415e-06, + "loss": 0.5762, + "num_tokens": 3227722134.0, + "step": 3080 + }, + { + "epoch": 4.881329113924051, + "grad_norm": 0.21576735952327258, + "learning_rate": 5.069267650847191e-06, + "loss": 0.5687, + "num_tokens": 3232960340.0, + "step": 3085 + }, + { + "epoch": 4.889240506329114, + "grad_norm": 0.21146979612133376, + "learning_rate": 5.060343813256054e-06, + "loss": 0.5749, + "num_tokens": 3238203220.0, + "step": 3090 + }, + { + "epoch": 4.897151898734177, + "grad_norm": 0.22343231609158218, + "learning_rate": 5.052034350736642e-06, + "loss": 0.577, + "num_tokens": 3243446100.0, + "step": 3095 + }, + { + "epoch": 4.905063291139241, + "grad_norm": 0.21750437307740444, + "learning_rate": 5.044339490793628e-06, + "loss": 0.5746, + "num_tokens": 3248682414.0, + "step": 3100 + }, + { + "epoch": 4.912974683544304, + "grad_norm": 0.2040207080561066, + "learning_rate": 5.037259444104488e-06, + "loss": 0.5752, + "num_tokens": 3253925294.0, + "step": 3105 + }, + { + "epoch": 4.920886075949367, + "grad_norm": 0.21492843271025994, + "learning_rate": 5.030794404513734e-06, + "loss": 0.5723, + "num_tokens": 3259168174.0, + "step": 3110 + }, + { + "epoch": 4.92879746835443, + "grad_norm": 0.22472996856568292, + "learning_rate": 5.024944549027605e-06, + "loss": 0.5757, + "num_tokens": 3264411054.0, + "step": 3115 + }, + { + "epoch": 4.936708860759493, + "grad_norm": 0.2105235983069695, + "learning_rate": 5.019710037809223e-06, + "loss": 0.5764, + "num_tokens": 3269653934.0, + "step": 3120 + }, + { + "epoch": 4.944620253164557, + "grad_norm": 0.20160858018191585, + "learning_rate": 5.015091014174209e-06, + "loss": 0.5676, + "num_tokens": 3274885682.0, + "step": 3125 + }, + { + "epoch": 4.9525316455696204, + "grad_norm": 0.2079284452125249, + "learning_rate": 5.0110876045867496e-06, + "loss": 0.5662, + "num_tokens": 3280128562.0, + "step": 3130 + }, + { + "epoch": 4.9604430379746836, + "grad_norm": 0.21834231623167524, + "learning_rate": 5.007699918656152e-06, + "loss": 0.5719, + "num_tokens": 3285371442.0, + "step": 3135 + }, + { + "epoch": 4.968354430379747, + "grad_norm": 0.21543773946569159, + "learning_rate": 5.004928049133823e-06, + "loss": 0.5694, + "num_tokens": 3290614322.0, + "step": 3140 + }, + { + "epoch": 4.97626582278481, + "grad_norm": 0.2150336576277292, + "learning_rate": 5.002772071910747e-06, + "loss": 0.5747, + "num_tokens": 3295843098.0, + "step": 3145 + }, + { + "epoch": 4.984177215189874, + "grad_norm": 0.20873534752474804, + "learning_rate": 5.0012320460153974e-06, + "loss": 0.5696, + "num_tokens": 3301070530.0, + "step": 3150 + }, + { + "epoch": 4.992088607594937, + "grad_norm": 0.21825458015635657, + "learning_rate": 5.000308013612126e-06, + "loss": 0.5762, + "num_tokens": 3306313410.0, + "step": 3155 + }, + { + "epoch": 5.0, + "grad_norm": 0.2206973186589871, + "learning_rate": 5e-06, + "loss": 0.5717, + "num_tokens": 3311556290.0, + "step": 3160 + }, + { + "epoch": 5.0, + "step": 3160, + "total_flos": 2880677515100160.0, + "train_loss": 0.6431828314744974, + "train_runtime": 46582.8084, + "train_samples_per_second": 4.34, + "train_steps_per_second": 0.068 + } + ], + "logging_steps": 5, + "max_steps": 3160, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2880677515100160.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}