{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 3160, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007911392405063292, "grad_norm": 2.3738328422587087, "learning_rate": 1.5822784810126583e-06, "loss": 1.1261, "num_tokens": 5242880.0, "step": 5 }, { "epoch": 0.015822784810126583, "grad_norm": 1.7722440301570659, "learning_rate": 3.1645569620253167e-06, "loss": 1.1058, "num_tokens": 10470029.0, "step": 10 }, { "epoch": 0.023734177215189875, "grad_norm": 1.083276838182905, "learning_rate": 4.746835443037975e-06, "loss": 1.0806, "num_tokens": 15712909.0, "step": 15 }, { "epoch": 0.03164556962025317, "grad_norm": 0.8611689905530553, "learning_rate": 6.329113924050633e-06, "loss": 1.02, "num_tokens": 20951872.0, "step": 20 }, { "epoch": 0.03955696202531646, "grad_norm": 0.7250421722694049, "learning_rate": 7.911392405063292e-06, "loss": 0.9766, "num_tokens": 26185341.0, "step": 25 }, { "epoch": 0.04746835443037975, "grad_norm": 0.6382913905764741, "learning_rate": 9.49367088607595e-06, "loss": 0.9456, "num_tokens": 31428221.0, "step": 30 }, { "epoch": 0.055379746835443035, "grad_norm": 0.5436432996737252, "learning_rate": 1.1075949367088608e-05, "loss": 0.9187, "num_tokens": 36669330.0, "step": 35 }, { "epoch": 0.06329113924050633, "grad_norm": 0.45175884987072906, "learning_rate": 1.2658227848101267e-05, "loss": 0.8956, "num_tokens": 41910897.0, "step": 40 }, { "epoch": 0.07120253164556962, "grad_norm": 0.3906460936286765, "learning_rate": 1.4240506329113925e-05, "loss": 0.8787, "num_tokens": 47153777.0, "step": 45 }, { "epoch": 0.07911392405063292, "grad_norm": 0.38060702422443193, "learning_rate": 1.5822784810126583e-05, "loss": 0.8674, "num_tokens": 52391193.0, "step": 50 }, { "epoch": 0.08702531645569621, "grad_norm": 0.3976893756810291, "learning_rate": 1.7405063291139243e-05, "loss": 0.8589, "num_tokens": 57634073.0, "step": 55 }, { "epoch": 0.0949367088607595, "grad_norm": 0.39610453911616206, "learning_rate": 1.89873417721519e-05, "loss": 0.8577, "num_tokens": 62876953.0, "step": 60 }, { "epoch": 0.10284810126582279, "grad_norm": 0.3232899679188804, "learning_rate": 2.056962025316456e-05, "loss": 0.8456, "num_tokens": 68119833.0, "step": 65 }, { "epoch": 0.11075949367088607, "grad_norm": 0.3785345815272384, "learning_rate": 2.2151898734177217e-05, "loss": 0.8354, "num_tokens": 73362713.0, "step": 70 }, { "epoch": 0.11867088607594936, "grad_norm": 0.42958423811773794, "learning_rate": 2.3734177215189873e-05, "loss": 0.8277, "num_tokens": 78605593.0, "step": 75 }, { "epoch": 0.12658227848101267, "grad_norm": 0.3759983838903421, "learning_rate": 2.5316455696202533e-05, "loss": 0.8257, "num_tokens": 83848473.0, "step": 80 }, { "epoch": 0.13449367088607594, "grad_norm": 0.4694604862178893, "learning_rate": 2.689873417721519e-05, "loss": 0.8212, "num_tokens": 89091353.0, "step": 85 }, { "epoch": 0.14240506329113925, "grad_norm": 0.41926661815869387, "learning_rate": 2.848101265822785e-05, "loss": 0.8132, "num_tokens": 94334233.0, "step": 90 }, { "epoch": 0.15031645569620253, "grad_norm": 0.8644371279587081, "learning_rate": 3.0063291139240506e-05, "loss": 0.8356, "num_tokens": 99577113.0, "step": 95 }, { "epoch": 0.15822784810126583, "grad_norm": 0.7239620838593109, "learning_rate": 3.1645569620253167e-05, "loss": 0.8115, "num_tokens": 104819993.0, "step": 100 }, { "epoch": 0.1661392405063291, "grad_norm": 0.6759442627573765, "learning_rate": 3.322784810126582e-05, "loss": 0.8131, "num_tokens": 110062873.0, "step": 105 }, { "epoch": 0.17405063291139242, "grad_norm": 0.5445796635610345, "learning_rate": 3.4810126582278487e-05, "loss": 0.8051, "num_tokens": 115305753.0, "step": 110 }, { "epoch": 0.1819620253164557, "grad_norm": 0.6030674874140043, "learning_rate": 3.639240506329114e-05, "loss": 0.8069, "num_tokens": 120539230.0, "step": 115 }, { "epoch": 0.189873417721519, "grad_norm": 0.5775646724321816, "learning_rate": 3.79746835443038e-05, "loss": 0.8001, "num_tokens": 125782110.0, "step": 120 }, { "epoch": 0.19778481012658228, "grad_norm": 1.1999946695600228, "learning_rate": 3.9556962025316456e-05, "loss": 0.7923, "num_tokens": 131014137.0, "step": 125 }, { "epoch": 0.20569620253164558, "grad_norm": 0.9396643672179282, "learning_rate": 4.113924050632912e-05, "loss": 0.8032, "num_tokens": 136257017.0, "step": 130 }, { "epoch": 0.21360759493670886, "grad_norm": 0.989562272615721, "learning_rate": 4.2721518987341776e-05, "loss": 0.7945, "num_tokens": 141499897.0, "step": 135 }, { "epoch": 0.22151898734177214, "grad_norm": 0.7938020654922184, "learning_rate": 4.430379746835443e-05, "loss": 0.7911, "num_tokens": 146712990.0, "step": 140 }, { "epoch": 0.22943037974683544, "grad_norm": 0.772547411213572, "learning_rate": 4.588607594936709e-05, "loss": 0.788, "num_tokens": 151927597.0, "step": 145 }, { "epoch": 0.23734177215189872, "grad_norm": 0.5427524570272351, "learning_rate": 4.7468354430379746e-05, "loss": 0.7931, "num_tokens": 157170477.0, "step": 150 }, { "epoch": 0.24525316455696203, "grad_norm": 0.520616997801594, "learning_rate": 4.905063291139241e-05, "loss": 0.7761, "num_tokens": 162397909.0, "step": 155 }, { "epoch": 0.25316455696202533, "grad_norm": 1.321171420646125, "learning_rate": 4.9999950717727614e-05, "loss": 0.8103, "num_tokens": 167639712.0, "step": 160 }, { "epoch": 0.2610759493670886, "grad_norm": 0.8179961728401967, "learning_rate": 4.999939629464255e-05, "loss": 0.7829, "num_tokens": 172882592.0, "step": 165 }, { "epoch": 0.2689873417721519, "grad_norm": 1.07950233433435, "learning_rate": 4.999822586086208e-05, "loss": 0.7815, "num_tokens": 178125472.0, "step": 170 }, { "epoch": 0.27689873417721517, "grad_norm": 0.764049579486957, "learning_rate": 4.9996439448431495e-05, "loss": 0.7936, "num_tokens": 183355239.0, "step": 175 }, { "epoch": 0.2848101265822785, "grad_norm": 0.6643682550526433, "learning_rate": 4.9994037106260965e-05, "loss": 0.777, "num_tokens": 188598119.0, "step": 180 }, { "epoch": 0.2927215189873418, "grad_norm": 0.4988533797702287, "learning_rate": 4.999101890012418e-05, "loss": 0.7755, "num_tokens": 193840999.0, "step": 185 }, { "epoch": 0.30063291139240506, "grad_norm": 0.49381791616490034, "learning_rate": 4.998738491265655e-05, "loss": 0.7646, "num_tokens": 199082063.0, "step": 190 }, { "epoch": 0.30854430379746833, "grad_norm": 0.5570710586204755, "learning_rate": 4.998313524335302e-05, "loss": 0.7698, "num_tokens": 204324943.0, "step": 195 }, { "epoch": 0.31645569620253167, "grad_norm": 0.4911267149039519, "learning_rate": 4.9978270008565184e-05, "loss": 0.7641, "num_tokens": 209561370.0, "step": 200 }, { "epoch": 0.32436708860759494, "grad_norm": 0.6409945115911554, "learning_rate": 4.997278934149828e-05, "loss": 0.7677, "num_tokens": 214804250.0, "step": 205 }, { "epoch": 0.3322784810126582, "grad_norm": 0.5401344460770882, "learning_rate": 4.996669339220741e-05, "loss": 0.7657, "num_tokens": 220047130.0, "step": 210 }, { "epoch": 0.3401898734177215, "grad_norm": 0.5361201681791574, "learning_rate": 4.995998232759349e-05, "loss": 0.7663, "num_tokens": 225290010.0, "step": 215 }, { "epoch": 0.34810126582278483, "grad_norm": 0.5803989669904568, "learning_rate": 4.995265633139869e-05, "loss": 0.7636, "num_tokens": 230532890.0, "step": 220 }, { "epoch": 0.3560126582278481, "grad_norm": 0.726211587240982, "learning_rate": 4.994471560420137e-05, "loss": 0.7631, "num_tokens": 235775770.0, "step": 225 }, { "epoch": 0.3639240506329114, "grad_norm": 0.698163213622165, "learning_rate": 4.9936160363410584e-05, "loss": 0.7545, "num_tokens": 241018650.0, "step": 230 }, { "epoch": 0.37183544303797467, "grad_norm": 0.5153045338635585, "learning_rate": 4.992699084326018e-05, "loss": 0.7541, "num_tokens": 246261530.0, "step": 235 }, { "epoch": 0.379746835443038, "grad_norm": 0.6959186613242923, "learning_rate": 4.9917207294802346e-05, "loss": 0.7635, "num_tokens": 251504410.0, "step": 240 }, { "epoch": 0.3876582278481013, "grad_norm": 0.475983385570461, "learning_rate": 4.990680998590071e-05, "loss": 0.755, "num_tokens": 256747290.0, "step": 245 }, { "epoch": 0.39556962025316456, "grad_norm": 0.6721509783638105, "learning_rate": 4.989579920122307e-05, "loss": 0.7532, "num_tokens": 261990170.0, "step": 250 }, { "epoch": 0.40348101265822783, "grad_norm": 0.4813755853658302, "learning_rate": 4.9884175242233585e-05, "loss": 0.7552, "num_tokens": 267233050.0, "step": 255 }, { "epoch": 0.41139240506329117, "grad_norm": 0.46711611820253635, "learning_rate": 4.987193842718448e-05, "loss": 0.7513, "num_tokens": 272475930.0, "step": 260 }, { "epoch": 0.41930379746835444, "grad_norm": 0.5972032077244416, "learning_rate": 4.985908909110735e-05, "loss": 0.7443, "num_tokens": 277718810.0, "step": 265 }, { "epoch": 0.4272151898734177, "grad_norm": 0.5059990764921701, "learning_rate": 4.9845627585804025e-05, "loss": 0.7458, "num_tokens": 282942349.0, "step": 270 }, { "epoch": 0.435126582278481, "grad_norm": 0.5582209701112856, "learning_rate": 4.9831554279836884e-05, "loss": 0.7521, "num_tokens": 288185229.0, "step": 275 }, { "epoch": 0.4430379746835443, "grad_norm": 0.4595435461264603, "learning_rate": 4.9816869558518796e-05, "loss": 0.7517, "num_tokens": 293428109.0, "step": 280 }, { "epoch": 0.4509493670886076, "grad_norm": 0.4675296642906588, "learning_rate": 4.980157382390258e-05, "loss": 0.7427, "num_tokens": 298670989.0, "step": 285 }, { "epoch": 0.4588607594936709, "grad_norm": 0.4659606979020265, "learning_rate": 4.978566749476995e-05, "loss": 0.7397, "num_tokens": 303913869.0, "step": 290 }, { "epoch": 0.46677215189873417, "grad_norm": 0.5143550840646993, "learning_rate": 4.9769151006620115e-05, "loss": 0.7488, "num_tokens": 309151588.0, "step": 295 }, { "epoch": 0.47468354430379744, "grad_norm": 0.5527231062409742, "learning_rate": 4.975202481165778e-05, "loss": 0.7528, "num_tokens": 314394468.0, "step": 300 }, { "epoch": 0.4825949367088608, "grad_norm": 0.6582518727489046, "learning_rate": 4.973428937878085e-05, "loss": 0.7388, "num_tokens": 319622567.0, "step": 305 }, { "epoch": 0.49050632911392406, "grad_norm": 0.5202900734758532, "learning_rate": 4.9715945193567506e-05, "loss": 0.7332, "num_tokens": 324865447.0, "step": 310 }, { "epoch": 0.49841772151898733, "grad_norm": 0.38453517395197584, "learning_rate": 4.969699275826298e-05, "loss": 0.7364, "num_tokens": 330108327.0, "step": 315 }, { "epoch": 0.5063291139240507, "grad_norm": 0.5632153181654556, "learning_rate": 4.967743259176575e-05, "loss": 0.7387, "num_tokens": 335351207.0, "step": 320 }, { "epoch": 0.5142405063291139, "grad_norm": 0.44887504740519496, "learning_rate": 4.96572652296134e-05, "loss": 0.7354, "num_tokens": 340594087.0, "step": 325 }, { "epoch": 0.5221518987341772, "grad_norm": 0.48735207873608405, "learning_rate": 4.963649122396788e-05, "loss": 0.7376, "num_tokens": 345825159.0, "step": 330 }, { "epoch": 0.5300632911392406, "grad_norm": 0.4193108797718534, "learning_rate": 4.961511114360043e-05, "loss": 0.7361, "num_tokens": 351068039.0, "step": 335 }, { "epoch": 0.5379746835443038, "grad_norm": 0.5662379674502074, "learning_rate": 4.959312557387601e-05, "loss": 0.739, "num_tokens": 356304704.0, "step": 340 }, { "epoch": 0.5458860759493671, "grad_norm": 0.5775103716382664, "learning_rate": 4.957053511673728e-05, "loss": 0.739, "num_tokens": 361547584.0, "step": 345 }, { "epoch": 0.5537974683544303, "grad_norm": 0.4784805413121098, "learning_rate": 4.954734039068806e-05, "loss": 0.7369, "num_tokens": 366775047.0, "step": 350 }, { "epoch": 0.5617088607594937, "grad_norm": 0.3826615225742145, "learning_rate": 4.95235420307765e-05, "loss": 0.7371, "num_tokens": 372017927.0, "step": 355 }, { "epoch": 0.569620253164557, "grad_norm": 0.5250012256720773, "learning_rate": 4.949914068857758e-05, "loss": 0.7401, "num_tokens": 377260807.0, "step": 360 }, { "epoch": 0.5775316455696202, "grad_norm": 0.5294573268848846, "learning_rate": 4.947413703217537e-05, "loss": 0.7303, "num_tokens": 382503687.0, "step": 365 }, { "epoch": 0.5854430379746836, "grad_norm": 0.4657426470359354, "learning_rate": 4.944853174614465e-05, "loss": 0.7276, "num_tokens": 387746567.0, "step": 370 }, { "epoch": 0.5933544303797469, "grad_norm": 0.6110352375166948, "learning_rate": 4.9422325531532254e-05, "loss": 0.7332, "num_tokens": 392989447.0, "step": 375 }, { "epoch": 0.6012658227848101, "grad_norm": 0.5173863425147587, "learning_rate": 4.9395519105837794e-05, "loss": 0.7345, "num_tokens": 398220435.0, "step": 380 }, { "epoch": 0.6091772151898734, "grad_norm": 0.44630851057828674, "learning_rate": 4.936811320299409e-05, "loss": 0.7344, "num_tokens": 403454091.0, "step": 385 }, { "epoch": 0.6170886075949367, "grad_norm": 0.5511830323483669, "learning_rate": 4.9340108573346994e-05, "loss": 0.7277, "num_tokens": 408695566.0, "step": 390 }, { "epoch": 0.625, "grad_norm": 0.488590082197144, "learning_rate": 4.931150598363494e-05, "loss": 0.7262, "num_tokens": 413938446.0, "step": 395 }, { "epoch": 0.6329113924050633, "grad_norm": 0.5980267797290539, "learning_rate": 4.9282306216967866e-05, "loss": 0.7367, "num_tokens": 419181326.0, "step": 400 }, { "epoch": 0.6408227848101266, "grad_norm": 0.4042639446511123, "learning_rate": 4.925251007280584e-05, "loss": 0.7277, "num_tokens": 424410675.0, "step": 405 }, { "epoch": 0.6487341772151899, "grad_norm": 0.5873987854936606, "learning_rate": 4.922211836693711e-05, "loss": 0.7216, "num_tokens": 429653555.0, "step": 410 }, { "epoch": 0.6566455696202531, "grad_norm": 0.49135762238140734, "learning_rate": 4.9191131931455836e-05, "loss": 0.7249, "num_tokens": 434896435.0, "step": 415 }, { "epoch": 0.6645569620253164, "grad_norm": 0.5977070057438132, "learning_rate": 4.915955161473925e-05, "loss": 0.7213, "num_tokens": 440139315.0, "step": 420 }, { "epoch": 0.6724683544303798, "grad_norm": 0.5865545939218848, "learning_rate": 4.9127378281424454e-05, "loss": 0.7291, "num_tokens": 445382195.0, "step": 425 }, { "epoch": 0.680379746835443, "grad_norm": 0.5141049741136753, "learning_rate": 4.909461281238475e-05, "loss": 0.7275, "num_tokens": 450625075.0, "step": 430 }, { "epoch": 0.6882911392405063, "grad_norm": 0.586173282775543, "learning_rate": 4.9061256104705514e-05, "loss": 0.7147, "num_tokens": 455867955.0, "step": 435 }, { "epoch": 0.6962025316455697, "grad_norm": 0.4884462198074093, "learning_rate": 4.902730907165962e-05, "loss": 0.7262, "num_tokens": 461110835.0, "step": 440 }, { "epoch": 0.7041139240506329, "grad_norm": 0.5155782485627577, "learning_rate": 4.899277264268249e-05, "loss": 0.726, "num_tokens": 466329951.0, "step": 445 }, { "epoch": 0.7120253164556962, "grad_norm": 0.5894348115226837, "learning_rate": 4.8957647763346575e-05, "loss": 0.7276, "num_tokens": 471572831.0, "step": 450 }, { "epoch": 0.7199367088607594, "grad_norm": 0.6178813430658108, "learning_rate": 4.892193539533553e-05, "loss": 0.7234, "num_tokens": 476815711.0, "step": 455 }, { "epoch": 0.7278481012658228, "grad_norm": 0.7787566109612348, "learning_rate": 4.888563651641784e-05, "loss": 0.7297, "num_tokens": 482044487.0, "step": 460 }, { "epoch": 0.7357594936708861, "grad_norm": 0.5460153094045689, "learning_rate": 4.884875212042005e-05, "loss": 0.7257, "num_tokens": 487287367.0, "step": 465 }, { "epoch": 0.7436708860759493, "grad_norm": 0.5019931926593302, "learning_rate": 4.88112832171996e-05, "loss": 0.7219, "num_tokens": 492530247.0, "step": 470 }, { "epoch": 0.7515822784810127, "grad_norm": 0.46922376523806203, "learning_rate": 4.877323083261713e-05, "loss": 0.7207, "num_tokens": 497773127.0, "step": 475 }, { "epoch": 0.759493670886076, "grad_norm": 0.4491131631584376, "learning_rate": 4.873459600850841e-05, "loss": 0.718, "num_tokens": 503013415.0, "step": 480 }, { "epoch": 0.7674050632911392, "grad_norm": 0.4576902411009356, "learning_rate": 4.869537980265581e-05, "loss": 0.7119, "num_tokens": 508255611.0, "step": 485 }, { "epoch": 0.7753164556962026, "grad_norm": 0.6179813478813936, "learning_rate": 4.865558328875937e-05, "loss": 0.7086, "num_tokens": 513498491.0, "step": 490 }, { "epoch": 0.7832278481012658, "grad_norm": 0.5471403707624125, "learning_rate": 4.8615207556407324e-05, "loss": 0.7242, "num_tokens": 518723580.0, "step": 495 }, { "epoch": 0.7911392405063291, "grad_norm": 0.4589250795263245, "learning_rate": 4.857425371104638e-05, "loss": 0.7159, "num_tokens": 523966460.0, "step": 500 }, { "epoch": 0.7990506329113924, "grad_norm": 0.4660626316391112, "learning_rate": 4.853272287395135e-05, "loss": 0.7251, "num_tokens": 529209340.0, "step": 505 }, { "epoch": 0.8069620253164557, "grad_norm": 0.5382100486924232, "learning_rate": 4.8490616182194534e-05, "loss": 0.7187, "num_tokens": 534429844.0, "step": 510 }, { "epoch": 0.814873417721519, "grad_norm": 0.5238715230138492, "learning_rate": 4.8447934788614515e-05, "loss": 0.7187, "num_tokens": 539672724.0, "step": 515 }, { "epoch": 0.8227848101265823, "grad_norm": 0.5194663159593585, "learning_rate": 4.840467986178464e-05, "loss": 0.7168, "num_tokens": 544915604.0, "step": 520 }, { "epoch": 0.8306962025316456, "grad_norm": 0.4135374652648048, "learning_rate": 4.8360852585981034e-05, "loss": 0.7236, "num_tokens": 550158484.0, "step": 525 }, { "epoch": 0.8386075949367089, "grad_norm": 0.47687000377153205, "learning_rate": 4.831645416115014e-05, "loss": 0.7091, "num_tokens": 555401364.0, "step": 530 }, { "epoch": 0.8465189873417721, "grad_norm": 0.4659031703902228, "learning_rate": 4.827148580287588e-05, "loss": 0.7042, "num_tokens": 560644244.0, "step": 535 }, { "epoch": 0.8544303797468354, "grad_norm": 0.432869438663762, "learning_rate": 4.82259487423464e-05, "loss": 0.7119, "num_tokens": 565887124.0, "step": 540 }, { "epoch": 0.8623417721518988, "grad_norm": 0.3989063394441302, "learning_rate": 4.81798442263203e-05, "loss": 0.7164, "num_tokens": 571114334.0, "step": 545 }, { "epoch": 0.870253164556962, "grad_norm": 0.5839119244408789, "learning_rate": 4.8133173517092575e-05, "loss": 0.7147, "num_tokens": 576351352.0, "step": 550 }, { "epoch": 0.8781645569620253, "grad_norm": 0.4989762196407589, "learning_rate": 4.808593789246e-05, "loss": 0.7116, "num_tokens": 581594232.0, "step": 555 }, { "epoch": 0.8860759493670886, "grad_norm": 0.6013182449359468, "learning_rate": 4.803813864568616e-05, "loss": 0.7101, "num_tokens": 586837112.0, "step": 560 }, { "epoch": 0.8939873417721519, "grad_norm": 0.47611561215232195, "learning_rate": 4.7989777085466054e-05, "loss": 0.7089, "num_tokens": 592079992.0, "step": 565 }, { "epoch": 0.9018987341772152, "grad_norm": 0.5649210203680972, "learning_rate": 4.794085453589022e-05, "loss": 0.7074, "num_tokens": 597322872.0, "step": 570 }, { "epoch": 0.9098101265822784, "grad_norm": 0.4954711968927705, "learning_rate": 4.789137233640858e-05, "loss": 0.7057, "num_tokens": 602545993.0, "step": 575 }, { "epoch": 0.9177215189873418, "grad_norm": 0.45694982466683615, "learning_rate": 4.7841331841793646e-05, "loss": 0.7099, "num_tokens": 607788873.0, "step": 580 }, { "epoch": 0.9256329113924051, "grad_norm": 0.4331078260141868, "learning_rate": 4.7790734422103503e-05, "loss": 0.7074, "num_tokens": 613031753.0, "step": 585 }, { "epoch": 0.9335443037974683, "grad_norm": 0.6029099335069901, "learning_rate": 4.77395814626443e-05, "loss": 0.7132, "num_tokens": 618274069.0, "step": 590 }, { "epoch": 0.9414556962025317, "grad_norm": 0.3964222914078214, "learning_rate": 4.7687874363932277e-05, "loss": 0.7107, "num_tokens": 623516949.0, "step": 595 }, { "epoch": 0.9493670886075949, "grad_norm": 0.40557860921547184, "learning_rate": 4.763561454165546e-05, "loss": 0.7114, "num_tokens": 628759829.0, "step": 600 }, { "epoch": 0.9572784810126582, "grad_norm": 0.3911037173148, "learning_rate": 4.758280342663488e-05, "loss": 0.7011, "num_tokens": 634002709.0, "step": 605 }, { "epoch": 0.9651898734177216, "grad_norm": 0.5205226454815663, "learning_rate": 4.7529442464785404e-05, "loss": 0.7035, "num_tokens": 639245589.0, "step": 610 }, { "epoch": 0.9731012658227848, "grad_norm": 0.6081952511481521, "learning_rate": 4.747553311707616e-05, "loss": 0.7109, "num_tokens": 644488469.0, "step": 615 }, { "epoch": 0.9810126582278481, "grad_norm": 0.5230701333389108, "learning_rate": 4.742107685949052e-05, "loss": 0.7054, "num_tokens": 649731349.0, "step": 620 }, { "epoch": 0.9889240506329114, "grad_norm": 0.45553268701488436, "learning_rate": 4.736607518298568e-05, "loss": 0.7084, "num_tokens": 654971226.0, "step": 625 }, { "epoch": 0.9968354430379747, "grad_norm": 0.4252255621189171, "learning_rate": 4.731052959345188e-05, "loss": 0.698, "num_tokens": 660214106.0, "step": 630 }, { "epoch": 1.004746835443038, "grad_norm": 0.5213909939853125, "learning_rate": 4.725444161167113e-05, "loss": 0.6984, "num_tokens": 665456986.0, "step": 635 }, { "epoch": 1.0126582278481013, "grad_norm": 0.6092766382306241, "learning_rate": 4.71978127732756e-05, "loss": 0.6834, "num_tokens": 670699866.0, "step": 640 }, { "epoch": 1.0205696202531644, "grad_norm": 0.6016879209190829, "learning_rate": 4.714064462870556e-05, "loss": 0.6799, "num_tokens": 675941341.0, "step": 645 }, { "epoch": 1.0284810126582278, "grad_norm": 0.46770946063736324, "learning_rate": 4.708293874316693e-05, "loss": 0.6718, "num_tokens": 681184221.0, "step": 650 }, { "epoch": 1.0363924050632911, "grad_norm": 0.5193625042396812, "learning_rate": 4.702469669658845e-05, "loss": 0.6824, "num_tokens": 686427101.0, "step": 655 }, { "epoch": 1.0443037974683544, "grad_norm": 0.4310707796867463, "learning_rate": 4.6965920083578406e-05, "loss": 0.6814, "num_tokens": 691643199.0, "step": 660 }, { "epoch": 1.0522151898734178, "grad_norm": 0.46509456309972086, "learning_rate": 4.690661051338096e-05, "loss": 0.6783, "num_tokens": 696885395.0, "step": 665 }, { "epoch": 1.0601265822784811, "grad_norm": 0.4290977849490662, "learning_rate": 4.6846769609832106e-05, "loss": 0.6745, "num_tokens": 702099431.0, "step": 670 }, { "epoch": 1.0680379746835442, "grad_norm": 0.44562761654453537, "learning_rate": 4.6786399011315215e-05, "loss": 0.6708, "num_tokens": 707335241.0, "step": 675 }, { "epoch": 1.0759493670886076, "grad_norm": 0.4027263571632704, "learning_rate": 4.672550037071616e-05, "loss": 0.6742, "num_tokens": 712578121.0, "step": 680 }, { "epoch": 1.0838607594936709, "grad_norm": 0.4509630202842952, "learning_rate": 4.6664075355378064e-05, "loss": 0.6787, "num_tokens": 717813308.0, "step": 685 }, { "epoch": 1.0917721518987342, "grad_norm": 0.579102128781878, "learning_rate": 4.660212564705569e-05, "loss": 0.6827, "num_tokens": 723056188.0, "step": 690 }, { "epoch": 1.0996835443037976, "grad_norm": 0.4750686074161037, "learning_rate": 4.653965294186933e-05, "loss": 0.6846, "num_tokens": 728299068.0, "step": 695 }, { "epoch": 1.1075949367088607, "grad_norm": 0.3720445468877555, "learning_rate": 4.647665895025842e-05, "loss": 0.68, "num_tokens": 733541948.0, "step": 700 }, { "epoch": 1.115506329113924, "grad_norm": 0.41896058368158273, "learning_rate": 4.6413145396934677e-05, "loss": 0.6727, "num_tokens": 738784828.0, "step": 705 }, { "epoch": 1.1234177215189873, "grad_norm": 0.41709918206048263, "learning_rate": 4.634911402083491e-05, "loss": 0.683, "num_tokens": 744027708.0, "step": 710 }, { "epoch": 1.1313291139240507, "grad_norm": 0.44679889502370607, "learning_rate": 4.628456657507336e-05, "loss": 0.6745, "num_tokens": 749270588.0, "step": 715 }, { "epoch": 1.139240506329114, "grad_norm": 0.4049444543836899, "learning_rate": 4.6219504826893774e-05, "loss": 0.6754, "num_tokens": 754513468.0, "step": 720 }, { "epoch": 1.1471518987341773, "grad_norm": 0.46788618407855487, "learning_rate": 4.615393055762095e-05, "loss": 0.6705, "num_tokens": 759755275.0, "step": 725 }, { "epoch": 1.1550632911392404, "grad_norm": 0.45889187811298193, "learning_rate": 4.6087845562612e-05, "loss": 0.6775, "num_tokens": 764993238.0, "step": 730 }, { "epoch": 1.1629746835443038, "grad_norm": 0.5053059074244526, "learning_rate": 4.602125165120721e-05, "loss": 0.6739, "num_tokens": 770236118.0, "step": 735 }, { "epoch": 1.1708860759493671, "grad_norm": 0.5163271050928327, "learning_rate": 4.595415064668044e-05, "loss": 0.6768, "num_tokens": 775478998.0, "step": 740 }, { "epoch": 1.1787974683544304, "grad_norm": 0.4282870421961532, "learning_rate": 4.5886544386189286e-05, "loss": 0.6681, "num_tokens": 780721878.0, "step": 745 }, { "epoch": 1.1867088607594938, "grad_norm": 0.46026364834735695, "learning_rate": 4.581843472072472e-05, "loss": 0.677, "num_tokens": 785964758.0, "step": 750 }, { "epoch": 1.1946202531645569, "grad_norm": 0.3647491363050659, "learning_rate": 4.574982351506044e-05, "loss": 0.673, "num_tokens": 791207638.0, "step": 755 }, { "epoch": 1.2025316455696202, "grad_norm": 0.40334584668377854, "learning_rate": 4.568071264770179e-05, "loss": 0.6733, "num_tokens": 796450518.0, "step": 760 }, { "epoch": 1.2104430379746836, "grad_norm": 0.6172425933341248, "learning_rate": 4.5611104010834384e-05, "loss": 0.6714, "num_tokens": 801691627.0, "step": 765 }, { "epoch": 1.2183544303797469, "grad_norm": 0.5660318449559009, "learning_rate": 4.554099951027223e-05, "loss": 0.6695, "num_tokens": 806934507.0, "step": 770 }, { "epoch": 1.2262658227848102, "grad_norm": 0.5013160792670336, "learning_rate": 4.5470401065405574e-05, "loss": 0.6733, "num_tokens": 812177387.0, "step": 775 }, { "epoch": 1.2341772151898733, "grad_norm": 0.48633536466547755, "learning_rate": 4.539931060914841e-05, "loss": 0.6775, "num_tokens": 817420267.0, "step": 780 }, { "epoch": 1.2420886075949367, "grad_norm": 0.5345960534136472, "learning_rate": 4.532773008788542e-05, "loss": 0.6742, "num_tokens": 822663147.0, "step": 785 }, { "epoch": 1.25, "grad_norm": 0.46751637317248607, "learning_rate": 4.5255661461418854e-05, "loss": 0.6683, "num_tokens": 827902185.0, "step": 790 }, { "epoch": 1.2579113924050633, "grad_norm": 0.40210236140036404, "learning_rate": 4.5183106702914744e-05, "loss": 0.6684, "num_tokens": 833145065.0, "step": 795 }, { "epoch": 1.2658227848101267, "grad_norm": 0.40800636198331647, "learning_rate": 4.511006779884894e-05, "loss": 0.6717, "num_tokens": 838387945.0, "step": 800 }, { "epoch": 1.2737341772151898, "grad_norm": 0.39143626003055415, "learning_rate": 4.503654674895268e-05, "loss": 0.6729, "num_tokens": 843630825.0, "step": 805 }, { "epoch": 1.2816455696202531, "grad_norm": 0.4114368812178738, "learning_rate": 4.49625455661579e-05, "loss": 0.6662, "num_tokens": 848873705.0, "step": 810 }, { "epoch": 1.2895569620253164, "grad_norm": 0.3850979319103424, "learning_rate": 4.4888066276542076e-05, "loss": 0.6676, "num_tokens": 854116585.0, "step": 815 }, { "epoch": 1.2974683544303798, "grad_norm": 0.3942547463552602, "learning_rate": 4.481311091927278e-05, "loss": 0.6697, "num_tokens": 859340010.0, "step": 820 }, { "epoch": 1.3053797468354431, "grad_norm": 0.3827683736050918, "learning_rate": 4.47376815465518e-05, "loss": 0.6648, "num_tokens": 864560605.0, "step": 825 }, { "epoch": 1.3132911392405062, "grad_norm": 0.4613241812192372, "learning_rate": 4.466178022355902e-05, "loss": 0.6757, "num_tokens": 869803485.0, "step": 830 }, { "epoch": 1.3212025316455696, "grad_norm": 0.3407656518079187, "learning_rate": 4.458540902839582e-05, "loss": 0.6643, "num_tokens": 875046365.0, "step": 835 }, { "epoch": 1.3291139240506329, "grad_norm": 0.3433476064598642, "learning_rate": 4.450857005202823e-05, "loss": 0.6684, "num_tokens": 880272119.0, "step": 840 }, { "epoch": 1.3370253164556962, "grad_norm": 0.4072649317968132, "learning_rate": 4.443126539822962e-05, "loss": 0.6758, "num_tokens": 885514999.0, "step": 845 }, { "epoch": 1.3449367088607596, "grad_norm": 0.38388206068514286, "learning_rate": 4.435349718352319e-05, "loss": 0.6651, "num_tokens": 890751664.0, "step": 850 }, { "epoch": 1.3528481012658227, "grad_norm": 0.4906281471488268, "learning_rate": 4.427526753712392e-05, "loss": 0.6758, "num_tokens": 895978998.0, "step": 855 }, { "epoch": 1.360759493670886, "grad_norm": 0.5690161031589217, "learning_rate": 4.419657860088033e-05, "loss": 0.6661, "num_tokens": 901206430.0, "step": 860 }, { "epoch": 1.3686708860759493, "grad_norm": 0.5239364843573313, "learning_rate": 4.411743252921587e-05, "loss": 0.6711, "num_tokens": 906449310.0, "step": 865 }, { "epoch": 1.3765822784810127, "grad_norm": 0.4590626844046868, "learning_rate": 4.403783148906984e-05, "loss": 0.6742, "num_tokens": 911692190.0, "step": 870 }, { "epoch": 1.384493670886076, "grad_norm": 0.6369825573287001, "learning_rate": 4.3957777659838156e-05, "loss": 0.6691, "num_tokens": 916923938.0, "step": 875 }, { "epoch": 1.3924050632911391, "grad_norm": 0.5309325547971587, "learning_rate": 4.387727323331362e-05, "loss": 0.6671, "num_tokens": 922166818.0, "step": 880 }, { "epoch": 1.4003164556962027, "grad_norm": 0.4182221238855379, "learning_rate": 4.379632041362594e-05, "loss": 0.6673, "num_tokens": 927409698.0, "step": 885 }, { "epoch": 1.4082278481012658, "grad_norm": 0.37453984190952516, "learning_rate": 4.371492141718138e-05, "loss": 0.6669, "num_tokens": 932652578.0, "step": 890 }, { "epoch": 1.4161392405063291, "grad_norm": 0.4472314832066788, "learning_rate": 4.363307847260206e-05, "loss": 0.6609, "num_tokens": 937883810.0, "step": 895 }, { "epoch": 1.4240506329113924, "grad_norm": 0.40759290291866934, "learning_rate": 4.3550793820664965e-05, "loss": 0.6657, "num_tokens": 943113159.0, "step": 900 }, { "epoch": 1.4319620253164558, "grad_norm": 0.38119165043031206, "learning_rate": 4.3468069714240556e-05, "loss": 0.6674, "num_tokens": 948346815.0, "step": 905 }, { "epoch": 1.439873417721519, "grad_norm": 0.3408121507908347, "learning_rate": 4.3384908418231144e-05, "loss": 0.6649, "num_tokens": 953589695.0, "step": 910 }, { "epoch": 1.4477848101265822, "grad_norm": 0.3824255603451359, "learning_rate": 4.330131220950883e-05, "loss": 0.664, "num_tokens": 958831498.0, "step": 915 }, { "epoch": 1.4556962025316456, "grad_norm": 0.42035687659685655, "learning_rate": 4.321728337685318e-05, "loss": 0.667, "num_tokens": 964074378.0, "step": 920 }, { "epoch": 1.4636075949367089, "grad_norm": 0.45598858937412523, "learning_rate": 4.313282422088859e-05, "loss": 0.6685, "num_tokens": 969317258.0, "step": 925 }, { "epoch": 1.4715189873417722, "grad_norm": 0.44573187825703264, "learning_rate": 4.304793705402124e-05, "loss": 0.6613, "num_tokens": 974546034.0, "step": 930 }, { "epoch": 1.4794303797468356, "grad_norm": 0.3946235668573835, "learning_rate": 4.2962624200375835e-05, "loss": 0.6671, "num_tokens": 979788914.0, "step": 935 }, { "epoch": 1.4873417721518987, "grad_norm": 0.3228711597439979, "learning_rate": 4.287688799573195e-05, "loss": 0.6686, "num_tokens": 985031794.0, "step": 940 }, { "epoch": 1.495253164556962, "grad_norm": 0.3636228559548819, "learning_rate": 4.2790730787460096e-05, "loss": 0.6657, "num_tokens": 990255694.0, "step": 945 }, { "epoch": 1.5031645569620253, "grad_norm": 0.36277524743915257, "learning_rate": 4.270415493445739e-05, "loss": 0.6618, "num_tokens": 995498574.0, "step": 950 }, { "epoch": 1.5110759493670884, "grad_norm": 0.449377328776007, "learning_rate": 4.2617162807083084e-05, "loss": 0.664, "num_tokens": 1000741454.0, "step": 955 }, { "epoch": 1.518987341772152, "grad_norm": 0.5226848678404952, "learning_rate": 4.252975678709354e-05, "loss": 0.664, "num_tokens": 1005981331.0, "step": 960 }, { "epoch": 1.5268987341772151, "grad_norm": 0.426961497295756, "learning_rate": 4.244193926757713e-05, "loss": 0.6672, "num_tokens": 1011224211.0, "step": 965 }, { "epoch": 1.5348101265822784, "grad_norm": 0.3336409027569924, "learning_rate": 4.235371265288864e-05, "loss": 0.6681, "num_tokens": 1016467091.0, "step": 970 }, { "epoch": 1.5427215189873418, "grad_norm": 0.4036839066490488, "learning_rate": 4.2265079358583455e-05, "loss": 0.6643, "num_tokens": 1021695784.0, "step": 975 }, { "epoch": 1.5506329113924051, "grad_norm": 0.7140710792273346, "learning_rate": 4.217604181135147e-05, "loss": 0.6697, "num_tokens": 1026938664.0, "step": 980 }, { "epoch": 1.5585443037974684, "grad_norm": 0.5962932136354121, "learning_rate": 4.20866024489506e-05, "loss": 0.6636, "num_tokens": 1032181544.0, "step": 985 }, { "epoch": 1.5664556962025316, "grad_norm": 0.5399117523136845, "learning_rate": 4.199676372014004e-05, "loss": 0.6642, "num_tokens": 1037424424.0, "step": 990 }, { "epoch": 1.5743670886075949, "grad_norm": 0.5727471985317703, "learning_rate": 4.190652808461326e-05, "loss": 0.6673, "num_tokens": 1042667304.0, "step": 995 }, { "epoch": 1.5822784810126582, "grad_norm": 0.4218917245287635, "learning_rate": 4.1815898012930603e-05, "loss": 0.6597, "num_tokens": 1047910184.0, "step": 1000 }, { "epoch": 1.5901898734177216, "grad_norm": 0.3919982695655947, "learning_rate": 4.172487598645171e-05, "loss": 0.6587, "num_tokens": 1053153064.0, "step": 1005 }, { "epoch": 1.5981012658227849, "grad_norm": 0.35433285883968835, "learning_rate": 4.163346449726752e-05, "loss": 0.6658, "num_tokens": 1058395944.0, "step": 1010 }, { "epoch": 1.606012658227848, "grad_norm": 0.40624024878352494, "learning_rate": 4.154166604813206e-05, "loss": 0.661, "num_tokens": 1063638824.0, "step": 1015 }, { "epoch": 1.6139240506329116, "grad_norm": 0.45360696653600874, "learning_rate": 4.144948315239397e-05, "loss": 0.6636, "num_tokens": 1068880391.0, "step": 1020 }, { "epoch": 1.6218354430379747, "grad_norm": 0.528510428398564, "learning_rate": 4.13569183339276e-05, "loss": 0.6561, "num_tokens": 1074123271.0, "step": 1025 }, { "epoch": 1.629746835443038, "grad_norm": 0.435117205930865, "learning_rate": 4.1263974127063994e-05, "loss": 0.6572, "num_tokens": 1079366151.0, "step": 1030 }, { "epoch": 1.6376582278481013, "grad_norm": 0.35010239698525997, "learning_rate": 4.117065307652145e-05, "loss": 0.6675, "num_tokens": 1084603567.0, "step": 1035 }, { "epoch": 1.6455696202531644, "grad_norm": 0.34340761605514514, "learning_rate": 4.107695773733586e-05, "loss": 0.6634, "num_tokens": 1089846447.0, "step": 1040 }, { "epoch": 1.653481012658228, "grad_norm": 0.3800590366717773, "learning_rate": 4.098289067479077e-05, "loss": 0.6529, "num_tokens": 1095089327.0, "step": 1045 }, { "epoch": 1.6613924050632911, "grad_norm": 0.5301647765231198, "learning_rate": 4.0888454464347156e-05, "loss": 0.6608, "num_tokens": 1100328290.0, "step": 1050 }, { "epoch": 1.6693037974683544, "grad_norm": 0.40371596065925885, "learning_rate": 4.079365169157283e-05, "loss": 0.6596, "num_tokens": 1105565945.0, "step": 1055 }, { "epoch": 1.6772151898734178, "grad_norm": 0.44878584304658214, "learning_rate": 4.069848495207176e-05, "loss": 0.6584, "num_tokens": 1110808825.0, "step": 1060 }, { "epoch": 1.685126582278481, "grad_norm": 0.42760590310826146, "learning_rate": 4.060295685141295e-05, "loss": 0.6601, "num_tokens": 1116051705.0, "step": 1065 }, { "epoch": 1.6930379746835444, "grad_norm": 0.42094155594543997, "learning_rate": 4.0507070005059086e-05, "loss": 0.6636, "num_tokens": 1121294585.0, "step": 1070 }, { "epoch": 1.7009493670886076, "grad_norm": 0.366450602326517, "learning_rate": 4.0410827038294966e-05, "loss": 0.6565, "num_tokens": 1126537465.0, "step": 1075 }, { "epoch": 1.7088607594936709, "grad_norm": 0.3382264432193073, "learning_rate": 4.031423058615559e-05, "loss": 0.6637, "num_tokens": 1131778529.0, "step": 1080 }, { "epoch": 1.7167721518987342, "grad_norm": 0.4135053453091331, "learning_rate": 4.0217283293354044e-05, "loss": 0.6598, "num_tokens": 1137021409.0, "step": 1085 }, { "epoch": 1.7246835443037973, "grad_norm": 0.32789223717337174, "learning_rate": 4.011998781420907e-05, "loss": 0.66, "num_tokens": 1142264289.0, "step": 1090 }, { "epoch": 1.7325949367088609, "grad_norm": 0.37159495208450527, "learning_rate": 4.002234681257239e-05, "loss": 0.6634, "num_tokens": 1147488495.0, "step": 1095 }, { "epoch": 1.740506329113924, "grad_norm": 0.41706053572545077, "learning_rate": 3.992436296175581e-05, "loss": 0.6554, "num_tokens": 1152731375.0, "step": 1100 }, { "epoch": 1.7484177215189873, "grad_norm": 0.3410031339365452, "learning_rate": 3.982603894445796e-05, "loss": 0.6574, "num_tokens": 1157974255.0, "step": 1105 }, { "epoch": 1.7563291139240507, "grad_norm": 0.40387170849869736, "learning_rate": 3.97273774526909e-05, "loss": 0.6518, "num_tokens": 1163196277.0, "step": 1110 }, { "epoch": 1.7642405063291138, "grad_norm": 0.41494353564073877, "learning_rate": 3.962838118770643e-05, "loss": 0.6521, "num_tokens": 1168438593.0, "step": 1115 }, { "epoch": 1.7721518987341773, "grad_norm": 0.33845287541155555, "learning_rate": 3.952905285992206e-05, "loss": 0.6636, "num_tokens": 1173674462.0, "step": 1120 }, { "epoch": 1.7800632911392404, "grad_norm": 0.3616932835037642, "learning_rate": 3.942939518884686e-05, "loss": 0.6566, "num_tokens": 1178914750.0, "step": 1125 }, { "epoch": 1.7879746835443038, "grad_norm": 0.38849401008078355, "learning_rate": 3.932941090300699e-05, "loss": 0.6546, "num_tokens": 1184150094.0, "step": 1130 }, { "epoch": 1.7958860759493671, "grad_norm": 0.42281182227916236, "learning_rate": 3.922910273987098e-05, "loss": 0.6548, "num_tokens": 1189392974.0, "step": 1135 }, { "epoch": 1.8037974683544302, "grad_norm": 0.4210942734836147, "learning_rate": 3.912847344577481e-05, "loss": 0.6555, "num_tokens": 1194635854.0, "step": 1140 }, { "epoch": 1.8117088607594938, "grad_norm": 0.5474931691520276, "learning_rate": 3.9027525775846666e-05, "loss": 0.6533, "num_tokens": 1199878734.0, "step": 1145 }, { "epoch": 1.8196202531645569, "grad_norm": 0.3732844050319297, "learning_rate": 3.892626249393159e-05, "loss": 0.6533, "num_tokens": 1205121614.0, "step": 1150 }, { "epoch": 1.8275316455696202, "grad_norm": 0.45982960737179485, "learning_rate": 3.882468637251573e-05, "loss": 0.658, "num_tokens": 1210364494.0, "step": 1155 }, { "epoch": 1.8354430379746836, "grad_norm": 0.3489396715842175, "learning_rate": 3.872280019265046e-05, "loss": 0.6583, "num_tokens": 1215607374.0, "step": 1160 }, { "epoch": 1.8433544303797469, "grad_norm": 0.35751783832753387, "learning_rate": 3.8620606743876264e-05, "loss": 0.6522, "num_tokens": 1220843801.0, "step": 1165 }, { "epoch": 1.8512658227848102, "grad_norm": 0.4188675688211132, "learning_rate": 3.851810882414632e-05, "loss": 0.6497, "num_tokens": 1226086681.0, "step": 1170 }, { "epoch": 1.8591772151898733, "grad_norm": 0.3704256158964576, "learning_rate": 3.841530923974991e-05, "loss": 0.6542, "num_tokens": 1231329561.0, "step": 1175 }, { "epoch": 1.8670886075949367, "grad_norm": 0.35739760458738973, "learning_rate": 3.83122108052356e-05, "loss": 0.6558, "num_tokens": 1236572441.0, "step": 1180 }, { "epoch": 1.875, "grad_norm": 0.4108797442075266, "learning_rate": 3.8208816343334156e-05, "loss": 0.6503, "num_tokens": 1241815321.0, "step": 1185 }, { "epoch": 1.8829113924050633, "grad_norm": 0.3989382930649431, "learning_rate": 3.810512868488129e-05, "loss": 0.6567, "num_tokens": 1247058201.0, "step": 1190 }, { "epoch": 1.8908227848101267, "grad_norm": 0.36123580918393156, "learning_rate": 3.800115066874014e-05, "loss": 0.6539, "num_tokens": 1252301081.0, "step": 1195 }, { "epoch": 1.8987341772151898, "grad_norm": 0.3372531720914223, "learning_rate": 3.789688514172353e-05, "loss": 0.6473, "num_tokens": 1257543961.0, "step": 1200 }, { "epoch": 1.9066455696202531, "grad_norm": 0.3512013159202198, "learning_rate": 3.779233495851604e-05, "loss": 0.6581, "num_tokens": 1262786841.0, "step": 1205 }, { "epoch": 1.9145569620253164, "grad_norm": 0.3542753269510345, "learning_rate": 3.768750298159587e-05, "loss": 0.6445, "num_tokens": 1268029721.0, "step": 1210 }, { "epoch": 1.9224683544303798, "grad_norm": 0.4031550450820841, "learning_rate": 3.75823920811564e-05, "loss": 0.6508, "num_tokens": 1273258360.0, "step": 1215 }, { "epoch": 1.9303797468354431, "grad_norm": 0.4403211137259473, "learning_rate": 3.747700513502772e-05, "loss": 0.6573, "num_tokens": 1278501240.0, "step": 1220 }, { "epoch": 1.9382911392405062, "grad_norm": 0.4812191269085378, "learning_rate": 3.737134502859772e-05, "loss": 0.6536, "num_tokens": 1283744120.0, "step": 1225 }, { "epoch": 1.9462025316455698, "grad_norm": 0.39299909706187625, "learning_rate": 3.726541465473317e-05, "loss": 0.649, "num_tokens": 1288987000.0, "step": 1230 }, { "epoch": 1.9541139240506329, "grad_norm": 0.39824695051848524, "learning_rate": 3.7159216913700456e-05, "loss": 0.6478, "num_tokens": 1294229880.0, "step": 1235 }, { "epoch": 1.9620253164556962, "grad_norm": 0.42039758141791633, "learning_rate": 3.7052754713086246e-05, "loss": 0.6534, "num_tokens": 1299472760.0, "step": 1240 }, { "epoch": 1.9699367088607596, "grad_norm": 0.3960672982967281, "learning_rate": 3.694603096771781e-05, "loss": 0.6568, "num_tokens": 1304715640.0, "step": 1245 }, { "epoch": 1.9778481012658227, "grad_norm": 0.3747599520232236, "learning_rate": 3.683904859958329e-05, "loss": 0.651, "num_tokens": 1309942452.0, "step": 1250 }, { "epoch": 1.9857594936708862, "grad_norm": 0.34963602368732627, "learning_rate": 3.673181053775162e-05, "loss": 0.6591, "num_tokens": 1315185332.0, "step": 1255 }, { "epoch": 1.9936708860759493, "grad_norm": 0.3336233742912806, "learning_rate": 3.662431971829237e-05, "loss": 0.6567, "num_tokens": 1320428212.0, "step": 1260 }, { "epoch": 2.0015822784810124, "grad_norm": 0.34626223476299406, "learning_rate": 3.651657908419537e-05, "loss": 0.6528, "num_tokens": 1325656988.0, "step": 1265 }, { "epoch": 2.009493670886076, "grad_norm": 0.3739591180690111, "learning_rate": 3.640859158529014e-05, "loss": 0.6228, "num_tokens": 1330899868.0, "step": 1270 }, { "epoch": 2.017405063291139, "grad_norm": 0.36163026913569957, "learning_rate": 3.6300360178165065e-05, "loss": 0.6198, "num_tokens": 1336142748.0, "step": 1275 }, { "epoch": 2.0253164556962027, "grad_norm": 0.3413456124109847, "learning_rate": 3.619188782608653e-05, "loss": 0.6249, "num_tokens": 1341371387.0, "step": 1280 }, { "epoch": 2.0332278481012658, "grad_norm": 0.41603416587113357, "learning_rate": 3.6083177498917745e-05, "loss": 0.6137, "num_tokens": 1346614267.0, "step": 1285 }, { "epoch": 2.041139240506329, "grad_norm": 0.39514773706497214, "learning_rate": 3.5974232173037385e-05, "loss": 0.6292, "num_tokens": 1351857147.0, "step": 1290 }, { "epoch": 2.0490506329113924, "grad_norm": 0.3949477431561823, "learning_rate": 3.586505483125823e-05, "loss": 0.6204, "num_tokens": 1357100027.0, "step": 1295 }, { "epoch": 2.0569620253164556, "grad_norm": 0.324048846717354, "learning_rate": 3.5755648462745366e-05, "loss": 0.6263, "num_tokens": 1362342907.0, "step": 1300 }, { "epoch": 2.064873417721519, "grad_norm": 0.3478785681280589, "learning_rate": 3.5646016062934413e-05, "loss": 0.6216, "num_tokens": 1367576563.0, "step": 1305 }, { "epoch": 2.0727848101265822, "grad_norm": 0.38436773833943666, "learning_rate": 3.553616063344951e-05, "loss": 0.6255, "num_tokens": 1372800486.0, "step": 1310 }, { "epoch": 2.0806962025316458, "grad_norm": 0.38048821813178435, "learning_rate": 3.5426085182021114e-05, "loss": 0.6208, "num_tokens": 1378043366.0, "step": 1315 }, { "epoch": 2.088607594936709, "grad_norm": 0.3868214542706034, "learning_rate": 3.531579272240366e-05, "loss": 0.6235, "num_tokens": 1383286246.0, "step": 1320 }, { "epoch": 2.096518987341772, "grad_norm": 0.29924119376240554, "learning_rate": 3.520528627429304e-05, "loss": 0.616, "num_tokens": 1388529126.0, "step": 1325 }, { "epoch": 2.1044303797468356, "grad_norm": 0.30143694906012813, "learning_rate": 3.509456886324395e-05, "loss": 0.6228, "num_tokens": 1393772006.0, "step": 1330 }, { "epoch": 2.1123417721518987, "grad_norm": 0.3430834549488217, "learning_rate": 3.498364352058703e-05, "loss": 0.6231, "num_tokens": 1399014886.0, "step": 1335 }, { "epoch": 2.1202531645569622, "grad_norm": 0.3918758750528519, "learning_rate": 3.487251328334588e-05, "loss": 0.6184, "num_tokens": 1404254509.0, "step": 1340 }, { "epoch": 2.1281645569620253, "grad_norm": 0.3860740822731293, "learning_rate": 3.47611811941539e-05, "loss": 0.6223, "num_tokens": 1409497389.0, "step": 1345 }, { "epoch": 2.1360759493670884, "grad_norm": 0.30507806139472327, "learning_rate": 3.464965030117099e-05, "loss": 0.6227, "num_tokens": 1414740269.0, "step": 1350 }, { "epoch": 2.143987341772152, "grad_norm": 0.31931935671102873, "learning_rate": 3.453792365800011e-05, "loss": 0.6213, "num_tokens": 1419976138.0, "step": 1355 }, { "epoch": 2.151898734177215, "grad_norm": 0.4296009660896887, "learning_rate": 3.4426004323603655e-05, "loss": 0.6195, "num_tokens": 1425219018.0, "step": 1360 }, { "epoch": 2.1598101265822787, "grad_norm": 0.36080348698318637, "learning_rate": 3.4313895362219704e-05, "loss": 0.6321, "num_tokens": 1430461898.0, "step": 1365 }, { "epoch": 2.1677215189873418, "grad_norm": 0.31266622534218946, "learning_rate": 3.420159984327814e-05, "loss": 0.6329, "num_tokens": 1435704778.0, "step": 1370 }, { "epoch": 2.175632911392405, "grad_norm": 0.29368200019966956, "learning_rate": 3.40891208413166e-05, "loss": 0.622, "num_tokens": 1440941205.0, "step": 1375 }, { "epoch": 2.1835443037974684, "grad_norm": 0.3318857805197203, "learning_rate": 3.397646143589629e-05, "loss": 0.6218, "num_tokens": 1446182176.0, "step": 1380 }, { "epoch": 2.1914556962025316, "grad_norm": 0.3469231890169258, "learning_rate": 3.38636247115177e-05, "loss": 0.6241, "num_tokens": 1451425056.0, "step": 1385 }, { "epoch": 2.199367088607595, "grad_norm": 0.3555477367643954, "learning_rate": 3.3750613757536134e-05, "loss": 0.6237, "num_tokens": 1456667936.0, "step": 1390 }, { "epoch": 2.2072784810126582, "grad_norm": 0.4308050328882218, "learning_rate": 3.36374316680771e-05, "loss": 0.6269, "num_tokens": 1461909000.0, "step": 1395 }, { "epoch": 2.2151898734177213, "grad_norm": 0.35870973114272886, "learning_rate": 3.3524081541951644e-05, "loss": 0.6222, "num_tokens": 1467150803.0, "step": 1400 }, { "epoch": 2.223101265822785, "grad_norm": 0.3606193086420081, "learning_rate": 3.3410566482571484e-05, "loss": 0.625, "num_tokens": 1472393683.0, "step": 1405 }, { "epoch": 2.231012658227848, "grad_norm": 0.3173225301332037, "learning_rate": 3.329688959786403e-05, "loss": 0.6225, "num_tokens": 1477635158.0, "step": 1410 }, { "epoch": 2.2389240506329116, "grad_norm": 0.3691990634713518, "learning_rate": 3.318305400018732e-05, "loss": 0.6231, "num_tokens": 1482878038.0, "step": 1415 }, { "epoch": 2.2468354430379747, "grad_norm": 0.4172329830772149, "learning_rate": 3.306906280624479e-05, "loss": 0.6286, "num_tokens": 1488105187.0, "step": 1420 }, { "epoch": 2.254746835443038, "grad_norm": 0.3279503525329187, "learning_rate": 3.295491913699994e-05, "loss": 0.6233, "num_tokens": 1493346994.0, "step": 1425 }, { "epoch": 2.2626582278481013, "grad_norm": 0.3632060308412657, "learning_rate": 3.284062611759089e-05, "loss": 0.6226, "num_tokens": 1498589874.0, "step": 1430 }, { "epoch": 2.2705696202531644, "grad_norm": 0.46252170164225265, "learning_rate": 3.272618687724481e-05, "loss": 0.6225, "num_tokens": 1503832754.0, "step": 1435 }, { "epoch": 2.278481012658228, "grad_norm": 0.3317711593176921, "learning_rate": 3.261160454919228e-05, "loss": 0.6236, "num_tokens": 1509075634.0, "step": 1440 }, { "epoch": 2.286392405063291, "grad_norm": 0.3470857725220561, "learning_rate": 3.249688227058145e-05, "loss": 0.6273, "num_tokens": 1514318514.0, "step": 1445 }, { "epoch": 2.2943037974683547, "grad_norm": 0.29280649749694565, "learning_rate": 3.238202318239216e-05, "loss": 0.6304, "num_tokens": 1519551983.0, "step": 1450 }, { "epoch": 2.3022151898734178, "grad_norm": 0.37241388988066654, "learning_rate": 3.226703042935e-05, "loss": 0.6184, "num_tokens": 1524794863.0, "step": 1455 }, { "epoch": 2.310126582278481, "grad_norm": 0.33976317790538424, "learning_rate": 3.215190715984013e-05, "loss": 0.622, "num_tokens": 1530031528.0, "step": 1460 }, { "epoch": 2.3180379746835444, "grad_norm": 0.29066911940116874, "learning_rate": 3.2036656525821144e-05, "loss": 0.6231, "num_tokens": 1535274408.0, "step": 1465 }, { "epoch": 2.3259493670886076, "grad_norm": 0.28162811956261474, "learning_rate": 3.1921281682738734e-05, "loss": 0.6182, "num_tokens": 1540517288.0, "step": 1470 }, { "epoch": 2.333860759493671, "grad_norm": 0.3524371457891882, "learning_rate": 3.180578578943933e-05, "loss": 0.6294, "num_tokens": 1545748994.0, "step": 1475 }, { "epoch": 2.3417721518987342, "grad_norm": 0.3468858213633566, "learning_rate": 3.1690172008083554e-05, "loss": 0.619, "num_tokens": 1550991874.0, "step": 1480 }, { "epoch": 2.3496835443037973, "grad_norm": 0.4224520445380759, "learning_rate": 3.157444350405972e-05, "loss": 0.6257, "num_tokens": 1556223622.0, "step": 1485 }, { "epoch": 2.357594936708861, "grad_norm": 0.3754485859450013, "learning_rate": 3.1458603445897136e-05, "loss": 0.6255, "num_tokens": 1561439202.0, "step": 1490 }, { "epoch": 2.365506329113924, "grad_norm": 0.3607328127982716, "learning_rate": 3.134265500517932e-05, "loss": 0.6184, "num_tokens": 1566661187.0, "step": 1495 }, { "epoch": 2.3734177215189876, "grad_norm": 0.3016702669665355, "learning_rate": 3.122660135645721e-05, "loss": 0.6226, "num_tokens": 1571904067.0, "step": 1500 }, { "epoch": 2.3813291139240507, "grad_norm": 0.3397158516715143, "learning_rate": 3.111044567716223e-05, "loss": 0.6237, "num_tokens": 1577146947.0, "step": 1505 }, { "epoch": 2.3892405063291138, "grad_norm": 0.34056042132518993, "learning_rate": 3.0994191147519296e-05, "loss": 0.6205, "num_tokens": 1582389827.0, "step": 1510 }, { "epoch": 2.3971518987341773, "grad_norm": 0.34821134798667835, "learning_rate": 3.0877840950459763e-05, "loss": 0.6217, "num_tokens": 1587632707.0, "step": 1515 }, { "epoch": 2.4050632911392404, "grad_norm": 0.29071305472990666, "learning_rate": 3.076139827153425e-05, "loss": 0.6212, "num_tokens": 1592875587.0, "step": 1520 }, { "epoch": 2.412974683544304, "grad_norm": 0.3213307109034431, "learning_rate": 3.064486629882543e-05, "loss": 0.6261, "num_tokens": 1598118467.0, "step": 1525 }, { "epoch": 2.420886075949367, "grad_norm": 0.317409439920391, "learning_rate": 3.052824822286075e-05, "loss": 0.6192, "num_tokens": 1603361347.0, "step": 1530 }, { "epoch": 2.4287974683544302, "grad_norm": 0.31888869908174583, "learning_rate": 3.041154723652509e-05, "loss": 0.6208, "num_tokens": 1608604227.0, "step": 1535 }, { "epoch": 2.4367088607594938, "grad_norm": 0.2946434998719226, "learning_rate": 3.0294766534973335e-05, "loss": 0.6291, "num_tokens": 1613847107.0, "step": 1540 }, { "epoch": 2.444620253164557, "grad_norm": 0.34443852085875215, "learning_rate": 3.0177909315542862e-05, "loss": 0.6256, "num_tokens": 1619089987.0, "step": 1545 }, { "epoch": 2.4525316455696204, "grad_norm": 0.35106351147093817, "learning_rate": 3.0060978777666054e-05, "loss": 0.6217, "num_tokens": 1624332867.0, "step": 1550 }, { "epoch": 2.4604430379746836, "grad_norm": 0.31920875041898267, "learning_rate": 2.994397812278269e-05, "loss": 0.6229, "num_tokens": 1629575747.0, "step": 1555 }, { "epoch": 2.4683544303797467, "grad_norm": 0.3877394978298516, "learning_rate": 2.9826910554252253e-05, "loss": 0.6272, "num_tokens": 1634818627.0, "step": 1560 }, { "epoch": 2.4762658227848102, "grad_norm": 0.31347418712210595, "learning_rate": 2.9709779277266258e-05, "loss": 0.6276, "num_tokens": 1640061507.0, "step": 1565 }, { "epoch": 2.4841772151898733, "grad_norm": 0.3663502650533136, "learning_rate": 2.959258749876052e-05, "loss": 0.6228, "num_tokens": 1645292739.0, "step": 1570 }, { "epoch": 2.492088607594937, "grad_norm": 0.4369820023540263, "learning_rate": 2.9475338427327298e-05, "loss": 0.622, "num_tokens": 1650520838.0, "step": 1575 }, { "epoch": 2.5, "grad_norm": 0.38676813237014673, "learning_rate": 2.9358035273127483e-05, "loss": 0.6269, "num_tokens": 1655763718.0, "step": 1580 }, { "epoch": 2.507911392405063, "grad_norm": 0.3206842281049778, "learning_rate": 2.924068124780269e-05, "loss": 0.625, "num_tokens": 1661006598.0, "step": 1585 }, { "epoch": 2.5158227848101267, "grad_norm": 0.3173236770874382, "learning_rate": 2.9123279564387355e-05, "loss": 0.6134, "num_tokens": 1666249478.0, "step": 1590 }, { "epoch": 2.5237341772151898, "grad_norm": 0.32043576387698464, "learning_rate": 2.900583343722072e-05, "loss": 0.6215, "num_tokens": 1671492358.0, "step": 1595 }, { "epoch": 2.5316455696202533, "grad_norm": 0.34279988720816984, "learning_rate": 2.8888346081858858e-05, "loss": 0.6176, "num_tokens": 1676735238.0, "step": 1600 }, { "epoch": 2.5395569620253164, "grad_norm": 0.3445942670476214, "learning_rate": 2.8770820714986647e-05, "loss": 0.6256, "num_tokens": 1681960978.0, "step": 1605 }, { "epoch": 2.5474683544303796, "grad_norm": 0.3249239294351048, "learning_rate": 2.8653260554329643e-05, "loss": 0.6229, "num_tokens": 1687203858.0, "step": 1610 }, { "epoch": 2.555379746835443, "grad_norm": 0.4195352553077635, "learning_rate": 2.8535668818566065e-05, "loss": 0.6209, "num_tokens": 1692446738.0, "step": 1615 }, { "epoch": 2.5632911392405062, "grad_norm": 0.3148180466101991, "learning_rate": 2.8418048727238585e-05, "loss": 0.6243, "num_tokens": 1697685701.0, "step": 1620 }, { "epoch": 2.5712025316455698, "grad_norm": 0.3145292653905644, "learning_rate": 2.8300403500666256e-05, "loss": 0.6118, "num_tokens": 1702928581.0, "step": 1625 }, { "epoch": 2.579113924050633, "grad_norm": 0.34661533020564694, "learning_rate": 2.81827363598563e-05, "loss": 0.6209, "num_tokens": 1708171461.0, "step": 1630 }, { "epoch": 2.587025316455696, "grad_norm": 0.32721458461346575, "learning_rate": 2.8065050526415916e-05, "loss": 0.6231, "num_tokens": 1713414341.0, "step": 1635 }, { "epoch": 2.5949367088607596, "grad_norm": 0.3293080910499085, "learning_rate": 2.794734922246408e-05, "loss": 0.6227, "num_tokens": 1718657221.0, "step": 1640 }, { "epoch": 2.6028481012658227, "grad_norm": 0.3006610821739955, "learning_rate": 2.7829635670543375e-05, "loss": 0.6198, "num_tokens": 1723897509.0, "step": 1645 }, { "epoch": 2.6107594936708862, "grad_norm": 0.3221265574540506, "learning_rate": 2.7711913093531682e-05, "loss": 0.6215, "num_tokens": 1729134925.0, "step": 1650 }, { "epoch": 2.6186708860759493, "grad_norm": 0.3629658305187627, "learning_rate": 2.7594184714554e-05, "loss": 0.6193, "num_tokens": 1734377805.0, "step": 1655 }, { "epoch": 2.6265822784810124, "grad_norm": 0.3305255086161955, "learning_rate": 2.7476453756894173e-05, "loss": 0.6163, "num_tokens": 1739620685.0, "step": 1660 }, { "epoch": 2.634493670886076, "grad_norm": 0.30385118713526643, "learning_rate": 2.7358723443906653e-05, "loss": 0.6168, "num_tokens": 1744863565.0, "step": 1665 }, { "epoch": 2.642405063291139, "grad_norm": 0.3729938271215438, "learning_rate": 2.724099699892822e-05, "loss": 0.6247, "num_tokens": 1750106445.0, "step": 1670 }, { "epoch": 2.6503164556962027, "grad_norm": 0.3363024823177763, "learning_rate": 2.712327764518979e-05, "loss": 0.6227, "num_tokens": 1755349325.0, "step": 1675 }, { "epoch": 2.6582278481012658, "grad_norm": 0.3055930853692746, "learning_rate": 2.7005568605728087e-05, "loss": 0.6229, "num_tokens": 1760592205.0, "step": 1680 }, { "epoch": 2.666139240506329, "grad_norm": 0.3298226038292838, "learning_rate": 2.6887873103297474e-05, "loss": 0.6214, "num_tokens": 1765835085.0, "step": 1685 }, { "epoch": 2.6740506329113924, "grad_norm": 0.2867154405096028, "learning_rate": 2.6770194360281653e-05, "loss": 0.6199, "num_tokens": 1771077965.0, "step": 1690 }, { "epoch": 2.6819620253164556, "grad_norm": 0.3195638173711784, "learning_rate": 2.6652535598605504e-05, "loss": 0.6191, "num_tokens": 1776320845.0, "step": 1695 }, { "epoch": 2.689873417721519, "grad_norm": 0.32434124204868575, "learning_rate": 2.6534900039646816e-05, "loss": 0.6181, "num_tokens": 1781532860.0, "step": 1700 }, { "epoch": 2.6977848101265822, "grad_norm": 0.2918992438055131, "learning_rate": 2.6417290904148116e-05, "loss": 0.6197, "num_tokens": 1786775740.0, "step": 1705 }, { "epoch": 2.7056962025316453, "grad_norm": 0.3055847287718812, "learning_rate": 2.629971141212848e-05, "loss": 0.6173, "num_tokens": 1792018620.0, "step": 1710 }, { "epoch": 2.713607594936709, "grad_norm": 0.2985953721458501, "learning_rate": 2.618216478279538e-05, "loss": 0.6165, "num_tokens": 1797261500.0, "step": 1715 }, { "epoch": 2.721518987341772, "grad_norm": 0.3212918053380019, "learning_rate": 2.6064654234456554e-05, "loss": 0.6214, "num_tokens": 1802504380.0, "step": 1720 }, { "epoch": 2.7294303797468356, "grad_norm": 0.2919099745172365, "learning_rate": 2.5947182984431827e-05, "loss": 0.6255, "num_tokens": 1807747260.0, "step": 1725 }, { "epoch": 2.7373417721518987, "grad_norm": 0.2860751140237954, "learning_rate": 2.5829754248965154e-05, "loss": 0.6189, "num_tokens": 1812990140.0, "step": 1730 }, { "epoch": 2.745253164556962, "grad_norm": 0.3019949199367483, "learning_rate": 2.571237124313643e-05, "loss": 0.6124, "num_tokens": 1818233020.0, "step": 1735 }, { "epoch": 2.7531645569620253, "grad_norm": 0.3187461502303022, "learning_rate": 2.5595037180773524e-05, "loss": 0.6196, "num_tokens": 1823475900.0, "step": 1740 }, { "epoch": 2.7610759493670884, "grad_norm": 0.31735001215836417, "learning_rate": 2.5477755274364308e-05, "loss": 0.623, "num_tokens": 1828718780.0, "step": 1745 }, { "epoch": 2.768987341772152, "grad_norm": 0.33876769709652277, "learning_rate": 2.5360528734968647e-05, "loss": 0.611, "num_tokens": 1833961660.0, "step": 1750 }, { "epoch": 2.776898734177215, "grad_norm": 0.3542233915904492, "learning_rate": 2.5243360772130532e-05, "loss": 0.625, "num_tokens": 1839200698.0, "step": 1755 }, { "epoch": 2.7848101265822782, "grad_norm": 0.3630368367495965, "learning_rate": 2.5126254593790182e-05, "loss": 0.6179, "num_tokens": 1844443578.0, "step": 1760 }, { "epoch": 2.7927215189873418, "grad_norm": 0.3826705315790488, "learning_rate": 2.500921340619621e-05, "loss": 0.6205, "num_tokens": 1849680596.0, "step": 1765 }, { "epoch": 2.8006329113924053, "grad_norm": 0.3644864833948713, "learning_rate": 2.489224041381787e-05, "loss": 0.6156, "num_tokens": 1854907408.0, "step": 1770 }, { "epoch": 2.8085443037974684, "grad_norm": 0.3386672412828899, "learning_rate": 2.477533881925727e-05, "loss": 0.6238, "num_tokens": 1860134742.0, "step": 1775 }, { "epoch": 2.8164556962025316, "grad_norm": 0.2921538507032039, "learning_rate": 2.465851182316176e-05, "loss": 0.6202, "num_tokens": 1865377622.0, "step": 1780 }, { "epoch": 2.8243670886075947, "grad_norm": 0.28494983081811437, "learning_rate": 2.4541762624136212e-05, "loss": 0.6176, "num_tokens": 1870620502.0, "step": 1785 }, { "epoch": 2.8322784810126582, "grad_norm": 0.29021546121305103, "learning_rate": 2.4425094418655526e-05, "loss": 0.6177, "num_tokens": 1875863382.0, "step": 1790 }, { "epoch": 2.8401898734177218, "grad_norm": 0.27953269634031763, "learning_rate": 2.430851040097707e-05, "loss": 0.6158, "num_tokens": 1881101588.0, "step": 1795 }, { "epoch": 2.848101265822785, "grad_norm": 0.28367003015841197, "learning_rate": 2.4192013763053222e-05, "loss": 0.6151, "num_tokens": 1886333356.0, "step": 1800 }, { "epoch": 2.856012658227848, "grad_norm": 0.30695047341331355, "learning_rate": 2.4075607694444014e-05, "loss": 0.615, "num_tokens": 1891576236.0, "step": 1805 }, { "epoch": 2.8639240506329116, "grad_norm": 0.3389129881606512, "learning_rate": 2.395929538222978e-05, "loss": 0.6194, "num_tokens": 1896816113.0, "step": 1810 }, { "epoch": 2.8718354430379747, "grad_norm": 0.2993951418009312, "learning_rate": 2.3843080010923873e-05, "loss": 0.6151, "num_tokens": 1902045462.0, "step": 1815 }, { "epoch": 2.879746835443038, "grad_norm": 0.2747247961261455, "learning_rate": 2.3726964762385546e-05, "loss": 0.6152, "num_tokens": 1907275468.0, "step": 1820 }, { "epoch": 2.8876582278481013, "grad_norm": 0.27656654436718103, "learning_rate": 2.361095281573274e-05, "loss": 0.6204, "num_tokens": 1912517035.0, "step": 1825 }, { "epoch": 2.8955696202531644, "grad_norm": 0.26450516299877097, "learning_rate": 2.3495047347255122e-05, "loss": 0.6141, "num_tokens": 1917752379.0, "step": 1830 }, { "epoch": 2.903481012658228, "grad_norm": 0.2680090640698674, "learning_rate": 2.3379251530327112e-05, "loss": 0.6098, "num_tokens": 1922995259.0, "step": 1835 }, { "epoch": 2.911392405063291, "grad_norm": 0.3020470058551071, "learning_rate": 2.326356853532094e-05, "loss": 0.6121, "num_tokens": 1928238139.0, "step": 1840 }, { "epoch": 2.9193037974683547, "grad_norm": 0.35207295628986174, "learning_rate": 2.3148001529519918e-05, "loss": 0.6176, "num_tokens": 1933481019.0, "step": 1845 }, { "epoch": 2.9272151898734178, "grad_norm": 0.31909804688235877, "learning_rate": 2.303255367703169e-05, "loss": 0.6206, "num_tokens": 1938714322.0, "step": 1850 }, { "epoch": 2.935126582278481, "grad_norm": 0.30462726559520964, "learning_rate": 2.2917228138701596e-05, "loss": 0.6154, "num_tokens": 1943957202.0, "step": 1855 }, { "epoch": 2.9430379746835444, "grad_norm": 0.32103060565165314, "learning_rate": 2.2802028072026115e-05, "loss": 0.6089, "num_tokens": 1949191447.0, "step": 1860 }, { "epoch": 2.9509493670886076, "grad_norm": 0.3062127051035162, "learning_rate": 2.268695663106648e-05, "loss": 0.6127, "num_tokens": 1954434327.0, "step": 1865 }, { "epoch": 2.958860759493671, "grad_norm": 0.3239852849229511, "learning_rate": 2.257201696636225e-05, "loss": 0.615, "num_tokens": 1959677207.0, "step": 1870 }, { "epoch": 2.9667721518987342, "grad_norm": 0.2893785533195971, "learning_rate": 2.2457212224845115e-05, "loss": 0.6182, "num_tokens": 1964920087.0, "step": 1875 }, { "epoch": 2.9746835443037973, "grad_norm": 0.3565583813030718, "learning_rate": 2.2342545549752658e-05, "loss": 0.6187, "num_tokens": 1970162967.0, "step": 1880 }, { "epoch": 2.982594936708861, "grad_norm": 0.2901982983640066, "learning_rate": 2.2228020080542394e-05, "loss": 0.6155, "num_tokens": 1975405847.0, "step": 1885 }, { "epoch": 2.990506329113924, "grad_norm": 0.2552747365991554, "learning_rate": 2.211363895280573e-05, "loss": 0.617, "num_tokens": 1980648163.0, "step": 1890 }, { "epoch": 2.9984177215189876, "grad_norm": 0.2925757603065985, "learning_rate": 2.1999405298182176e-05, "loss": 0.6114, "num_tokens": 1985885198.0, "step": 1895 }, { "epoch": 3.0063291139240507, "grad_norm": 0.3169901185514265, "learning_rate": 2.1885322244273537e-05, "loss": 0.6001, "num_tokens": 1991128078.0, "step": 1900 }, { "epoch": 3.0142405063291138, "grad_norm": 0.2733791677119331, "learning_rate": 2.177139291455837e-05, "loss": 0.5909, "num_tokens": 1996370958.0, "step": 1905 }, { "epoch": 3.0221518987341773, "grad_norm": 0.2969117337975924, "learning_rate": 2.1657620428306386e-05, "loss": 0.5979, "num_tokens": 2001613838.0, "step": 1910 }, { "epoch": 3.0300632911392404, "grad_norm": 0.259222996285033, "learning_rate": 2.154400790049309e-05, "loss": 0.5935, "num_tokens": 2006856718.0, "step": 1915 }, { "epoch": 3.037974683544304, "grad_norm": 0.29751625615948685, "learning_rate": 2.1430558441714488e-05, "loss": 0.5892, "num_tokens": 2012099598.0, "step": 1920 }, { "epoch": 3.045886075949367, "grad_norm": 0.2808339111842118, "learning_rate": 2.131727515810193e-05, "loss": 0.5881, "num_tokens": 2017342478.0, "step": 1925 }, { "epoch": 3.0537974683544302, "grad_norm": 0.2665237486778169, "learning_rate": 2.1204161151237044e-05, "loss": 0.5869, "num_tokens": 2022585358.0, "step": 1930 }, { "epoch": 3.0617088607594938, "grad_norm": 0.28147532434254685, "learning_rate": 2.1091219518066846e-05, "loss": 0.5857, "num_tokens": 2027828238.0, "step": 1935 }, { "epoch": 3.069620253164557, "grad_norm": 0.27827204435611375, "learning_rate": 2.0978453350818923e-05, "loss": 0.5892, "num_tokens": 2033071118.0, "step": 1940 }, { "epoch": 3.0775316455696204, "grad_norm": 0.27504620542469616, "learning_rate": 2.0865865736916767e-05, "loss": 0.5927, "num_tokens": 2038313998.0, "step": 1945 }, { "epoch": 3.0854430379746836, "grad_norm": 0.2625667359010651, "learning_rate": 2.0753459758895305e-05, "loss": 0.5878, "num_tokens": 2043556878.0, "step": 1950 }, { "epoch": 3.0933544303797467, "grad_norm": 0.31412193685640266, "learning_rate": 2.0641238494316412e-05, "loss": 0.5954, "num_tokens": 2048795916.0, "step": 1955 }, { "epoch": 3.1012658227848102, "grad_norm": 0.2551742932124429, "learning_rate": 2.0529205015684733e-05, "loss": 0.5896, "num_tokens": 2054038796.0, "step": 1960 }, { "epoch": 3.1091772151898733, "grad_norm": 0.27107418641432374, "learning_rate": 2.0417362390363497e-05, "loss": 0.5949, "num_tokens": 2059281676.0, "step": 1965 }, { "epoch": 3.117088607594937, "grad_norm": 0.2910292647945281, "learning_rate": 2.030571368049058e-05, "loss": 0.593, "num_tokens": 2064524556.0, "step": 1970 }, { "epoch": 3.125, "grad_norm": 0.2702497087199772, "learning_rate": 2.0194261942894628e-05, "loss": 0.5913, "num_tokens": 2069763519.0, "step": 1975 }, { "epoch": 3.132911392405063, "grad_norm": 0.27879096153789196, "learning_rate": 2.0083010229011405e-05, "loss": 0.5871, "num_tokens": 2075006399.0, "step": 1980 }, { "epoch": 3.1408227848101267, "grad_norm": 0.2760119851379016, "learning_rate": 1.9971961584800218e-05, "loss": 0.5889, "num_tokens": 2080249279.0, "step": 1985 }, { "epoch": 3.1487341772151898, "grad_norm": 0.2980189742569214, "learning_rate": 1.986111905066055e-05, "loss": 0.5936, "num_tokens": 2085492159.0, "step": 1990 }, { "epoch": 3.1566455696202533, "grad_norm": 0.27221033878781475, "learning_rate": 1.9750485661348762e-05, "loss": 0.5925, "num_tokens": 2090735039.0, "step": 1995 }, { "epoch": 3.1645569620253164, "grad_norm": 0.282708489024635, "learning_rate": 1.964006444589509e-05, "loss": 0.5906, "num_tokens": 2095977919.0, "step": 2000 }, { "epoch": 3.1724683544303796, "grad_norm": 0.2689336862752935, "learning_rate": 1.952985842752062e-05, "loss": 0.5908, "num_tokens": 2101206695.0, "step": 2005 }, { "epoch": 3.180379746835443, "grad_norm": 0.26260733949618775, "learning_rate": 1.941987062355458e-05, "loss": 0.5907, "num_tokens": 2106449575.0, "step": 2010 }, { "epoch": 3.1882911392405062, "grad_norm": 0.2829230017612488, "learning_rate": 1.9310104045351724e-05, "loss": 0.5869, "num_tokens": 2111692455.0, "step": 2015 }, { "epoch": 3.1962025316455698, "grad_norm": 0.2496685091023237, "learning_rate": 1.9200561698209828e-05, "loss": 0.5895, "num_tokens": 2116935335.0, "step": 2020 }, { "epoch": 3.204113924050633, "grad_norm": 0.30588754751321845, "learning_rate": 1.9091246581287487e-05, "loss": 0.5911, "num_tokens": 2122178215.0, "step": 2025 }, { "epoch": 3.212025316455696, "grad_norm": 0.3013756834853421, "learning_rate": 1.8982161687521938e-05, "loss": 0.5926, "num_tokens": 2127416178.0, "step": 2030 }, { "epoch": 3.2199367088607596, "grad_norm": 0.3260497257360817, "learning_rate": 1.8873310003547152e-05, "loss": 0.5957, "num_tokens": 2132659058.0, "step": 2035 }, { "epoch": 3.2278481012658227, "grad_norm": 18.37765842944507, "learning_rate": 1.8764694509612057e-05, "loss": 0.6045, "num_tokens": 2137900029.0, "step": 2040 }, { "epoch": 3.2357594936708862, "grad_norm": 0.2844360272669497, "learning_rate": 1.865631817949893e-05, "loss": 0.5961, "num_tokens": 2143139652.0, "step": 2045 }, { "epoch": 3.2436708860759493, "grad_norm": 0.25708276328077995, "learning_rate": 1.8548183980441985e-05, "loss": 0.5988, "num_tokens": 2148382532.0, "step": 2050 }, { "epoch": 3.2515822784810124, "grad_norm": 0.27434743791141486, "learning_rate": 1.8440294873046126e-05, "loss": 0.5904, "num_tokens": 2153625412.0, "step": 2055 }, { "epoch": 3.259493670886076, "grad_norm": 0.25975458422652054, "learning_rate": 1.8332653811205906e-05, "loss": 0.5928, "num_tokens": 2158854105.0, "step": 2060 }, { "epoch": 3.267405063291139, "grad_norm": 0.26925828595298595, "learning_rate": 1.8225263742024638e-05, "loss": 0.5833, "num_tokens": 2164096985.0, "step": 2065 }, { "epoch": 3.2753164556962027, "grad_norm": 0.282185762273433, "learning_rate": 1.8118127605733697e-05, "loss": 0.5981, "num_tokens": 2169337273.0, "step": 2070 }, { "epoch": 3.2832278481012658, "grad_norm": 0.25816071189825335, "learning_rate": 1.801124833561208e-05, "loss": 0.5961, "num_tokens": 2174564736.0, "step": 2075 }, { "epoch": 3.291139240506329, "grad_norm": 0.24619197376885332, "learning_rate": 1.7904628857905982e-05, "loss": 0.6011, "num_tokens": 2179797828.0, "step": 2080 }, { "epoch": 3.2990506329113924, "grad_norm": 0.24421414534208322, "learning_rate": 1.7798272091748785e-05, "loss": 0.5865, "num_tokens": 2185040708.0, "step": 2085 }, { "epoch": 3.3069620253164556, "grad_norm": 0.2524526157340073, "learning_rate": 1.7692180949081093e-05, "loss": 0.5959, "num_tokens": 2190283588.0, "step": 2090 }, { "epoch": 3.314873417721519, "grad_norm": 0.27130450376385096, "learning_rate": 1.758635833457099e-05, "loss": 0.5848, "num_tokens": 2195513594.0, "step": 2095 }, { "epoch": 3.3227848101265822, "grad_norm": 0.26771998653275103, "learning_rate": 1.748080714553455e-05, "loss": 0.5879, "num_tokens": 2200745342.0, "step": 2100 }, { "epoch": 3.3306962025316453, "grad_norm": 0.26138164229192307, "learning_rate": 1.7375530271856487e-05, "loss": 0.5922, "num_tokens": 2205988222.0, "step": 2105 }, { "epoch": 3.338607594936709, "grad_norm": 0.2715626154228517, "learning_rate": 1.727053059591101e-05, "loss": 0.5971, "num_tokens": 2211231102.0, "step": 2110 }, { "epoch": 3.346518987341772, "grad_norm": 0.2455227895466599, "learning_rate": 1.7165810992482994e-05, "loss": 0.5887, "num_tokens": 2216460869.0, "step": 2115 }, { "epoch": 3.3544303797468356, "grad_norm": 0.2777380961416323, "learning_rate": 1.706137432868917e-05, "loss": 0.5943, "num_tokens": 2221703749.0, "step": 2120 }, { "epoch": 3.3623417721518987, "grad_norm": 0.25563580174980555, "learning_rate": 1.6957223463899658e-05, "loss": 0.5897, "num_tokens": 2226946629.0, "step": 2125 }, { "epoch": 3.370253164556962, "grad_norm": 0.25164624412979186, "learning_rate": 1.6853361249659722e-05, "loss": 0.5929, "num_tokens": 2232189509.0, "step": 2130 }, { "epoch": 3.3781645569620253, "grad_norm": 0.2806108194410442, "learning_rate": 1.674979052961166e-05, "loss": 0.5975, "num_tokens": 2237432389.0, "step": 2135 }, { "epoch": 3.3860759493670884, "grad_norm": 0.2605837907585764, "learning_rate": 1.664651413941693e-05, "loss": 0.5907, "num_tokens": 2242665858.0, "step": 2140 }, { "epoch": 3.393987341772152, "grad_norm": 0.2598679967531081, "learning_rate": 1.6543534906678552e-05, "loss": 0.592, "num_tokens": 2247908738.0, "step": 2145 }, { "epoch": 3.401898734177215, "grad_norm": 0.25797499837806687, "learning_rate": 1.6440855650863695e-05, "loss": 0.5894, "num_tokens": 2253151618.0, "step": 2150 }, { "epoch": 3.4098101265822787, "grad_norm": 0.2427493258825347, "learning_rate": 1.6338479183226436e-05, "loss": 0.5924, "num_tokens": 2258394498.0, "step": 2155 }, { "epoch": 3.4177215189873418, "grad_norm": 0.2359671364270129, "learning_rate": 1.6236408306730828e-05, "loss": 0.5916, "num_tokens": 2263637378.0, "step": 2160 }, { "epoch": 3.425632911392405, "grad_norm": 0.23492984902262662, "learning_rate": 1.6134645815974153e-05, "loss": 0.588, "num_tokens": 2268880258.0, "step": 2165 }, { "epoch": 3.4335443037974684, "grad_norm": 0.2684129651198138, "learning_rate": 1.6033194497110387e-05, "loss": 0.5863, "num_tokens": 2274107070.0, "step": 2170 }, { "epoch": 3.4414556962025316, "grad_norm": 0.2551028754173729, "learning_rate": 1.5932057127773956e-05, "loss": 0.5912, "num_tokens": 2279335709.0, "step": 2175 }, { "epoch": 3.449367088607595, "grad_norm": 0.27552281820576113, "learning_rate": 1.5831236477003657e-05, "loss": 0.5901, "num_tokens": 2284578589.0, "step": 2180 }, { "epoch": 3.4572784810126582, "grad_norm": 0.245355717229245, "learning_rate": 1.573073530516683e-05, "loss": 0.5984, "num_tokens": 2289805799.0, "step": 2185 }, { "epoch": 3.4651898734177213, "grad_norm": 0.2711024543904534, "learning_rate": 1.563055636388385e-05, "loss": 0.5901, "num_tokens": 2295041668.0, "step": 2190 }, { "epoch": 3.473101265822785, "grad_norm": 0.25534456241862863, "learning_rate": 1.5530702395952726e-05, "loss": 0.5885, "num_tokens": 2300284548.0, "step": 2195 }, { "epoch": 3.481012658227848, "grad_norm": 0.28472473513191726, "learning_rate": 1.5431176135274004e-05, "loss": 0.5875, "num_tokens": 2305527428.0, "step": 2200 }, { "epoch": 3.4889240506329116, "grad_norm": 0.26401107803555085, "learning_rate": 1.533198030677596e-05, "loss": 0.587, "num_tokens": 2310770308.0, "step": 2205 }, { "epoch": 3.4968354430379747, "grad_norm": 0.23581382627455794, "learning_rate": 1.5233117626339988e-05, "loss": 0.5829, "num_tokens": 2316013188.0, "step": 2210 }, { "epoch": 3.504746835443038, "grad_norm": 0.2919647262861195, "learning_rate": 1.513459080072617e-05, "loss": 0.593, "num_tokens": 2321256068.0, "step": 2215 }, { "epoch": 3.5126582278481013, "grad_norm": 0.27795974121603223, "learning_rate": 1.503640252749928e-05, "loss": 0.5948, "num_tokens": 2326498948.0, "step": 2220 }, { "epoch": 3.5205696202531644, "grad_norm": 0.2513485774758495, "learning_rate": 1.4938555494954838e-05, "loss": 0.5928, "num_tokens": 2331740515.0, "step": 2225 }, { "epoch": 3.528481012658228, "grad_norm": 0.27468633702239875, "learning_rate": 1.484105238204555e-05, "loss": 0.592, "num_tokens": 2336983395.0, "step": 2230 }, { "epoch": 3.536392405063291, "grad_norm": 0.24973282593959645, "learning_rate": 1.4743895858307951e-05, "loss": 0.5858, "num_tokens": 2342226275.0, "step": 2235 }, { "epoch": 3.5443037974683547, "grad_norm": 0.236224838159416, "learning_rate": 1.4647088583789325e-05, "loss": 0.5943, "num_tokens": 2347461462.0, "step": 2240 }, { "epoch": 3.5522151898734178, "grad_norm": 0.25744574090772937, "learning_rate": 1.4550633208974832e-05, "loss": 0.5888, "num_tokens": 2352704342.0, "step": 2245 }, { "epoch": 3.560126582278481, "grad_norm": 0.24904139861794714, "learning_rate": 1.445453237471501e-05, "loss": 0.5937, "num_tokens": 2357947222.0, "step": 2250 }, { "epoch": 3.5680379746835444, "grad_norm": 0.26268941423266656, "learning_rate": 1.435878871215342e-05, "loss": 0.5903, "num_tokens": 2363190102.0, "step": 2255 }, { "epoch": 3.5759493670886076, "grad_norm": 0.24790895208789265, "learning_rate": 1.4263404842654638e-05, "loss": 0.5972, "num_tokens": 2368432982.0, "step": 2260 }, { "epoch": 3.583860759493671, "grad_norm": 0.2613242366100201, "learning_rate": 1.416838337773243e-05, "loss": 0.595, "num_tokens": 2373661081.0, "step": 2265 }, { "epoch": 3.5917721518987342, "grad_norm": 0.2922830460241977, "learning_rate": 1.4073726918978347e-05, "loss": 0.594, "num_tokens": 2378903961.0, "step": 2270 }, { "epoch": 3.5996835443037973, "grad_norm": 0.27101780547058585, "learning_rate": 1.3979438057990399e-05, "loss": 0.592, "num_tokens": 2384146841.0, "step": 2275 }, { "epoch": 3.607594936708861, "grad_norm": 0.26512432923141993, "learning_rate": 1.3885519376302159e-05, "loss": 0.5878, "num_tokens": 2389389721.0, "step": 2280 }, { "epoch": 3.615506329113924, "grad_norm": 0.23952238274180007, "learning_rate": 1.3791973445312058e-05, "loss": 0.589, "num_tokens": 2394632601.0, "step": 2285 }, { "epoch": 3.6234177215189876, "grad_norm": 0.25423971405929885, "learning_rate": 1.3698802826213008e-05, "loss": 0.5908, "num_tokens": 2399860042.0, "step": 2290 }, { "epoch": 3.6313291139240507, "grad_norm": 0.24740192193842994, "learning_rate": 1.3606010069922232e-05, "loss": 0.5882, "num_tokens": 2405102922.0, "step": 2295 }, { "epoch": 3.6392405063291138, "grad_norm": 0.5035309107226975, "learning_rate": 1.351359771701147e-05, "loss": 0.5966, "num_tokens": 2410345802.0, "step": 2300 }, { "epoch": 3.6471518987341773, "grad_norm": 0.231935472046041, "learning_rate": 1.3421568297637404e-05, "loss": 0.5873, "num_tokens": 2415588682.0, "step": 2305 }, { "epoch": 3.6550632911392404, "grad_norm": 0.2584802973832586, "learning_rate": 1.3329924331472376e-05, "loss": 0.5925, "num_tokens": 2420806919.0, "step": 2310 }, { "epoch": 3.662974683544304, "grad_norm": 0.31314246833170745, "learning_rate": 1.3238668327635422e-05, "loss": 0.5922, "num_tokens": 2426049799.0, "step": 2315 }, { "epoch": 3.670886075949367, "grad_norm": 0.2663616787106021, "learning_rate": 1.3147802784623536e-05, "loss": 0.5962, "num_tokens": 2431292679.0, "step": 2320 }, { "epoch": 3.6787974683544302, "grad_norm": 0.23711780957648615, "learning_rate": 1.3057330190243317e-05, "loss": 0.5894, "num_tokens": 2436523336.0, "step": 2325 }, { "epoch": 3.6867088607594938, "grad_norm": 0.2549540114874716, "learning_rate": 1.2967253021542813e-05, "loss": 0.5957, "num_tokens": 2441766216.0, "step": 2330 }, { "epoch": 3.694620253164557, "grad_norm": 0.24052219589182647, "learning_rate": 1.2877573744743737e-05, "loss": 0.5935, "num_tokens": 2447009096.0, "step": 2335 }, { "epoch": 3.7025316455696204, "grad_norm": 0.23451782912483585, "learning_rate": 1.278829481517388e-05, "loss": 0.5907, "num_tokens": 2452248973.0, "step": 2340 }, { "epoch": 3.7104430379746836, "grad_norm": 0.2572531172651155, "learning_rate": 1.2699418677199992e-05, "loss": 0.5871, "num_tokens": 2457491853.0, "step": 2345 }, { "epoch": 3.7183544303797467, "grad_norm": 0.22205564327904748, "learning_rate": 1.2610947764160743e-05, "loss": 0.5911, "num_tokens": 2462732233.0, "step": 2350 }, { "epoch": 3.7262658227848102, "grad_norm": 0.25061314597163403, "learning_rate": 1.2522884498300185e-05, "loss": 0.589, "num_tokens": 2467975113.0, "step": 2355 }, { "epoch": 3.7341772151898733, "grad_norm": 0.24341650567160555, "learning_rate": 1.2435231290701398e-05, "loss": 0.5839, "num_tokens": 2473217993.0, "step": 2360 }, { "epoch": 3.742088607594937, "grad_norm": 0.23957280785298024, "learning_rate": 1.234799054122048e-05, "loss": 0.5856, "num_tokens": 2478460873.0, "step": 2365 }, { "epoch": 3.75, "grad_norm": 0.23496985031896744, "learning_rate": 1.2261164638420832e-05, "loss": 0.5943, "num_tokens": 2483695118.0, "step": 2370 }, { "epoch": 3.757911392405063, "grad_norm": 0.2562958360759268, "learning_rate": 1.2174755959507785e-05, "loss": 0.5905, "num_tokens": 2488915999.0, "step": 2375 }, { "epoch": 3.7658227848101267, "grad_norm": 0.23818568022151865, "learning_rate": 1.2088766870263504e-05, "loss": 0.5833, "num_tokens": 2494158879.0, "step": 2380 }, { "epoch": 3.7737341772151898, "grad_norm": 0.271749751046968, "learning_rate": 1.2003199724982206e-05, "loss": 0.5898, "num_tokens": 2499392356.0, "step": 2385 }, { "epoch": 3.7816455696202533, "grad_norm": 0.2768085008184621, "learning_rate": 1.1918056866405717e-05, "loss": 0.5964, "num_tokens": 2504635236.0, "step": 2390 }, { "epoch": 3.7895569620253164, "grad_norm": 0.24209821208112645, "learning_rate": 1.1833340625659325e-05, "loss": 0.5907, "num_tokens": 2509878116.0, "step": 2395 }, { "epoch": 3.7974683544303796, "grad_norm": 0.2587091544484298, "learning_rate": 1.1749053322187933e-05, "loss": 0.5887, "num_tokens": 2515098012.0, "step": 2400 }, { "epoch": 3.805379746835443, "grad_norm": 0.241562313900529, "learning_rate": 1.1665197263692593e-05, "loss": 0.5914, "num_tokens": 2520340892.0, "step": 2405 }, { "epoch": 3.8132911392405062, "grad_norm": 0.21894193923034058, "learning_rate": 1.1581774746067315e-05, "loss": 0.5861, "num_tokens": 2525582367.0, "step": 2410 }, { "epoch": 3.8212025316455698, "grad_norm": 0.241635293617956, "learning_rate": 1.149878805333616e-05, "loss": 0.5875, "num_tokens": 2530819032.0, "step": 2415 }, { "epoch": 3.829113924050633, "grad_norm": 0.2377026625673294, "learning_rate": 1.1416239457590797e-05, "loss": 0.5895, "num_tokens": 2536061912.0, "step": 2420 }, { "epoch": 3.837025316455696, "grad_norm": 0.2500766928834093, "learning_rate": 1.1334131218928215e-05, "loss": 0.5955, "num_tokens": 2541294406.0, "step": 2425 }, { "epoch": 3.8449367088607596, "grad_norm": 0.24157809368048716, "learning_rate": 1.1252465585388861e-05, "loss": 0.593, "num_tokens": 2546537286.0, "step": 2430 }, { "epoch": 3.8528481012658227, "grad_norm": 0.23659644820965803, "learning_rate": 1.1171244792895122e-05, "loss": 0.5875, "num_tokens": 2551780166.0, "step": 2435 }, { "epoch": 3.8607594936708862, "grad_norm": 0.22665381320819925, "learning_rate": 1.1090471065190087e-05, "loss": 0.5895, "num_tokens": 2557023046.0, "step": 2440 }, { "epoch": 3.8686708860759493, "grad_norm": 0.2531032288030383, "learning_rate": 1.1010146613776646e-05, "loss": 0.5911, "num_tokens": 2562249118.0, "step": 2445 }, { "epoch": 3.8765822784810124, "grad_norm": 0.2375978750430171, "learning_rate": 1.0930273637856969e-05, "loss": 0.5913, "num_tokens": 2567491998.0, "step": 2450 }, { "epoch": 3.884493670886076, "grad_norm": 0.24286465990334033, "learning_rate": 1.0850854324272289e-05, "loss": 0.5877, "num_tokens": 2572734878.0, "step": 2455 }, { "epoch": 3.892405063291139, "grad_norm": 0.2575357182094295, "learning_rate": 1.0771890847443022e-05, "loss": 0.596, "num_tokens": 2577975987.0, "step": 2460 }, { "epoch": 3.9003164556962027, "grad_norm": 0.27638888789609667, "learning_rate": 1.069338536930923e-05, "loss": 0.5865, "num_tokens": 2583218867.0, "step": 2465 }, { "epoch": 3.9082278481012658, "grad_norm": 0.2344539051269508, "learning_rate": 1.0615340039271443e-05, "loss": 0.5892, "num_tokens": 2588461747.0, "step": 2470 }, { "epoch": 3.916139240506329, "grad_norm": 0.27080111151637076, "learning_rate": 1.0537756994131783e-05, "loss": 0.5895, "num_tokens": 2593704627.0, "step": 2475 }, { "epoch": 3.9240506329113924, "grad_norm": 0.2249920243150278, "learning_rate": 1.0460638358035496e-05, "loss": 0.5924, "num_tokens": 2598947507.0, "step": 2480 }, { "epoch": 3.9319620253164556, "grad_norm": 0.28780985267539305, "learning_rate": 1.0383986242412785e-05, "loss": 0.5871, "num_tokens": 2604190387.0, "step": 2485 }, { "epoch": 3.939873417721519, "grad_norm": 0.23427765653838825, "learning_rate": 1.0307802745920964e-05, "loss": 0.5839, "num_tokens": 2609433267.0, "step": 2490 }, { "epoch": 3.9477848101265822, "grad_norm": 0.2511572467034619, "learning_rate": 1.0232089954387068e-05, "loss": 0.591, "num_tokens": 2614676147.0, "step": 2495 }, { "epoch": 3.9556962025316453, "grad_norm": 0.24133671906166224, "learning_rate": 1.0156849940750688e-05, "loss": 0.5873, "num_tokens": 2619919027.0, "step": 2500 }, { "epoch": 3.963607594936709, "grad_norm": 0.24130871799729056, "learning_rate": 1.0082084765007226e-05, "loss": 0.5919, "num_tokens": 2625150643.0, "step": 2505 }, { "epoch": 3.971518987341772, "grad_norm": 0.23545792709918428, "learning_rate": 1.0007796474151514e-05, "loss": 0.5972, "num_tokens": 2630393523.0, "step": 2510 }, { "epoch": 3.9794303797468356, "grad_norm": 0.2506437876450052, "learning_rate": 9.933987102121764e-06, "loss": 0.5814, "num_tokens": 2635636403.0, "step": 2515 }, { "epoch": 3.9873417721518987, "grad_norm": 0.2284653998147006, "learning_rate": 9.860658669743861e-06, "loss": 0.5897, "num_tokens": 2640879283.0, "step": 2520 }, { "epoch": 3.995253164556962, "grad_norm": 0.22307394105006934, "learning_rate": 9.787813184676056e-06, "loss": 0.5861, "num_tokens": 2646099304.0, "step": 2525 }, { "epoch": 4.003164556962025, "grad_norm": 0.28089029948590827, "learning_rate": 9.715452641353992e-06, "loss": 0.5826, "num_tokens": 2651342184.0, "step": 2530 }, { "epoch": 4.011075949367089, "grad_norm": 0.23002963489023126, "learning_rate": 9.643579020936106e-06, "loss": 0.5662, "num_tokens": 2656585064.0, "step": 2535 }, { "epoch": 4.018987341772152, "grad_norm": 0.24076823558688779, "learning_rate": 9.572194291249362e-06, "loss": 0.5767, "num_tokens": 2661827944.0, "step": 2540 }, { "epoch": 4.026898734177215, "grad_norm": 0.22483920300780846, "learning_rate": 9.501300406735406e-06, "loss": 0.5743, "num_tokens": 2667070260.0, "step": 2545 }, { "epoch": 4.034810126582278, "grad_norm": 0.2425268841198098, "learning_rate": 9.430899308397024e-06, "loss": 0.5691, "num_tokens": 2672313140.0, "step": 2550 }, { "epoch": 4.042721518987341, "grad_norm": 0.22463870074019276, "learning_rate": 9.360992923745032e-06, "loss": 0.5727, "num_tokens": 2677556020.0, "step": 2555 }, { "epoch": 4.050632911392405, "grad_norm": 0.2257862359365115, "learning_rate": 9.29158316674548e-06, "loss": 0.5771, "num_tokens": 2682798900.0, "step": 2560 }, { "epoch": 4.0585443037974684, "grad_norm": 0.21855947078972188, "learning_rate": 9.222671937767247e-06, "loss": 0.5693, "num_tokens": 2688041780.0, "step": 2565 }, { "epoch": 4.0664556962025316, "grad_norm": 0.22655998729700472, "learning_rate": 9.154261123530024e-06, "loss": 0.5688, "num_tokens": 2693280743.0, "step": 2570 }, { "epoch": 4.074367088607595, "grad_norm": 0.2417803758293443, "learning_rate": 9.086352597052674e-06, "loss": 0.5699, "num_tokens": 2698523623.0, "step": 2575 }, { "epoch": 4.082278481012658, "grad_norm": 0.23491529993846316, "learning_rate": 9.018948217601894e-06, "loss": 0.5776, "num_tokens": 2703760050.0, "step": 2580 }, { "epoch": 4.090189873417722, "grad_norm": 0.21683310271118783, "learning_rate": 8.952049830641368e-06, "loss": 0.569, "num_tokens": 2709001853.0, "step": 2585 }, { "epoch": 4.098101265822785, "grad_norm": 0.22683758792329442, "learning_rate": 8.885659267781218e-06, "loss": 0.5723, "num_tokens": 2714244733.0, "step": 2590 }, { "epoch": 4.106012658227848, "grad_norm": 0.23440417433660363, "learning_rate": 8.819778346727839e-06, "loss": 0.571, "num_tokens": 2719487613.0, "step": 2595 }, { "epoch": 4.113924050632911, "grad_norm": 0.2192458553426772, "learning_rate": 8.754408871234168e-06, "loss": 0.5733, "num_tokens": 2724715076.0, "step": 2600 }, { "epoch": 4.121835443037975, "grad_norm": 0.2530876360164053, "learning_rate": 8.689552631050274e-06, "loss": 0.5714, "num_tokens": 2729957956.0, "step": 2605 }, { "epoch": 4.129746835443038, "grad_norm": 0.21492163566185005, "learning_rate": 8.625211401874362e-06, "loss": 0.5694, "num_tokens": 2735200836.0, "step": 2610 }, { "epoch": 4.137658227848101, "grad_norm": 0.22363359389581403, "learning_rate": 8.561386945304155e-06, "loss": 0.5727, "num_tokens": 2740443716.0, "step": 2615 }, { "epoch": 4.1455696202531644, "grad_norm": 0.2663268042449171, "learning_rate": 8.498081008788677e-06, "loss": 0.5689, "num_tokens": 2745686596.0, "step": 2620 }, { "epoch": 4.1534810126582276, "grad_norm": 0.2379773928525528, "learning_rate": 8.435295325580373e-06, "loss": 0.5693, "num_tokens": 2750929476.0, "step": 2625 }, { "epoch": 4.1613924050632916, "grad_norm": 0.2227866522568964, "learning_rate": 8.373031614687699e-06, "loss": 0.5698, "num_tokens": 2756172356.0, "step": 2630 }, { "epoch": 4.169303797468355, "grad_norm": 0.23722314775734296, "learning_rate": 8.311291580828034e-06, "loss": 0.5723, "num_tokens": 2761415236.0, "step": 2635 }, { "epoch": 4.177215189873418, "grad_norm": 0.2122410135791533, "learning_rate": 8.250076914381017e-06, "loss": 0.5661, "num_tokens": 2766658116.0, "step": 2640 }, { "epoch": 4.185126582278481, "grad_norm": 0.21872455435466664, "learning_rate": 8.189389291342229e-06, "loss": 0.5668, "num_tokens": 2771900996.0, "step": 2645 }, { "epoch": 4.193037974683544, "grad_norm": 0.22736670887055646, "learning_rate": 8.12923037327738e-06, "loss": 0.5723, "num_tokens": 2777143876.0, "step": 2650 }, { "epoch": 4.200949367088608, "grad_norm": 0.24119729706862408, "learning_rate": 8.06960180727674e-06, "loss": 0.5689, "num_tokens": 2782386756.0, "step": 2655 }, { "epoch": 4.208860759493671, "grad_norm": 0.22351050734609607, "learning_rate": 8.010505225910083e-06, "loss": 0.5677, "num_tokens": 2787629636.0, "step": 2660 }, { "epoch": 4.216772151898734, "grad_norm": 0.265494072724875, "learning_rate": 7.951942247181992e-06, "loss": 0.5718, "num_tokens": 2792872516.0, "step": 2665 }, { "epoch": 4.224683544303797, "grad_norm": 0.21820442695718212, "learning_rate": 7.89391447448755e-06, "loss": 0.5682, "num_tokens": 2798109181.0, "step": 2670 }, { "epoch": 4.2325949367088604, "grad_norm": 0.2420612432104863, "learning_rate": 7.836423496568418e-06, "loss": 0.5753, "num_tokens": 2803337280.0, "step": 2675 }, { "epoch": 4.2405063291139244, "grad_norm": 0.23517468870764166, "learning_rate": 7.779470887469387e-06, "loss": 0.5716, "num_tokens": 2808580160.0, "step": 2680 }, { "epoch": 4.2484177215189876, "grad_norm": 0.2267612123573776, "learning_rate": 7.723058206495242e-06, "loss": 0.5746, "num_tokens": 2813823040.0, "step": 2685 }, { "epoch": 4.256329113924051, "grad_norm": 0.22002454461146628, "learning_rate": 7.667186998168082e-06, "loss": 0.5684, "num_tokens": 2819065920.0, "step": 2690 }, { "epoch": 4.264240506329114, "grad_norm": 0.22912652065839756, "learning_rate": 7.611858792185038e-06, "loss": 0.5751, "num_tokens": 2824308800.0, "step": 2695 }, { "epoch": 4.272151898734177, "grad_norm": 0.22408651950567107, "learning_rate": 7.557075103376383e-06, "loss": 0.5724, "num_tokens": 2829542456.0, "step": 2700 }, { "epoch": 4.280063291139241, "grad_norm": 0.22342018834903565, "learning_rate": 7.502837431664059e-06, "loss": 0.5731, "num_tokens": 2834785336.0, "step": 2705 }, { "epoch": 4.287974683544304, "grad_norm": 0.22864174140723967, "learning_rate": 7.449147262020616e-06, "loss": 0.5679, "num_tokens": 2840015342.0, "step": 2710 }, { "epoch": 4.295886075949367, "grad_norm": 0.22865644771070281, "learning_rate": 7.396006064428554e-06, "loss": 0.5627, "num_tokens": 2845248811.0, "step": 2715 }, { "epoch": 4.30379746835443, "grad_norm": 0.2674648148165605, "learning_rate": 7.34341529384006e-06, "loss": 0.5765, "num_tokens": 2850491691.0, "step": 2720 }, { "epoch": 4.311708860759493, "grad_norm": 0.22345156469242552, "learning_rate": 7.291376390137214e-06, "loss": 0.5805, "num_tokens": 2855734571.0, "step": 2725 }, { "epoch": 4.319620253164557, "grad_norm": 0.2122765810320669, "learning_rate": 7.239890778092509e-06, "loss": 0.5712, "num_tokens": 2860977451.0, "step": 2730 }, { "epoch": 4.3275316455696204, "grad_norm": 0.2477019061431497, "learning_rate": 7.188959867329893e-06, "loss": 0.5814, "num_tokens": 2866220331.0, "step": 2735 }, { "epoch": 4.3354430379746836, "grad_norm": 0.21235713827113287, "learning_rate": 7.138585052286142e-06, "loss": 0.5732, "num_tokens": 2871451563.0, "step": 2740 }, { "epoch": 4.343354430379747, "grad_norm": 0.2146641271693942, "learning_rate": 7.088767712172711e-06, "loss": 0.5717, "num_tokens": 2876694443.0, "step": 2745 }, { "epoch": 4.35126582278481, "grad_norm": 0.23309761275535962, "learning_rate": 7.039509210937932e-06, "loss": 0.5732, "num_tokens": 2881937323.0, "step": 2750 }, { "epoch": 4.359177215189874, "grad_norm": 0.2234172047939556, "learning_rate": 6.990810897229709e-06, "loss": 0.57, "num_tokens": 2887179519.0, "step": 2755 }, { "epoch": 4.367088607594937, "grad_norm": 0.22237797838934822, "learning_rate": 6.942674104358582e-06, "loss": 0.5738, "num_tokens": 2892422399.0, "step": 2760 }, { "epoch": 4.375, "grad_norm": 0.21593708745753062, "learning_rate": 6.8951001502612065e-06, "loss": 0.5658, "num_tokens": 2897636988.0, "step": 2765 }, { "epoch": 4.382911392405063, "grad_norm": 0.22987570961197548, "learning_rate": 6.848090337464294e-06, "loss": 0.5682, "num_tokens": 2902879868.0, "step": 2770 }, { "epoch": 4.390822784810126, "grad_norm": 0.22029484227034318, "learning_rate": 6.8016459530489335e-06, "loss": 0.5746, "num_tokens": 2908122748.0, "step": 2775 }, { "epoch": 4.39873417721519, "grad_norm": 0.2171355743200665, "learning_rate": 6.755768268615354e-06, "loss": 0.567, "num_tokens": 2913363036.0, "step": 2780 }, { "epoch": 4.406645569620253, "grad_norm": 0.2213627866085991, "learning_rate": 6.710458540248109e-06, "loss": 0.5725, "num_tokens": 2918605916.0, "step": 2785 }, { "epoch": 4.4145569620253164, "grad_norm": 0.21863302278329563, "learning_rate": 6.66571800848171e-06, "loss": 0.5733, "num_tokens": 2923835265.0, "step": 2790 }, { "epoch": 4.4224683544303796, "grad_norm": 0.2114144591578047, "learning_rate": 6.621547898266615e-06, "loss": 0.5731, "num_tokens": 2929076236.0, "step": 2795 }, { "epoch": 4.430379746835443, "grad_norm": 0.22099052084655507, "learning_rate": 6.577949418935732e-06, "loss": 0.5741, "num_tokens": 2934319116.0, "step": 2800 }, { "epoch": 4.438291139240507, "grad_norm": 0.23002587290313806, "learning_rate": 6.534923764171305e-06, "loss": 0.572, "num_tokens": 2939556532.0, "step": 2805 }, { "epoch": 4.44620253164557, "grad_norm": 0.2525898694118869, "learning_rate": 6.492472111972193e-06, "loss": 0.5758, "num_tokens": 2944799412.0, "step": 2810 }, { "epoch": 4.454113924050633, "grad_norm": 0.22052197863760098, "learning_rate": 6.450595624621672e-06, "loss": 0.5687, "num_tokens": 2950042292.0, "step": 2815 }, { "epoch": 4.462025316455696, "grad_norm": 0.22012961127429426, "learning_rate": 6.409295448655572e-06, "loss": 0.5701, "num_tokens": 2955275769.0, "step": 2820 }, { "epoch": 4.469936708860759, "grad_norm": 0.22397593983188469, "learning_rate": 6.3685727148309014e-06, "loss": 0.5708, "num_tokens": 2960518649.0, "step": 2825 }, { "epoch": 4.477848101265823, "grad_norm": 0.21586061078083138, "learning_rate": 6.328428538094895e-06, "loss": 0.5763, "num_tokens": 2965750136.0, "step": 2830 }, { "epoch": 4.485759493670886, "grad_norm": 0.24077949998848439, "learning_rate": 6.288864017554471e-06, "loss": 0.5733, "num_tokens": 2970993016.0, "step": 2835 }, { "epoch": 4.493670886075949, "grad_norm": 0.225684038542863, "learning_rate": 6.249880236446157e-06, "loss": 0.5733, "num_tokens": 2976232054.0, "step": 2840 }, { "epoch": 4.5015822784810124, "grad_norm": 0.212864190216861, "learning_rate": 6.211478262106416e-06, "loss": 0.5757, "num_tokens": 2981474934.0, "step": 2845 }, { "epoch": 4.509493670886076, "grad_norm": 0.2240303672598969, "learning_rate": 6.173659145942439e-06, "loss": 0.5715, "num_tokens": 2986716501.0, "step": 2850 }, { "epoch": 4.5174050632911396, "grad_norm": 0.21964401447515855, "learning_rate": 6.136423923403332e-06, "loss": 0.5729, "num_tokens": 2991948269.0, "step": 2855 }, { "epoch": 4.525316455696203, "grad_norm": 0.21745396683364604, "learning_rate": 6.099773613951805e-06, "loss": 0.5691, "num_tokens": 2997172161.0, "step": 2860 }, { "epoch": 4.533227848101266, "grad_norm": 0.2051009997691557, "learning_rate": 6.063709221036231e-06, "loss": 0.5729, "num_tokens": 3002415041.0, "step": 2865 }, { "epoch": 4.541139240506329, "grad_norm": 0.2082354488706409, "learning_rate": 6.028231732063179e-06, "loss": 0.5683, "num_tokens": 3007657921.0, "step": 2870 }, { "epoch": 4.549050632911392, "grad_norm": 0.20333733144092334, "learning_rate": 5.993342118370382e-06, "loss": 0.5728, "num_tokens": 3012900801.0, "step": 2875 }, { "epoch": 4.556962025316456, "grad_norm": 0.22876886064683172, "learning_rate": 5.959041335200154e-06, "loss": 0.5713, "num_tokens": 3018143681.0, "step": 2880 }, { "epoch": 4.564873417721519, "grad_norm": 0.23107232316277548, "learning_rate": 5.925330321673209e-06, "loss": 0.5695, "num_tokens": 3023386561.0, "step": 2885 }, { "epoch": 4.572784810126582, "grad_norm": 0.2134849519354519, "learning_rate": 5.892210000762978e-06, "loss": 0.5669, "num_tokens": 3028629441.0, "step": 2890 }, { "epoch": 4.580696202531645, "grad_norm": 0.22796244574283725, "learning_rate": 5.859681279270323e-06, "loss": 0.5737, "num_tokens": 3033867096.0, "step": 2895 }, { "epoch": 4.588607594936709, "grad_norm": 0.2241995743982485, "learning_rate": 5.827745047798706e-06, "loss": 0.574, "num_tokens": 3039109976.0, "step": 2900 }, { "epoch": 4.5965189873417724, "grad_norm": 0.2432096355202257, "learning_rate": 5.796402180729816e-06, "loss": 0.5727, "num_tokens": 3044352856.0, "step": 2905 }, { "epoch": 4.6044303797468356, "grad_norm": 0.22578199424608314, "learning_rate": 5.7656535361996335e-06, "loss": 0.5732, "num_tokens": 3049595736.0, "step": 2910 }, { "epoch": 4.612341772151899, "grad_norm": 0.237878631640341, "learning_rate": 5.735499956074914e-06, "loss": 0.5694, "num_tokens": 3054838616.0, "step": 2915 }, { "epoch": 4.620253164556962, "grad_norm": 0.20655029863391322, "learning_rate": 5.705942265930159e-06, "loss": 0.568, "num_tokens": 3060076579.0, "step": 2920 }, { "epoch": 4.628164556962025, "grad_norm": 0.2102514850204724, "learning_rate": 5.676981275025011e-06, "loss": 0.5716, "num_tokens": 3065319459.0, "step": 2925 }, { "epoch": 4.636075949367089, "grad_norm": 0.21521201282981267, "learning_rate": 5.648617776282077e-06, "loss": 0.5727, "num_tokens": 3070562339.0, "step": 2930 }, { "epoch": 4.643987341772152, "grad_norm": 0.22220846495501412, "learning_rate": 5.620852546265245e-06, "loss": 0.5734, "num_tokens": 3075805219.0, "step": 2935 }, { "epoch": 4.651898734177215, "grad_norm": 0.20514030112007264, "learning_rate": 5.5936863451584056e-06, "loss": 0.5746, "num_tokens": 3081039464.0, "step": 2940 }, { "epoch": 4.659810126582278, "grad_norm": 0.1984930661211808, "learning_rate": 5.56711991674465e-06, "loss": 0.5775, "num_tokens": 3086282344.0, "step": 2945 }, { "epoch": 4.667721518987342, "grad_norm": 0.21848535912224024, "learning_rate": 5.54115398838589e-06, "loss": 0.5747, "num_tokens": 3091525224.0, "step": 2950 }, { "epoch": 4.675632911392405, "grad_norm": 0.21623565097615416, "learning_rate": 5.515789271002967e-06, "loss": 0.572, "num_tokens": 3096751361.0, "step": 2955 }, { "epoch": 4.6835443037974684, "grad_norm": 0.2067122434127961, "learning_rate": 5.491026459056163e-06, "loss": 0.5702, "num_tokens": 3101978695.0, "step": 2960 }, { "epoch": 4.6914556962025316, "grad_norm": 0.23059832785126186, "learning_rate": 5.466866230526202e-06, "loss": 0.5763, "num_tokens": 3107221575.0, "step": 2965 }, { "epoch": 4.699367088607595, "grad_norm": 0.21776141057424536, "learning_rate": 5.443309246895688e-06, "loss": 0.5724, "num_tokens": 3112464455.0, "step": 2970 }, { "epoch": 4.707278481012658, "grad_norm": 0.2086382988283168, "learning_rate": 5.420356153130986e-06, "loss": 0.5756, "num_tokens": 3117704332.0, "step": 2975 }, { "epoch": 4.715189873417722, "grad_norm": 0.22426323834508063, "learning_rate": 5.398007577664566e-06, "loss": 0.5766, "num_tokens": 3122940201.0, "step": 2980 }, { "epoch": 4.723101265822785, "grad_norm": 0.2134984091805231, "learning_rate": 5.376264132377806e-06, "loss": 0.5715, "num_tokens": 3128183081.0, "step": 2985 }, { "epoch": 4.731012658227848, "grad_norm": 0.2078865591432323, "learning_rate": 5.355126412584226e-06, "loss": 0.5796, "num_tokens": 3133425961.0, "step": 2990 }, { "epoch": 4.738924050632911, "grad_norm": 0.21033825281525734, "learning_rate": 5.334594997013194e-06, "loss": 0.5722, "num_tokens": 3138668841.0, "step": 2995 }, { "epoch": 4.746835443037975, "grad_norm": 0.20218242486649357, "learning_rate": 5.314670447794094e-06, "loss": 0.5697, "num_tokens": 3143911721.0, "step": 3000 }, { "epoch": 4.754746835443038, "grad_norm": 0.20528211055221243, "learning_rate": 5.295353310440908e-06, "loss": 0.5775, "num_tokens": 3149154601.0, "step": 3005 }, { "epoch": 4.762658227848101, "grad_norm": 0.20386433324381634, "learning_rate": 5.276644113837305e-06, "loss": 0.5703, "num_tokens": 3154397481.0, "step": 3010 }, { "epoch": 4.7705696202531644, "grad_norm": 0.21879527292522297, "learning_rate": 5.258543370222146e-06, "loss": 0.5739, "num_tokens": 3159640361.0, "step": 3015 }, { "epoch": 4.7784810126582276, "grad_norm": 0.238651659867707, "learning_rate": 5.241051575175476e-06, "loss": 0.567, "num_tokens": 3164869054.0, "step": 3020 }, { "epoch": 4.786392405063291, "grad_norm": 0.21930826191732458, "learning_rate": 5.2241692076049294e-06, "loss": 0.5713, "num_tokens": 3170097693.0, "step": 3025 }, { "epoch": 4.794303797468355, "grad_norm": 0.24549638164106027, "learning_rate": 5.207896729732644e-06, "loss": 0.5784, "num_tokens": 3175338802.0, "step": 3030 }, { "epoch": 4.802215189873418, "grad_norm": 0.20497121752386122, "learning_rate": 5.1922345870825936e-06, "loss": 0.5718, "num_tokens": 3180581682.0, "step": 3035 }, { "epoch": 4.810126582278481, "grad_norm": 0.21556589673067728, "learning_rate": 5.1771832084683874e-06, "loss": 0.5755, "num_tokens": 3185824562.0, "step": 3040 }, { "epoch": 4.818037974683544, "grad_norm": 0.2177509045329286, "learning_rate": 5.162743005981538e-06, "loss": 0.5693, "num_tokens": 3191054329.0, "step": 3045 }, { "epoch": 4.825949367088608, "grad_norm": 0.22258143026660726, "learning_rate": 5.148914374980175e-06, "loss": 0.5697, "num_tokens": 3196297209.0, "step": 3050 }, { "epoch": 4.833860759493671, "grad_norm": 0.21877472366506615, "learning_rate": 5.135697694078209e-06, "loss": 0.5732, "num_tokens": 3201522963.0, "step": 3055 }, { "epoch": 4.841772151898734, "grad_norm": 0.2194924983288656, "learning_rate": 5.123093325134992e-06, "loss": 0.577, "num_tokens": 3206765843.0, "step": 3060 }, { "epoch": 4.849683544303797, "grad_norm": 0.24236174132097704, "learning_rate": 5.111101613245384e-06, "loss": 0.5734, "num_tokens": 3212008723.0, "step": 3065 }, { "epoch": 4.8575949367088604, "grad_norm": 0.22136019033385568, "learning_rate": 5.099722886730315e-06, "loss": 0.5738, "num_tokens": 3217244067.0, "step": 3070 }, { "epoch": 4.865506329113924, "grad_norm": 0.21908628318619908, "learning_rate": 5.088957457127804e-06, "loss": 0.5657, "num_tokens": 3222479254.0, "step": 3075 }, { "epoch": 4.8734177215189876, "grad_norm": 0.23149848485508415, "learning_rate": 5.078805619184415e-06, "loss": 0.5762, "num_tokens": 3227722134.0, "step": 3080 }, { "epoch": 4.881329113924051, "grad_norm": 0.21576735952327258, "learning_rate": 5.069267650847191e-06, "loss": 0.5687, "num_tokens": 3232960340.0, "step": 3085 }, { "epoch": 4.889240506329114, "grad_norm": 0.21146979612133376, "learning_rate": 5.060343813256054e-06, "loss": 0.5749, "num_tokens": 3238203220.0, "step": 3090 }, { "epoch": 4.897151898734177, "grad_norm": 0.22343231609158218, "learning_rate": 5.052034350736642e-06, "loss": 0.577, "num_tokens": 3243446100.0, "step": 3095 }, { "epoch": 4.905063291139241, "grad_norm": 0.21750437307740444, "learning_rate": 5.044339490793628e-06, "loss": 0.5746, "num_tokens": 3248682414.0, "step": 3100 }, { "epoch": 4.912974683544304, "grad_norm": 0.2040207080561066, "learning_rate": 5.037259444104488e-06, "loss": 0.5752, "num_tokens": 3253925294.0, "step": 3105 }, { "epoch": 4.920886075949367, "grad_norm": 0.21492843271025994, "learning_rate": 5.030794404513734e-06, "loss": 0.5723, "num_tokens": 3259168174.0, "step": 3110 }, { "epoch": 4.92879746835443, "grad_norm": 0.22472996856568292, "learning_rate": 5.024944549027605e-06, "loss": 0.5757, "num_tokens": 3264411054.0, "step": 3115 }, { "epoch": 4.936708860759493, "grad_norm": 0.2105235983069695, "learning_rate": 5.019710037809223e-06, "loss": 0.5764, "num_tokens": 3269653934.0, "step": 3120 }, { "epoch": 4.944620253164557, "grad_norm": 0.20160858018191585, "learning_rate": 5.015091014174209e-06, "loss": 0.5676, "num_tokens": 3274885682.0, "step": 3125 }, { "epoch": 4.9525316455696204, "grad_norm": 0.2079284452125249, "learning_rate": 5.0110876045867496e-06, "loss": 0.5662, "num_tokens": 3280128562.0, "step": 3130 }, { "epoch": 4.9604430379746836, "grad_norm": 0.21834231623167524, "learning_rate": 5.007699918656152e-06, "loss": 0.5719, "num_tokens": 3285371442.0, "step": 3135 }, { "epoch": 4.968354430379747, "grad_norm": 0.21543773946569159, "learning_rate": 5.004928049133823e-06, "loss": 0.5694, "num_tokens": 3290614322.0, "step": 3140 }, { "epoch": 4.97626582278481, "grad_norm": 0.2150336576277292, "learning_rate": 5.002772071910747e-06, "loss": 0.5747, "num_tokens": 3295843098.0, "step": 3145 }, { "epoch": 4.984177215189874, "grad_norm": 0.20873534752474804, "learning_rate": 5.0012320460153974e-06, "loss": 0.5696, "num_tokens": 3301070530.0, "step": 3150 }, { "epoch": 4.992088607594937, "grad_norm": 0.21825458015635657, "learning_rate": 5.000308013612126e-06, "loss": 0.5762, "num_tokens": 3306313410.0, "step": 3155 }, { "epoch": 5.0, "grad_norm": 0.2206973186589871, "learning_rate": 5e-06, "loss": 0.5717, "num_tokens": 3311556290.0, "step": 3160 }, { "epoch": 5.0, "step": 3160, "total_flos": 2880677515100160.0, "train_loss": 0.6431828314744974, "train_runtime": 46582.8084, "train_samples_per_second": 4.34, "train_steps_per_second": 0.068 } ], "logging_steps": 5, "max_steps": 3160, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2880677515100160.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }