{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2094, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.014344629729245113, "grad_norm": 3.3866311674749072, "learning_rate": 4.285714285714286e-06, "loss": 1.5266, "step": 10 }, { "epoch": 0.028689259458490227, "grad_norm": 1.8728384521077752, "learning_rate": 9.047619047619047e-06, "loss": 1.3066, "step": 20 }, { "epoch": 0.04303388918773534, "grad_norm": 1.1126841032649692, "learning_rate": 1.3809523809523811e-05, "loss": 1.1436, "step": 30 }, { "epoch": 0.05737851891698045, "grad_norm": 1.1207104465369435, "learning_rate": 1.8571428571428572e-05, "loss": 1.0535, "step": 40 }, { "epoch": 0.07172314864622557, "grad_norm": 1.1391226479086436, "learning_rate": 2.3333333333333336e-05, "loss": 1.0172, "step": 50 }, { "epoch": 0.08606777837547068, "grad_norm": 1.146545389971907, "learning_rate": 2.8095238095238096e-05, "loss": 0.9859, "step": 60 }, { "epoch": 0.1004124081047158, "grad_norm": 1.1888015259911366, "learning_rate": 3.285714285714286e-05, "loss": 0.9528, "step": 70 }, { "epoch": 0.1147570378339609, "grad_norm": 1.2903823343286538, "learning_rate": 3.761904761904762e-05, "loss": 0.9397, "step": 80 }, { "epoch": 0.129101667563206, "grad_norm": 1.2943713329149742, "learning_rate": 4.2380952380952385e-05, "loss": 0.9282, "step": 90 }, { "epoch": 0.14344629729245115, "grad_norm": 1.2718909502396083, "learning_rate": 4.714285714285714e-05, "loss": 0.9186, "step": 100 }, { "epoch": 0.15779092702169625, "grad_norm": 1.4791012823315328, "learning_rate": 5.1904761904761913e-05, "loss": 0.9074, "step": 110 }, { "epoch": 0.17213555675094136, "grad_norm": 1.336532460075517, "learning_rate": 5.666666666666667e-05, "loss": 0.9002, "step": 120 }, { "epoch": 0.1864801864801865, "grad_norm": 1.6094057366302217, "learning_rate": 6.142857142857143e-05, "loss": 0.8887, "step": 130 }, { "epoch": 0.2008248162094316, "grad_norm": 1.070539230983606, "learning_rate": 6.619047619047619e-05, "loss": 0.884, "step": 140 }, { "epoch": 0.2151694459386767, "grad_norm": 1.1200792510458928, "learning_rate": 7.095238095238096e-05, "loss": 0.8809, "step": 150 }, { "epoch": 0.2295140756679218, "grad_norm": 1.087537431311211, "learning_rate": 7.571428571428571e-05, "loss": 0.8799, "step": 160 }, { "epoch": 0.24385870539716695, "grad_norm": 0.9999716597796505, "learning_rate": 8.047619047619048e-05, "loss": 0.8802, "step": 170 }, { "epoch": 0.258203335126412, "grad_norm": 1.2050995451812405, "learning_rate": 8.523809523809524e-05, "loss": 0.8692, "step": 180 }, { "epoch": 0.2725479648556572, "grad_norm": 0.9483149209977663, "learning_rate": 9e-05, "loss": 0.8675, "step": 190 }, { "epoch": 0.2868925945849023, "grad_norm": 0.9047494721372505, "learning_rate": 9.476190476190476e-05, "loss": 0.8651, "step": 200 }, { "epoch": 0.3012372243141474, "grad_norm": 0.9614336061705705, "learning_rate": 9.952380952380953e-05, "loss": 0.8743, "step": 210 }, { "epoch": 0.3155818540433925, "grad_norm": 0.840798357905234, "learning_rate": 9.999436939807164e-05, "loss": 0.8831, "step": 220 }, { "epoch": 0.3299264837726376, "grad_norm": 0.894805005888125, "learning_rate": 9.99749072170404e-05, "loss": 0.8753, "step": 230 }, { "epoch": 0.3442711135018827, "grad_norm": 0.9621522988496558, "learning_rate": 9.994154935353517e-05, "loss": 0.8553, "step": 240 }, { "epoch": 0.3586157432311278, "grad_norm": 0.8338221919665476, "learning_rate": 9.98943050828164e-05, "loss": 0.8577, "step": 250 }, { "epoch": 0.372960372960373, "grad_norm": 0.8401540225438021, "learning_rate": 9.983318754130435e-05, "loss": 0.842, "step": 260 }, { "epoch": 0.3873050026896181, "grad_norm": 0.7967631358112257, "learning_rate": 9.975821372292653e-05, "loss": 0.8346, "step": 270 }, { "epoch": 0.4016496324188632, "grad_norm": 0.7572726108984663, "learning_rate": 9.966940447439245e-05, "loss": 0.8272, "step": 280 }, { "epoch": 0.4159942621481083, "grad_norm": 0.7862046452006615, "learning_rate": 9.956678448939718e-05, "loss": 0.8161, "step": 290 }, { "epoch": 0.4303388918773534, "grad_norm": 0.8007930832524255, "learning_rate": 9.945038230175509e-05, "loss": 0.8263, "step": 300 }, { "epoch": 0.4446835216065985, "grad_norm": 0.6882798627741967, "learning_rate": 9.932023027746602e-05, "loss": 0.8119, "step": 310 }, { "epoch": 0.4590281513358436, "grad_norm": 0.7191586339334445, "learning_rate": 9.917636460571578e-05, "loss": 0.8195, "step": 320 }, { "epoch": 0.47337278106508873, "grad_norm": 0.6479226154771014, "learning_rate": 9.901882528881363e-05, "loss": 0.8152, "step": 330 }, { "epoch": 0.4877174107943339, "grad_norm": 0.7520756414640943, "learning_rate": 9.884765613106948e-05, "loss": 0.8056, "step": 340 }, { "epoch": 0.502062040523579, "grad_norm": 0.7496882676872738, "learning_rate": 9.866290472661406e-05, "loss": 0.8143, "step": 350 }, { "epoch": 0.516406670252824, "grad_norm": 0.8849980877887388, "learning_rate": 9.846462244616508e-05, "loss": 0.8139, "step": 360 }, { "epoch": 0.5307512999820692, "grad_norm": 0.689970676364689, "learning_rate": 9.825286442274357e-05, "loss": 0.8052, "step": 370 }, { "epoch": 0.5450959297113144, "grad_norm": 0.7192552702117557, "learning_rate": 9.802768953634388e-05, "loss": 0.7918, "step": 380 }, { "epoch": 0.5594405594405595, "grad_norm": 0.6975024662772501, "learning_rate": 9.778916039756193e-05, "loss": 0.7859, "step": 390 }, { "epoch": 0.5737851891698046, "grad_norm": 0.6553934247634706, "learning_rate": 9.753734333018616e-05, "loss": 0.776, "step": 400 }, { "epoch": 0.5881298188990497, "grad_norm": 0.757809076489788, "learning_rate": 9.727230835275598e-05, "loss": 0.7811, "step": 410 }, { "epoch": 0.6024744486282948, "grad_norm": 0.7082602077432336, "learning_rate": 9.699412915909284e-05, "loss": 0.7819, "step": 420 }, { "epoch": 0.6168190783575399, "grad_norm": 0.6348735880274488, "learning_rate": 9.670288309780953e-05, "loss": 0.7679, "step": 430 }, { "epoch": 0.631163708086785, "grad_norm": 0.6405240862226912, "learning_rate": 9.639865115080304e-05, "loss": 0.779, "step": 440 }, { "epoch": 0.6455083378160301, "grad_norm": 0.6634767779062684, "learning_rate": 9.608151791073737e-05, "loss": 0.7739, "step": 450 }, { "epoch": 0.6598529675452752, "grad_norm": 0.5759274474197833, "learning_rate": 9.575157155752222e-05, "loss": 0.7643, "step": 460 }, { "epoch": 0.6741975972745203, "grad_norm": 0.6640199963749864, "learning_rate": 9.54089038337943e-05, "loss": 0.7556, "step": 470 }, { "epoch": 0.6885422270037654, "grad_norm": 0.6203004765144989, "learning_rate": 9.5053610019408e-05, "loss": 0.7596, "step": 480 }, { "epoch": 0.7028868567330105, "grad_norm": 0.5775087991410168, "learning_rate": 9.468578890494256e-05, "loss": 0.7534, "step": 490 }, { "epoch": 0.7172314864622557, "grad_norm": 0.6714966253079345, "learning_rate": 9.430554276423292e-05, "loss": 0.7552, "step": 500 }, { "epoch": 0.7315761161915008, "grad_norm": 0.5888288333456809, "learning_rate": 9.391297732593229e-05, "loss": 0.7526, "step": 510 }, { "epoch": 0.745920745920746, "grad_norm": 0.6259226612543665, "learning_rate": 9.350820174411386e-05, "loss": 0.7537, "step": 520 }, { "epoch": 0.7602653756499911, "grad_norm": 0.5499968625723676, "learning_rate": 9.309132856792023e-05, "loss": 0.7499, "step": 530 }, { "epoch": 0.7746100053792362, "grad_norm": 0.6500285934875327, "learning_rate": 9.266247371026873e-05, "loss": 0.7447, "step": 540 }, { "epoch": 0.7889546351084813, "grad_norm": 0.5672171674827838, "learning_rate": 9.222175641562143e-05, "loss": 0.7496, "step": 550 }, { "epoch": 0.8032992648377264, "grad_norm": 0.6465733807975256, "learning_rate": 9.176929922682891e-05, "loss": 0.74, "step": 560 }, { "epoch": 0.8176438945669715, "grad_norm": 0.6101304041389433, "learning_rate": 9.130522795105676e-05, "loss": 0.7411, "step": 570 }, { "epoch": 0.8319885242962166, "grad_norm": 0.5489063036723607, "learning_rate": 9.082967162480459e-05, "loss": 0.7323, "step": 580 }, { "epoch": 0.8463331540254617, "grad_norm": 0.5516558559909698, "learning_rate": 9.034276247802688e-05, "loss": 0.7231, "step": 590 }, { "epoch": 0.8606777837547068, "grad_norm": 0.6477118047968917, "learning_rate": 8.984463589736614e-05, "loss": 0.7288, "step": 600 }, { "epoch": 0.8750224134839519, "grad_norm": 0.5962486006428408, "learning_rate": 8.933543038850816e-05, "loss": 0.732, "step": 610 }, { "epoch": 0.889367043213197, "grad_norm": 0.5713715183887016, "learning_rate": 8.881528753767007e-05, "loss": 0.728, "step": 620 }, { "epoch": 0.9037116729424421, "grad_norm": 0.6301745705273146, "learning_rate": 8.82843519722319e-05, "loss": 0.7263, "step": 630 }, { "epoch": 0.9180563026716873, "grad_norm": 0.5819941044165688, "learning_rate": 8.774277132052237e-05, "loss": 0.7189, "step": 640 }, { "epoch": 0.9324009324009324, "grad_norm": 0.5336149445366645, "learning_rate": 8.719069617077046e-05, "loss": 0.7254, "step": 650 }, { "epoch": 0.9467455621301775, "grad_norm": 0.6083401813961216, "learning_rate": 8.662828002923378e-05, "loss": 0.7218, "step": 660 }, { "epoch": 0.9610901918594227, "grad_norm": 0.5371262960983777, "learning_rate": 8.605567927751576e-05, "loss": 0.719, "step": 670 }, { "epoch": 0.9754348215886678, "grad_norm": 0.5551868136565808, "learning_rate": 8.547305312908318e-05, "loss": 0.7089, "step": 680 }, { "epoch": 0.9897794513179129, "grad_norm": 0.5899796664828114, "learning_rate": 8.48805635849964e-05, "loss": 0.7075, "step": 690 }, { "epoch": 1.002868925945849, "grad_norm": 0.608174630463498, "learning_rate": 8.427837538886437e-05, "loss": 0.6894, "step": 700 }, { "epoch": 1.017213555675094, "grad_norm": 0.6116990279567732, "learning_rate": 8.366665598103727e-05, "loss": 0.5865, "step": 710 }, { "epoch": 1.0315581854043392, "grad_norm": 0.5768126568975168, "learning_rate": 8.304557545204908e-05, "loss": 0.5931, "step": 720 }, { "epoch": 1.0459028151335843, "grad_norm": 0.5727241793491287, "learning_rate": 8.241530649532339e-05, "loss": 0.5859, "step": 730 }, { "epoch": 1.0602474448628294, "grad_norm": 0.5834066080830065, "learning_rate": 8.177602435915546e-05, "loss": 0.5833, "step": 740 }, { "epoch": 1.0745920745920745, "grad_norm": 0.6168394884857555, "learning_rate": 8.11279067979839e-05, "loss": 0.5805, "step": 750 }, { "epoch": 1.0889367043213196, "grad_norm": 0.4885078172344663, "learning_rate": 8.04711340229654e-05, "loss": 0.5733, "step": 760 }, { "epoch": 1.1032813340505647, "grad_norm": 0.5289947493535304, "learning_rate": 7.980588865186649e-05, "loss": 0.5812, "step": 770 }, { "epoch": 1.11762596377981, "grad_norm": 0.5672904537751536, "learning_rate": 7.913235565828613e-05, "loss": 0.5811, "step": 780 }, { "epoch": 1.1319705935090552, "grad_norm": 0.6107186114378813, "learning_rate": 7.845072232022311e-05, "loss": 0.5755, "step": 790 }, { "epoch": 1.1463152232383003, "grad_norm": 0.6696900109358123, "learning_rate": 7.776117816800288e-05, "loss": 0.5916, "step": 800 }, { "epoch": 1.1606598529675454, "grad_norm": 0.5412095989289454, "learning_rate": 7.706391493157805e-05, "loss": 0.5822, "step": 810 }, { "epoch": 1.1750044826967905, "grad_norm": 0.5408330429359739, "learning_rate": 7.635912648721718e-05, "loss": 0.5811, "step": 820 }, { "epoch": 1.1893491124260356, "grad_norm": 0.5952577265784902, "learning_rate": 7.564700880359696e-05, "loss": 0.5799, "step": 830 }, { "epoch": 1.2036937421552807, "grad_norm": 0.5026109808475954, "learning_rate": 7.492775988731243e-05, "loss": 0.5673, "step": 840 }, { "epoch": 1.2180383718845258, "grad_norm": 0.5302543029902266, "learning_rate": 7.420157972782063e-05, "loss": 0.5754, "step": 850 }, { "epoch": 1.232383001613771, "grad_norm": 0.5532266667153597, "learning_rate": 7.346867024183291e-05, "loss": 0.579, "step": 860 }, { "epoch": 1.246727631343016, "grad_norm": 0.5669382050811871, "learning_rate": 7.272923521717133e-05, "loss": 0.5756, "step": 870 }, { "epoch": 1.2610722610722611, "grad_norm": 0.5338735116713776, "learning_rate": 7.198348025610481e-05, "loss": 0.5769, "step": 880 }, { "epoch": 1.2754168908015062, "grad_norm": 0.5863122040109188, "learning_rate": 7.12316127181808e-05, "loss": 0.5685, "step": 890 }, { "epoch": 1.2897615205307513, "grad_norm": 0.5541827820189409, "learning_rate": 7.047384166256815e-05, "loss": 0.5801, "step": 900 }, { "epoch": 1.3041061502599964, "grad_norm": 0.5753955062923616, "learning_rate": 6.971037778992775e-05, "loss": 0.5795, "step": 910 }, { "epoch": 1.3184507799892415, "grad_norm": 0.47433414173616173, "learning_rate": 6.894143338382639e-05, "loss": 0.5768, "step": 920 }, { "epoch": 1.3327954097184866, "grad_norm": 0.5274791181928579, "learning_rate": 6.81672222517107e-05, "loss": 0.578, "step": 930 }, { "epoch": 1.3471400394477318, "grad_norm": 0.5774970902851654, "learning_rate": 6.73879596654573e-05, "loss": 0.5682, "step": 940 }, { "epoch": 1.3614846691769769, "grad_norm": 0.5662099451295628, "learning_rate": 6.660386230151571e-05, "loss": 0.5735, "step": 950 }, { "epoch": 1.375829298906222, "grad_norm": 0.5574084243877258, "learning_rate": 6.581514818066088e-05, "loss": 0.5739, "step": 960 }, { "epoch": 1.390173928635467, "grad_norm": 0.539589642853108, "learning_rate": 6.502203660737169e-05, "loss": 0.5688, "step": 970 }, { "epoch": 1.4045185583647122, "grad_norm": 0.49293926542133193, "learning_rate": 6.422474810885278e-05, "loss": 0.5669, "step": 980 }, { "epoch": 1.4188631880939573, "grad_norm": 0.49620916916034896, "learning_rate": 6.342350437371614e-05, "loss": 0.5702, "step": 990 }, { "epoch": 1.4332078178232024, "grad_norm": 0.5152950617397968, "learning_rate": 6.26185281903399e-05, "loss": 0.5685, "step": 1000 }, { "epoch": 1.4475524475524475, "grad_norm": 0.545911912027696, "learning_rate": 6.181004338492141e-05, "loss": 0.5651, "step": 1010 }, { "epoch": 1.4618970772816926, "grad_norm": 0.5074263813146201, "learning_rate": 6.09982747592415e-05, "loss": 0.5615, "step": 1020 }, { "epoch": 1.4762417070109377, "grad_norm": 0.5094909684862107, "learning_rate": 6.018344802815778e-05, "loss": 0.5679, "step": 1030 }, { "epoch": 1.4905863367401828, "grad_norm": 0.5103259058296251, "learning_rate": 5.936578975684378e-05, "loss": 0.5701, "step": 1040 }, { "epoch": 1.504930966469428, "grad_norm": 0.5278709818430807, "learning_rate": 5.854552729779184e-05, "loss": 0.5627, "step": 1050 }, { "epoch": 1.519275596198673, "grad_norm": 0.510686482301255, "learning_rate": 5.772288872759702e-05, "loss": 0.5625, "step": 1060 }, { "epoch": 1.5336202259279181, "grad_norm": 0.5278526048784437, "learning_rate": 5.6898102783539665e-05, "loss": 0.5606, "step": 1070 }, { "epoch": 1.5479648556571632, "grad_norm": 0.5559268236737518, "learning_rate": 5.607139879998427e-05, "loss": 0.5595, "step": 1080 }, { "epoch": 1.5623094853864083, "grad_norm": 0.5583305948616533, "learning_rate": 5.524300664461235e-05, "loss": 0.5575, "step": 1090 }, { "epoch": 1.5766541151156535, "grad_norm": 0.5186185461448056, "learning_rate": 5.441315665450697e-05, "loss": 0.5589, "step": 1100 }, { "epoch": 1.5909987448448986, "grad_norm": 0.5168193702732372, "learning_rate": 5.3582079572106794e-05, "loss": 0.5534, "step": 1110 }, { "epoch": 1.6053433745741437, "grad_norm": 0.5037158486433893, "learning_rate": 5.275000648104743e-05, "loss": 0.5569, "step": 1120 }, { "epoch": 1.6196880043033888, "grad_norm": 0.49884254214328216, "learning_rate": 5.191716874190785e-05, "loss": 0.5564, "step": 1130 }, { "epoch": 1.6340326340326339, "grad_norm": 0.5071669705468655, "learning_rate": 5.1083797927879896e-05, "loss": 0.5552, "step": 1140 }, { "epoch": 1.648377263761879, "grad_norm": 0.5410568223255828, "learning_rate": 5.025012576037855e-05, "loss": 0.564, "step": 1150 }, { "epoch": 1.6627218934911243, "grad_norm": 0.4850296619715352, "learning_rate": 4.9416384044611124e-05, "loss": 0.5579, "step": 1160 }, { "epoch": 1.6770665232203694, "grad_norm": 0.5039355867374112, "learning_rate": 4.858280460512302e-05, "loss": 0.5551, "step": 1170 }, { "epoch": 1.6914111529496145, "grad_norm": 0.5354526153870778, "learning_rate": 4.7749619221338227e-05, "loss": 0.5553, "step": 1180 }, { "epoch": 1.7057557826788596, "grad_norm": 0.5149848285267117, "learning_rate": 4.691705956311225e-05, "loss": 0.554, "step": 1190 }, { "epoch": 1.7201004124081047, "grad_norm": 0.5076867052953596, "learning_rate": 4.608535712631566e-05, "loss": 0.553, "step": 1200 }, { "epoch": 1.7344450421373498, "grad_norm": 0.5245034600250137, "learning_rate": 4.525474316846581e-05, "loss": 0.5455, "step": 1210 }, { "epoch": 1.748789671866595, "grad_norm": 0.5044035057693634, "learning_rate": 4.4425448644425066e-05, "loss": 0.5487, "step": 1220 }, { "epoch": 1.76313430159584, "grad_norm": 0.5182301310051464, "learning_rate": 4.359770414218296e-05, "loss": 0.5401, "step": 1230 }, { "epoch": 1.7774789313250852, "grad_norm": 0.5066048824304953, "learning_rate": 4.2771739818740565e-05, "loss": 0.5496, "step": 1240 }, { "epoch": 1.7918235610543303, "grad_norm": 0.5337687629976401, "learning_rate": 4.194778533611451e-05, "loss": 0.5413, "step": 1250 }, { "epoch": 1.8061681907835754, "grad_norm": 0.5207289329535878, "learning_rate": 4.112606979747881e-05, "loss": 0.544, "step": 1260 }, { "epoch": 1.8205128205128205, "grad_norm": 0.5450234045888507, "learning_rate": 4.030682168346192e-05, "loss": 0.5358, "step": 1270 }, { "epoch": 1.8348574502420656, "grad_norm": 0.48504143892984203, "learning_rate": 3.949026878861704e-05, "loss": 0.539, "step": 1280 }, { "epoch": 1.8492020799713107, "grad_norm": 0.49111142317889844, "learning_rate": 3.867663815808303e-05, "loss": 0.5389, "step": 1290 }, { "epoch": 1.8635467097005558, "grad_norm": 0.5008380643048362, "learning_rate": 3.78661560244539e-05, "loss": 0.5363, "step": 1300 }, { "epoch": 1.8778913394298011, "grad_norm": 0.4952709952315973, "learning_rate": 3.705904774487396e-05, "loss": 0.5319, "step": 1310 }, { "epoch": 1.8922359691590462, "grad_norm": 0.5164716516662802, "learning_rate": 3.6255537738376706e-05, "loss": 0.5402, "step": 1320 }, { "epoch": 1.9065805988882913, "grad_norm": 0.5240016137645628, "learning_rate": 3.545584942348426e-05, "loss": 0.5309, "step": 1330 }, { "epoch": 1.9209252286175365, "grad_norm": 0.5058856022275582, "learning_rate": 3.466020515608525e-05, "loss": 0.5298, "step": 1340 }, { "epoch": 1.9352698583467816, "grad_norm": 0.5019184399441088, "learning_rate": 3.386882616760794e-05, "loss": 0.5319, "step": 1350 }, { "epoch": 1.9496144880760267, "grad_norm": 0.471920743317965, "learning_rate": 3.30819325035062e-05, "loss": 0.5287, "step": 1360 }, { "epoch": 1.9639591178052718, "grad_norm": 0.5094525188470136, "learning_rate": 3.229974296207513e-05, "loss": 0.5385, "step": 1370 }, { "epoch": 1.9783037475345169, "grad_norm": 0.5606826371060244, "learning_rate": 3.152247503361353e-05, "loss": 0.527, "step": 1380 }, { "epoch": 1.992648377263762, "grad_norm": 0.5085670056810321, "learning_rate": 3.075034483994997e-05, "loss": 0.5257, "step": 1390 }, { "epoch": 2.005737851891698, "grad_norm": 0.5719992437170487, "learning_rate": 2.998356707434947e-05, "loss": 0.4623, "step": 1400 }, { "epoch": 2.020082481620943, "grad_norm": 0.5387964274909228, "learning_rate": 2.9222354941817375e-05, "loss": 0.3606, "step": 1410 }, { "epoch": 2.034427111350188, "grad_norm": 0.5698349863645911, "learning_rate": 2.846692009981693e-05, "loss": 0.3527, "step": 1420 }, { "epoch": 2.0487717410794333, "grad_norm": 0.5324199413342471, "learning_rate": 2.771747259941734e-05, "loss": 0.3462, "step": 1430 }, { "epoch": 2.0631163708086784, "grad_norm": 0.53215786873291, "learning_rate": 2.6974220826888374e-05, "loss": 0.342, "step": 1440 }, { "epoch": 2.0774610005379235, "grad_norm": 0.5110033203467198, "learning_rate": 2.623737144575787e-05, "loss": 0.3462, "step": 1450 }, { "epoch": 2.0918056302671686, "grad_norm": 0.5246624681955846, "learning_rate": 2.5507129339348335e-05, "loss": 0.3487, "step": 1460 }, { "epoch": 2.1061502599964137, "grad_norm": 0.5070575905086573, "learning_rate": 2.478369755380839e-05, "loss": 0.3465, "step": 1470 }, { "epoch": 2.120494889725659, "grad_norm": 0.5132555087524301, "learning_rate": 2.406727724165524e-05, "loss": 0.3451, "step": 1480 }, { "epoch": 2.134839519454904, "grad_norm": 0.5143913782142077, "learning_rate": 2.3358067605843537e-05, "loss": 0.3442, "step": 1490 }, { "epoch": 2.149184149184149, "grad_norm": 0.4894675520933593, "learning_rate": 2.2656265844376367e-05, "loss": 0.3449, "step": 1500 }, { "epoch": 2.163528778913394, "grad_norm": 0.5118938379622366, "learning_rate": 2.1962067095473648e-05, "loss": 0.3443, "step": 1510 }, { "epoch": 2.1778734086426392, "grad_norm": 0.5264199925745078, "learning_rate": 2.127566438331345e-05, "loss": 0.3443, "step": 1520 }, { "epoch": 2.1922180383718843, "grad_norm": 0.5517034745554297, "learning_rate": 2.059724856436092e-05, "loss": 0.34, "step": 1530 }, { "epoch": 2.2065626681011294, "grad_norm": 0.5153912300862288, "learning_rate": 1.992700827430007e-05, "loss": 0.3447, "step": 1540 }, { "epoch": 2.2209072978303745, "grad_norm": 0.49999210085621454, "learning_rate": 1.9265129875582954e-05, "loss": 0.3402, "step": 1550 }, { "epoch": 2.23525192755962, "grad_norm": 0.5117332103404209, "learning_rate": 1.8611797405611097e-05, "loss": 0.3434, "step": 1560 }, { "epoch": 2.249596557288865, "grad_norm": 0.5334345620530477, "learning_rate": 1.7967192525563254e-05, "loss": 0.3429, "step": 1570 }, { "epoch": 2.2639411870181103, "grad_norm": 0.5052797787039551, "learning_rate": 1.733149446988394e-05, "loss": 0.3402, "step": 1580 }, { "epoch": 2.2782858167473554, "grad_norm": 0.5115033156660913, "learning_rate": 1.670487999644669e-05, "loss": 0.3413, "step": 1590 }, { "epoch": 2.2926304464766005, "grad_norm": 0.4961174768714891, "learning_rate": 1.6087523337406024e-05, "loss": 0.3407, "step": 1600 }, { "epoch": 2.3069750762058456, "grad_norm": 0.5181202897712733, "learning_rate": 1.547959615075164e-05, "loss": 0.3343, "step": 1610 }, { "epoch": 2.3213197059350907, "grad_norm": 0.5076475142323975, "learning_rate": 1.4881267472578325e-05, "loss": 0.3415, "step": 1620 }, { "epoch": 2.335664335664336, "grad_norm": 0.4921735423850851, "learning_rate": 1.4292703670084916e-05, "loss": 0.3392, "step": 1630 }, { "epoch": 2.350008965393581, "grad_norm": 0.5151996030919073, "learning_rate": 1.3714068395315427e-05, "loss": 0.3349, "step": 1640 }, { "epoch": 2.364353595122826, "grad_norm": 0.5104074354031792, "learning_rate": 1.3145522539655041e-05, "loss": 0.3354, "step": 1650 }, { "epoch": 2.378698224852071, "grad_norm": 0.5380960601841327, "learning_rate": 1.2587224189093755e-05, "loss": 0.3302, "step": 1660 }, { "epoch": 2.3930428545813163, "grad_norm": 0.5049847193289703, "learning_rate": 1.2039328580270065e-05, "loss": 0.3333, "step": 1670 }, { "epoch": 2.4073874843105614, "grad_norm": 0.5040716684214055, "learning_rate": 1.150198805730689e-05, "loss": 0.3329, "step": 1680 }, { "epoch": 2.4217321140398065, "grad_norm": 0.5040857331959766, "learning_rate": 1.0975352029451863e-05, "loss": 0.3337, "step": 1690 }, { "epoch": 2.4360767437690516, "grad_norm": 0.5091871791617659, "learning_rate": 1.0459566929533588e-05, "loss": 0.3345, "step": 1700 }, { "epoch": 2.4504213734982967, "grad_norm": 0.4986671760440932, "learning_rate": 9.954776173245511e-06, "loss": 0.3369, "step": 1710 }, { "epoch": 2.464766003227542, "grad_norm": 0.5146311314437344, "learning_rate": 9.461120119268713e-06, "loss": 0.3338, "step": 1720 }, { "epoch": 2.479110632956787, "grad_norm": 0.5052152754810867, "learning_rate": 8.978736030244783e-06, "loss": 0.331, "step": 1730 }, { "epoch": 2.493455262686032, "grad_norm": 0.5060074098914322, "learning_rate": 8.50775803460948e-06, "loss": 0.3336, "step": 1740 }, { "epoch": 2.507799892415277, "grad_norm": 0.5047637166351485, "learning_rate": 8.048317089297875e-06, "loss": 0.331, "step": 1750 }, { "epoch": 2.5221445221445222, "grad_norm": 0.5051085483525463, "learning_rate": 7.600540943331347e-06, "loss": 0.3316, "step": 1760 }, { "epoch": 2.5364891518737673, "grad_norm": 0.5291232815833027, "learning_rate": 7.164554102296617e-06, "loss": 0.3322, "step": 1770 }, { "epoch": 2.5508337816030124, "grad_norm": 0.4950112167812311, "learning_rate": 6.740477793726529e-06, "loss": 0.3302, "step": 1780 }, { "epoch": 2.5651784113322575, "grad_norm": 0.5171790095780296, "learning_rate": 6.32842993339236e-06, "loss": 0.3278, "step": 1790 }, { "epoch": 2.5795230410615027, "grad_norm": 0.5041803659831317, "learning_rate": 5.928525092516934e-06, "loss": 0.3334, "step": 1800 }, { "epoch": 2.5938676707907478, "grad_norm": 0.5073754782012652, "learning_rate": 5.540874465917778e-06, "loss": 0.3325, "step": 1810 }, { "epoch": 2.608212300519993, "grad_norm": 0.48715355421051987, "learning_rate": 5.165585841089021e-06, "loss": 0.3246, "step": 1820 }, { "epoch": 2.622556930249238, "grad_norm": 0.5162975920861064, "learning_rate": 4.8027635682307445e-06, "loss": 0.325, "step": 1830 }, { "epoch": 2.636901559978483, "grad_norm": 0.5030085273093482, "learning_rate": 4.45250853123404e-06, "loss": 0.3267, "step": 1840 }, { "epoch": 2.651246189707728, "grad_norm": 0.5261872357667398, "learning_rate": 4.1149181196299905e-06, "loss": 0.3277, "step": 1850 }, { "epoch": 2.6655908194369733, "grad_norm": 0.49678770417028423, "learning_rate": 3.7900862015101457e-06, "loss": 0.3285, "step": 1860 }, { "epoch": 2.6799354491662184, "grad_norm": 0.4889164865698146, "learning_rate": 3.4781030974262108e-06, "loss": 0.3268, "step": 1870 }, { "epoch": 2.6942800788954635, "grad_norm": 0.4884408260663131, "learning_rate": 3.1790555552761615e-06, "loss": 0.3246, "step": 1880 }, { "epoch": 2.7086247086247086, "grad_norm": 0.49292610830996564, "learning_rate": 2.8930267261836395e-06, "loss": 0.3231, "step": 1890 }, { "epoch": 2.7229693383539537, "grad_norm": 0.4890311152459734, "learning_rate": 2.6200961413776094e-06, "loss": 0.324, "step": 1900 }, { "epoch": 2.737313968083199, "grad_norm": 0.5165429385071639, "learning_rate": 2.3603396900783724e-06, "loss": 0.3251, "step": 1910 }, { "epoch": 2.751658597812444, "grad_norm": 0.5024981727502739, "learning_rate": 2.113829598396383e-06, "loss": 0.322, "step": 1920 }, { "epoch": 2.766003227541689, "grad_norm": 0.49038298209467507, "learning_rate": 1.8806344092494932e-06, "loss": 0.3248, "step": 1930 }, { "epoch": 2.780347857270934, "grad_norm": 0.4917027026533285, "learning_rate": 1.6608189633044113e-06, "loss": 0.325, "step": 1940 }, { "epoch": 2.7946924870001792, "grad_norm": 0.4869252686371276, "learning_rate": 1.4544443809475561e-06, "loss": 0.3247, "step": 1950 }, { "epoch": 2.8090371167294244, "grad_norm": 0.48253239623178734, "learning_rate": 1.2615680452903067e-06, "loss": 0.3234, "step": 1960 }, { "epoch": 2.8233817464586695, "grad_norm": 0.5096600620224961, "learning_rate": 1.082243586213455e-06, "loss": 0.3219, "step": 1970 }, { "epoch": 2.8377263761879146, "grad_norm": 0.48655701646749633, "learning_rate": 9.165208654552671e-07, "loss": 0.321, "step": 1980 }, { "epoch": 2.8520710059171597, "grad_norm": 0.47440137878879957, "learning_rate": 7.6444596274724e-07, "loss": 0.3229, "step": 1990 }, { "epoch": 2.866415635646405, "grad_norm": 0.5034996226878241, "learning_rate": 6.260611630015067e-07, "loss": 0.3221, "step": 2000 }, { "epoch": 2.88076026537565, "grad_norm": 0.47683170425958754, "learning_rate": 5.01404944553363e-07, "loss": 0.3219, "step": 2010 }, { "epoch": 2.895104895104895, "grad_norm": 0.4909218835510976, "learning_rate": 3.9051196846225935e-07, "loss": 0.3251, "step": 2020 }, { "epoch": 2.90944952483414, "grad_norm": 0.48456217220790176, "learning_rate": 2.9341306887417653e-07, "loss": 0.3213, "step": 2030 }, { "epoch": 2.923794154563385, "grad_norm": 0.4779578566731188, "learning_rate": 2.101352444480842e-07, "loss": 0.3241, "step": 2040 }, { "epoch": 2.9381387842926303, "grad_norm": 0.49966345259743034, "learning_rate": 1.4070165084889008e-07, "loss": 0.3199, "step": 2050 }, { "epoch": 2.9524834140218754, "grad_norm": 0.4754785997240394, "learning_rate": 8.513159430892925e-08, "loss": 0.3225, "step": 2060 }, { "epoch": 2.9668280437511205, "grad_norm": 0.4903628137644153, "learning_rate": 4.344052625981365e-08, "loss": 0.3219, "step": 2070 }, { "epoch": 2.9811726734803656, "grad_norm": 0.4992880216120249, "learning_rate": 1.5640039036085575e-08, "loss": 0.3187, "step": 2080 }, { "epoch": 2.9955173032096107, "grad_norm": 0.4963785468111048, "learning_rate": 1.7378626519626296e-09, "loss": 0.3238, "step": 2090 }, { "epoch": 3.0, "step": 2094, "total_flos": 306663055425536.0, "train_loss": 0.5762376924754328, "train_runtime": 90627.2697, "train_samples_per_second": 1.477, "train_steps_per_second": 0.023 } ], "logging_steps": 10, "max_steps": 2094, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 306663055425536.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }