{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 934, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010718113612004287, "grad_norm": 1.3203599452972412, "learning_rate": 1.0256410256410257e-06, "loss": 1.3095, "step": 5 }, { "epoch": 0.021436227224008574, "grad_norm": 1.005972981452942, "learning_rate": 2.307692307692308e-06, "loss": 1.3756, "step": 10 }, { "epoch": 0.03215434083601286, "grad_norm": 0.8182684183120728, "learning_rate": 3.58974358974359e-06, "loss": 1.3238, "step": 15 }, { "epoch": 0.04287245444801715, "grad_norm": 0.791577160358429, "learning_rate": 4.871794871794872e-06, "loss": 1.2938, "step": 20 }, { "epoch": 0.05359056806002144, "grad_norm": 0.6379374861717224, "learning_rate": 6.153846153846154e-06, "loss": 1.2856, "step": 25 }, { "epoch": 0.06430868167202572, "grad_norm": 0.6668587327003479, "learning_rate": 7.435897435897436e-06, "loss": 1.2841, "step": 30 }, { "epoch": 0.07502679528403002, "grad_norm": 0.716699481010437, "learning_rate": 8.717948717948719e-06, "loss": 1.2718, "step": 35 }, { "epoch": 0.0857449088960343, "grad_norm": 0.5486516952514648, "learning_rate": 9.999999999999999e-06, "loss": 1.273, "step": 40 }, { "epoch": 0.09646302250803858, "grad_norm": 0.46251392364501953, "learning_rate": 1.1282051282051283e-05, "loss": 1.312, "step": 45 }, { "epoch": 0.10718113612004287, "grad_norm": 0.4610588848590851, "learning_rate": 1.2564102564102565e-05, "loss": 1.2015, "step": 50 }, { "epoch": 0.11789924973204716, "grad_norm": 0.4434472918510437, "learning_rate": 1.3846153846153847e-05, "loss": 1.2457, "step": 55 }, { "epoch": 0.12861736334405144, "grad_norm": 0.43525955080986023, "learning_rate": 1.5128205128205129e-05, "loss": 1.2111, "step": 60 }, { "epoch": 0.13933547695605572, "grad_norm": 0.45412635803222656, "learning_rate": 1.641025641025641e-05, "loss": 1.2055, "step": 65 }, { "epoch": 0.15005359056806003, "grad_norm": 0.47974953055381775, "learning_rate": 1.7692307692307694e-05, "loss": 1.1836, "step": 70 }, { "epoch": 0.1607717041800643, "grad_norm": 0.5585247278213501, "learning_rate": 1.8974358974358975e-05, "loss": 1.1456, "step": 75 }, { "epoch": 0.1714898177920686, "grad_norm": 0.4740090072154999, "learning_rate": 2.025641025641026e-05, "loss": 1.1697, "step": 80 }, { "epoch": 0.18220793140407288, "grad_norm": 0.49104949831962585, "learning_rate": 2.153846153846154e-05, "loss": 1.1723, "step": 85 }, { "epoch": 0.19292604501607716, "grad_norm": 0.47726204991340637, "learning_rate": 2.282051282051282e-05, "loss": 1.1729, "step": 90 }, { "epoch": 0.20364415862808147, "grad_norm": 0.5089394450187683, "learning_rate": 2.4102564102564103e-05, "loss": 1.1313, "step": 95 }, { "epoch": 0.21436227224008575, "grad_norm": 0.4856749475002289, "learning_rate": 2.5384615384615386e-05, "loss": 1.1221, "step": 100 }, { "epoch": 0.22508038585209003, "grad_norm": 0.4452688992023468, "learning_rate": 2.6666666666666667e-05, "loss": 1.1334, "step": 105 }, { "epoch": 0.2357984994640943, "grad_norm": 0.5895005464553833, "learning_rate": 2.794871794871795e-05, "loss": 1.1453, "step": 110 }, { "epoch": 0.2465166130760986, "grad_norm": 0.5326249003410339, "learning_rate": 2.923076923076923e-05, "loss": 1.0856, "step": 115 }, { "epoch": 0.2572347266881029, "grad_norm": 0.588718056678772, "learning_rate": 2.9999939813702703e-05, "loss": 1.0957, "step": 120 }, { "epoch": 0.2679528403001072, "grad_norm": 0.4917924106121063, "learning_rate": 2.9999262723404875e-05, "loss": 1.0265, "step": 125 }, { "epoch": 0.27867095391211144, "grad_norm": 0.624920666217804, "learning_rate": 2.9997833344010312e-05, "loss": 1.0845, "step": 130 }, { "epoch": 0.28938906752411575, "grad_norm": 0.627967119216919, "learning_rate": 2.9995651747209637e-05, "loss": 1.1037, "step": 135 }, { "epoch": 0.30010718113612006, "grad_norm": 0.5074423551559448, "learning_rate": 2.9992718042420993e-05, "loss": 1.0788, "step": 140 }, { "epoch": 0.3108252947481243, "grad_norm": 0.5567290186882019, "learning_rate": 2.9989032376784556e-05, "loss": 1.1003, "step": 145 }, { "epoch": 0.3215434083601286, "grad_norm": 0.5891143083572388, "learning_rate": 2.998459493515515e-05, "loss": 1.0162, "step": 150 }, { "epoch": 0.3322615219721329, "grad_norm": 0.6841644048690796, "learning_rate": 2.9979405940092978e-05, "loss": 1.0204, "step": 155 }, { "epoch": 0.3429796355841372, "grad_norm": 0.6039224863052368, "learning_rate": 2.997346565185246e-05, "loss": 1.0219, "step": 160 }, { "epoch": 0.3536977491961415, "grad_norm": 0.5960545539855957, "learning_rate": 2.9966774368369175e-05, "loss": 1.0065, "step": 165 }, { "epoch": 0.36441586280814575, "grad_norm": 0.6452131867408752, "learning_rate": 2.995933242524493e-05, "loss": 0.9917, "step": 170 }, { "epoch": 0.37513397642015006, "grad_norm": 0.5966811776161194, "learning_rate": 2.995114019573091e-05, "loss": 0.9868, "step": 175 }, { "epoch": 0.3858520900321543, "grad_norm": 0.6129615306854248, "learning_rate": 2.9942198090708976e-05, "loss": 1.034, "step": 180 }, { "epoch": 0.3965702036441586, "grad_norm": 0.6630488038063049, "learning_rate": 2.993250655867105e-05, "loss": 1.0303, "step": 185 }, { "epoch": 0.40728831725616294, "grad_norm": 0.6165331602096558, "learning_rate": 2.9922066085696613e-05, "loss": 0.987, "step": 190 }, { "epoch": 0.4180064308681672, "grad_norm": 0.7296874523162842, "learning_rate": 2.9910877195428335e-05, "loss": 0.9805, "step": 195 }, { "epoch": 0.4287245444801715, "grad_norm": 0.6991822123527527, "learning_rate": 2.989894044904581e-05, "loss": 0.9377, "step": 200 }, { "epoch": 0.43944265809217575, "grad_norm": 0.7727455496788025, "learning_rate": 2.9886256445237405e-05, "loss": 0.9726, "step": 205 }, { "epoch": 0.45016077170418006, "grad_norm": 0.7376512289047241, "learning_rate": 2.987282582017024e-05, "loss": 0.946, "step": 210 }, { "epoch": 0.4608788853161844, "grad_norm": 0.819678544998169, "learning_rate": 2.9858649247458287e-05, "loss": 0.9861, "step": 215 }, { "epoch": 0.4715969989281886, "grad_norm": 0.7300285696983337, "learning_rate": 2.984372743812855e-05, "loss": 1.0026, "step": 220 }, { "epoch": 0.48231511254019294, "grad_norm": 0.7369091510772705, "learning_rate": 2.9828061140585463e-05, "loss": 0.9064, "step": 225 }, { "epoch": 0.4930332261521972, "grad_norm": 0.6813873052597046, "learning_rate": 2.9811651140573287e-05, "loss": 0.8859, "step": 230 }, { "epoch": 0.5037513397642015, "grad_norm": 0.7347345352172852, "learning_rate": 2.979449826113676e-05, "loss": 0.8983, "step": 235 }, { "epoch": 0.5144694533762058, "grad_norm": 0.8283259272575378, "learning_rate": 2.9776603362579775e-05, "loss": 0.8876, "step": 240 }, { "epoch": 0.5251875669882101, "grad_norm": 0.7388147711753845, "learning_rate": 2.9757967342422246e-05, "loss": 0.899, "step": 245 }, { "epoch": 0.5359056806002144, "grad_norm": 0.7657457590103149, "learning_rate": 2.9738591135355108e-05, "loss": 0.9412, "step": 250 }, { "epoch": 0.5466237942122186, "grad_norm": 0.8804264664649963, "learning_rate": 2.971847571319341e-05, "loss": 0.8838, "step": 255 }, { "epoch": 0.5573419078242229, "grad_norm": 0.771865963935852, "learning_rate": 2.969762208482759e-05, "loss": 0.8697, "step": 260 }, { "epoch": 0.5680600214362272, "grad_norm": 0.7006550431251526, "learning_rate": 2.967603129617288e-05, "loss": 0.8801, "step": 265 }, { "epoch": 0.5787781350482315, "grad_norm": 0.7909011244773865, "learning_rate": 2.9653704430116827e-05, "loss": 0.9199, "step": 270 }, { "epoch": 0.5894962486602358, "grad_norm": 0.9315905570983887, "learning_rate": 2.9630642606464998e-05, "loss": 0.8249, "step": 275 }, { "epoch": 0.6002143622722401, "grad_norm": 0.7528437376022339, "learning_rate": 2.9606846981884812e-05, "loss": 0.8656, "step": 280 }, { "epoch": 0.6109324758842444, "grad_norm": 0.7914732098579407, "learning_rate": 2.958231874984753e-05, "loss": 0.8618, "step": 285 }, { "epoch": 0.6216505894962486, "grad_norm": 0.7694675922393799, "learning_rate": 2.955705914056838e-05, "loss": 0.8546, "step": 290 }, { "epoch": 0.632368703108253, "grad_norm": 0.8651600480079651, "learning_rate": 2.9531069420944876e-05, "loss": 0.8654, "step": 295 }, { "epoch": 0.6430868167202572, "grad_norm": 0.7878230214118958, "learning_rate": 2.9504350894493272e-05, "loss": 0.8624, "step": 300 }, { "epoch": 0.6538049303322615, "grad_norm": 0.9521771669387817, "learning_rate": 2.9476904901283175e-05, "loss": 0.8175, "step": 305 }, { "epoch": 0.6645230439442658, "grad_norm": 0.8132469058036804, "learning_rate": 2.944873281787034e-05, "loss": 0.8087, "step": 310 }, { "epoch": 0.6752411575562701, "grad_norm": 0.8626110553741455, "learning_rate": 2.9419836057227634e-05, "loss": 0.8033, "step": 315 }, { "epoch": 0.6859592711682744, "grad_norm": 0.858932614326477, "learning_rate": 2.9390216068674154e-05, "loss": 0.7535, "step": 320 }, { "epoch": 0.6966773847802786, "grad_norm": 1.0335079431533813, "learning_rate": 2.9359874337802542e-05, "loss": 0.8272, "step": 325 }, { "epoch": 0.707395498392283, "grad_norm": 0.819476842880249, "learning_rate": 2.9328812386404503e-05, "loss": 0.8184, "step": 330 }, { "epoch": 0.7181136120042872, "grad_norm": 0.9225292205810547, "learning_rate": 2.9297031772394412e-05, "loss": 0.8009, "step": 335 }, { "epoch": 0.7288317256162915, "grad_norm": 0.9197941422462463, "learning_rate": 2.9264534089731262e-05, "loss": 0.7841, "step": 340 }, { "epoch": 0.7395498392282959, "grad_norm": 0.8611735701560974, "learning_rate": 2.9231320968338644e-05, "loss": 0.7975, "step": 345 }, { "epoch": 0.7502679528403001, "grad_norm": 0.8244771361351013, "learning_rate": 2.9197394074023042e-05, "loss": 0.7895, "step": 350 }, { "epoch": 0.7609860664523044, "grad_norm": 0.8496798872947693, "learning_rate": 2.9162755108390278e-05, "loss": 0.7931, "step": 355 }, { "epoch": 0.7717041800643086, "grad_norm": 0.9810552000999451, "learning_rate": 2.9127405808760152e-05, "loss": 0.7507, "step": 360 }, { "epoch": 0.782422293676313, "grad_norm": 0.9915773272514343, "learning_rate": 2.9091347948079318e-05, "loss": 0.8251, "step": 365 }, { "epoch": 0.7931404072883173, "grad_norm": 1.0386714935302734, "learning_rate": 2.9054583334832366e-05, "loss": 0.8063, "step": 370 }, { "epoch": 0.8038585209003215, "grad_norm": 0.9649940729141235, "learning_rate": 2.9017113812951105e-05, "loss": 0.7797, "step": 375 }, { "epoch": 0.8145766345123259, "grad_norm": 1.0805562734603882, "learning_rate": 2.897894126172209e-05, "loss": 0.7873, "step": 380 }, { "epoch": 0.8252947481243301, "grad_norm": 0.9175759553909302, "learning_rate": 2.8940067595692367e-05, "loss": 0.7306, "step": 385 }, { "epoch": 0.8360128617363344, "grad_norm": 1.0049532651901245, "learning_rate": 2.890049476457344e-05, "loss": 0.7357, "step": 390 }, { "epoch": 0.8467309753483387, "grad_norm": 0.8982154726982117, "learning_rate": 2.8860224753143486e-05, "loss": 0.7478, "step": 395 }, { "epoch": 0.857449088960343, "grad_norm": 0.898063600063324, "learning_rate": 2.881925958114782e-05, "loss": 0.7576, "step": 400 }, { "epoch": 0.8681672025723473, "grad_norm": 0.9208119511604309, "learning_rate": 2.877760130319757e-05, "loss": 0.7413, "step": 405 }, { "epoch": 0.8788853161843515, "grad_norm": 0.9576625823974609, "learning_rate": 2.8735252008666663e-05, "loss": 0.6863, "step": 410 }, { "epoch": 0.8896034297963559, "grad_norm": 0.9474856853485107, "learning_rate": 2.8692213821586988e-05, "loss": 0.7466, "step": 415 }, { "epoch": 0.9003215434083601, "grad_norm": 1.0231027603149414, "learning_rate": 2.8648488900541915e-05, "loss": 0.7399, "step": 420 }, { "epoch": 0.9110396570203644, "grad_norm": 0.9832124710083008, "learning_rate": 2.8604079438557985e-05, "loss": 0.7037, "step": 425 }, { "epoch": 0.9217577706323687, "grad_norm": 1.0980058908462524, "learning_rate": 2.8558987662994948e-05, "loss": 0.7189, "step": 430 }, { "epoch": 0.932475884244373, "grad_norm": 1.1283589601516724, "learning_rate": 2.851321583543404e-05, "loss": 0.6951, "step": 435 }, { "epoch": 0.9431939978563773, "grad_norm": 0.9087287187576294, "learning_rate": 2.8466766251564554e-05, "loss": 0.6932, "step": 440 }, { "epoch": 0.9539121114683816, "grad_norm": 1.0392457246780396, "learning_rate": 2.8419641241068687e-05, "loss": 0.7065, "step": 445 }, { "epoch": 0.9646302250803859, "grad_norm": 0.9226632118225098, "learning_rate": 2.837184316750473e-05, "loss": 0.7066, "step": 450 }, { "epoch": 0.9753483386923901, "grad_norm": 0.9685462713241577, "learning_rate": 2.832337442818848e-05, "loss": 0.6805, "step": 455 }, { "epoch": 0.9860664523043944, "grad_norm": 0.9703472852706909, "learning_rate": 2.8274237454073034e-05, "loss": 0.668, "step": 460 }, { "epoch": 0.9967845659163987, "grad_norm": 0.902520477771759, "learning_rate": 2.822443470962685e-05, "loss": 0.6379, "step": 465 }, { "epoch": 1.0064308681672025, "grad_norm": 1.2527391910552979, "learning_rate": 2.8173968692710142e-05, "loss": 0.6063, "step": 470 }, { "epoch": 1.0171489817792068, "grad_norm": 1.0625535249710083, "learning_rate": 2.8122841934449607e-05, "loss": 0.5916, "step": 475 }, { "epoch": 1.0278670953912112, "grad_norm": 1.1103384494781494, "learning_rate": 2.807105699911147e-05, "loss": 0.6323, "step": 480 }, { "epoch": 1.0385852090032155, "grad_norm": 1.0214422941207886, "learning_rate": 2.8018616483972874e-05, "loss": 0.573, "step": 485 }, { "epoch": 1.0493033226152197, "grad_norm": 0.9643238186836243, "learning_rate": 2.7965523019191606e-05, "loss": 0.5675, "step": 490 }, { "epoch": 1.060021436227224, "grad_norm": 0.9685465097427368, "learning_rate": 2.7911779267674207e-05, "loss": 0.5708, "step": 495 }, { "epoch": 1.0707395498392283, "grad_norm": 1.0887079238891602, "learning_rate": 2.7857387924942387e-05, "loss": 0.5954, "step": 500 }, { "epoch": 1.0814576634512325, "grad_norm": 0.9755250811576843, "learning_rate": 2.7802351718997825e-05, "loss": 0.5384, "step": 505 }, { "epoch": 1.092175777063237, "grad_norm": 0.9574642181396484, "learning_rate": 2.7746673410185387e-05, "loss": 0.5627, "step": 510 }, { "epoch": 1.1028938906752412, "grad_norm": 1.0391875505447388, "learning_rate": 2.7690355791054632e-05, "loss": 0.5621, "step": 515 }, { "epoch": 1.1136120042872455, "grad_norm": 0.9786142706871033, "learning_rate": 2.7633401686219793e-05, "loss": 0.5991, "step": 520 }, { "epoch": 1.1243301178992497, "grad_norm": 1.0193188190460205, "learning_rate": 2.7575813952218075e-05, "loss": 0.5723, "step": 525 }, { "epoch": 1.135048231511254, "grad_norm": 1.0080389976501465, "learning_rate": 2.75175954773664e-05, "loss": 0.5206, "step": 530 }, { "epoch": 1.1457663451232583, "grad_norm": 0.9439852833747864, "learning_rate": 2.7458749181616545e-05, "loss": 0.571, "step": 535 }, { "epoch": 1.1564844587352625, "grad_norm": 0.9455674290657043, "learning_rate": 2.7399278016408695e-05, "loss": 0.547, "step": 540 }, { "epoch": 1.167202572347267, "grad_norm": 1.138099193572998, "learning_rate": 2.7339184964523404e-05, "loss": 0.5942, "step": 545 }, { "epoch": 1.1779206859592712, "grad_norm": 1.1031590700149536, "learning_rate": 2.7278473039931992e-05, "loss": 0.5378, "step": 550 }, { "epoch": 1.1886387995712755, "grad_norm": 1.1105057001113892, "learning_rate": 2.7217145287645393e-05, "loss": 0.5186, "step": 555 }, { "epoch": 1.1993569131832797, "grad_norm": 1.2202413082122803, "learning_rate": 2.715520478356141e-05, "loss": 0.5721, "step": 560 }, { "epoch": 1.210075026795284, "grad_norm": 0.9721954464912415, "learning_rate": 2.7092654634310468e-05, "loss": 0.5433, "step": 565 }, { "epoch": 1.2207931404072883, "grad_norm": 1.2260032892227173, "learning_rate": 2.7029497977099784e-05, "loss": 0.543, "step": 570 }, { "epoch": 1.2315112540192925, "grad_norm": 0.9644677639007568, "learning_rate": 2.6965737979556025e-05, "loss": 0.552, "step": 575 }, { "epoch": 1.242229367631297, "grad_norm": 0.9937833547592163, "learning_rate": 2.6901377839566443e-05, "loss": 0.535, "step": 580 }, { "epoch": 1.2529474812433012, "grad_norm": 1.031482219696045, "learning_rate": 2.683642078511846e-05, "loss": 0.5322, "step": 585 }, { "epoch": 1.2636655948553055, "grad_norm": 0.9676262140274048, "learning_rate": 2.6770870074137808e-05, "loss": 0.5006, "step": 590 }, { "epoch": 1.2743837084673098, "grad_norm": 1.3593294620513916, "learning_rate": 2.670472899432509e-05, "loss": 0.5064, "step": 595 }, { "epoch": 1.285101822079314, "grad_norm": 1.0694024562835693, "learning_rate": 2.66380008629909e-05, "loss": 0.5848, "step": 600 }, { "epoch": 1.2958199356913183, "grad_norm": 1.0443016290664673, "learning_rate": 2.657068902688945e-05, "loss": 0.538, "step": 605 }, { "epoch": 1.3065380493033225, "grad_norm": 1.023464560508728, "learning_rate": 2.65027968620507e-05, "loss": 0.5587, "step": 610 }, { "epoch": 1.317256162915327, "grad_norm": 0.9948901534080505, "learning_rate": 2.643432777361105e-05, "loss": 0.534, "step": 615 }, { "epoch": 1.3279742765273312, "grad_norm": 1.0392403602600098, "learning_rate": 2.636528519564253e-05, "loss": 0.5365, "step": 620 }, { "epoch": 1.3386923901393355, "grad_norm": 1.0413520336151123, "learning_rate": 2.6295672590980586e-05, "loss": 0.4708, "step": 625 }, { "epoch": 1.3494105037513398, "grad_norm": 1.1629996299743652, "learning_rate": 2.6225493451050397e-05, "loss": 0.5554, "step": 630 }, { "epoch": 1.360128617363344, "grad_norm": 1.023750901222229, "learning_rate": 2.6154751295691756e-05, "loss": 0.4827, "step": 635 }, { "epoch": 1.3708467309753483, "grad_norm": 1.13310706615448, "learning_rate": 2.608344967298253e-05, "loss": 0.4657, "step": 640 }, { "epoch": 1.3815648445873525, "grad_norm": 1.083983063697815, "learning_rate": 2.6011592159060717e-05, "loss": 0.5213, "step": 645 }, { "epoch": 1.392282958199357, "grad_norm": 1.048169493675232, "learning_rate": 2.593918235794508e-05, "loss": 0.4819, "step": 650 }, { "epoch": 1.4030010718113612, "grad_norm": 0.9783421158790588, "learning_rate": 2.5866223901354372e-05, "loss": 0.5378, "step": 655 }, { "epoch": 1.4137191854233655, "grad_norm": 1.2764405012130737, "learning_rate": 2.579272044852523e-05, "loss": 0.4782, "step": 660 }, { "epoch": 1.4244372990353698, "grad_norm": 1.0567833185195923, "learning_rate": 2.5718675686028582e-05, "loss": 0.5, "step": 665 }, { "epoch": 1.435155412647374, "grad_norm": 1.150490164756775, "learning_rate": 2.56440933275848e-05, "loss": 0.4415, "step": 670 }, { "epoch": 1.4458735262593785, "grad_norm": 1.0718622207641602, "learning_rate": 2.5568977113877426e-05, "loss": 0.5022, "step": 675 }, { "epoch": 1.4565916398713825, "grad_norm": 1.0396467447280884, "learning_rate": 2.549333081236553e-05, "loss": 0.4582, "step": 680 }, { "epoch": 1.467309753483387, "grad_norm": 1.0076903104782104, "learning_rate": 2.5417158217094788e-05, "loss": 0.4865, "step": 685 }, { "epoch": 1.4780278670953912, "grad_norm": 1.1110858917236328, "learning_rate": 2.5340463148507178e-05, "loss": 0.4707, "step": 690 }, { "epoch": 1.4887459807073955, "grad_norm": 1.0835309028625488, "learning_rate": 2.5263249453249356e-05, "loss": 0.4569, "step": 695 }, { "epoch": 1.4994640943193998, "grad_norm": 1.1380281448364258, "learning_rate": 2.5185521003979746e-05, "loss": 0.4699, "step": 700 }, { "epoch": 1.510182207931404, "grad_norm": 1.2212399244308472, "learning_rate": 2.5107281699174287e-05, "loss": 0.4579, "step": 705 }, { "epoch": 1.5209003215434085, "grad_norm": 1.080427646636963, "learning_rate": 2.5028535462930924e-05, "loss": 0.4509, "step": 710 }, { "epoch": 1.5316184351554125, "grad_norm": 1.0146664381027222, "learning_rate": 2.494928624477277e-05, "loss": 0.5209, "step": 715 }, { "epoch": 1.542336548767417, "grad_norm": 1.023618221282959, "learning_rate": 2.4869538019450045e-05, "loss": 0.459, "step": 720 }, { "epoch": 1.5530546623794212, "grad_norm": 1.0760775804519653, "learning_rate": 2.4789294786740705e-05, "loss": 0.4522, "step": 725 }, { "epoch": 1.5637727759914255, "grad_norm": 1.0976240634918213, "learning_rate": 2.4708560571249845e-05, "loss": 0.4853, "step": 730 }, { "epoch": 1.5744908896034298, "grad_norm": 1.2561285495758057, "learning_rate": 2.4627339422207805e-05, "loss": 0.4584, "step": 735 }, { "epoch": 1.585209003215434, "grad_norm": 1.1148028373718262, "learning_rate": 2.454563541326716e-05, "loss": 0.4492, "step": 740 }, { "epoch": 1.5959271168274385, "grad_norm": 1.085665225982666, "learning_rate": 2.4463452642298324e-05, "loss": 0.4545, "step": 745 }, { "epoch": 1.6066452304394425, "grad_norm": 0.9975435137748718, "learning_rate": 2.438079523118406e-05, "loss": 0.4392, "step": 750 }, { "epoch": 1.617363344051447, "grad_norm": 1.1030360460281372, "learning_rate": 2.4297667325612756e-05, "loss": 0.4255, "step": 755 }, { "epoch": 1.6280814576634512, "grad_norm": 1.02976655960083, "learning_rate": 2.421407309487046e-05, "loss": 0.4452, "step": 760 }, { "epoch": 1.6387995712754555, "grad_norm": 1.0580244064331055, "learning_rate": 2.4130016731631815e-05, "loss": 0.4275, "step": 765 }, { "epoch": 1.6495176848874598, "grad_norm": 1.1036502122879028, "learning_rate": 2.4045502451749736e-05, "loss": 0.4672, "step": 770 }, { "epoch": 1.660235798499464, "grad_norm": 1.1784788370132446, "learning_rate": 2.3960534494043993e-05, "loss": 0.4036, "step": 775 }, { "epoch": 1.6709539121114685, "grad_norm": 1.1079624891281128, "learning_rate": 2.387511712008859e-05, "loss": 0.4294, "step": 780 }, { "epoch": 1.6816720257234725, "grad_norm": 1.0606523752212524, "learning_rate": 2.3789254613998043e-05, "loss": 0.449, "step": 785 }, { "epoch": 1.692390139335477, "grad_norm": 0.9858099818229675, "learning_rate": 2.3702951282212492e-05, "loss": 0.3832, "step": 790 }, { "epoch": 1.7031082529474812, "grad_norm": 1.1735355854034424, "learning_rate": 2.3616211453281726e-05, "loss": 0.445, "step": 795 }, { "epoch": 1.7138263665594855, "grad_norm": 1.083713412284851, "learning_rate": 2.3529039477648087e-05, "loss": 0.4306, "step": 800 }, { "epoch": 1.72454448017149, "grad_norm": 1.1281569004058838, "learning_rate": 2.344143972742826e-05, "loss": 0.3807, "step": 805 }, { "epoch": 1.735262593783494, "grad_norm": 1.0552798509597778, "learning_rate": 2.3353416596193985e-05, "loss": 0.4013, "step": 810 }, { "epoch": 1.7459807073954985, "grad_norm": 1.0200108289718628, "learning_rate": 2.326497449875173e-05, "loss": 0.369, "step": 815 }, { "epoch": 1.7566988210075025, "grad_norm": 1.0658568143844604, "learning_rate": 2.3176117870921218e-05, "loss": 0.3814, "step": 820 }, { "epoch": 1.767416934619507, "grad_norm": 1.164661169052124, "learning_rate": 2.3086851169312992e-05, "loss": 0.4044, "step": 825 }, { "epoch": 1.7781350482315113, "grad_norm": 1.178032636642456, "learning_rate": 2.299717887110487e-05, "loss": 0.3985, "step": 830 }, { "epoch": 1.7888531618435155, "grad_norm": 1.1956833600997925, "learning_rate": 2.290710547381739e-05, "loss": 0.3905, "step": 835 }, { "epoch": 1.79957127545552, "grad_norm": 1.205641269683838, "learning_rate": 2.2816635495088244e-05, "loss": 0.3915, "step": 840 }, { "epoch": 1.810289389067524, "grad_norm": 1.208314299583435, "learning_rate": 2.272577347244571e-05, "loss": 0.403, "step": 845 }, { "epoch": 1.8210075026795285, "grad_norm": 1.1529418230056763, "learning_rate": 2.2634523963081033e-05, "loss": 0.3932, "step": 850 }, { "epoch": 1.8317256162915327, "grad_norm": 1.1101915836334229, "learning_rate": 2.25428915436199e-05, "loss": 0.3763, "step": 855 }, { "epoch": 1.842443729903537, "grad_norm": 1.071894884109497, "learning_rate": 2.245088080989287e-05, "loss": 0.3824, "step": 860 }, { "epoch": 1.8531618435155413, "grad_norm": 1.1817514896392822, "learning_rate": 2.2358496376704898e-05, "loss": 0.4261, "step": 865 }, { "epoch": 1.8638799571275455, "grad_norm": 1.1303304433822632, "learning_rate": 2.2265742877603838e-05, "loss": 0.3604, "step": 870 }, { "epoch": 1.87459807073955, "grad_norm": 1.0623246431350708, "learning_rate": 2.2172624964648094e-05, "loss": 0.4049, "step": 875 }, { "epoch": 1.885316184351554, "grad_norm": 1.171410322189331, "learning_rate": 2.2079147308173258e-05, "loss": 0.3633, "step": 880 }, { "epoch": 1.8960342979635585, "grad_norm": 1.20453941822052, "learning_rate": 2.1985314596557887e-05, "loss": 0.4022, "step": 885 }, { "epoch": 1.9067524115755627, "grad_norm": 1.171126365661621, "learning_rate": 2.1891131535988364e-05, "loss": 0.3967, "step": 890 }, { "epoch": 1.917470525187567, "grad_norm": 1.1599109172821045, "learning_rate": 2.1796602850222832e-05, "loss": 0.349, "step": 895 }, { "epoch": 1.9281886387995713, "grad_norm": 1.056682825088501, "learning_rate": 2.170173328035431e-05, "loss": 0.3705, "step": 900 }, { "epoch": 1.9389067524115755, "grad_norm": 1.1990083456039429, "learning_rate": 2.1606527584572854e-05, "loss": 0.367, "step": 905 }, { "epoch": 1.94962486602358, "grad_norm": 1.1245200634002686, "learning_rate": 2.1510990537926983e-05, "loss": 0.3734, "step": 910 }, { "epoch": 1.960342979635584, "grad_norm": 1.2177960872650146, "learning_rate": 2.1415126932084104e-05, "loss": 0.4062, "step": 915 }, { "epoch": 1.9710610932475885, "grad_norm": 1.1364030838012695, "learning_rate": 2.1318941575090233e-05, "loss": 0.3492, "step": 920 }, { "epoch": 1.9817792068595927, "grad_norm": 1.1492586135864258, "learning_rate": 2.1222439291128855e-05, "loss": 0.3597, "step": 925 }, { "epoch": 1.992497320471597, "grad_norm": 1.0529663562774658, "learning_rate": 2.112562492027893e-05, "loss": 0.3996, "step": 930 } ], "logging_steps": 5, "max_steps": 2335, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.394156316393472e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }