| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 934, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.010718113612004287, | |
| "grad_norm": 1.3203599452972412, | |
| "learning_rate": 1.0256410256410257e-06, | |
| "loss": 1.3095, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.021436227224008574, | |
| "grad_norm": 1.005972981452942, | |
| "learning_rate": 2.307692307692308e-06, | |
| "loss": 1.3756, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03215434083601286, | |
| "grad_norm": 0.8182684183120728, | |
| "learning_rate": 3.58974358974359e-06, | |
| "loss": 1.3238, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.04287245444801715, | |
| "grad_norm": 0.791577160358429, | |
| "learning_rate": 4.871794871794872e-06, | |
| "loss": 1.2938, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05359056806002144, | |
| "grad_norm": 0.6379374861717224, | |
| "learning_rate": 6.153846153846154e-06, | |
| "loss": 1.2856, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.06430868167202572, | |
| "grad_norm": 0.6668587327003479, | |
| "learning_rate": 7.435897435897436e-06, | |
| "loss": 1.2841, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07502679528403002, | |
| "grad_norm": 0.716699481010437, | |
| "learning_rate": 8.717948717948719e-06, | |
| "loss": 1.2718, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.0857449088960343, | |
| "grad_norm": 0.5486516952514648, | |
| "learning_rate": 9.999999999999999e-06, | |
| "loss": 1.273, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.09646302250803858, | |
| "grad_norm": 0.46251392364501953, | |
| "learning_rate": 1.1282051282051283e-05, | |
| "loss": 1.312, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.10718113612004287, | |
| "grad_norm": 0.4610588848590851, | |
| "learning_rate": 1.2564102564102565e-05, | |
| "loss": 1.2015, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.11789924973204716, | |
| "grad_norm": 0.4434472918510437, | |
| "learning_rate": 1.3846153846153847e-05, | |
| "loss": 1.2457, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12861736334405144, | |
| "grad_norm": 0.43525955080986023, | |
| "learning_rate": 1.5128205128205129e-05, | |
| "loss": 1.2111, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.13933547695605572, | |
| "grad_norm": 0.45412635803222656, | |
| "learning_rate": 1.641025641025641e-05, | |
| "loss": 1.2055, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.15005359056806003, | |
| "grad_norm": 0.47974953055381775, | |
| "learning_rate": 1.7692307692307694e-05, | |
| "loss": 1.1836, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1607717041800643, | |
| "grad_norm": 0.5585247278213501, | |
| "learning_rate": 1.8974358974358975e-05, | |
| "loss": 1.1456, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.1714898177920686, | |
| "grad_norm": 0.4740090072154999, | |
| "learning_rate": 2.025641025641026e-05, | |
| "loss": 1.1697, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.18220793140407288, | |
| "grad_norm": 0.49104949831962585, | |
| "learning_rate": 2.153846153846154e-05, | |
| "loss": 1.1723, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.19292604501607716, | |
| "grad_norm": 0.47726204991340637, | |
| "learning_rate": 2.282051282051282e-05, | |
| "loss": 1.1729, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.20364415862808147, | |
| "grad_norm": 0.5089394450187683, | |
| "learning_rate": 2.4102564102564103e-05, | |
| "loss": 1.1313, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.21436227224008575, | |
| "grad_norm": 0.4856749475002289, | |
| "learning_rate": 2.5384615384615386e-05, | |
| "loss": 1.1221, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.22508038585209003, | |
| "grad_norm": 0.4452688992023468, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 1.1334, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.2357984994640943, | |
| "grad_norm": 0.5895005464553833, | |
| "learning_rate": 2.794871794871795e-05, | |
| "loss": 1.1453, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2465166130760986, | |
| "grad_norm": 0.5326249003410339, | |
| "learning_rate": 2.923076923076923e-05, | |
| "loss": 1.0856, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.2572347266881029, | |
| "grad_norm": 0.588718056678772, | |
| "learning_rate": 2.9999939813702703e-05, | |
| "loss": 1.0957, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2679528403001072, | |
| "grad_norm": 0.4917924106121063, | |
| "learning_rate": 2.9999262723404875e-05, | |
| "loss": 1.0265, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.27867095391211144, | |
| "grad_norm": 0.624920666217804, | |
| "learning_rate": 2.9997833344010312e-05, | |
| "loss": 1.0845, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.28938906752411575, | |
| "grad_norm": 0.627967119216919, | |
| "learning_rate": 2.9995651747209637e-05, | |
| "loss": 1.1037, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.30010718113612006, | |
| "grad_norm": 0.5074423551559448, | |
| "learning_rate": 2.9992718042420993e-05, | |
| "loss": 1.0788, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3108252947481243, | |
| "grad_norm": 0.5567290186882019, | |
| "learning_rate": 2.9989032376784556e-05, | |
| "loss": 1.1003, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.3215434083601286, | |
| "grad_norm": 0.5891143083572388, | |
| "learning_rate": 2.998459493515515e-05, | |
| "loss": 1.0162, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3322615219721329, | |
| "grad_norm": 0.6841644048690796, | |
| "learning_rate": 2.9979405940092978e-05, | |
| "loss": 1.0204, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.3429796355841372, | |
| "grad_norm": 0.6039224863052368, | |
| "learning_rate": 2.997346565185246e-05, | |
| "loss": 1.0219, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3536977491961415, | |
| "grad_norm": 0.5960545539855957, | |
| "learning_rate": 2.9966774368369175e-05, | |
| "loss": 1.0065, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.36441586280814575, | |
| "grad_norm": 0.6452131867408752, | |
| "learning_rate": 2.995933242524493e-05, | |
| "loss": 0.9917, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.37513397642015006, | |
| "grad_norm": 0.5966811776161194, | |
| "learning_rate": 2.995114019573091e-05, | |
| "loss": 0.9868, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.3858520900321543, | |
| "grad_norm": 0.6129615306854248, | |
| "learning_rate": 2.9942198090708976e-05, | |
| "loss": 1.034, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3965702036441586, | |
| "grad_norm": 0.6630488038063049, | |
| "learning_rate": 2.993250655867105e-05, | |
| "loss": 1.0303, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.40728831725616294, | |
| "grad_norm": 0.6165331602096558, | |
| "learning_rate": 2.9922066085696613e-05, | |
| "loss": 0.987, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4180064308681672, | |
| "grad_norm": 0.7296874523162842, | |
| "learning_rate": 2.9910877195428335e-05, | |
| "loss": 0.9805, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.4287245444801715, | |
| "grad_norm": 0.6991822123527527, | |
| "learning_rate": 2.989894044904581e-05, | |
| "loss": 0.9377, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.43944265809217575, | |
| "grad_norm": 0.7727455496788025, | |
| "learning_rate": 2.9886256445237405e-05, | |
| "loss": 0.9726, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.45016077170418006, | |
| "grad_norm": 0.7376512289047241, | |
| "learning_rate": 2.987282582017024e-05, | |
| "loss": 0.946, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.4608788853161844, | |
| "grad_norm": 0.819678544998169, | |
| "learning_rate": 2.9858649247458287e-05, | |
| "loss": 0.9861, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.4715969989281886, | |
| "grad_norm": 0.7300285696983337, | |
| "learning_rate": 2.984372743812855e-05, | |
| "loss": 1.0026, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.48231511254019294, | |
| "grad_norm": 0.7369091510772705, | |
| "learning_rate": 2.9828061140585463e-05, | |
| "loss": 0.9064, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4930332261521972, | |
| "grad_norm": 0.6813873052597046, | |
| "learning_rate": 2.9811651140573287e-05, | |
| "loss": 0.8859, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5037513397642015, | |
| "grad_norm": 0.7347345352172852, | |
| "learning_rate": 2.979449826113676e-05, | |
| "loss": 0.8983, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.5144694533762058, | |
| "grad_norm": 0.8283259272575378, | |
| "learning_rate": 2.9776603362579775e-05, | |
| "loss": 0.8876, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5251875669882101, | |
| "grad_norm": 0.7388147711753845, | |
| "learning_rate": 2.9757967342422246e-05, | |
| "loss": 0.899, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.5359056806002144, | |
| "grad_norm": 0.7657457590103149, | |
| "learning_rate": 2.9738591135355108e-05, | |
| "loss": 0.9412, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5466237942122186, | |
| "grad_norm": 0.8804264664649963, | |
| "learning_rate": 2.971847571319341e-05, | |
| "loss": 0.8838, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.5573419078242229, | |
| "grad_norm": 0.771865963935852, | |
| "learning_rate": 2.969762208482759e-05, | |
| "loss": 0.8697, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5680600214362272, | |
| "grad_norm": 0.7006550431251526, | |
| "learning_rate": 2.967603129617288e-05, | |
| "loss": 0.8801, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.5787781350482315, | |
| "grad_norm": 0.7909011244773865, | |
| "learning_rate": 2.9653704430116827e-05, | |
| "loss": 0.9199, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5894962486602358, | |
| "grad_norm": 0.9315905570983887, | |
| "learning_rate": 2.9630642606464998e-05, | |
| "loss": 0.8249, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6002143622722401, | |
| "grad_norm": 0.7528437376022339, | |
| "learning_rate": 2.9606846981884812e-05, | |
| "loss": 0.8656, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6109324758842444, | |
| "grad_norm": 0.7914732098579407, | |
| "learning_rate": 2.958231874984753e-05, | |
| "loss": 0.8618, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.6216505894962486, | |
| "grad_norm": 0.7694675922393799, | |
| "learning_rate": 2.955705914056838e-05, | |
| "loss": 0.8546, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.632368703108253, | |
| "grad_norm": 0.8651600480079651, | |
| "learning_rate": 2.9531069420944876e-05, | |
| "loss": 0.8654, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.6430868167202572, | |
| "grad_norm": 0.7878230214118958, | |
| "learning_rate": 2.9504350894493272e-05, | |
| "loss": 0.8624, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6538049303322615, | |
| "grad_norm": 0.9521771669387817, | |
| "learning_rate": 2.9476904901283175e-05, | |
| "loss": 0.8175, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.6645230439442658, | |
| "grad_norm": 0.8132469058036804, | |
| "learning_rate": 2.944873281787034e-05, | |
| "loss": 0.8087, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.6752411575562701, | |
| "grad_norm": 0.8626110553741455, | |
| "learning_rate": 2.9419836057227634e-05, | |
| "loss": 0.8033, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.6859592711682744, | |
| "grad_norm": 0.858932614326477, | |
| "learning_rate": 2.9390216068674154e-05, | |
| "loss": 0.7535, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.6966773847802786, | |
| "grad_norm": 1.0335079431533813, | |
| "learning_rate": 2.9359874337802542e-05, | |
| "loss": 0.8272, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.707395498392283, | |
| "grad_norm": 0.819476842880249, | |
| "learning_rate": 2.9328812386404503e-05, | |
| "loss": 0.8184, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7181136120042872, | |
| "grad_norm": 0.9225292205810547, | |
| "learning_rate": 2.9297031772394412e-05, | |
| "loss": 0.8009, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7288317256162915, | |
| "grad_norm": 0.9197941422462463, | |
| "learning_rate": 2.9264534089731262e-05, | |
| "loss": 0.7841, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7395498392282959, | |
| "grad_norm": 0.8611735701560974, | |
| "learning_rate": 2.9231320968338644e-05, | |
| "loss": 0.7975, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.7502679528403001, | |
| "grad_norm": 0.8244771361351013, | |
| "learning_rate": 2.9197394074023042e-05, | |
| "loss": 0.7895, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7609860664523044, | |
| "grad_norm": 0.8496798872947693, | |
| "learning_rate": 2.9162755108390278e-05, | |
| "loss": 0.7931, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.7717041800643086, | |
| "grad_norm": 0.9810552000999451, | |
| "learning_rate": 2.9127405808760152e-05, | |
| "loss": 0.7507, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.782422293676313, | |
| "grad_norm": 0.9915773272514343, | |
| "learning_rate": 2.9091347948079318e-05, | |
| "loss": 0.8251, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.7931404072883173, | |
| "grad_norm": 1.0386714935302734, | |
| "learning_rate": 2.9054583334832366e-05, | |
| "loss": 0.8063, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.8038585209003215, | |
| "grad_norm": 0.9649940729141235, | |
| "learning_rate": 2.9017113812951105e-05, | |
| "loss": 0.7797, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.8145766345123259, | |
| "grad_norm": 1.0805562734603882, | |
| "learning_rate": 2.897894126172209e-05, | |
| "loss": 0.7873, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8252947481243301, | |
| "grad_norm": 0.9175759553909302, | |
| "learning_rate": 2.8940067595692367e-05, | |
| "loss": 0.7306, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8360128617363344, | |
| "grad_norm": 1.0049532651901245, | |
| "learning_rate": 2.890049476457344e-05, | |
| "loss": 0.7357, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8467309753483387, | |
| "grad_norm": 0.8982154726982117, | |
| "learning_rate": 2.8860224753143486e-05, | |
| "loss": 0.7478, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.857449088960343, | |
| "grad_norm": 0.898063600063324, | |
| "learning_rate": 2.881925958114782e-05, | |
| "loss": 0.7576, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8681672025723473, | |
| "grad_norm": 0.9208119511604309, | |
| "learning_rate": 2.877760130319757e-05, | |
| "loss": 0.7413, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.8788853161843515, | |
| "grad_norm": 0.9576625823974609, | |
| "learning_rate": 2.8735252008666663e-05, | |
| "loss": 0.6863, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.8896034297963559, | |
| "grad_norm": 0.9474856853485107, | |
| "learning_rate": 2.8692213821586988e-05, | |
| "loss": 0.7466, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.9003215434083601, | |
| "grad_norm": 1.0231027603149414, | |
| "learning_rate": 2.8648488900541915e-05, | |
| "loss": 0.7399, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9110396570203644, | |
| "grad_norm": 0.9832124710083008, | |
| "learning_rate": 2.8604079438557985e-05, | |
| "loss": 0.7037, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9217577706323687, | |
| "grad_norm": 1.0980058908462524, | |
| "learning_rate": 2.8558987662994948e-05, | |
| "loss": 0.7189, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.932475884244373, | |
| "grad_norm": 1.1283589601516724, | |
| "learning_rate": 2.851321583543404e-05, | |
| "loss": 0.6951, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.9431939978563773, | |
| "grad_norm": 0.9087287187576294, | |
| "learning_rate": 2.8466766251564554e-05, | |
| "loss": 0.6932, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.9539121114683816, | |
| "grad_norm": 1.0392457246780396, | |
| "learning_rate": 2.8419641241068687e-05, | |
| "loss": 0.7065, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.9646302250803859, | |
| "grad_norm": 0.9226632118225098, | |
| "learning_rate": 2.837184316750473e-05, | |
| "loss": 0.7066, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9753483386923901, | |
| "grad_norm": 0.9685462713241577, | |
| "learning_rate": 2.832337442818848e-05, | |
| "loss": 0.6805, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.9860664523043944, | |
| "grad_norm": 0.9703472852706909, | |
| "learning_rate": 2.8274237454073034e-05, | |
| "loss": 0.668, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.9967845659163987, | |
| "grad_norm": 0.902520477771759, | |
| "learning_rate": 2.822443470962685e-05, | |
| "loss": 0.6379, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.0064308681672025, | |
| "grad_norm": 1.2527391910552979, | |
| "learning_rate": 2.8173968692710142e-05, | |
| "loss": 0.6063, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.0171489817792068, | |
| "grad_norm": 1.0625535249710083, | |
| "learning_rate": 2.8122841934449607e-05, | |
| "loss": 0.5916, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.0278670953912112, | |
| "grad_norm": 1.1103384494781494, | |
| "learning_rate": 2.807105699911147e-05, | |
| "loss": 0.6323, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.0385852090032155, | |
| "grad_norm": 1.0214422941207886, | |
| "learning_rate": 2.8018616483972874e-05, | |
| "loss": 0.573, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.0493033226152197, | |
| "grad_norm": 0.9643238186836243, | |
| "learning_rate": 2.7965523019191606e-05, | |
| "loss": 0.5675, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.060021436227224, | |
| "grad_norm": 0.9685465097427368, | |
| "learning_rate": 2.7911779267674207e-05, | |
| "loss": 0.5708, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.0707395498392283, | |
| "grad_norm": 1.0887079238891602, | |
| "learning_rate": 2.7857387924942387e-05, | |
| "loss": 0.5954, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.0814576634512325, | |
| "grad_norm": 0.9755250811576843, | |
| "learning_rate": 2.7802351718997825e-05, | |
| "loss": 0.5384, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.092175777063237, | |
| "grad_norm": 0.9574642181396484, | |
| "learning_rate": 2.7746673410185387e-05, | |
| "loss": 0.5627, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.1028938906752412, | |
| "grad_norm": 1.0391875505447388, | |
| "learning_rate": 2.7690355791054632e-05, | |
| "loss": 0.5621, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.1136120042872455, | |
| "grad_norm": 0.9786142706871033, | |
| "learning_rate": 2.7633401686219793e-05, | |
| "loss": 0.5991, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.1243301178992497, | |
| "grad_norm": 1.0193188190460205, | |
| "learning_rate": 2.7575813952218075e-05, | |
| "loss": 0.5723, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.135048231511254, | |
| "grad_norm": 1.0080389976501465, | |
| "learning_rate": 2.75175954773664e-05, | |
| "loss": 0.5206, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.1457663451232583, | |
| "grad_norm": 0.9439852833747864, | |
| "learning_rate": 2.7458749181616545e-05, | |
| "loss": 0.571, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.1564844587352625, | |
| "grad_norm": 0.9455674290657043, | |
| "learning_rate": 2.7399278016408695e-05, | |
| "loss": 0.547, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.167202572347267, | |
| "grad_norm": 1.138099193572998, | |
| "learning_rate": 2.7339184964523404e-05, | |
| "loss": 0.5942, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.1779206859592712, | |
| "grad_norm": 1.1031590700149536, | |
| "learning_rate": 2.7278473039931992e-05, | |
| "loss": 0.5378, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.1886387995712755, | |
| "grad_norm": 1.1105057001113892, | |
| "learning_rate": 2.7217145287645393e-05, | |
| "loss": 0.5186, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.1993569131832797, | |
| "grad_norm": 1.2202413082122803, | |
| "learning_rate": 2.715520478356141e-05, | |
| "loss": 0.5721, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.210075026795284, | |
| "grad_norm": 0.9721954464912415, | |
| "learning_rate": 2.7092654634310468e-05, | |
| "loss": 0.5433, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.2207931404072883, | |
| "grad_norm": 1.2260032892227173, | |
| "learning_rate": 2.7029497977099784e-05, | |
| "loss": 0.543, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.2315112540192925, | |
| "grad_norm": 0.9644677639007568, | |
| "learning_rate": 2.6965737979556025e-05, | |
| "loss": 0.552, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.242229367631297, | |
| "grad_norm": 0.9937833547592163, | |
| "learning_rate": 2.6901377839566443e-05, | |
| "loss": 0.535, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.2529474812433012, | |
| "grad_norm": 1.031482219696045, | |
| "learning_rate": 2.683642078511846e-05, | |
| "loss": 0.5322, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.2636655948553055, | |
| "grad_norm": 0.9676262140274048, | |
| "learning_rate": 2.6770870074137808e-05, | |
| "loss": 0.5006, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.2743837084673098, | |
| "grad_norm": 1.3593294620513916, | |
| "learning_rate": 2.670472899432509e-05, | |
| "loss": 0.5064, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.285101822079314, | |
| "grad_norm": 1.0694024562835693, | |
| "learning_rate": 2.66380008629909e-05, | |
| "loss": 0.5848, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.2958199356913183, | |
| "grad_norm": 1.0443016290664673, | |
| "learning_rate": 2.657068902688945e-05, | |
| "loss": 0.538, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.3065380493033225, | |
| "grad_norm": 1.023464560508728, | |
| "learning_rate": 2.65027968620507e-05, | |
| "loss": 0.5587, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.317256162915327, | |
| "grad_norm": 0.9948901534080505, | |
| "learning_rate": 2.643432777361105e-05, | |
| "loss": 0.534, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.3279742765273312, | |
| "grad_norm": 1.0392403602600098, | |
| "learning_rate": 2.636528519564253e-05, | |
| "loss": 0.5365, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.3386923901393355, | |
| "grad_norm": 1.0413520336151123, | |
| "learning_rate": 2.6295672590980586e-05, | |
| "loss": 0.4708, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.3494105037513398, | |
| "grad_norm": 1.1629996299743652, | |
| "learning_rate": 2.6225493451050397e-05, | |
| "loss": 0.5554, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.360128617363344, | |
| "grad_norm": 1.023750901222229, | |
| "learning_rate": 2.6154751295691756e-05, | |
| "loss": 0.4827, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.3708467309753483, | |
| "grad_norm": 1.13310706615448, | |
| "learning_rate": 2.608344967298253e-05, | |
| "loss": 0.4657, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.3815648445873525, | |
| "grad_norm": 1.083983063697815, | |
| "learning_rate": 2.6011592159060717e-05, | |
| "loss": 0.5213, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.392282958199357, | |
| "grad_norm": 1.048169493675232, | |
| "learning_rate": 2.593918235794508e-05, | |
| "loss": 0.4819, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.4030010718113612, | |
| "grad_norm": 0.9783421158790588, | |
| "learning_rate": 2.5866223901354372e-05, | |
| "loss": 0.5378, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.4137191854233655, | |
| "grad_norm": 1.2764405012130737, | |
| "learning_rate": 2.579272044852523e-05, | |
| "loss": 0.4782, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.4244372990353698, | |
| "grad_norm": 1.0567833185195923, | |
| "learning_rate": 2.5718675686028582e-05, | |
| "loss": 0.5, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.435155412647374, | |
| "grad_norm": 1.150490164756775, | |
| "learning_rate": 2.56440933275848e-05, | |
| "loss": 0.4415, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.4458735262593785, | |
| "grad_norm": 1.0718622207641602, | |
| "learning_rate": 2.5568977113877426e-05, | |
| "loss": 0.5022, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.4565916398713825, | |
| "grad_norm": 1.0396467447280884, | |
| "learning_rate": 2.549333081236553e-05, | |
| "loss": 0.4582, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.467309753483387, | |
| "grad_norm": 1.0076903104782104, | |
| "learning_rate": 2.5417158217094788e-05, | |
| "loss": 0.4865, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.4780278670953912, | |
| "grad_norm": 1.1110858917236328, | |
| "learning_rate": 2.5340463148507178e-05, | |
| "loss": 0.4707, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.4887459807073955, | |
| "grad_norm": 1.0835309028625488, | |
| "learning_rate": 2.5263249453249356e-05, | |
| "loss": 0.4569, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.4994640943193998, | |
| "grad_norm": 1.1380281448364258, | |
| "learning_rate": 2.5185521003979746e-05, | |
| "loss": 0.4699, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.510182207931404, | |
| "grad_norm": 1.2212399244308472, | |
| "learning_rate": 2.5107281699174287e-05, | |
| "loss": 0.4579, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.5209003215434085, | |
| "grad_norm": 1.080427646636963, | |
| "learning_rate": 2.5028535462930924e-05, | |
| "loss": 0.4509, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.5316184351554125, | |
| "grad_norm": 1.0146664381027222, | |
| "learning_rate": 2.494928624477277e-05, | |
| "loss": 0.5209, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.542336548767417, | |
| "grad_norm": 1.023618221282959, | |
| "learning_rate": 2.4869538019450045e-05, | |
| "loss": 0.459, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.5530546623794212, | |
| "grad_norm": 1.0760775804519653, | |
| "learning_rate": 2.4789294786740705e-05, | |
| "loss": 0.4522, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.5637727759914255, | |
| "grad_norm": 1.0976240634918213, | |
| "learning_rate": 2.4708560571249845e-05, | |
| "loss": 0.4853, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.5744908896034298, | |
| "grad_norm": 1.2561285495758057, | |
| "learning_rate": 2.4627339422207805e-05, | |
| "loss": 0.4584, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.585209003215434, | |
| "grad_norm": 1.1148028373718262, | |
| "learning_rate": 2.454563541326716e-05, | |
| "loss": 0.4492, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.5959271168274385, | |
| "grad_norm": 1.085665225982666, | |
| "learning_rate": 2.4463452642298324e-05, | |
| "loss": 0.4545, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.6066452304394425, | |
| "grad_norm": 0.9975435137748718, | |
| "learning_rate": 2.438079523118406e-05, | |
| "loss": 0.4392, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.617363344051447, | |
| "grad_norm": 1.1030360460281372, | |
| "learning_rate": 2.4297667325612756e-05, | |
| "loss": 0.4255, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.6280814576634512, | |
| "grad_norm": 1.02976655960083, | |
| "learning_rate": 2.421407309487046e-05, | |
| "loss": 0.4452, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.6387995712754555, | |
| "grad_norm": 1.0580244064331055, | |
| "learning_rate": 2.4130016731631815e-05, | |
| "loss": 0.4275, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.6495176848874598, | |
| "grad_norm": 1.1036502122879028, | |
| "learning_rate": 2.4045502451749736e-05, | |
| "loss": 0.4672, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.660235798499464, | |
| "grad_norm": 1.1784788370132446, | |
| "learning_rate": 2.3960534494043993e-05, | |
| "loss": 0.4036, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.6709539121114685, | |
| "grad_norm": 1.1079624891281128, | |
| "learning_rate": 2.387511712008859e-05, | |
| "loss": 0.4294, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.6816720257234725, | |
| "grad_norm": 1.0606523752212524, | |
| "learning_rate": 2.3789254613998043e-05, | |
| "loss": 0.449, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.692390139335477, | |
| "grad_norm": 0.9858099818229675, | |
| "learning_rate": 2.3702951282212492e-05, | |
| "loss": 0.3832, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.7031082529474812, | |
| "grad_norm": 1.1735355854034424, | |
| "learning_rate": 2.3616211453281726e-05, | |
| "loss": 0.445, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.7138263665594855, | |
| "grad_norm": 1.083713412284851, | |
| "learning_rate": 2.3529039477648087e-05, | |
| "loss": 0.4306, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.72454448017149, | |
| "grad_norm": 1.1281569004058838, | |
| "learning_rate": 2.344143972742826e-05, | |
| "loss": 0.3807, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.735262593783494, | |
| "grad_norm": 1.0552798509597778, | |
| "learning_rate": 2.3353416596193985e-05, | |
| "loss": 0.4013, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.7459807073954985, | |
| "grad_norm": 1.0200108289718628, | |
| "learning_rate": 2.326497449875173e-05, | |
| "loss": 0.369, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.7566988210075025, | |
| "grad_norm": 1.0658568143844604, | |
| "learning_rate": 2.3176117870921218e-05, | |
| "loss": 0.3814, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.767416934619507, | |
| "grad_norm": 1.164661169052124, | |
| "learning_rate": 2.3086851169312992e-05, | |
| "loss": 0.4044, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.7781350482315113, | |
| "grad_norm": 1.178032636642456, | |
| "learning_rate": 2.299717887110487e-05, | |
| "loss": 0.3985, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.7888531618435155, | |
| "grad_norm": 1.1956833600997925, | |
| "learning_rate": 2.290710547381739e-05, | |
| "loss": 0.3905, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.79957127545552, | |
| "grad_norm": 1.205641269683838, | |
| "learning_rate": 2.2816635495088244e-05, | |
| "loss": 0.3915, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.810289389067524, | |
| "grad_norm": 1.208314299583435, | |
| "learning_rate": 2.272577347244571e-05, | |
| "loss": 0.403, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.8210075026795285, | |
| "grad_norm": 1.1529418230056763, | |
| "learning_rate": 2.2634523963081033e-05, | |
| "loss": 0.3932, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.8317256162915327, | |
| "grad_norm": 1.1101915836334229, | |
| "learning_rate": 2.25428915436199e-05, | |
| "loss": 0.3763, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.842443729903537, | |
| "grad_norm": 1.071894884109497, | |
| "learning_rate": 2.245088080989287e-05, | |
| "loss": 0.3824, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.8531618435155413, | |
| "grad_norm": 1.1817514896392822, | |
| "learning_rate": 2.2358496376704898e-05, | |
| "loss": 0.4261, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.8638799571275455, | |
| "grad_norm": 1.1303304433822632, | |
| "learning_rate": 2.2265742877603838e-05, | |
| "loss": 0.3604, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.87459807073955, | |
| "grad_norm": 1.0623246431350708, | |
| "learning_rate": 2.2172624964648094e-05, | |
| "loss": 0.4049, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.885316184351554, | |
| "grad_norm": 1.171410322189331, | |
| "learning_rate": 2.2079147308173258e-05, | |
| "loss": 0.3633, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.8960342979635585, | |
| "grad_norm": 1.20453941822052, | |
| "learning_rate": 2.1985314596557887e-05, | |
| "loss": 0.4022, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.9067524115755627, | |
| "grad_norm": 1.171126365661621, | |
| "learning_rate": 2.1891131535988364e-05, | |
| "loss": 0.3967, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.917470525187567, | |
| "grad_norm": 1.1599109172821045, | |
| "learning_rate": 2.1796602850222832e-05, | |
| "loss": 0.349, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.9281886387995713, | |
| "grad_norm": 1.056682825088501, | |
| "learning_rate": 2.170173328035431e-05, | |
| "loss": 0.3705, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.9389067524115755, | |
| "grad_norm": 1.1990083456039429, | |
| "learning_rate": 2.1606527584572854e-05, | |
| "loss": 0.367, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.94962486602358, | |
| "grad_norm": 1.1245200634002686, | |
| "learning_rate": 2.1510990537926983e-05, | |
| "loss": 0.3734, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.960342979635584, | |
| "grad_norm": 1.2177960872650146, | |
| "learning_rate": 2.1415126932084104e-05, | |
| "loss": 0.4062, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.9710610932475885, | |
| "grad_norm": 1.1364030838012695, | |
| "learning_rate": 2.1318941575090233e-05, | |
| "loss": 0.3492, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.9817792068595927, | |
| "grad_norm": 1.1492586135864258, | |
| "learning_rate": 2.1222439291128855e-05, | |
| "loss": 0.3597, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.992497320471597, | |
| "grad_norm": 1.0529663562774658, | |
| "learning_rate": 2.112562492027893e-05, | |
| "loss": 0.3996, | |
| "step": 930 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 2335, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 2000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.394156316393472e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |