| { | |
| "best_global_step": 7028, | |
| "best_metric": 0.767158567905426, | |
| "best_model_checkpoint": "./mcqa_qwen3_letter/checkpoint-7028", | |
| "epoch": 0.9998932953939178, | |
| "eval_steps": 500, | |
| "global_step": 7028, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.007113640405477503, | |
| "grad_norm": 20.374778747558594, | |
| "learning_rate": 1.337126600284495e-06, | |
| "loss": 2.342, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.014227280810955006, | |
| "grad_norm": 23.720001220703125, | |
| "learning_rate": 2.7596017069701283e-06, | |
| "loss": 1.6092, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.02134092121643251, | |
| "grad_norm": 14.003034591674805, | |
| "learning_rate": 4.182076813655762e-06, | |
| "loss": 1.4564, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.028454561621910013, | |
| "grad_norm": 11.145376205444336, | |
| "learning_rate": 5.604551920341395e-06, | |
| "loss": 1.481, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.035568202027387516, | |
| "grad_norm": 14.064179420471191, | |
| "learning_rate": 7.027027027027028e-06, | |
| "loss": 1.4483, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.04268184243286502, | |
| "grad_norm": 14.6947021484375, | |
| "learning_rate": 8.44950213371266e-06, | |
| "loss": 1.4355, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.04979548283834252, | |
| "grad_norm": 11.520552635192871, | |
| "learning_rate": 9.871977240398294e-06, | |
| "loss": 1.4314, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.056909123243820026, | |
| "grad_norm": 11.234253883361816, | |
| "learning_rate": 1.1294452347083926e-05, | |
| "loss": 1.3494, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.06402276364929753, | |
| "grad_norm": 9.99629020690918, | |
| "learning_rate": 1.271692745376956e-05, | |
| "loss": 1.3342, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.07113640405477503, | |
| "grad_norm": 8.978431701660156, | |
| "learning_rate": 1.4139402560455193e-05, | |
| "loss": 1.2951, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.07825004446025254, | |
| "grad_norm": 9.42021369934082, | |
| "learning_rate": 1.5561877667140826e-05, | |
| "loss": 1.3058, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.08536368486573004, | |
| "grad_norm": 9.690632820129395, | |
| "learning_rate": 1.698435277382646e-05, | |
| "loss": 1.2561, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.09247732527120754, | |
| "grad_norm": 10.151556015014648, | |
| "learning_rate": 1.8406827880512092e-05, | |
| "loss": 1.2415, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.09959096567668504, | |
| "grad_norm": 8.446418762207031, | |
| "learning_rate": 1.9829302987197725e-05, | |
| "loss": 1.3303, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.10670460608216255, | |
| "grad_norm": 9.451847076416016, | |
| "learning_rate": 1.9860869565217395e-05, | |
| "loss": 1.2735, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.11381824648764005, | |
| "grad_norm": 9.343132019042969, | |
| "learning_rate": 1.9702766798418973e-05, | |
| "loss": 1.3471, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.12093188689311755, | |
| "grad_norm": 9.069653511047363, | |
| "learning_rate": 1.9544664031620554e-05, | |
| "loss": 1.2535, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.12804552729859506, | |
| "grad_norm": 5.554582595825195, | |
| "learning_rate": 1.9386561264822135e-05, | |
| "loss": 1.2887, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.13515916770407255, | |
| "grad_norm": 8.713581085205078, | |
| "learning_rate": 1.9228458498023717e-05, | |
| "loss": 1.1681, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.14227280810955006, | |
| "grad_norm": 8.742165565490723, | |
| "learning_rate": 1.9070355731225298e-05, | |
| "loss": 1.2271, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.14938644851502755, | |
| "grad_norm": 8.755561828613281, | |
| "learning_rate": 1.891225296442688e-05, | |
| "loss": 1.2557, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.15650008892050507, | |
| "grad_norm": 7.881649017333984, | |
| "learning_rate": 1.875415019762846e-05, | |
| "loss": 1.1999, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.16361372932598256, | |
| "grad_norm": 9.955330848693848, | |
| "learning_rate": 1.8596047430830042e-05, | |
| "loss": 1.1375, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.17072736973146008, | |
| "grad_norm": 9.453360557556152, | |
| "learning_rate": 1.843794466403162e-05, | |
| "loss": 1.2173, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.17784101013693757, | |
| "grad_norm": 6.806760787963867, | |
| "learning_rate": 1.8279841897233205e-05, | |
| "loss": 1.2116, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.18495465054241508, | |
| "grad_norm": 13.604923248291016, | |
| "learning_rate": 1.8121739130434783e-05, | |
| "loss": 1.0788, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.19206829094789257, | |
| "grad_norm": 10.025015830993652, | |
| "learning_rate": 1.7963636363636364e-05, | |
| "loss": 1.149, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.1991819313533701, | |
| "grad_norm": 9.246174812316895, | |
| "learning_rate": 1.7805533596837945e-05, | |
| "loss": 1.1171, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.20629557175884758, | |
| "grad_norm": 9.556463241577148, | |
| "learning_rate": 1.7647430830039527e-05, | |
| "loss": 1.1744, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.2134092121643251, | |
| "grad_norm": 13.477895736694336, | |
| "learning_rate": 1.7489328063241108e-05, | |
| "loss": 1.0723, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.22052285256980259, | |
| "grad_norm": 7.583223342895508, | |
| "learning_rate": 1.733122529644269e-05, | |
| "loss": 1.1004, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.2276364929752801, | |
| "grad_norm": 9.056886672973633, | |
| "learning_rate": 1.717312252964427e-05, | |
| "loss": 1.1413, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.2347501333807576, | |
| "grad_norm": 8.818636894226074, | |
| "learning_rate": 1.701501976284585e-05, | |
| "loss": 1.0248, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.2418637737862351, | |
| "grad_norm": 11.717618942260742, | |
| "learning_rate": 1.6856916996047434e-05, | |
| "loss": 1.0386, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.2489774141917126, | |
| "grad_norm": 7.414971828460693, | |
| "learning_rate": 1.6698814229249015e-05, | |
| "loss": 1.0885, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.2560910545971901, | |
| "grad_norm": 6.960204601287842, | |
| "learning_rate": 1.6540711462450593e-05, | |
| "loss": 1.053, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.26320469500266763, | |
| "grad_norm": 6.861100673675537, | |
| "learning_rate": 1.6382608695652174e-05, | |
| "loss": 1.0398, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.2703183354081451, | |
| "grad_norm": 7.645370960235596, | |
| "learning_rate": 1.6224505928853756e-05, | |
| "loss": 0.9453, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.2774319758136226, | |
| "grad_norm": 7.169046878814697, | |
| "learning_rate": 1.6066403162055337e-05, | |
| "loss": 0.9847, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.28454561621910013, | |
| "grad_norm": 6.290506362915039, | |
| "learning_rate": 1.5908300395256918e-05, | |
| "loss": 0.9592, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.29165925662457765, | |
| "grad_norm": 8.001459121704102, | |
| "learning_rate": 1.57501976284585e-05, | |
| "loss": 0.9674, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.2987728970300551, | |
| "grad_norm": 8.543235778808594, | |
| "learning_rate": 1.559525691699605e-05, | |
| "loss": 1.003, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.3058865374355326, | |
| "grad_norm": 10.702069282531738, | |
| "learning_rate": 1.543715415019763e-05, | |
| "loss": 0.9914, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.31300017784101014, | |
| "grad_norm": 5.234061241149902, | |
| "learning_rate": 1.527905138339921e-05, | |
| "loss": 1.0809, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.32011381824648766, | |
| "grad_norm": 8.00466251373291, | |
| "learning_rate": 1.5120948616600793e-05, | |
| "loss": 0.9713, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.3272274586519651, | |
| "grad_norm": 7.484270095825195, | |
| "learning_rate": 1.4962845849802372e-05, | |
| "loss": 0.906, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.33434109905744264, | |
| "grad_norm": 17.150543212890625, | |
| "learning_rate": 1.4804743083003954e-05, | |
| "loss": 1.0206, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.34145473946292015, | |
| "grad_norm": 8.58785343170166, | |
| "learning_rate": 1.4646640316205533e-05, | |
| "loss": 0.9868, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.34856837986839767, | |
| "grad_norm": 7.040744304656982, | |
| "learning_rate": 1.4488537549407116e-05, | |
| "loss": 0.9225, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.35568202027387513, | |
| "grad_norm": 8.559785842895508, | |
| "learning_rate": 1.4330434782608698e-05, | |
| "loss": 0.9345, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.36279566067935265, | |
| "grad_norm": 9.140971183776855, | |
| "learning_rate": 1.4172332015810277e-05, | |
| "loss": 1.004, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.36990930108483017, | |
| "grad_norm": 8.379377365112305, | |
| "learning_rate": 1.4014229249011859e-05, | |
| "loss": 1.0199, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.3770229414903077, | |
| "grad_norm": 6.753004550933838, | |
| "learning_rate": 1.3856126482213438e-05, | |
| "loss": 0.9463, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.38413658189578515, | |
| "grad_norm": 7.884097576141357, | |
| "learning_rate": 1.3698023715415021e-05, | |
| "loss": 0.9367, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.39125022230126266, | |
| "grad_norm": 8.405110359191895, | |
| "learning_rate": 1.3539920948616603e-05, | |
| "loss": 0.9036, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.3983638627067402, | |
| "grad_norm": 7.3707661628723145, | |
| "learning_rate": 1.3381818181818182e-05, | |
| "loss": 0.9868, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.4054775031122177, | |
| "grad_norm": 9.472613334655762, | |
| "learning_rate": 1.3223715415019764e-05, | |
| "loss": 0.9809, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.41259114351769516, | |
| "grad_norm": 8.191240310668945, | |
| "learning_rate": 1.3065612648221343e-05, | |
| "loss": 0.9249, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.4197047839231727, | |
| "grad_norm": 7.100245952606201, | |
| "learning_rate": 1.2907509881422926e-05, | |
| "loss": 0.9303, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.4268184243286502, | |
| "grad_norm": 6.30331563949585, | |
| "learning_rate": 1.2749407114624508e-05, | |
| "loss": 0.9459, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.4339320647341277, | |
| "grad_norm": 5.755526065826416, | |
| "learning_rate": 1.2591304347826087e-05, | |
| "loss": 0.8818, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.44104570513960517, | |
| "grad_norm": 7.514272689819336, | |
| "learning_rate": 1.243320158102767e-05, | |
| "loss": 0.8817, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.4481593455450827, | |
| "grad_norm": 6.345920562744141, | |
| "learning_rate": 1.227509881422925e-05, | |
| "loss": 0.8257, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.4552729859505602, | |
| "grad_norm": 5.915317058563232, | |
| "learning_rate": 1.2120158102766799e-05, | |
| "loss": 0.9691, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.4623866263560377, | |
| "grad_norm": 4.785401821136475, | |
| "learning_rate": 1.1962055335968382e-05, | |
| "loss": 0.97, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.4695002667615152, | |
| "grad_norm": 8.403907775878906, | |
| "learning_rate": 1.1803952569169962e-05, | |
| "loss": 0.8699, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.4766139071669927, | |
| "grad_norm": 8.583006858825684, | |
| "learning_rate": 1.1645849802371543e-05, | |
| "loss": 0.9233, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.4837275475724702, | |
| "grad_norm": 7.836755752563477, | |
| "learning_rate": 1.1487747035573123e-05, | |
| "loss": 0.8716, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.49084118797794773, | |
| "grad_norm": 4.714875221252441, | |
| "learning_rate": 1.1329644268774704e-05, | |
| "loss": 0.8499, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.4979548283834252, | |
| "grad_norm": 7.097923755645752, | |
| "learning_rate": 1.1171541501976287e-05, | |
| "loss": 0.912, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.5050684687889028, | |
| "grad_norm": 4.853347301483154, | |
| "learning_rate": 1.1013438735177867e-05, | |
| "loss": 0.8411, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.5121821091943802, | |
| "grad_norm": 6.642279148101807, | |
| "learning_rate": 1.0855335968379448e-05, | |
| "loss": 0.9325, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.5192957495998577, | |
| "grad_norm": 6.702445030212402, | |
| "learning_rate": 1.0697233201581028e-05, | |
| "loss": 0.8354, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.5264093900053353, | |
| "grad_norm": 6.634207725524902, | |
| "learning_rate": 1.0539130434782609e-05, | |
| "loss": 0.8813, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.5335230304108127, | |
| "grad_norm": 4.122372627258301, | |
| "learning_rate": 1.0381027667984192e-05, | |
| "loss": 0.7811, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.5406366708162902, | |
| "grad_norm": 8.899633407592773, | |
| "learning_rate": 1.0222924901185772e-05, | |
| "loss": 0.8707, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.5477503112217678, | |
| "grad_norm": 5.374997615814209, | |
| "learning_rate": 1.0064822134387353e-05, | |
| "loss": 0.8854, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.5548639516272452, | |
| "grad_norm": 7.9275102615356445, | |
| "learning_rate": 9.906719367588934e-06, | |
| "loss": 0.9255, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.5619775920327228, | |
| "grad_norm": 6.925314903259277, | |
| "learning_rate": 9.748616600790514e-06, | |
| "loss": 0.9027, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.5690912324382003, | |
| "grad_norm": 9.117793083190918, | |
| "learning_rate": 9.590513833992095e-06, | |
| "loss": 0.8831, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.5762048728436777, | |
| "grad_norm": 5.861302852630615, | |
| "learning_rate": 9.432411067193677e-06, | |
| "loss": 0.8329, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.5833185132491553, | |
| "grad_norm": 6.011074066162109, | |
| "learning_rate": 9.274308300395256e-06, | |
| "loss": 0.8766, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.5904321536546328, | |
| "grad_norm": 5.763038158416748, | |
| "learning_rate": 9.11620553359684e-06, | |
| "loss": 0.9374, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.5975457940601102, | |
| "grad_norm": 5.520111083984375, | |
| "learning_rate": 8.95810276679842e-06, | |
| "loss": 0.8813, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.6046594344655878, | |
| "grad_norm": 7.351500511169434, | |
| "learning_rate": 8.8e-06, | |
| "loss": 0.8677, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.6117730748710652, | |
| "grad_norm": 5.451608657836914, | |
| "learning_rate": 8.641897233201582e-06, | |
| "loss": 0.8606, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.6188867152765428, | |
| "grad_norm": 5.458895683288574, | |
| "learning_rate": 8.483794466403163e-06, | |
| "loss": 0.8522, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.6260003556820203, | |
| "grad_norm": 5.554731369018555, | |
| "learning_rate": 8.325691699604743e-06, | |
| "loss": 0.9042, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.6331139960874977, | |
| "grad_norm": 3.6643424034118652, | |
| "learning_rate": 8.167588932806326e-06, | |
| "loss": 0.8696, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.6402276364929753, | |
| "grad_norm": 5.593344688415527, | |
| "learning_rate": 8.009486166007906e-06, | |
| "loss": 0.7663, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.6473412768984528, | |
| "grad_norm": 7.597021102905273, | |
| "learning_rate": 7.851383399209487e-06, | |
| "loss": 0.7752, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.6544549173039302, | |
| "grad_norm": 4.095972537994385, | |
| "learning_rate": 7.693280632411068e-06, | |
| "loss": 0.897, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.6615685577094078, | |
| "grad_norm": 7.817249298095703, | |
| "learning_rate": 7.535177865612649e-06, | |
| "loss": 0.7448, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.6686821981148853, | |
| "grad_norm": 3.296755313873291, | |
| "learning_rate": 7.37707509881423e-06, | |
| "loss": 0.8065, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.6757958385203628, | |
| "grad_norm": 5.186858177185059, | |
| "learning_rate": 7.218972332015811e-06, | |
| "loss": 0.8618, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.6829094789258403, | |
| "grad_norm": 6.484861373901367, | |
| "learning_rate": 7.060869565217392e-06, | |
| "loss": 0.8549, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.6900231193313178, | |
| "grad_norm": 4.90764856338501, | |
| "learning_rate": 6.902766798418972e-06, | |
| "loss": 0.9441, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.6971367597367953, | |
| "grad_norm": 3.6083099842071533, | |
| "learning_rate": 6.744664031620554e-06, | |
| "loss": 0.7467, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.7042504001422728, | |
| "grad_norm": 4.79904317855835, | |
| "learning_rate": 6.586561264822134e-06, | |
| "loss": 0.7974, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.7113640405477503, | |
| "grad_norm": 8.30154800415039, | |
| "learning_rate": 6.4284584980237165e-06, | |
| "loss": 0.8927, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.7184776809532278, | |
| "grad_norm": 5.582029819488525, | |
| "learning_rate": 6.270355731225297e-06, | |
| "loss": 0.8348, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.7255913213587053, | |
| "grad_norm": 7.575035572052002, | |
| "learning_rate": 6.112252964426878e-06, | |
| "loss": 0.8691, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.7327049617641829, | |
| "grad_norm": 6.100114345550537, | |
| "learning_rate": 5.954150197628459e-06, | |
| "loss": 0.788, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.7398186021696603, | |
| "grad_norm": 7.05925178527832, | |
| "learning_rate": 5.796047430830039e-06, | |
| "loss": 0.8325, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.7469322425751378, | |
| "grad_norm": 6.5520920753479, | |
| "learning_rate": 5.6379446640316215e-06, | |
| "loss": 0.7792, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.7540458829806154, | |
| "grad_norm": 6.9081807136535645, | |
| "learning_rate": 5.479841897233202e-06, | |
| "loss": 0.805, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.7611595233860928, | |
| "grad_norm": 7.752345085144043, | |
| "learning_rate": 5.321739130434783e-06, | |
| "loss": 0.7998, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.7682731637915703, | |
| "grad_norm": 3.8013064861297607, | |
| "learning_rate": 5.163636363636364e-06, | |
| "loss": 0.7748, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.7753868041970479, | |
| "grad_norm": 5.316600322723389, | |
| "learning_rate": 5.005533596837945e-06, | |
| "loss": 0.7644, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.7825004446025253, | |
| "grad_norm": 8.250207901000977, | |
| "learning_rate": 4.8474308300395266e-06, | |
| "loss": 0.7766, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.7896140850080029, | |
| "grad_norm": 5.134970664978027, | |
| "learning_rate": 4.689328063241107e-06, | |
| "loss": 0.79, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.7967277254134804, | |
| "grad_norm": 5.832414627075195, | |
| "learning_rate": 4.531225296442688e-06, | |
| "loss": 0.9154, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.8038413658189578, | |
| "grad_norm": 6.164233684539795, | |
| "learning_rate": 4.373122529644269e-06, | |
| "loss": 0.8288, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.8109550062244354, | |
| "grad_norm": 11.60009479522705, | |
| "learning_rate": 4.21501976284585e-06, | |
| "loss": 0.7047, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.8180686466299129, | |
| "grad_norm": 3.422287702560425, | |
| "learning_rate": 4.056916996047432e-06, | |
| "loss": 0.7387, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.8251822870353903, | |
| "grad_norm": 3.786428689956665, | |
| "learning_rate": 3.898814229249012e-06, | |
| "loss": 0.806, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.8322959274408679, | |
| "grad_norm": 3.7693772315979004, | |
| "learning_rate": 3.740711462450593e-06, | |
| "loss": 0.8136, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.8394095678463454, | |
| "grad_norm": 3.2537221908569336, | |
| "learning_rate": 3.5826086956521744e-06, | |
| "loss": 0.7834, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.8465232082518229, | |
| "grad_norm": 2.938720226287842, | |
| "learning_rate": 3.4245059288537553e-06, | |
| "loss": 0.7607, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.8536368486573004, | |
| "grad_norm": 6.700362205505371, | |
| "learning_rate": 3.266403162055336e-06, | |
| "loss": 0.8713, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.8607504890627778, | |
| "grad_norm": 4.768782615661621, | |
| "learning_rate": 3.1083003952569176e-06, | |
| "loss": 0.8086, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.8678641294682554, | |
| "grad_norm": 6.537806034088135, | |
| "learning_rate": 2.950197628458498e-06, | |
| "loss": 0.8722, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.8749777698737329, | |
| "grad_norm": 6.817727088928223, | |
| "learning_rate": 2.792094861660079e-06, | |
| "loss": 0.7126, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.8820914102792103, | |
| "grad_norm": 5.932295799255371, | |
| "learning_rate": 2.6339920948616604e-06, | |
| "loss": 0.778, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.8892050506846879, | |
| "grad_norm": 6.475405693054199, | |
| "learning_rate": 2.4758893280632413e-06, | |
| "loss": 0.7812, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.8963186910901654, | |
| "grad_norm": 8.57321834564209, | |
| "learning_rate": 2.3177865612648222e-06, | |
| "loss": 0.7654, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.903432331495643, | |
| "grad_norm": 1.9723538160324097, | |
| "learning_rate": 2.1596837944664036e-06, | |
| "loss": 0.7701, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.9105459719011204, | |
| "grad_norm": 6.585085391998291, | |
| "learning_rate": 2.001581027667984e-06, | |
| "loss": 0.8419, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.9176596123065979, | |
| "grad_norm": 4.754767894744873, | |
| "learning_rate": 1.8434782608695654e-06, | |
| "loss": 0.9006, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.9247732527120754, | |
| "grad_norm": 4.661940574645996, | |
| "learning_rate": 1.6853754940711464e-06, | |
| "loss": 0.7784, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.9318868931175529, | |
| "grad_norm": 4.972583770751953, | |
| "learning_rate": 1.5272727272727275e-06, | |
| "loss": 0.8112, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.9390005335230304, | |
| "grad_norm": 6.641800403594971, | |
| "learning_rate": 1.3691699604743084e-06, | |
| "loss": 0.7466, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.9461141739285079, | |
| "grad_norm": 5.098995208740234, | |
| "learning_rate": 1.2110671936758893e-06, | |
| "loss": 0.824, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.9532278143339854, | |
| "grad_norm": 7.060574054718018, | |
| "learning_rate": 1.0529644268774705e-06, | |
| "loss": 0.8502, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.960341454739463, | |
| "grad_norm": 4.678645610809326, | |
| "learning_rate": 8.948616600790515e-07, | |
| "loss": 0.8004, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.9674550951449404, | |
| "grad_norm": 5.895313739776611, | |
| "learning_rate": 7.367588932806324e-07, | |
| "loss": 0.6811, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.9745687355504179, | |
| "grad_norm": 5.316034317016602, | |
| "learning_rate": 5.786561264822135e-07, | |
| "loss": 0.7955, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.9816823759558955, | |
| "grad_norm": 5.032801628112793, | |
| "learning_rate": 4.205533596837945e-07, | |
| "loss": 0.8206, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.9887960163613729, | |
| "grad_norm": 5.486656188964844, | |
| "learning_rate": 2.6245059288537554e-07, | |
| "loss": 0.795, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.9959096567668504, | |
| "grad_norm": 6.862821102142334, | |
| "learning_rate": 1.0434782608695654e-07, | |
| "loss": 0.676, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.9998932953939178, | |
| "eval_loss": 0.767158567905426, | |
| "eval_runtime": 341.0802, | |
| "eval_samples_per_second": 15.454, | |
| "eval_steps_per_second": 1.932, | |
| "step": 7028 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 7028, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.803879769007718e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |