| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.0, |
| "eval_steps": 500, |
| "global_step": 4184, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04780114722753346, |
| "grad_norm": 0.0546465665102005, |
| "learning_rate": 4.941443594646272e-05, |
| "loss": 2.2922, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.09560229445506692, |
| "grad_norm": 0.08099879324436188, |
| "learning_rate": 4.881692160611855e-05, |
| "loss": 2.2658, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.14340344168260039, |
| "grad_norm": 0.08135072886943817, |
| "learning_rate": 4.821940726577438e-05, |
| "loss": 2.1959, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.19120458891013384, |
| "grad_norm": 0.08765063434839249, |
| "learning_rate": 4.7621892925430214e-05, |
| "loss": 2.1285, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.2390057361376673, |
| "grad_norm": 0.10307478904724121, |
| "learning_rate": 4.7024378585086046e-05, |
| "loss": 2.0899, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.28680688336520077, |
| "grad_norm": 0.10948552936315536, |
| "learning_rate": 4.642686424474187e-05, |
| "loss": 2.0663, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.33460803059273425, |
| "grad_norm": 0.1135614812374115, |
| "learning_rate": 4.5829349904397704e-05, |
| "loss": 2.0508, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.3824091778202677, |
| "grad_norm": 0.12696176767349243, |
| "learning_rate": 4.5231835564053536e-05, |
| "loss": 1.9819, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.43021032504780116, |
| "grad_norm": 0.13437969982624054, |
| "learning_rate": 4.463432122370937e-05, |
| "loss": 2.0004, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.4780114722753346, |
| "grad_norm": 0.1275995969772339, |
| "learning_rate": 4.40368068833652e-05, |
| "loss": 1.9961, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.5258126195028681, |
| "grad_norm": 0.14496305584907532, |
| "learning_rate": 4.343929254302104e-05, |
| "loss": 1.9855, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.5736137667304015, |
| "grad_norm": 0.13401177525520325, |
| "learning_rate": 4.284177820267687e-05, |
| "loss": 1.976, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.621414913957935, |
| "grad_norm": 0.15063685178756714, |
| "learning_rate": 4.2244263862332704e-05, |
| "loss": 1.9428, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.6692160611854685, |
| "grad_norm": 0.1525646150112152, |
| "learning_rate": 4.164674952198853e-05, |
| "loss": 1.9565, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.7170172084130019, |
| "grad_norm": 0.1553465574979782, |
| "learning_rate": 4.104923518164436e-05, |
| "loss": 1.9499, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.7648183556405354, |
| "grad_norm": 0.178068146109581, |
| "learning_rate": 4.0451720841300193e-05, |
| "loss": 1.9342, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.8126195028680688, |
| "grad_norm": 0.16074113547801971, |
| "learning_rate": 3.9854206500956026e-05, |
| "loss": 1.9304, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.8604206500956023, |
| "grad_norm": 0.17012304067611694, |
| "learning_rate": 3.925669216061186e-05, |
| "loss": 1.9358, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.9082217973231358, |
| "grad_norm": 0.1635637879371643, |
| "learning_rate": 3.865917782026769e-05, |
| "loss": 1.9145, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.9560229445506692, |
| "grad_norm": 0.18287719786167145, |
| "learning_rate": 3.806166347992352e-05, |
| "loss": 1.9186, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.0038240917782026, |
| "grad_norm": 0.18034423887729645, |
| "learning_rate": 3.7464149139579354e-05, |
| "loss": 1.8995, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.0516252390057361, |
| "grad_norm": 0.16354230046272278, |
| "learning_rate": 3.6866634799235186e-05, |
| "loss": 1.9068, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.0994263862332696, |
| "grad_norm": 0.236989825963974, |
| "learning_rate": 3.626912045889101e-05, |
| "loss": 1.9141, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.147227533460803, |
| "grad_norm": 0.2356785088777542, |
| "learning_rate": 3.5671606118546844e-05, |
| "loss": 1.9088, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.1950286806883366, |
| "grad_norm": 0.1805247962474823, |
| "learning_rate": 3.5074091778202676e-05, |
| "loss": 1.8923, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.24282982791587, |
| "grad_norm": 0.21303710341453552, |
| "learning_rate": 3.447657743785851e-05, |
| "loss": 1.898, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.2906309751434035, |
| "grad_norm": 0.18716047704219818, |
| "learning_rate": 3.387906309751434e-05, |
| "loss": 1.8986, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.338432122370937, |
| "grad_norm": 0.16454172134399414, |
| "learning_rate": 3.328154875717017e-05, |
| "loss": 1.9149, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.3862332695984705, |
| "grad_norm": 0.18206411600112915, |
| "learning_rate": 3.2684034416826005e-05, |
| "loss": 1.8662, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.4340344168260037, |
| "grad_norm": 0.19803249835968018, |
| "learning_rate": 3.208652007648184e-05, |
| "loss": 1.9015, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.4818355640535372, |
| "grad_norm": 0.19260399043560028, |
| "learning_rate": 3.148900573613767e-05, |
| "loss": 1.905, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.5296367112810707, |
| "grad_norm": 0.20765070617198944, |
| "learning_rate": 3.08914913957935e-05, |
| "loss": 1.8981, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.5774378585086042, |
| "grad_norm": 0.1981675922870636, |
| "learning_rate": 3.029397705544933e-05, |
| "loss": 1.8957, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.6252390057361377, |
| "grad_norm": 0.19996145367622375, |
| "learning_rate": 2.9696462715105166e-05, |
| "loss": 1.8797, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.6730401529636711, |
| "grad_norm": 0.21547172963619232, |
| "learning_rate": 2.9098948374760998e-05, |
| "loss": 1.8919, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.7208413001912046, |
| "grad_norm": 0.21214577555656433, |
| "learning_rate": 2.850143403441683e-05, |
| "loss": 1.8942, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.7686424474187379, |
| "grad_norm": 0.19430747628211975, |
| "learning_rate": 2.7903919694072662e-05, |
| "loss": 1.8744, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.8164435946462714, |
| "grad_norm": 0.20672687888145447, |
| "learning_rate": 2.730640535372849e-05, |
| "loss": 1.8784, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.8642447418738048, |
| "grad_norm": 0.204985573887825, |
| "learning_rate": 2.6708891013384323e-05, |
| "loss": 1.8902, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.9120458891013383, |
| "grad_norm": 0.20450404286384583, |
| "learning_rate": 2.6111376673040155e-05, |
| "loss": 1.8649, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.9598470363288718, |
| "grad_norm": 0.20468485355377197, |
| "learning_rate": 2.5513862332695987e-05, |
| "loss": 1.8899, |
| "step": 2050 |
| }, |
| { |
| "epoch": 2.0076481835564053, |
| "grad_norm": 0.206816628575325, |
| "learning_rate": 2.491634799235182e-05, |
| "loss": 1.8607, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.0554493307839388, |
| "grad_norm": 0.2078491896390915, |
| "learning_rate": 2.431883365200765e-05, |
| "loss": 1.8737, |
| "step": 2150 |
| }, |
| { |
| "epoch": 2.1032504780114722, |
| "grad_norm": 0.19321344792842865, |
| "learning_rate": 2.372131931166348e-05, |
| "loss": 1.8597, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.1510516252390057, |
| "grad_norm": 0.22135953605175018, |
| "learning_rate": 2.3123804971319313e-05, |
| "loss": 1.88, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.198852772466539, |
| "grad_norm": 0.20404009521007538, |
| "learning_rate": 2.2526290630975145e-05, |
| "loss": 1.8908, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.2466539196940727, |
| "grad_norm": 0.21296437084674835, |
| "learning_rate": 2.1928776290630977e-05, |
| "loss": 1.8632, |
| "step": 2350 |
| }, |
| { |
| "epoch": 2.294455066921606, |
| "grad_norm": 0.2082318514585495, |
| "learning_rate": 2.1331261950286806e-05, |
| "loss": 1.8819, |
| "step": 2400 |
| }, |
| { |
| "epoch": 2.3422562141491396, |
| "grad_norm": 0.2119467705488205, |
| "learning_rate": 2.0733747609942638e-05, |
| "loss": 1.8679, |
| "step": 2450 |
| }, |
| { |
| "epoch": 2.390057361376673, |
| "grad_norm": 0.2162160575389862, |
| "learning_rate": 2.013623326959847e-05, |
| "loss": 1.8698, |
| "step": 2500 |
| }, |
| { |
| "epoch": 2.4378585086042066, |
| "grad_norm": 0.19639310240745544, |
| "learning_rate": 1.9538718929254306e-05, |
| "loss": 1.8808, |
| "step": 2550 |
| }, |
| { |
| "epoch": 2.48565965583174, |
| "grad_norm": 0.2149832397699356, |
| "learning_rate": 1.8941204588910135e-05, |
| "loss": 1.8743, |
| "step": 2600 |
| }, |
| { |
| "epoch": 2.5334608030592736, |
| "grad_norm": 0.20657892525196075, |
| "learning_rate": 1.8343690248565967e-05, |
| "loss": 1.8611, |
| "step": 2650 |
| }, |
| { |
| "epoch": 2.581261950286807, |
| "grad_norm": 0.22579824924468994, |
| "learning_rate": 1.77461759082218e-05, |
| "loss": 1.8834, |
| "step": 2700 |
| }, |
| { |
| "epoch": 2.62906309751434, |
| "grad_norm": 0.214285746216774, |
| "learning_rate": 1.714866156787763e-05, |
| "loss": 1.8712, |
| "step": 2750 |
| }, |
| { |
| "epoch": 2.676864244741874, |
| "grad_norm": 0.2056400179862976, |
| "learning_rate": 1.6551147227533463e-05, |
| "loss": 1.8515, |
| "step": 2800 |
| }, |
| { |
| "epoch": 2.724665391969407, |
| "grad_norm": 0.21018265187740326, |
| "learning_rate": 1.5953632887189292e-05, |
| "loss": 1.8537, |
| "step": 2850 |
| }, |
| { |
| "epoch": 2.772466539196941, |
| "grad_norm": 0.1982487589120865, |
| "learning_rate": 1.5356118546845124e-05, |
| "loss": 1.8637, |
| "step": 2900 |
| }, |
| { |
| "epoch": 2.820267686424474, |
| "grad_norm": 0.22870182991027832, |
| "learning_rate": 1.4758604206500956e-05, |
| "loss": 1.857, |
| "step": 2950 |
| }, |
| { |
| "epoch": 2.8680688336520075, |
| "grad_norm": 0.2098233997821808, |
| "learning_rate": 1.4161089866156787e-05, |
| "loss": 1.8433, |
| "step": 3000 |
| }, |
| { |
| "epoch": 2.915869980879541, |
| "grad_norm": 0.23259004950523376, |
| "learning_rate": 1.3563575525812619e-05, |
| "loss": 1.8934, |
| "step": 3050 |
| }, |
| { |
| "epoch": 2.9636711281070744, |
| "grad_norm": 0.2191799432039261, |
| "learning_rate": 1.2966061185468451e-05, |
| "loss": 1.8603, |
| "step": 3100 |
| }, |
| { |
| "epoch": 3.011472275334608, |
| "grad_norm": 0.21849119663238525, |
| "learning_rate": 1.2368546845124283e-05, |
| "loss": 1.8603, |
| "step": 3150 |
| }, |
| { |
| "epoch": 3.0592734225621414, |
| "grad_norm": 0.22140590846538544, |
| "learning_rate": 1.1771032504780115e-05, |
| "loss": 1.8696, |
| "step": 3200 |
| }, |
| { |
| "epoch": 3.107074569789675, |
| "grad_norm": 0.2533087432384491, |
| "learning_rate": 1.1173518164435948e-05, |
| "loss": 1.8491, |
| "step": 3250 |
| }, |
| { |
| "epoch": 3.1548757170172084, |
| "grad_norm": 0.22291821241378784, |
| "learning_rate": 1.057600382409178e-05, |
| "loss": 1.8664, |
| "step": 3300 |
| }, |
| { |
| "epoch": 3.202676864244742, |
| "grad_norm": 0.23389829695224762, |
| "learning_rate": 9.97848948374761e-06, |
| "loss": 1.8633, |
| "step": 3350 |
| }, |
| { |
| "epoch": 3.2504780114722753, |
| "grad_norm": 0.2284611165523529, |
| "learning_rate": 9.380975143403442e-06, |
| "loss": 1.8739, |
| "step": 3400 |
| }, |
| { |
| "epoch": 3.298279158699809, |
| "grad_norm": 0.2357141673564911, |
| "learning_rate": 8.783460803059273e-06, |
| "loss": 1.8837, |
| "step": 3450 |
| }, |
| { |
| "epoch": 3.3460803059273423, |
| "grad_norm": 0.23133474588394165, |
| "learning_rate": 8.185946462715105e-06, |
| "loss": 1.8284, |
| "step": 3500 |
| }, |
| { |
| "epoch": 3.3938814531548758, |
| "grad_norm": 0.24003466963768005, |
| "learning_rate": 7.588432122370938e-06, |
| "loss": 1.8557, |
| "step": 3550 |
| }, |
| { |
| "epoch": 3.4416826003824093, |
| "grad_norm": 0.22754019498825073, |
| "learning_rate": 6.990917782026769e-06, |
| "loss": 1.8726, |
| "step": 3600 |
| }, |
| { |
| "epoch": 3.4894837476099427, |
| "grad_norm": 0.24840685725212097, |
| "learning_rate": 6.393403441682601e-06, |
| "loss": 1.8683, |
| "step": 3650 |
| }, |
| { |
| "epoch": 3.537284894837476, |
| "grad_norm": 0.23282456398010254, |
| "learning_rate": 5.795889101338432e-06, |
| "loss": 1.8624, |
| "step": 3700 |
| }, |
| { |
| "epoch": 3.5850860420650097, |
| "grad_norm": 0.22449611127376556, |
| "learning_rate": 5.198374760994264e-06, |
| "loss": 1.8355, |
| "step": 3750 |
| }, |
| { |
| "epoch": 3.632887189292543, |
| "grad_norm": 0.22073350846767426, |
| "learning_rate": 4.6008604206500955e-06, |
| "loss": 1.8706, |
| "step": 3800 |
| }, |
| { |
| "epoch": 3.6806883365200767, |
| "grad_norm": 0.2504482567310333, |
| "learning_rate": 4.003346080305928e-06, |
| "loss": 1.8607, |
| "step": 3850 |
| }, |
| { |
| "epoch": 3.7284894837476097, |
| "grad_norm": 0.23835700750350952, |
| "learning_rate": 3.4058317399617594e-06, |
| "loss": 1.8495, |
| "step": 3900 |
| }, |
| { |
| "epoch": 3.7762906309751436, |
| "grad_norm": 0.2028919905424118, |
| "learning_rate": 2.808317399617591e-06, |
| "loss": 1.8423, |
| "step": 3950 |
| }, |
| { |
| "epoch": 3.8240917782026767, |
| "grad_norm": 0.23366276919841766, |
| "learning_rate": 2.210803059273423e-06, |
| "loss": 1.8431, |
| "step": 4000 |
| }, |
| { |
| "epoch": 3.8718929254302106, |
| "grad_norm": 0.261652410030365, |
| "learning_rate": 1.6132887189292542e-06, |
| "loss": 1.8639, |
| "step": 4050 |
| }, |
| { |
| "epoch": 3.9196940726577436, |
| "grad_norm": 0.2397291362285614, |
| "learning_rate": 1.015774378585086e-06, |
| "loss": 1.8626, |
| "step": 4100 |
| }, |
| { |
| "epoch": 3.967495219885277, |
| "grad_norm": 0.24784159660339355, |
| "learning_rate": 4.1826003824091784e-07, |
| "loss": 1.8817, |
| "step": 4150 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 4184, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 4, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.603041003909939e+16, |
| "train_batch_size": 10, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|