{ "best_metric": null, "best_model_checkpoint": null, "epoch": 8.330556481172943, "eval_steps": 500, "global_step": 25000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.16661112962345884, "grad_norm": 2.0231502056121826, "learning_rate": 2.776852160390981e-06, "loss": 8.4699, "step": 500 }, { "epoch": 0.3332222592469177, "grad_norm": 1.640729546546936, "learning_rate": 5.553704320781962e-06, "loss": 7.446, "step": 1000 }, { "epoch": 0.49983338887037654, "grad_norm": 1.6770328283309937, "learning_rate": 8.330556481172942e-06, "loss": 6.6034, "step": 1500 }, { "epoch": 0.6664445184938353, "grad_norm": 2.338146924972534, "learning_rate": 1.1107408641563923e-05, "loss": 5.9787, "step": 2000 }, { "epoch": 0.8330556481172943, "grad_norm": 2.557213068008423, "learning_rate": 1.3884260801954905e-05, "loss": 5.5732, "step": 2500 }, { "epoch": 0.9996667777407531, "grad_norm": 1.7546786069869995, "learning_rate": 1.6661112962345883e-05, "loss": 5.2881, "step": 3000 }, { "epoch": 1.166277907364212, "grad_norm": 1.7593083381652832, "learning_rate": 1.9437965122736867e-05, "loss": 5.0771, "step": 3500 }, { "epoch": 1.3328890369876707, "grad_norm": 1.860612392425537, "learning_rate": 2.2214817283127847e-05, "loss": 4.8991, "step": 4000 }, { "epoch": 1.4995001666111296, "grad_norm": 2.036371946334839, "learning_rate": 2.4991669443518826e-05, "loss": 4.7465, "step": 4500 }, { "epoch": 1.6661112962345883, "grad_norm": 1.8470864295959473, "learning_rate": 2.776852160390981e-05, "loss": 4.5925, "step": 5000 }, { "epoch": 1.8327224258580475, "grad_norm": 1.9861804246902466, "learning_rate": 3.0545373764300786e-05, "loss": 4.4767, "step": 5500 }, { "epoch": 1.9993335554815062, "grad_norm": 2.018094062805176, "learning_rate": 3.3322225924691766e-05, "loss": 4.3708, "step": 6000 }, { "epoch": 2.165944685104965, "grad_norm": 1.8410370349884033, "learning_rate": 3.609352438076197e-05, "loss": 4.2741, "step": 6500 }, { "epoch": 2.332555814728424, "grad_norm": 1.7712337970733643, "learning_rate": 3.887037654115295e-05, "loss": 4.192, "step": 7000 }, { "epoch": 2.4991669443518827, "grad_norm": 1.7352423667907715, "learning_rate": 4.164722870154393e-05, "loss": 4.1192, "step": 7500 }, { "epoch": 2.6657780739753414, "grad_norm": 1.8459163904190063, "learning_rate": 4.442408086193491e-05, "loss": 4.0351, "step": 8000 }, { "epoch": 2.8323892035988005, "grad_norm": 2.009974956512451, "learning_rate": 4.719537931800511e-05, "loss": 3.9709, "step": 8500 }, { "epoch": 2.9990003332222592, "grad_norm": 1.7068768739700317, "learning_rate": 4.997223147839609e-05, "loss": 3.899, "step": 9000 }, { "epoch": 3.165611462845718, "grad_norm": 1.6234753131866455, "learning_rate": 5.274908363878708e-05, "loss": 3.8237, "step": 9500 }, { "epoch": 3.332222592469177, "grad_norm": 1.786528468132019, "learning_rate": 5.552593579917805e-05, "loss": 3.7704, "step": 10000 }, { "epoch": 3.4988337220926358, "grad_norm": 1.7539526224136353, "learning_rate": 5.829723425524826e-05, "loss": 3.7183, "step": 10500 }, { "epoch": 3.665444851716095, "grad_norm": 1.6715501546859741, "learning_rate": 6.107408641563923e-05, "loss": 3.6671, "step": 11000 }, { "epoch": 3.8320559813395536, "grad_norm": 1.7561357021331787, "learning_rate": 6.385093857603022e-05, "loss": 3.6214, "step": 11500 }, { "epoch": 3.9986671109630123, "grad_norm": 1.5936201810836792, "learning_rate": 6.662779073642119e-05, "loss": 3.5782, "step": 12000 }, { "epoch": 4.1652782405864714, "grad_norm": 1.7363481521606445, "learning_rate": 6.940464289681218e-05, "loss": 3.5291, "step": 12500 }, { "epoch": 4.33188937020993, "grad_norm": 1.5552008152008057, "learning_rate": 7.217594135288238e-05, "loss": 3.5017, "step": 13000 }, { "epoch": 4.498500499833389, "grad_norm": 1.4608962535858154, "learning_rate": 7.495279351327336e-05, "loss": 3.4685, "step": 13500 }, { "epoch": 4.665111629456848, "grad_norm": 1.470942735671997, "learning_rate": 7.772964567366434e-05, "loss": 3.4407, "step": 14000 }, { "epoch": 4.831722759080306, "grad_norm": 1.4744900465011597, "learning_rate": 8.050649783405532e-05, "loss": 3.4156, "step": 14500 }, { "epoch": 4.998333888703765, "grad_norm": 1.449549913406372, "learning_rate": 8.32833499944463e-05, "loss": 3.3933, "step": 15000 }, { "epoch": 5.1649450183272245, "grad_norm": 1.4103004932403564, "learning_rate": 8.606020215483728e-05, "loss": 3.355, "step": 15500 }, { "epoch": 5.331556147950683, "grad_norm": 1.3490815162658691, "learning_rate": 8.883705431522826e-05, "loss": 3.335, "step": 16000 }, { "epoch": 5.498167277574142, "grad_norm": 1.3762222528457642, "learning_rate": 9.161390647561924e-05, "loss": 3.3183, "step": 16500 }, { "epoch": 5.664778407197601, "grad_norm": 1.3008846044540405, "learning_rate": 9.438520493168944e-05, "loss": 3.3013, "step": 17000 }, { "epoch": 5.831389536821059, "grad_norm": 1.3132846355438232, "learning_rate": 9.716205709208043e-05, "loss": 3.2846, "step": 17500 }, { "epoch": 5.9980006664445185, "grad_norm": 1.194407343864441, "learning_rate": 9.99389092524714e-05, "loss": 3.2706, "step": 18000 }, { "epoch": 6.164611796067978, "grad_norm": 1.1745057106018066, "learning_rate": 9.996657887827809e-05, "loss": 3.2324, "step": 18500 }, { "epoch": 6.331222925691436, "grad_norm": 1.1769245862960815, "learning_rate": 9.986361454903414e-05, "loss": 3.2237, "step": 19000 }, { "epoch": 6.497834055314895, "grad_norm": 1.1529046297073364, "learning_rate": 9.969082137309689e-05, "loss": 3.2103, "step": 19500 }, { "epoch": 6.664445184938354, "grad_norm": 1.1225526332855225, "learning_rate": 9.944857739312099e-05, "loss": 3.1936, "step": 20000 }, { "epoch": 6.831056314561812, "grad_norm": 1.1167833805084229, "learning_rate": 9.913722118348934e-05, "loss": 3.1818, "step": 20500 }, { "epoch": 6.9976674441852715, "grad_norm": 1.093309998512268, "learning_rate": 9.875801618300155e-05, "loss": 3.1733, "step": 21000 }, { "epoch": 7.164278573808731, "grad_norm": 1.1539673805236816, "learning_rate": 9.83099727084671e-05, "loss": 3.1342, "step": 21500 }, { "epoch": 7.330889703432189, "grad_norm": 1.1930997371673584, "learning_rate": 9.779440838491111e-05, "loss": 3.1356, "step": 22000 }, { "epoch": 7.497500833055648, "grad_norm": 1.0827438831329346, "learning_rate": 9.721204379524642e-05, "loss": 3.1231, "step": 22500 }, { "epoch": 7.664111962679107, "grad_norm": 1.0850555896759033, "learning_rate": 9.656505485544595e-05, "loss": 3.1126, "step": 23000 }, { "epoch": 7.8307230923025655, "grad_norm": 1.0119590759277344, "learning_rate": 9.585175299579849e-05, "loss": 3.1049, "step": 23500 }, { "epoch": 7.997334221926025, "grad_norm": 1.0027412176132202, "learning_rate": 9.507436603906176e-05, "loss": 3.1004, "step": 24000 }, { "epoch": 8.163945351549483, "grad_norm": 1.0899256467819214, "learning_rate": 9.423398050680459e-05, "loss": 3.0591, "step": 24500 }, { "epoch": 8.330556481172943, "grad_norm": 1.0424436330795288, "learning_rate": 9.333363626660517e-05, "loss": 3.0582, "step": 25000 } ], "logging_steps": 500, "max_steps": 60020, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.046659236626432e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }