| { | |
| "best_metric": 1.1517876386642456, | |
| "best_model_checkpoint": "/media/user/Expansion/flan-t5-small-simplifier/checkpoint-80000", | |
| "epoch": 2.0, | |
| "eval_steps": 10000, | |
| "global_step": 89920, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01112099644128114, | |
| "grad_norm": 3.2175512313842773, | |
| "learning_rate": 4.972197508896798e-05, | |
| "loss": 1.6509, | |
| "num_input_tokens_seen": 183488, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02224199288256228, | |
| "grad_norm": 3.9029898643493652, | |
| "learning_rate": 4.9443950177935946e-05, | |
| "loss": 1.583, | |
| "num_input_tokens_seen": 363744, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.03336298932384341, | |
| "grad_norm": 3.4091885089874268, | |
| "learning_rate": 4.9165925266903915e-05, | |
| "loss": 1.5553, | |
| "num_input_tokens_seen": 547584, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.04448398576512456, | |
| "grad_norm": 2.5965495109558105, | |
| "learning_rate": 4.888790035587189e-05, | |
| "loss": 1.5444, | |
| "num_input_tokens_seen": 732480, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.055604982206405695, | |
| "grad_norm": 4.170108318328857, | |
| "learning_rate": 4.860987544483986e-05, | |
| "loss": 1.5137, | |
| "num_input_tokens_seen": 916672, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.06672597864768683, | |
| "grad_norm": 3.025068759918213, | |
| "learning_rate": 4.8331850533807835e-05, | |
| "loss": 1.5057, | |
| "num_input_tokens_seen": 1095952, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.07784697508896797, | |
| "grad_norm": 2.7047712802886963, | |
| "learning_rate": 4.80538256227758e-05, | |
| "loss": 1.5019, | |
| "num_input_tokens_seen": 1276936, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.08896797153024912, | |
| "grad_norm": 2.844285488128662, | |
| "learning_rate": 4.777580071174377e-05, | |
| "loss": 1.5074, | |
| "num_input_tokens_seen": 1464080, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.10008896797153025, | |
| "grad_norm": 3.055643081665039, | |
| "learning_rate": 4.749777580071175e-05, | |
| "loss": 1.4959, | |
| "num_input_tokens_seen": 1647288, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.11120996441281139, | |
| "grad_norm": 2.6245856285095215, | |
| "learning_rate": 4.721975088967972e-05, | |
| "loss": 1.4777, | |
| "num_input_tokens_seen": 1836832, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.12233096085409252, | |
| "grad_norm": 3.1758244037628174, | |
| "learning_rate": 4.694172597864769e-05, | |
| "loss": 1.4778, | |
| "num_input_tokens_seen": 2020728, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.13345195729537365, | |
| "grad_norm": 2.518728494644165, | |
| "learning_rate": 4.666370106761566e-05, | |
| "loss": 1.4758, | |
| "num_input_tokens_seen": 2198968, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.1445729537366548, | |
| "grad_norm": 3.8143937587738037, | |
| "learning_rate": 4.638567615658363e-05, | |
| "loss": 1.4846, | |
| "num_input_tokens_seen": 2381968, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.15569395017793594, | |
| "grad_norm": 2.766146421432495, | |
| "learning_rate": 4.6107651245551604e-05, | |
| "loss": 1.4468, | |
| "num_input_tokens_seen": 2562176, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.16681494661921709, | |
| "grad_norm": 3.891373634338379, | |
| "learning_rate": 4.582962633451958e-05, | |
| "loss": 1.4586, | |
| "num_input_tokens_seen": 2744688, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.17793594306049823, | |
| "grad_norm": 3.277316093444824, | |
| "learning_rate": 4.555160142348754e-05, | |
| "loss": 1.4606, | |
| "num_input_tokens_seen": 2926000, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.18905693950177935, | |
| "grad_norm": 3.242478132247925, | |
| "learning_rate": 4.5273576512455517e-05, | |
| "loss": 1.446, | |
| "num_input_tokens_seen": 3108520, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.2001779359430605, | |
| "grad_norm": 2.3061795234680176, | |
| "learning_rate": 4.499555160142349e-05, | |
| "loss": 1.4348, | |
| "num_input_tokens_seen": 3289352, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.21129893238434164, | |
| "grad_norm": 3.4106180667877197, | |
| "learning_rate": 4.471752669039146e-05, | |
| "loss": 1.4345, | |
| "num_input_tokens_seen": 3473784, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.22241992882562278, | |
| "grad_norm": 2.88779354095459, | |
| "learning_rate": 4.443950177935943e-05, | |
| "loss": 1.4423, | |
| "num_input_tokens_seen": 3655312, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.22241992882562278, | |
| "eval_loss": 1.2430843114852905, | |
| "eval_runtime": 2.8218, | |
| "eval_samples_per_second": 885.962, | |
| "eval_steps_per_second": 110.922, | |
| "num_input_tokens_seen": 3655312, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.23354092526690393, | |
| "grad_norm": 2.829268217086792, | |
| "learning_rate": 4.4161476868327405e-05, | |
| "loss": 1.4321, | |
| "num_input_tokens_seen": 3847328, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.24466192170818504, | |
| "grad_norm": 3.807185173034668, | |
| "learning_rate": 4.388345195729537e-05, | |
| "loss": 1.4487, | |
| "num_input_tokens_seen": 4025576, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.2557829181494662, | |
| "grad_norm": 2.449007511138916, | |
| "learning_rate": 4.360542704626335e-05, | |
| "loss": 1.413, | |
| "num_input_tokens_seen": 4213832, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.2669039145907473, | |
| "grad_norm": 3.363555431365967, | |
| "learning_rate": 4.3327402135231324e-05, | |
| "loss": 1.4313, | |
| "num_input_tokens_seen": 4395352, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.2780249110320285, | |
| "grad_norm": 2.702747344970703, | |
| "learning_rate": 4.3049377224199286e-05, | |
| "loss": 1.4179, | |
| "num_input_tokens_seen": 4582368, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.2891459074733096, | |
| "grad_norm": 3.286044120788574, | |
| "learning_rate": 4.277135231316726e-05, | |
| "loss": 1.4137, | |
| "num_input_tokens_seen": 4770640, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.30026690391459077, | |
| "grad_norm": 2.6410391330718994, | |
| "learning_rate": 4.249332740213524e-05, | |
| "loss": 1.4303, | |
| "num_input_tokens_seen": 4951280, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.3113879003558719, | |
| "grad_norm": 3.239133358001709, | |
| "learning_rate": 4.2215302491103205e-05, | |
| "loss": 1.4185, | |
| "num_input_tokens_seen": 5133552, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.322508896797153, | |
| "grad_norm": 2.943094253540039, | |
| "learning_rate": 4.1937277580071174e-05, | |
| "loss": 1.4065, | |
| "num_input_tokens_seen": 5319992, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.33362989323843417, | |
| "grad_norm": 2.180136203765869, | |
| "learning_rate": 4.165925266903915e-05, | |
| "loss": 1.4096, | |
| "num_input_tokens_seen": 5501592, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.3447508896797153, | |
| "grad_norm": 2.4302403926849365, | |
| "learning_rate": 4.138122775800712e-05, | |
| "loss": 1.4203, | |
| "num_input_tokens_seen": 5689312, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.35587188612099646, | |
| "grad_norm": 2.850964069366455, | |
| "learning_rate": 4.1103202846975093e-05, | |
| "loss": 1.4291, | |
| "num_input_tokens_seen": 5870384, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.3669928825622776, | |
| "grad_norm": 1.9641114473342896, | |
| "learning_rate": 4.082517793594306e-05, | |
| "loss": 1.4054, | |
| "num_input_tokens_seen": 6048624, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.3781138790035587, | |
| "grad_norm": 2.287353754043579, | |
| "learning_rate": 4.054715302491103e-05, | |
| "loss": 1.4118, | |
| "num_input_tokens_seen": 6229728, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.38923487544483987, | |
| "grad_norm": 3.4425182342529297, | |
| "learning_rate": 4.0269128113879006e-05, | |
| "loss": 1.4061, | |
| "num_input_tokens_seen": 6410968, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.400355871886121, | |
| "grad_norm": 2.0604770183563232, | |
| "learning_rate": 3.9991103202846975e-05, | |
| "loss": 1.3871, | |
| "num_input_tokens_seen": 6591312, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.41147686832740216, | |
| "grad_norm": 2.671599864959717, | |
| "learning_rate": 3.971307829181495e-05, | |
| "loss": 1.42, | |
| "num_input_tokens_seen": 6777912, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.4225978647686833, | |
| "grad_norm": 2.176579475402832, | |
| "learning_rate": 3.943505338078292e-05, | |
| "loss": 1.417, | |
| "num_input_tokens_seen": 6964240, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.4337188612099644, | |
| "grad_norm": 3.2378785610198975, | |
| "learning_rate": 3.915702846975089e-05, | |
| "loss": 1.3992, | |
| "num_input_tokens_seen": 7147288, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.44483985765124556, | |
| "grad_norm": 3.0986084938049316, | |
| "learning_rate": 3.887900355871886e-05, | |
| "loss": 1.3884, | |
| "num_input_tokens_seen": 7331520, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.44483985765124556, | |
| "eval_loss": 1.2093411684036255, | |
| "eval_runtime": 2.7984, | |
| "eval_samples_per_second": 893.355, | |
| "eval_steps_per_second": 111.848, | |
| "num_input_tokens_seen": 7331520, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.4559608540925267, | |
| "grad_norm": 3.4998202323913574, | |
| "learning_rate": 3.860097864768684e-05, | |
| "loss": 1.3915, | |
| "num_input_tokens_seen": 7515512, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.46708185053380785, | |
| "grad_norm": 3.0249533653259277, | |
| "learning_rate": 3.832295373665481e-05, | |
| "loss": 1.3967, | |
| "num_input_tokens_seen": 7696752, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.47820284697508897, | |
| "grad_norm": 6.8868513107299805, | |
| "learning_rate": 3.8044928825622775e-05, | |
| "loss": 1.4106, | |
| "num_input_tokens_seen": 7878832, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.4893238434163701, | |
| "grad_norm": 2.2134385108947754, | |
| "learning_rate": 3.776690391459075e-05, | |
| "loss": 1.3847, | |
| "num_input_tokens_seen": 8059592, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.5004448398576512, | |
| "grad_norm": 2.2698676586151123, | |
| "learning_rate": 3.748887900355872e-05, | |
| "loss": 1.3941, | |
| "num_input_tokens_seen": 8245432, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.5115658362989324, | |
| "grad_norm": 2.4593448638916016, | |
| "learning_rate": 3.7210854092526695e-05, | |
| "loss": 1.3716, | |
| "num_input_tokens_seen": 8429400, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.5226868327402135, | |
| "grad_norm": 2.5121207237243652, | |
| "learning_rate": 3.6932829181494664e-05, | |
| "loss": 1.3733, | |
| "num_input_tokens_seen": 8618216, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.5338078291814946, | |
| "grad_norm": 2.2493703365325928, | |
| "learning_rate": 3.665480427046263e-05, | |
| "loss": 1.3754, | |
| "num_input_tokens_seen": 8801640, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.5449288256227758, | |
| "grad_norm": 3.08921217918396, | |
| "learning_rate": 3.637677935943061e-05, | |
| "loss": 1.3694, | |
| "num_input_tokens_seen": 8977800, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.556049822064057, | |
| "grad_norm": 2.215222120285034, | |
| "learning_rate": 3.609875444839858e-05, | |
| "loss": 1.3674, | |
| "num_input_tokens_seen": 9160904, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.5671708185053381, | |
| "grad_norm": 2.553903818130493, | |
| "learning_rate": 3.582072953736655e-05, | |
| "loss": 1.3735, | |
| "num_input_tokens_seen": 9348344, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.5782918149466192, | |
| "grad_norm": 2.546022891998291, | |
| "learning_rate": 3.554270462633452e-05, | |
| "loss": 1.382, | |
| "num_input_tokens_seen": 9532664, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.5894128113879004, | |
| "grad_norm": 2.917534112930298, | |
| "learning_rate": 3.5264679715302496e-05, | |
| "loss": 1.3654, | |
| "num_input_tokens_seen": 9717800, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.6005338078291815, | |
| "grad_norm": 3.355299472808838, | |
| "learning_rate": 3.4986654804270464e-05, | |
| "loss": 1.3876, | |
| "num_input_tokens_seen": 9902536, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.6116548042704626, | |
| "grad_norm": 2.67924427986145, | |
| "learning_rate": 3.470862989323844e-05, | |
| "loss": 1.3575, | |
| "num_input_tokens_seen": 10082216, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.6227758007117438, | |
| "grad_norm": 3.040212392807007, | |
| "learning_rate": 3.44306049822064e-05, | |
| "loss": 1.3644, | |
| "num_input_tokens_seen": 10263992, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.6338967971530249, | |
| "grad_norm": 3.726254940032959, | |
| "learning_rate": 3.415258007117438e-05, | |
| "loss": 1.3616, | |
| "num_input_tokens_seen": 10444128, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.645017793594306, | |
| "grad_norm": 4.716592788696289, | |
| "learning_rate": 3.387455516014235e-05, | |
| "loss": 1.3662, | |
| "num_input_tokens_seen": 10626928, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.6561387900355872, | |
| "grad_norm": 2.9317166805267334, | |
| "learning_rate": 3.359653024911032e-05, | |
| "loss": 1.38, | |
| "num_input_tokens_seen": 10808960, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.6672597864768683, | |
| "grad_norm": 2.411684989929199, | |
| "learning_rate": 3.331850533807829e-05, | |
| "loss": 1.3782, | |
| "num_input_tokens_seen": 10990432, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.6672597864768683, | |
| "eval_loss": 1.185857892036438, | |
| "eval_runtime": 2.9618, | |
| "eval_samples_per_second": 844.079, | |
| "eval_steps_per_second": 105.679, | |
| "num_input_tokens_seen": 10990432, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.6783807829181495, | |
| "grad_norm": 2.765089273452759, | |
| "learning_rate": 3.3040480427046265e-05, | |
| "loss": 1.3808, | |
| "num_input_tokens_seen": 11172032, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.6895017793594306, | |
| "grad_norm": 2.808806896209717, | |
| "learning_rate": 3.2762455516014234e-05, | |
| "loss": 1.3925, | |
| "num_input_tokens_seen": 11356608, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.7006227758007118, | |
| "grad_norm": 2.6583220958709717, | |
| "learning_rate": 3.248443060498221e-05, | |
| "loss": 1.3716, | |
| "num_input_tokens_seen": 11538056, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.7117437722419929, | |
| "grad_norm": 2.2725088596343994, | |
| "learning_rate": 3.2206405693950184e-05, | |
| "loss": 1.3689, | |
| "num_input_tokens_seen": 11721264, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.722864768683274, | |
| "grad_norm": 2.927656412124634, | |
| "learning_rate": 3.1928380782918146e-05, | |
| "loss": 1.3722, | |
| "num_input_tokens_seen": 11896688, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.7339857651245552, | |
| "grad_norm": 2.0601186752319336, | |
| "learning_rate": 3.165035587188612e-05, | |
| "loss": 1.3408, | |
| "num_input_tokens_seen": 12084440, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.7451067615658363, | |
| "grad_norm": 2.5769150257110596, | |
| "learning_rate": 3.13723309608541e-05, | |
| "loss": 1.3874, | |
| "num_input_tokens_seen": 12264224, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.7562277580071174, | |
| "grad_norm": 2.845653772354126, | |
| "learning_rate": 3.1094306049822066e-05, | |
| "loss": 1.3755, | |
| "num_input_tokens_seen": 12446200, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.7673487544483986, | |
| "grad_norm": 2.3848676681518555, | |
| "learning_rate": 3.0816281138790034e-05, | |
| "loss": 1.3463, | |
| "num_input_tokens_seen": 12628992, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.7784697508896797, | |
| "grad_norm": 3.2360849380493164, | |
| "learning_rate": 3.053825622775801e-05, | |
| "loss": 1.3678, | |
| "num_input_tokens_seen": 12809808, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.7895907473309609, | |
| "grad_norm": 2.3211023807525635, | |
| "learning_rate": 3.026023131672598e-05, | |
| "loss": 1.3732, | |
| "num_input_tokens_seen": 12989912, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.800711743772242, | |
| "grad_norm": 3.599958658218384, | |
| "learning_rate": 2.9982206405693954e-05, | |
| "loss": 1.3606, | |
| "num_input_tokens_seen": 13170560, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.8118327402135231, | |
| "grad_norm": 2.0861263275146484, | |
| "learning_rate": 2.9704181494661926e-05, | |
| "loss": 1.3475, | |
| "num_input_tokens_seen": 13350424, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.8229537366548043, | |
| "grad_norm": 2.043938159942627, | |
| "learning_rate": 2.9426156583629895e-05, | |
| "loss": 1.3753, | |
| "num_input_tokens_seen": 13538640, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.8340747330960854, | |
| "grad_norm": 2.4880995750427246, | |
| "learning_rate": 2.9148131672597867e-05, | |
| "loss": 1.3497, | |
| "num_input_tokens_seen": 13730096, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.8451957295373665, | |
| "grad_norm": 2.535860300064087, | |
| "learning_rate": 2.8870106761565835e-05, | |
| "loss": 1.3708, | |
| "num_input_tokens_seen": 13909256, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.8563167259786477, | |
| "grad_norm": 2.499455213546753, | |
| "learning_rate": 2.8592081850533807e-05, | |
| "loss": 1.3546, | |
| "num_input_tokens_seen": 14101536, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.8674377224199288, | |
| "grad_norm": 2.3696117401123047, | |
| "learning_rate": 2.8314056939501783e-05, | |
| "loss": 1.3593, | |
| "num_input_tokens_seen": 14281392, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.87855871886121, | |
| "grad_norm": 3.260430097579956, | |
| "learning_rate": 2.803603202846975e-05, | |
| "loss": 1.3506, | |
| "num_input_tokens_seen": 14466320, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.8896797153024911, | |
| "grad_norm": 4.11997652053833, | |
| "learning_rate": 2.7758007117437723e-05, | |
| "loss": 1.3595, | |
| "num_input_tokens_seen": 14653328, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.8896797153024911, | |
| "eval_loss": 1.1787019968032837, | |
| "eval_runtime": 2.9698, | |
| "eval_samples_per_second": 841.809, | |
| "eval_steps_per_second": 105.394, | |
| "num_input_tokens_seen": 14653328, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.9008007117437722, | |
| "grad_norm": 1.932468056678772, | |
| "learning_rate": 2.7479982206405695e-05, | |
| "loss": 1.3559, | |
| "num_input_tokens_seen": 14837624, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.9119217081850534, | |
| "grad_norm": 2.6026477813720703, | |
| "learning_rate": 2.7201957295373664e-05, | |
| "loss": 1.3454, | |
| "num_input_tokens_seen": 15020192, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.9230427046263345, | |
| "grad_norm": 2.3455870151519775, | |
| "learning_rate": 2.692393238434164e-05, | |
| "loss": 1.3563, | |
| "num_input_tokens_seen": 15204288, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.9341637010676157, | |
| "grad_norm": 2.8757784366607666, | |
| "learning_rate": 2.664590747330961e-05, | |
| "loss": 1.3283, | |
| "num_input_tokens_seen": 15393744, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.9452846975088968, | |
| "grad_norm": 2.3972697257995605, | |
| "learning_rate": 2.636788256227758e-05, | |
| "loss": 1.3612, | |
| "num_input_tokens_seen": 15569824, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.9564056939501779, | |
| "grad_norm": 3.187290906906128, | |
| "learning_rate": 2.6089857651245552e-05, | |
| "loss": 1.3525, | |
| "num_input_tokens_seen": 15753768, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.9675266903914591, | |
| "grad_norm": 2.447659969329834, | |
| "learning_rate": 2.5811832740213527e-05, | |
| "loss": 1.3532, | |
| "num_input_tokens_seen": 15934952, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.9786476868327402, | |
| "grad_norm": 2.037935495376587, | |
| "learning_rate": 2.5533807829181493e-05, | |
| "loss": 1.3318, | |
| "num_input_tokens_seen": 16117896, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.9897686832740213, | |
| "grad_norm": 2.7559268474578857, | |
| "learning_rate": 2.5255782918149468e-05, | |
| "loss": 1.3325, | |
| "num_input_tokens_seen": 16298928, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 1.0008896797153024, | |
| "grad_norm": 2.2017595767974854, | |
| "learning_rate": 2.4977758007117437e-05, | |
| "loss": 1.3605, | |
| "num_input_tokens_seen": 16481072, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 1.0120106761565837, | |
| "grad_norm": 2.3097991943359375, | |
| "learning_rate": 2.4699733096085412e-05, | |
| "loss": 1.3404, | |
| "num_input_tokens_seen": 16664336, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 1.0231316725978647, | |
| "grad_norm": 2.6227993965148926, | |
| "learning_rate": 2.4421708185053384e-05, | |
| "loss": 1.3235, | |
| "num_input_tokens_seen": 16845800, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 1.0342526690391458, | |
| "grad_norm": 2.24474835395813, | |
| "learning_rate": 2.4143683274021353e-05, | |
| "loss": 1.3376, | |
| "num_input_tokens_seen": 17029008, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 1.045373665480427, | |
| "grad_norm": 2.7171192169189453, | |
| "learning_rate": 2.3865658362989325e-05, | |
| "loss": 1.3188, | |
| "num_input_tokens_seen": 17215936, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 1.0564946619217082, | |
| "grad_norm": 2.5323736667633057, | |
| "learning_rate": 2.3587633451957297e-05, | |
| "loss": 1.3369, | |
| "num_input_tokens_seen": 17401232, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 1.0676156583629894, | |
| "grad_norm": 2.267789363861084, | |
| "learning_rate": 2.330960854092527e-05, | |
| "loss": 1.3215, | |
| "num_input_tokens_seen": 17583344, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 1.0787366548042705, | |
| "grad_norm": 3.399862289428711, | |
| "learning_rate": 2.3031583629893237e-05, | |
| "loss": 1.3464, | |
| "num_input_tokens_seen": 17771600, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 1.0898576512455516, | |
| "grad_norm": 2.8749985694885254, | |
| "learning_rate": 2.2753558718861213e-05, | |
| "loss": 1.2928, | |
| "num_input_tokens_seen": 17957200, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 1.1009786476868326, | |
| "grad_norm": 2.6826517581939697, | |
| "learning_rate": 2.247553380782918e-05, | |
| "loss": 1.3191, | |
| "num_input_tokens_seen": 18138560, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 1.112099644128114, | |
| "grad_norm": 2.2963333129882812, | |
| "learning_rate": 2.2197508896797153e-05, | |
| "loss": 1.3059, | |
| "num_input_tokens_seen": 18326104, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 1.112099644128114, | |
| "eval_loss": 1.1665468215942383, | |
| "eval_runtime": 3.0497, | |
| "eval_samples_per_second": 819.748, | |
| "eval_steps_per_second": 102.632, | |
| "num_input_tokens_seen": 18326104, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 1.123220640569395, | |
| "grad_norm": 2.2697386741638184, | |
| "learning_rate": 2.1919483985765125e-05, | |
| "loss": 1.321, | |
| "num_input_tokens_seen": 18508560, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 1.1343416370106763, | |
| "grad_norm": 2.436851739883423, | |
| "learning_rate": 2.1641459074733097e-05, | |
| "loss": 1.3528, | |
| "num_input_tokens_seen": 18689264, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 1.1454626334519573, | |
| "grad_norm": 2.297527313232422, | |
| "learning_rate": 2.136343416370107e-05, | |
| "loss": 1.2987, | |
| "num_input_tokens_seen": 18874328, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 1.1565836298932384, | |
| "grad_norm": 2.5088889598846436, | |
| "learning_rate": 2.1085409252669038e-05, | |
| "loss": 1.3139, | |
| "num_input_tokens_seen": 19060472, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 1.1677046263345197, | |
| "grad_norm": 2.067575454711914, | |
| "learning_rate": 2.0807384341637014e-05, | |
| "loss": 1.2961, | |
| "num_input_tokens_seen": 19247416, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 1.1788256227758007, | |
| "grad_norm": 2.467543363571167, | |
| "learning_rate": 2.0529359430604982e-05, | |
| "loss": 1.3279, | |
| "num_input_tokens_seen": 19436888, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 1.1899466192170818, | |
| "grad_norm": 3.4245800971984863, | |
| "learning_rate": 2.0251334519572954e-05, | |
| "loss": 1.3303, | |
| "num_input_tokens_seen": 19616320, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 1.201067615658363, | |
| "grad_norm": 2.759120464324951, | |
| "learning_rate": 1.9973309608540926e-05, | |
| "loss": 1.3201, | |
| "num_input_tokens_seen": 19793576, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 1.2121886120996441, | |
| "grad_norm": 2.7749531269073486, | |
| "learning_rate": 1.9695284697508898e-05, | |
| "loss": 1.3194, | |
| "num_input_tokens_seen": 19980880, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 1.2233096085409252, | |
| "grad_norm": 2.4467661380767822, | |
| "learning_rate": 1.9417259786476867e-05, | |
| "loss": 1.3293, | |
| "num_input_tokens_seen": 20163048, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 1.2344306049822065, | |
| "grad_norm": 3.4420840740203857, | |
| "learning_rate": 1.9139234875444842e-05, | |
| "loss": 1.3015, | |
| "num_input_tokens_seen": 20340904, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 1.2455516014234875, | |
| "grad_norm": 2.4761664867401123, | |
| "learning_rate": 1.8861209964412814e-05, | |
| "loss": 1.3292, | |
| "num_input_tokens_seen": 20524328, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 1.2566725978647688, | |
| "grad_norm": 3.0505285263061523, | |
| "learning_rate": 1.8583185053380783e-05, | |
| "loss": 1.317, | |
| "num_input_tokens_seen": 20707808, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 1.2677935943060499, | |
| "grad_norm": 2.361429214477539, | |
| "learning_rate": 1.8305160142348755e-05, | |
| "loss": 1.3176, | |
| "num_input_tokens_seen": 20892744, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 1.278914590747331, | |
| "grad_norm": 1.9151511192321777, | |
| "learning_rate": 1.8027135231316727e-05, | |
| "loss": 1.3267, | |
| "num_input_tokens_seen": 21072216, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 1.290035587188612, | |
| "grad_norm": 3.0513691902160645, | |
| "learning_rate": 1.77491103202847e-05, | |
| "loss": 1.2948, | |
| "num_input_tokens_seen": 21256064, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 1.3011565836298933, | |
| "grad_norm": 1.7151504755020142, | |
| "learning_rate": 1.7471085409252668e-05, | |
| "loss": 1.2967, | |
| "num_input_tokens_seen": 21439648, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 1.3122775800711743, | |
| "grad_norm": 2.3011133670806885, | |
| "learning_rate": 1.7193060498220643e-05, | |
| "loss": 1.3199, | |
| "num_input_tokens_seen": 21621776, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 1.3233985765124556, | |
| "grad_norm": 2.1768672466278076, | |
| "learning_rate": 1.691503558718861e-05, | |
| "loss": 1.3233, | |
| "num_input_tokens_seen": 21810728, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 1.3345195729537367, | |
| "grad_norm": 3.249089241027832, | |
| "learning_rate": 1.6637010676156584e-05, | |
| "loss": 1.3298, | |
| "num_input_tokens_seen": 21991016, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 1.3345195729537367, | |
| "eval_loss": 1.1589475870132446, | |
| "eval_runtime": 3.0361, | |
| "eval_samples_per_second": 823.416, | |
| "eval_steps_per_second": 103.092, | |
| "num_input_tokens_seen": 21991016, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 1.3456405693950177, | |
| "grad_norm": 1.7282594442367554, | |
| "learning_rate": 1.6358985765124556e-05, | |
| "loss": 1.325, | |
| "num_input_tokens_seen": 22178168, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 1.3567615658362988, | |
| "grad_norm": 2.2659966945648193, | |
| "learning_rate": 1.6080960854092528e-05, | |
| "loss": 1.3321, | |
| "num_input_tokens_seen": 22359360, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 1.36788256227758, | |
| "grad_norm": 2.155791759490967, | |
| "learning_rate": 1.58029359430605e-05, | |
| "loss": 1.2995, | |
| "num_input_tokens_seen": 22543952, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 1.3790035587188612, | |
| "grad_norm": 2.902367353439331, | |
| "learning_rate": 1.5524911032028472e-05, | |
| "loss": 1.3061, | |
| "num_input_tokens_seen": 22728064, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 1.3901245551601424, | |
| "grad_norm": 2.168686866760254, | |
| "learning_rate": 1.5246886120996442e-05, | |
| "loss": 1.3231, | |
| "num_input_tokens_seen": 22913608, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 1.4012455516014235, | |
| "grad_norm": 2.682150363922119, | |
| "learning_rate": 1.4968861209964412e-05, | |
| "loss": 1.3091, | |
| "num_input_tokens_seen": 23101416, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 1.4123665480427046, | |
| "grad_norm": 2.82143497467041, | |
| "learning_rate": 1.4690836298932384e-05, | |
| "loss": 1.2922, | |
| "num_input_tokens_seen": 23281200, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 1.4234875444839858, | |
| "grad_norm": 2.1155049800872803, | |
| "learning_rate": 1.4412811387900358e-05, | |
| "loss": 1.3015, | |
| "num_input_tokens_seen": 23462552, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 1.434608540925267, | |
| "grad_norm": 2.6742069721221924, | |
| "learning_rate": 1.4134786476868328e-05, | |
| "loss": 1.3108, | |
| "num_input_tokens_seen": 23640424, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 1.445729537366548, | |
| "grad_norm": 2.575198173522949, | |
| "learning_rate": 1.3856761565836299e-05, | |
| "loss": 1.2987, | |
| "num_input_tokens_seen": 23825392, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 1.4568505338078293, | |
| "grad_norm": 3.2627415657043457, | |
| "learning_rate": 1.3578736654804272e-05, | |
| "loss": 1.3096, | |
| "num_input_tokens_seen": 24004872, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 1.4679715302491103, | |
| "grad_norm": 3.2727510929107666, | |
| "learning_rate": 1.3300711743772243e-05, | |
| "loss": 1.3108, | |
| "num_input_tokens_seen": 24186064, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 1.4790925266903914, | |
| "grad_norm": 2.187281608581543, | |
| "learning_rate": 1.3022686832740213e-05, | |
| "loss": 1.3239, | |
| "num_input_tokens_seen": 24367968, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 1.4902135231316727, | |
| "grad_norm": 3.5734856128692627, | |
| "learning_rate": 1.2744661921708187e-05, | |
| "loss": 1.3241, | |
| "num_input_tokens_seen": 24551184, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 1.5013345195729537, | |
| "grad_norm": 2.756578207015991, | |
| "learning_rate": 1.2466637010676157e-05, | |
| "loss": 1.307, | |
| "num_input_tokens_seen": 24730288, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 1.512455516014235, | |
| "grad_norm": 2.3752739429473877, | |
| "learning_rate": 1.2188612099644127e-05, | |
| "loss": 1.2963, | |
| "num_input_tokens_seen": 24914816, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 1.523576512455516, | |
| "grad_norm": 3.7108139991760254, | |
| "learning_rate": 1.1910587188612101e-05, | |
| "loss": 1.3126, | |
| "num_input_tokens_seen": 25100648, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 1.5346975088967971, | |
| "grad_norm": 2.5954089164733887, | |
| "learning_rate": 1.1632562277580072e-05, | |
| "loss": 1.3053, | |
| "num_input_tokens_seen": 25282312, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 1.5458185053380782, | |
| "grad_norm": 3.0645289421081543, | |
| "learning_rate": 1.1354537366548044e-05, | |
| "loss": 1.2976, | |
| "num_input_tokens_seen": 25465504, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 1.5569395017793595, | |
| "grad_norm": 2.3734166622161865, | |
| "learning_rate": 1.1076512455516016e-05, | |
| "loss": 1.2994, | |
| "num_input_tokens_seen": 25656600, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 1.5569395017793595, | |
| "eval_loss": 1.1562061309814453, | |
| "eval_runtime": 2.9007, | |
| "eval_samples_per_second": 861.851, | |
| "eval_steps_per_second": 107.904, | |
| "num_input_tokens_seen": 25656600, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 1.5680604982206405, | |
| "grad_norm": 2.8284573554992676, | |
| "learning_rate": 1.0798487544483986e-05, | |
| "loss": 1.3125, | |
| "num_input_tokens_seen": 25844216, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 1.5791814946619218, | |
| "grad_norm": 1.8697574138641357, | |
| "learning_rate": 1.0520462633451958e-05, | |
| "loss": 1.3247, | |
| "num_input_tokens_seen": 26026824, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 1.5903024911032029, | |
| "grad_norm": 2.4746646881103516, | |
| "learning_rate": 1.024243772241993e-05, | |
| "loss": 1.3295, | |
| "num_input_tokens_seen": 26211760, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 1.601423487544484, | |
| "grad_norm": 2.415778398513794, | |
| "learning_rate": 9.9644128113879e-06, | |
| "loss": 1.3191, | |
| "num_input_tokens_seen": 26393624, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 1.612544483985765, | |
| "grad_norm": 2.263882875442505, | |
| "learning_rate": 9.686387900355872e-06, | |
| "loss": 1.3136, | |
| "num_input_tokens_seen": 26576872, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 1.6236654804270463, | |
| "grad_norm": 2.436645269393921, | |
| "learning_rate": 9.408362989323843e-06, | |
| "loss": 1.286, | |
| "num_input_tokens_seen": 26763752, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 1.6347864768683276, | |
| "grad_norm": 2.739278793334961, | |
| "learning_rate": 9.130338078291816e-06, | |
| "loss": 1.2712, | |
| "num_input_tokens_seen": 26943856, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 1.6459074733096086, | |
| "grad_norm": 2.406345844268799, | |
| "learning_rate": 8.852313167259788e-06, | |
| "loss": 1.305, | |
| "num_input_tokens_seen": 27122192, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 1.6570284697508897, | |
| "grad_norm": 2.1659858226776123, | |
| "learning_rate": 8.574288256227759e-06, | |
| "loss": 1.312, | |
| "num_input_tokens_seen": 27305912, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 1.6681494661921707, | |
| "grad_norm": 2.7831106185913086, | |
| "learning_rate": 8.29626334519573e-06, | |
| "loss": 1.2997, | |
| "num_input_tokens_seen": 27485504, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 1.6792704626334518, | |
| "grad_norm": 2.9885916709899902, | |
| "learning_rate": 8.018238434163701e-06, | |
| "loss": 1.3187, | |
| "num_input_tokens_seen": 27665584, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 1.690391459074733, | |
| "grad_norm": 2.5562667846679688, | |
| "learning_rate": 7.740213523131673e-06, | |
| "loss": 1.3142, | |
| "num_input_tokens_seen": 27851448, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 1.7015124555160144, | |
| "grad_norm": 2.9897525310516357, | |
| "learning_rate": 7.462188612099645e-06, | |
| "loss": 1.3071, | |
| "num_input_tokens_seen": 28033824, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 1.7126334519572954, | |
| "grad_norm": 2.6760592460632324, | |
| "learning_rate": 7.184163701067615e-06, | |
| "loss": 1.3049, | |
| "num_input_tokens_seen": 28220000, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 1.7237544483985765, | |
| "grad_norm": 2.314532995223999, | |
| "learning_rate": 6.906138790035588e-06, | |
| "loss": 1.3124, | |
| "num_input_tokens_seen": 28403400, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 1.7348754448398576, | |
| "grad_norm": 2.2853899002075195, | |
| "learning_rate": 6.6281138790035586e-06, | |
| "loss": 1.3188, | |
| "num_input_tokens_seen": 28585168, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 1.7459964412811388, | |
| "grad_norm": 2.462369918823242, | |
| "learning_rate": 6.3500889679715306e-06, | |
| "loss": 1.3146, | |
| "num_input_tokens_seen": 28767960, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 1.75711743772242, | |
| "grad_norm": 2.626847505569458, | |
| "learning_rate": 6.072064056939502e-06, | |
| "loss": 1.3082, | |
| "num_input_tokens_seen": 28950344, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 1.7682384341637012, | |
| "grad_norm": 3.058187484741211, | |
| "learning_rate": 5.794039145907473e-06, | |
| "loss": 1.301, | |
| "num_input_tokens_seen": 29133248, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 1.7793594306049823, | |
| "grad_norm": 1.932924747467041, | |
| "learning_rate": 5.516014234875446e-06, | |
| "loss": 1.2952, | |
| "num_input_tokens_seen": 29314808, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 1.7793594306049823, | |
| "eval_loss": 1.1517876386642456, | |
| "eval_runtime": 3.0679, | |
| "eval_samples_per_second": 814.88, | |
| "eval_steps_per_second": 102.023, | |
| "num_input_tokens_seen": 29314808, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 1.7904804270462633, | |
| "grad_norm": 3.326586961746216, | |
| "learning_rate": 5.237989323843417e-06, | |
| "loss": 1.3161, | |
| "num_input_tokens_seen": 29504040, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 1.8016014234875444, | |
| "grad_norm": 3.139636754989624, | |
| "learning_rate": 4.959964412811388e-06, | |
| "loss": 1.3, | |
| "num_input_tokens_seen": 29686712, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 1.8127224199288257, | |
| "grad_norm": 2.5767743587493896, | |
| "learning_rate": 4.681939501779359e-06, | |
| "loss": 1.2934, | |
| "num_input_tokens_seen": 29868136, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 1.8238434163701067, | |
| "grad_norm": 2.1930339336395264, | |
| "learning_rate": 4.4039145907473305e-06, | |
| "loss": 1.3104, | |
| "num_input_tokens_seen": 30050016, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 1.834964412811388, | |
| "grad_norm": 2.9890389442443848, | |
| "learning_rate": 4.125889679715303e-06, | |
| "loss": 1.303, | |
| "num_input_tokens_seen": 30234304, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 1.846085409252669, | |
| "grad_norm": 2.9597034454345703, | |
| "learning_rate": 3.8478647686832745e-06, | |
| "loss": 1.3024, | |
| "num_input_tokens_seen": 30419248, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 1.8572064056939501, | |
| "grad_norm": 3.17301082611084, | |
| "learning_rate": 3.5698398576512457e-06, | |
| "loss": 1.2926, | |
| "num_input_tokens_seen": 30601776, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 1.8683274021352312, | |
| "grad_norm": 2.677340269088745, | |
| "learning_rate": 3.291814946619217e-06, | |
| "loss": 1.3184, | |
| "num_input_tokens_seen": 30783800, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 1.8794483985765125, | |
| "grad_norm": 3.5369062423706055, | |
| "learning_rate": 3.013790035587189e-06, | |
| "loss": 1.2858, | |
| "num_input_tokens_seen": 30968472, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 1.8905693950177938, | |
| "grad_norm": 2.870908737182617, | |
| "learning_rate": 2.73576512455516e-06, | |
| "loss": 1.3107, | |
| "num_input_tokens_seen": 31147072, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 1.9016903914590748, | |
| "grad_norm": 2.1105244159698486, | |
| "learning_rate": 2.457740213523132e-06, | |
| "loss": 1.2797, | |
| "num_input_tokens_seen": 31330536, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 1.9128113879003559, | |
| "grad_norm": 4.080565452575684, | |
| "learning_rate": 2.1797153024911032e-06, | |
| "loss": 1.2991, | |
| "num_input_tokens_seen": 31507752, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 1.923932384341637, | |
| "grad_norm": 3.036339044570923, | |
| "learning_rate": 1.901690391459075e-06, | |
| "loss": 1.2884, | |
| "num_input_tokens_seen": 31685504, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 1.935053380782918, | |
| "grad_norm": 2.3314597606658936, | |
| "learning_rate": 1.6236654804270462e-06, | |
| "loss": 1.2935, | |
| "num_input_tokens_seen": 31865648, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 1.9461743772241993, | |
| "grad_norm": 2.4721710681915283, | |
| "learning_rate": 1.3456405693950178e-06, | |
| "loss": 1.3124, | |
| "num_input_tokens_seen": 32046704, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 1.9572953736654806, | |
| "grad_norm": 2.4747238159179688, | |
| "learning_rate": 1.0676156583629894e-06, | |
| "loss": 1.2941, | |
| "num_input_tokens_seen": 32235800, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 1.9684163701067616, | |
| "grad_norm": 2.4421122074127197, | |
| "learning_rate": 7.89590747330961e-07, | |
| "loss": 1.3187, | |
| "num_input_tokens_seen": 32415672, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 1.9795373665480427, | |
| "grad_norm": 3.469534158706665, | |
| "learning_rate": 5.115658362989324e-07, | |
| "loss": 1.2897, | |
| "num_input_tokens_seen": 32597136, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 1.9906583629893237, | |
| "grad_norm": 3.1344103813171387, | |
| "learning_rate": 2.335409252669039e-07, | |
| "loss": 1.2902, | |
| "num_input_tokens_seen": 32782688, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "num_input_tokens_seen": 32939232, | |
| "step": 89920, | |
| "total_flos": 1.1959161056722944e+16, | |
| "train_loss": 1.358397777055082, | |
| "train_runtime": 3698.2072, | |
| "train_samples_per_second": 194.516, | |
| "train_steps_per_second": 24.314, | |
| "train_tokens_per_second": 8905.5 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 89920, | |
| "num_input_tokens_seen": 32939232, | |
| "num_train_epochs": 2, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.1959161056722944e+16, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |