| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 500, | |
| "global_step": 21930, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.11399908800729594, | |
| "grad_norm": 1.4812531471252441, | |
| "learning_rate": 1.9544915640674875e-05, | |
| "loss": 0.1287, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.22799817601459188, | |
| "grad_norm": 2.198870897293091, | |
| "learning_rate": 1.9088919288645692e-05, | |
| "loss": 0.1045, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.34199726402188785, | |
| "grad_norm": 1.3517733812332153, | |
| "learning_rate": 1.863292293661651e-05, | |
| "loss": 0.0895, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.45599635202918376, | |
| "grad_norm": 1.352580189704895, | |
| "learning_rate": 1.8176926584587324e-05, | |
| "loss": 0.0766, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.5699954400364797, | |
| "grad_norm": 1.960016131401062, | |
| "learning_rate": 1.772093023255814e-05, | |
| "loss": 0.0713, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.6839945280437757, | |
| "grad_norm": 0.5876455903053284, | |
| "learning_rate": 1.7264933880528958e-05, | |
| "loss": 0.0657, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.7979936160510716, | |
| "grad_norm": 0.9299254417419434, | |
| "learning_rate": 1.6808937528499772e-05, | |
| "loss": 0.0623, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.9119927040583675, | |
| "grad_norm": 1.3892526626586914, | |
| "learning_rate": 1.635294117647059e-05, | |
| "loss": 0.0621, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.0259917920656634, | |
| "grad_norm": 1.7030184268951416, | |
| "learning_rate": 1.5896944824441403e-05, | |
| "loss": 0.0585, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.1399908800729595, | |
| "grad_norm": 0.9616603255271912, | |
| "learning_rate": 1.5440948472412224e-05, | |
| "loss": 0.0448, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.2539899680802553, | |
| "grad_norm": 1.245125412940979, | |
| "learning_rate": 1.4984952120383038e-05, | |
| "loss": 0.0429, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.3679890560875512, | |
| "grad_norm": 1.2846943140029907, | |
| "learning_rate": 1.4528955768353854e-05, | |
| "loss": 0.0408, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.4819881440948472, | |
| "grad_norm": 1.0010948181152344, | |
| "learning_rate": 1.407295941632467e-05, | |
| "loss": 0.0386, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.5959872321021433, | |
| "grad_norm": 1.1298205852508545, | |
| "learning_rate": 1.3616963064295488e-05, | |
| "loss": 0.0364, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.7099863201094392, | |
| "grad_norm": 0.5932161211967468, | |
| "learning_rate": 1.3160966712266304e-05, | |
| "loss": 0.0344, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.823985408116735, | |
| "grad_norm": 1.195202112197876, | |
| "learning_rate": 1.270497036023712e-05, | |
| "loss": 0.033, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.937984496124031, | |
| "grad_norm": 1.2507325410842896, | |
| "learning_rate": 1.2248974008207935e-05, | |
| "loss": 0.0321, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.0519835841313268, | |
| "grad_norm": 2.5148258209228516, | |
| "learning_rate": 1.1792977656178753e-05, | |
| "loss": 0.0278, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.165982672138623, | |
| "grad_norm": 0.7329283356666565, | |
| "learning_rate": 1.1336981304149568e-05, | |
| "loss": 0.0233, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 2.279981760145919, | |
| "grad_norm": 0.42915332317352295, | |
| "learning_rate": 1.0880984952120384e-05, | |
| "loss": 0.0226, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.3939808481532148, | |
| "grad_norm": 0.7090346813201904, | |
| "learning_rate": 1.04249886000912e-05, | |
| "loss": 0.0223, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.5079799361605106, | |
| "grad_norm": 0.6480665802955627, | |
| "learning_rate": 9.968992248062017e-06, | |
| "loss": 0.0204, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 2.621979024167807, | |
| "grad_norm": 1.7029985189437866, | |
| "learning_rate": 9.512995896032832e-06, | |
| "loss": 0.0221, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 2.7359781121751023, | |
| "grad_norm": 0.9722331166267395, | |
| "learning_rate": 9.056999544003648e-06, | |
| "loss": 0.0208, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 2.8499772001823986, | |
| "grad_norm": 0.3082011938095093, | |
| "learning_rate": 8.601003191974465e-06, | |
| "loss": 0.0204, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 2.9639762881896945, | |
| "grad_norm": 0.33128827810287476, | |
| "learning_rate": 8.145006839945281e-06, | |
| "loss": 0.0207, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 3.0779753761969904, | |
| "grad_norm": 0.7698410749435425, | |
| "learning_rate": 7.689010487916098e-06, | |
| "loss": 0.0171, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 3.191974464204286, | |
| "grad_norm": 0.7875366806983948, | |
| "learning_rate": 7.233014135886913e-06, | |
| "loss": 0.0156, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 3.305973552211582, | |
| "grad_norm": 0.6971263289451599, | |
| "learning_rate": 6.77701778385773e-06, | |
| "loss": 0.0157, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 3.4199726402188784, | |
| "grad_norm": 0.464100182056427, | |
| "learning_rate": 6.321021431828546e-06, | |
| "loss": 0.0143, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 3.5339717282261742, | |
| "grad_norm": 0.5803436636924744, | |
| "learning_rate": 5.8650250797993626e-06, | |
| "loss": 0.0147, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 3.64797081623347, | |
| "grad_norm": 0.7879688739776611, | |
| "learning_rate": 5.409028727770178e-06, | |
| "loss": 0.0137, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 3.761969904240766, | |
| "grad_norm": 1.0888534784317017, | |
| "learning_rate": 4.953032375740995e-06, | |
| "loss": 0.0139, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 3.875968992248062, | |
| "grad_norm": 0.3893286883831024, | |
| "learning_rate": 4.49703602371181e-06, | |
| "loss": 0.0137, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 3.989968080255358, | |
| "grad_norm": 0.4917721748352051, | |
| "learning_rate": 4.041039671682627e-06, | |
| "loss": 0.0136, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 4.1039671682626535, | |
| "grad_norm": 1.25238835811615, | |
| "learning_rate": 3.585043319653443e-06, | |
| "loss": 0.0117, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 4.21796625626995, | |
| "grad_norm": 0.9064317941665649, | |
| "learning_rate": 3.129046967624259e-06, | |
| "loss": 0.0108, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 4.331965344277246, | |
| "grad_norm": 0.4522368311882019, | |
| "learning_rate": 2.6730506155950754e-06, | |
| "loss": 0.0105, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 4.4459644322845415, | |
| "grad_norm": 1.0190762281417847, | |
| "learning_rate": 2.217054263565892e-06, | |
| "loss": 0.0108, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 4.559963520291838, | |
| "grad_norm": 0.23909200727939606, | |
| "learning_rate": 1.7610579115367079e-06, | |
| "loss": 0.0114, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 4.673962608299133, | |
| "grad_norm": 0.8679990768432617, | |
| "learning_rate": 1.3050615595075241e-06, | |
| "loss": 0.0105, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 4.7879616963064295, | |
| "grad_norm": 0.4421948790550232, | |
| "learning_rate": 8.490652074783402e-07, | |
| "loss": 0.0109, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 4.901960784313726, | |
| "grad_norm": 1.2229384183883667, | |
| "learning_rate": 3.930688554491564e-07, | |
| "loss": 0.0103, | |
| "step": 21500 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 21930, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2352410932543488.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |