| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 250, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.04024144869215292, | |
| "grad_norm": 128.68338012695312, | |
| "learning_rate": 6.34920634920635e-07, | |
| "loss": 5.7142, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.08048289738430583, | |
| "grad_norm": 124.65970611572266, | |
| "learning_rate": 1.4285714285714286e-06, | |
| "loss": 5.8126, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.12072434607645875, | |
| "grad_norm": 63.15693664550781, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 4.9144, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.16096579476861167, | |
| "grad_norm": 26.409881591796875, | |
| "learning_rate": 3.015873015873016e-06, | |
| "loss": 4.1026, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2012072434607646, | |
| "grad_norm": 18.272987365722656, | |
| "learning_rate": 3.80952380952381e-06, | |
| "loss": 3.8405, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.2414486921529175, | |
| "grad_norm": 13.232375144958496, | |
| "learning_rate": 4.603174603174604e-06, | |
| "loss": 3.5374, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.28169014084507044, | |
| "grad_norm": 50.732887268066406, | |
| "learning_rate": 5.396825396825397e-06, | |
| "loss": 3.3776, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.32193158953722334, | |
| "grad_norm": 11.886248588562012, | |
| "learning_rate": 6.1904761904761914e-06, | |
| "loss": 3.2729, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.36217303822937624, | |
| "grad_norm": 12.157098770141602, | |
| "learning_rate": 6.984126984126984e-06, | |
| "loss": 3.0134, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.4024144869215292, | |
| "grad_norm": 11.598234176635742, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 2.9235, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.4426559356136821, | |
| "grad_norm": 11.506558418273926, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 2.584, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.482897384305835, | |
| "grad_norm": 14.368203163146973, | |
| "learning_rate": 9.365079365079366e-06, | |
| "loss": 2.4279, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.5231388329979879, | |
| "grad_norm": 12.335148811340332, | |
| "learning_rate": 9.999921879324127e-06, | |
| "loss": 2.244, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.5633802816901409, | |
| "grad_norm": 17.981578826904297, | |
| "learning_rate": 9.997187911979252e-06, | |
| "loss": 2.0648, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.6036217303822937, | |
| "grad_norm": 12.924208641052246, | |
| "learning_rate": 9.990550351633784e-06, | |
| "loss": 2.0333, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.6438631790744467, | |
| "grad_norm": 36.98970031738281, | |
| "learning_rate": 9.980014383270668e-06, | |
| "loss": 1.7114, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.6841046277665996, | |
| "grad_norm": 11.499068260192871, | |
| "learning_rate": 9.965588237145219e-06, | |
| "loss": 1.725, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.7243460764587525, | |
| "grad_norm": 13.697612762451172, | |
| "learning_rate": 9.947283182355982e-06, | |
| "loss": 1.7076, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.7645875251509054, | |
| "grad_norm": 11.36514663696289, | |
| "learning_rate": 9.925113518041796e-06, | |
| "loss": 1.4663, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.8048289738430584, | |
| "grad_norm": 11.20559024810791, | |
| "learning_rate": 9.899096562211902e-06, | |
| "loss": 1.1067, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.8450704225352113, | |
| "grad_norm": 10.782366752624512, | |
| "learning_rate": 9.869252638217846e-06, | |
| "loss": 1.2299, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.8853118712273642, | |
| "grad_norm": 12.094294548034668, | |
| "learning_rate": 9.83560505887773e-06, | |
| "loss": 0.9908, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.9255533199195171, | |
| "grad_norm": 10.271987915039062, | |
| "learning_rate": 9.798180108265218e-06, | |
| "loss": 0.8628, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.96579476861167, | |
| "grad_norm": 13.298333168029785, | |
| "learning_rate": 9.757007021177529e-06, | |
| "loss": 0.7913, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 7.651159286499023, | |
| "learning_rate": 9.712117960298433e-06, | |
| "loss": 0.6121, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.040241448692153, | |
| "grad_norm": 15.86595630645752, | |
| "learning_rate": 9.663547991074129e-06, | |
| "loss": 0.4909, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.080482897384306, | |
| "grad_norm": 9.43641185760498, | |
| "learning_rate": 9.611335054321576e-06, | |
| "loss": 0.3678, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.1207243460764587, | |
| "grad_norm": 8.751080513000488, | |
| "learning_rate": 9.555519936590739e-06, | |
| "loss": 0.3778, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.1609657947686116, | |
| "grad_norm": 6.040717601776123, | |
| "learning_rate": 9.496146238303846e-06, | |
| "loss": 0.3935, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.2012072434607646, | |
| "grad_norm": 8.183173179626465, | |
| "learning_rate": 9.433260339696564e-06, | |
| "loss": 0.3653, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.2414486921529175, | |
| "grad_norm": 7.956486701965332, | |
| "learning_rate": 9.366911364587726e-06, | |
| "loss": 0.3123, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.2816901408450705, | |
| "grad_norm": 11.906622886657715, | |
| "learning_rate": 9.297151142005852e-06, | |
| "loss": 0.335, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.3219315895372232, | |
| "grad_norm": 6.626152515411377, | |
| "learning_rate": 9.224034165702506e-06, | |
| "loss": 0.3211, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.3621730382293762, | |
| "grad_norm": 7.222238063812256, | |
| "learning_rate": 9.147617551584066e-06, | |
| "loss": 0.2317, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.4024144869215291, | |
| "grad_norm": 6.305289268493652, | |
| "learning_rate": 9.067960993095176e-06, | |
| "loss": 0.2209, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.442655935613682, | |
| "grad_norm": 7.212320804595947, | |
| "learning_rate": 8.985126714588739e-06, | |
| "loss": 0.2155, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.482897384305835, | |
| "grad_norm": 6.313644886016846, | |
| "learning_rate": 8.899179422718877e-06, | |
| "loss": 0.2051, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.5231388329979878, | |
| "grad_norm": 7.122276782989502, | |
| "learning_rate": 8.810186255894804e-06, | |
| "loss": 0.2192, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.563380281690141, | |
| "grad_norm": 5.341350555419922, | |
| "learning_rate": 8.718216731835131e-06, | |
| "loss": 0.1937, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.6036217303822937, | |
| "grad_norm": 3.3275582790374756, | |
| "learning_rate": 8.623342693263549e-06, | |
| "loss": 0.1281, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.6438631790744467, | |
| "grad_norm": 6.510648727416992, | |
| "learning_rate": 8.525638251788312e-06, | |
| "loss": 0.1872, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.6841046277665996, | |
| "grad_norm": 14.44485855102539, | |
| "learning_rate": 8.425179730009368e-06, | |
| "loss": 0.1722, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.7243460764587524, | |
| "grad_norm": 5.061001777648926, | |
| "learning_rate": 8.322045601898354e-06, | |
| "loss": 0.1602, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.7645875251509056, | |
| "grad_norm": 4.675739288330078, | |
| "learning_rate": 8.216316431498028e-06, | |
| "loss": 0.1034, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.8048289738430583, | |
| "grad_norm": 4.156198501586914, | |
| "learning_rate": 8.108074809989032e-06, | |
| "loss": 0.1249, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.8450704225352113, | |
| "grad_norm": 6.796616554260254, | |
| "learning_rate": 7.99740529117313e-06, | |
| "loss": 0.1205, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.8853118712273642, | |
| "grad_norm": 6.765830039978027, | |
| "learning_rate": 7.88439432542334e-06, | |
| "loss": 0.1609, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.925553319919517, | |
| "grad_norm": 3.9120049476623535, | |
| "learning_rate": 7.769130192152538e-06, | |
| "loss": 0.0915, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.9657947686116701, | |
| "grad_norm": 3.127553701400757, | |
| "learning_rate": 7.651702930853287e-06, | |
| "loss": 0.0684, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 3.80415415763855, | |
| "learning_rate": 7.532204270762786e-06, | |
| "loss": 0.1202, | |
| "step": 250 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 625, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.9107777718528246e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |