| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 37.0, |
| "eval_steps": 0, |
| "global_step": 37, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "learning_rate": 1e-05, |
| "loss": 0.002, |
| "step": 1 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.5070388375349495, |
| "learning_rate": 9.722222222222223e-06, |
| "loss": 0.0119, |
| "step": 2 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.5070388375349495, |
| "learning_rate": 9.444444444444445e-06, |
| "loss": 0.0032, |
| "step": 3 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.31938507733027843, |
| "learning_rate": 9.166666666666666e-06, |
| "loss": -0.0015, |
| "step": 4 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 0.31938507733027843, |
| "learning_rate": 8.888888888888888e-06, |
| "loss": 0.0073, |
| "step": 5 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 0.43541534478865934, |
| "learning_rate": 8.611111111111112e-06, |
| "loss": -0.0001, |
| "step": 6 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 0.43541534478865934, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 0.0014, |
| "step": 7 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.5428982959575133, |
| "learning_rate": 8.055555555555557e-06, |
| "loss": -0.0026, |
| "step": 8 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 0.5428982959575133, |
| "learning_rate": 7.77777777777778e-06, |
| "loss": 0.0012, |
| "step": 9 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 0.9728643220805211, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 0.0027, |
| "step": 10 |
| }, |
| { |
| "epoch": 11.0, |
| "grad_norm": 0.9728643220805211, |
| "learning_rate": 7.222222222222223e-06, |
| "loss": -0.0017, |
| "step": 11 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 0.8713152262560704, |
| "learning_rate": 6.944444444444445e-06, |
| "loss": -0.0037, |
| "step": 12 |
| }, |
| { |
| "epoch": 13.0, |
| "grad_norm": 0.8713152262560704, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": -0.0057, |
| "step": 13 |
| }, |
| { |
| "epoch": 14.0, |
| "grad_norm": 1.2483582246181377, |
| "learning_rate": 6.3888888888888885e-06, |
| "loss": -0.0042, |
| "step": 14 |
| }, |
| { |
| "epoch": 15.0, |
| "grad_norm": 1.2483582246181377, |
| "learning_rate": 6.111111111111112e-06, |
| "loss": -0.0059, |
| "step": 15 |
| }, |
| { |
| "epoch": 16.0, |
| "grad_norm": 1.338516376367072, |
| "learning_rate": 5.833333333333334e-06, |
| "loss": -0.0054, |
| "step": 16 |
| }, |
| { |
| "epoch": 17.0, |
| "grad_norm": 1.338516376367072, |
| "learning_rate": 5.555555555555557e-06, |
| "loss": -0.0068, |
| "step": 17 |
| }, |
| { |
| "epoch": 18.0, |
| "grad_norm": 1.8582031188698118, |
| "learning_rate": 5.2777777777777785e-06, |
| "loss": -0.0046, |
| "step": 18 |
| }, |
| { |
| "epoch": 19.0, |
| "grad_norm": 1.8582031188698118, |
| "learning_rate": 5e-06, |
| "loss": -0.0113, |
| "step": 19 |
| }, |
| { |
| "epoch": 20.0, |
| "grad_norm": 2.091199450117605, |
| "learning_rate": 4.722222222222222e-06, |
| "loss": -0.0107, |
| "step": 20 |
| }, |
| { |
| "epoch": 21.0, |
| "grad_norm": 2.091199450117605, |
| "learning_rate": 4.444444444444444e-06, |
| "loss": -0.0096, |
| "step": 21 |
| }, |
| { |
| "epoch": 22.0, |
| "grad_norm": 2.314746448196812, |
| "learning_rate": 4.166666666666667e-06, |
| "loss": -0.012, |
| "step": 22 |
| }, |
| { |
| "epoch": 23.0, |
| "grad_norm": 2.314746448196812, |
| "learning_rate": 3.88888888888889e-06, |
| "loss": -0.0153, |
| "step": 23 |
| }, |
| { |
| "epoch": 24.0, |
| "grad_norm": 2.2454811354700968, |
| "learning_rate": 3.6111111111111115e-06, |
| "loss": -0.0111, |
| "step": 24 |
| }, |
| { |
| "epoch": 25.0, |
| "grad_norm": 2.2454811354700968, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": -0.0088, |
| "step": 25 |
| }, |
| { |
| "epoch": 26.0, |
| "grad_norm": 2.811270150717608, |
| "learning_rate": 3.055555555555556e-06, |
| "loss": -0.0149, |
| "step": 26 |
| }, |
| { |
| "epoch": 27.0, |
| "grad_norm": 2.811270150717608, |
| "learning_rate": 2.7777777777777783e-06, |
| "loss": -0.019, |
| "step": 27 |
| }, |
| { |
| "epoch": 28.0, |
| "grad_norm": 2.678124329751125, |
| "learning_rate": 2.5e-06, |
| "loss": -0.0171, |
| "step": 28 |
| }, |
| { |
| "epoch": 29.0, |
| "grad_norm": 2.678124329751125, |
| "learning_rate": 2.222222222222222e-06, |
| "loss": -0.0129, |
| "step": 29 |
| }, |
| { |
| "epoch": 30.0, |
| "grad_norm": 2.596705407383602, |
| "learning_rate": 1.944444444444445e-06, |
| "loss": -0.0182, |
| "step": 30 |
| }, |
| { |
| "epoch": 31.0, |
| "grad_norm": 2.596705407383602, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": -0.0202, |
| "step": 31 |
| }, |
| { |
| "epoch": 32.0, |
| "grad_norm": 3.201965282085345, |
| "learning_rate": 1.3888888888888892e-06, |
| "loss": -0.023, |
| "step": 32 |
| }, |
| { |
| "epoch": 33.0, |
| "grad_norm": 3.201965282085345, |
| "learning_rate": 1.111111111111111e-06, |
| "loss": -0.0188, |
| "step": 33 |
| }, |
| { |
| "epoch": 34.0, |
| "grad_norm": 2.7138214701675016, |
| "learning_rate": 8.333333333333333e-07, |
| "loss": -0.0216, |
| "step": 34 |
| }, |
| { |
| "epoch": 35.0, |
| "grad_norm": 2.7138214701675016, |
| "learning_rate": 5.555555555555555e-07, |
| "loss": -0.025, |
| "step": 35 |
| }, |
| { |
| "epoch": 36.0, |
| "grad_norm": 3.3327313390783866, |
| "learning_rate": 2.7777777777777776e-07, |
| "loss": -0.0233, |
| "step": 36 |
| }, |
| { |
| "epoch": 37.0, |
| "grad_norm": 3.3327313390783866, |
| "learning_rate": 0.0, |
| "loss": -0.0247, |
| "step": 37 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 37, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 37, |
| "save_steps": 3, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|