| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 37.0, |
| "eval_steps": 0, |
| "global_step": 37, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "learning_rate": 1e-05, |
| "loss": 0.0018, |
| "step": 1 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.41081874815204006, |
| "learning_rate": 9.722222222222223e-06, |
| "loss": 0.0168, |
| "step": 2 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.41081874815204006, |
| "learning_rate": 9.444444444444445e-06, |
| "loss": -0.0013, |
| "step": 3 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.21691827337332767, |
| "learning_rate": 9.166666666666666e-06, |
| "loss": -0.0011, |
| "step": 4 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 0.21691827337332767, |
| "learning_rate": 8.888888888888888e-06, |
| "loss": -0.0008, |
| "step": 5 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 0.30914843458215285, |
| "learning_rate": 8.611111111111112e-06, |
| "loss": 0.0063, |
| "step": 6 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 0.30914843458215285, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 0.0021, |
| "step": 7 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.24907676722982752, |
| "learning_rate": 8.055555555555557e-06, |
| "loss": -0.0016, |
| "step": 8 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 0.24907676722982752, |
| "learning_rate": 7.77777777777778e-06, |
| "loss": -0.001, |
| "step": 9 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 0.5628992153072051, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 0.0035, |
| "step": 10 |
| }, |
| { |
| "epoch": 11.0, |
| "grad_norm": 0.5628992153072051, |
| "learning_rate": 7.222222222222223e-06, |
| "loss": -0.0018, |
| "step": 11 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 0.34356051372386226, |
| "learning_rate": 6.944444444444445e-06, |
| "loss": -0.0037, |
| "step": 12 |
| }, |
| { |
| "epoch": 13.0, |
| "grad_norm": 0.34356051372386226, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": -0.0047, |
| "step": 13 |
| }, |
| { |
| "epoch": 14.0, |
| "grad_norm": 0.4417329548789222, |
| "learning_rate": 6.3888888888888885e-06, |
| "loss": -0.0004, |
| "step": 14 |
| }, |
| { |
| "epoch": 15.0, |
| "grad_norm": 0.4417329548789222, |
| "learning_rate": 6.111111111111112e-06, |
| "loss": -0.0033, |
| "step": 15 |
| }, |
| { |
| "epoch": 16.0, |
| "grad_norm": 0.5371843961219906, |
| "learning_rate": 5.833333333333334e-06, |
| "loss": -0.0032, |
| "step": 16 |
| }, |
| { |
| "epoch": 17.0, |
| "grad_norm": 0.5371843961219906, |
| "learning_rate": 5.555555555555557e-06, |
| "loss": -0.0049, |
| "step": 17 |
| }, |
| { |
| "epoch": 18.0, |
| "grad_norm": 0.650124697585636, |
| "learning_rate": 5.2777777777777785e-06, |
| "loss": 0.0031, |
| "step": 18 |
| }, |
| { |
| "epoch": 19.0, |
| "grad_norm": 0.650124697585636, |
| "learning_rate": 5e-06, |
| "loss": -0.008, |
| "step": 19 |
| }, |
| { |
| "epoch": 20.0, |
| "grad_norm": 0.7037784230649523, |
| "learning_rate": 4.722222222222222e-06, |
| "loss": -0.0083, |
| "step": 20 |
| }, |
| { |
| "epoch": 21.0, |
| "grad_norm": 0.7037784230649523, |
| "learning_rate": 4.444444444444444e-06, |
| "loss": -0.0077, |
| "step": 21 |
| }, |
| { |
| "epoch": 22.0, |
| "grad_norm": 0.843411211123133, |
| "learning_rate": 4.166666666666667e-06, |
| "loss": -0.0095, |
| "step": 22 |
| }, |
| { |
| "epoch": 23.0, |
| "grad_norm": 0.843411211123133, |
| "learning_rate": 3.88888888888889e-06, |
| "loss": -0.0114, |
| "step": 23 |
| }, |
| { |
| "epoch": 24.0, |
| "grad_norm": 0.9388026732090895, |
| "learning_rate": 3.6111111111111115e-06, |
| "loss": -0.0115, |
| "step": 24 |
| }, |
| { |
| "epoch": 25.0, |
| "grad_norm": 0.9388026732090895, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": -0.0065, |
| "step": 25 |
| }, |
| { |
| "epoch": 26.0, |
| "grad_norm": 1.0928962196063456, |
| "learning_rate": 3.055555555555556e-06, |
| "loss": -0.0126, |
| "step": 26 |
| }, |
| { |
| "epoch": 27.0, |
| "grad_norm": 1.0928962196063456, |
| "learning_rate": 2.7777777777777783e-06, |
| "loss": -0.0158, |
| "step": 27 |
| }, |
| { |
| "epoch": 28.0, |
| "grad_norm": 1.1652418300050935, |
| "learning_rate": 2.5e-06, |
| "loss": -0.0131, |
| "step": 28 |
| }, |
| { |
| "epoch": 29.0, |
| "grad_norm": 1.1652418300050935, |
| "learning_rate": 2.222222222222222e-06, |
| "loss": -0.0138, |
| "step": 29 |
| }, |
| { |
| "epoch": 30.0, |
| "grad_norm": 1.2543616879035648, |
| "learning_rate": 1.944444444444445e-06, |
| "loss": -0.0137, |
| "step": 30 |
| }, |
| { |
| "epoch": 31.0, |
| "grad_norm": 1.2543616879035648, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": -0.0159, |
| "step": 31 |
| }, |
| { |
| "epoch": 32.0, |
| "grad_norm": 1.3658554215296144, |
| "learning_rate": 1.3888888888888892e-06, |
| "loss": 0.0005, |
| "step": 32 |
| }, |
| { |
| "epoch": 33.0, |
| "grad_norm": 1.3658554215296144, |
| "learning_rate": 1.111111111111111e-06, |
| "loss": -0.0166, |
| "step": 33 |
| }, |
| { |
| "epoch": 34.0, |
| "grad_norm": 1.328571156329741, |
| "learning_rate": 8.333333333333333e-07, |
| "loss": -0.019, |
| "step": 34 |
| }, |
| { |
| "epoch": 35.0, |
| "grad_norm": 1.328571156329741, |
| "learning_rate": 5.555555555555555e-07, |
| "loss": -0.02, |
| "step": 35 |
| }, |
| { |
| "epoch": 36.0, |
| "grad_norm": 1.3972218245865853, |
| "learning_rate": 2.7777777777777776e-07, |
| "loss": -0.0206, |
| "step": 36 |
| }, |
| { |
| "epoch": 37.0, |
| "grad_norm": 1.3972218245865853, |
| "learning_rate": 0.0, |
| "loss": -0.0216, |
| "step": 37 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 37, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 37, |
| "save_steps": 3, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|