| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.6442705936493327, |
| "eval_steps": 500, |
| "global_step": 2800, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05752416014726185, |
| "grad_norm": 0.5827537775039673, |
| "learning_rate": 0.00016513000460193283, |
| "loss": 5.0544, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.1150483202945237, |
| "grad_norm": 0.5847251415252686, |
| "learning_rate": 0.00016024045098941557, |
| "loss": 4.2362, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.1150483202945237, |
| "eval_loss": 3.4322922229766846, |
| "eval_runtime": 35.9031, |
| "eval_samples_per_second": 109.74, |
| "eval_steps_per_second": 3.231, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.17257248044178555, |
| "grad_norm": 0.6557937264442444, |
| "learning_rate": 0.0001553508973768983, |
| "loss": 4.0566, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.2300966405890474, |
| "grad_norm": 0.6734243631362915, |
| "learning_rate": 0.00015046134376438104, |
| "loss": 3.9349, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.2300966405890474, |
| "eval_loss": 3.1976470947265625, |
| "eval_runtime": 35.3562, |
| "eval_samples_per_second": 111.437, |
| "eval_steps_per_second": 3.281, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.28762080073630925, |
| "grad_norm": 0.6826881766319275, |
| "learning_rate": 0.0001455717901518638, |
| "loss": 3.856, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.3451449608835711, |
| "grad_norm": 0.6653180122375488, |
| "learning_rate": 0.00014068223653934653, |
| "loss": 3.7903, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.3451449608835711, |
| "eval_loss": 3.0786449909210205, |
| "eval_runtime": 36.0086, |
| "eval_samples_per_second": 109.418, |
| "eval_steps_per_second": 3.221, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.40266912103083297, |
| "grad_norm": 0.7236284613609314, |
| "learning_rate": 0.0001357926829268293, |
| "loss": 3.7295, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.4601932811780948, |
| "grad_norm": 0.7614400386810303, |
| "learning_rate": 0.000130903129314312, |
| "loss": 3.692, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.4601932811780948, |
| "eval_loss": 3.0163052082061768, |
| "eval_runtime": 35.3697, |
| "eval_samples_per_second": 111.395, |
| "eval_steps_per_second": 3.28, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.5177174413253567, |
| "grad_norm": 0.7231118083000183, |
| "learning_rate": 0.00012601357570179476, |
| "loss": 3.6578, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.5752416014726185, |
| "grad_norm": 0.7403096556663513, |
| "learning_rate": 0.00012112402208927751, |
| "loss": 3.6348, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5752416014726185, |
| "eval_loss": 2.9728314876556396, |
| "eval_runtime": 35.3987, |
| "eval_samples_per_second": 111.303, |
| "eval_steps_per_second": 3.277, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.6327657616198803, |
| "grad_norm": 0.8307600021362305, |
| "learning_rate": 0.00011623446847676025, |
| "loss": 3.5985, |
| "step": 2750 |
| } |
| ], |
| "logging_steps": 250, |
| "max_steps": 8692, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 350, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2148271757721600.0, |
| "train_batch_size": 34, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|