| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0469397146801658, |
| "eval_steps": 500, |
| "global_step": 4550, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05752416014726185, |
| "grad_norm": 0.5827537775039673, |
| "learning_rate": 0.00016513000460193283, |
| "loss": 5.0544, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.1150483202945237, |
| "grad_norm": 0.5847251415252686, |
| "learning_rate": 0.00016024045098941557, |
| "loss": 4.2362, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.1150483202945237, |
| "eval_loss": 3.4322922229766846, |
| "eval_runtime": 35.9031, |
| "eval_samples_per_second": 109.74, |
| "eval_steps_per_second": 3.231, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.17257248044178555, |
| "grad_norm": 0.6557937264442444, |
| "learning_rate": 0.0001553508973768983, |
| "loss": 4.0566, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.2300966405890474, |
| "grad_norm": 0.6734243631362915, |
| "learning_rate": 0.00015046134376438104, |
| "loss": 3.9349, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.2300966405890474, |
| "eval_loss": 3.1976470947265625, |
| "eval_runtime": 35.3562, |
| "eval_samples_per_second": 111.437, |
| "eval_steps_per_second": 3.281, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.28762080073630925, |
| "grad_norm": 0.6826881766319275, |
| "learning_rate": 0.0001455717901518638, |
| "loss": 3.856, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.3451449608835711, |
| "grad_norm": 0.6653180122375488, |
| "learning_rate": 0.00014068223653934653, |
| "loss": 3.7903, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.3451449608835711, |
| "eval_loss": 3.0786449909210205, |
| "eval_runtime": 36.0086, |
| "eval_samples_per_second": 109.418, |
| "eval_steps_per_second": 3.221, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.40266912103083297, |
| "grad_norm": 0.7236284613609314, |
| "learning_rate": 0.0001357926829268293, |
| "loss": 3.7295, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.4601932811780948, |
| "grad_norm": 0.7614400386810303, |
| "learning_rate": 0.000130903129314312, |
| "loss": 3.692, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.4601932811780948, |
| "eval_loss": 3.0163052082061768, |
| "eval_runtime": 35.3697, |
| "eval_samples_per_second": 111.395, |
| "eval_steps_per_second": 3.28, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.5177174413253567, |
| "grad_norm": 0.7231118083000183, |
| "learning_rate": 0.00012601357570179476, |
| "loss": 3.6578, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.5752416014726185, |
| "grad_norm": 0.7403096556663513, |
| "learning_rate": 0.00012112402208927751, |
| "loss": 3.6348, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5752416014726185, |
| "eval_loss": 2.9728314876556396, |
| "eval_runtime": 35.3987, |
| "eval_samples_per_second": 111.303, |
| "eval_steps_per_second": 3.277, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.6327657616198803, |
| "grad_norm": 0.8307600021362305, |
| "learning_rate": 0.00011623446847676025, |
| "loss": 3.5985, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.6902899217671422, |
| "grad_norm": 0.692379891872406, |
| "learning_rate": 0.000111344914864243, |
| "loss": 3.5853, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.6902899217671422, |
| "eval_loss": 2.949444055557251, |
| "eval_runtime": 35.5037, |
| "eval_samples_per_second": 110.974, |
| "eval_steps_per_second": 3.267, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.747814081914404, |
| "grad_norm": 0.6413500905036926, |
| "learning_rate": 0.00010645536125172573, |
| "loss": 3.553, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.8053382420616659, |
| "grad_norm": 0.747322678565979, |
| "learning_rate": 0.00010156580763920848, |
| "loss": 3.5418, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.8053382420616659, |
| "eval_loss": 2.9215950965881348, |
| "eval_runtime": 35.7029, |
| "eval_samples_per_second": 110.355, |
| "eval_steps_per_second": 3.249, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.8628624022089277, |
| "grad_norm": 0.7153074741363525, |
| "learning_rate": 9.667625402669122e-05, |
| "loss": 3.5174, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.9203865623561897, |
| "grad_norm": 0.7129898071289062, |
| "learning_rate": 9.178670041417396e-05, |
| "loss": 3.5023, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.9203865623561897, |
| "eval_loss": 2.8931143283843994, |
| "eval_runtime": 35.6334, |
| "eval_samples_per_second": 110.57, |
| "eval_steps_per_second": 3.255, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.9779107225034515, |
| "grad_norm": 0.6831061244010925, |
| "learning_rate": 8.68971468016567e-05, |
| "loss": 3.4839, |
| "step": 4250 |
| }, |
| { |
| "epoch": 1.0354348826507134, |
| "grad_norm": 0.6215568780899048, |
| "learning_rate": 8.200759318913945e-05, |
| "loss": 3.4785, |
| "step": 4500 |
| }, |
| { |
| "epoch": 1.0354348826507134, |
| "eval_loss": 2.888288974761963, |
| "eval_runtime": 35.5637, |
| "eval_samples_per_second": 110.787, |
| "eval_steps_per_second": 3.262, |
| "step": 4500 |
| } |
| ], |
| "logging_steps": 250, |
| "max_steps": 8692, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 350, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3490625683980288.0, |
| "train_batch_size": 34, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|