| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 252, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.11904761904761904, | |
| "grad_norm": 26.617824687216622, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7508, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.23809523809523808, | |
| "grad_norm": 1.8689351819227027, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6727, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 1.5792334038916935, | |
| "learning_rate": 5e-06, | |
| "loss": 0.633, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.47619047619047616, | |
| "grad_norm": 0.6218650960930915, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6164, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.5952380952380952, | |
| "grad_norm": 1.140712919876654, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6066, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.8602592078396757, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5966, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.8333333333333334, | |
| "grad_norm": 0.8320424935867776, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5944, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.9523809523809523, | |
| "grad_norm": 0.49165318108555545, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5861, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.5798346996307373, | |
| "eval_runtime": 8.4592, | |
| "eval_samples_per_second": 266.928, | |
| "eval_steps_per_second": 1.064, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.0714285714285714, | |
| "grad_norm": 0.6573425533819024, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5663, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.1904761904761905, | |
| "grad_norm": 0.4670077635391962, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5517, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.3095238095238095, | |
| "grad_norm": 0.5237829600316058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5546, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 0.5547024135708416, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5526, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.5476190476190477, | |
| "grad_norm": 0.47689471851329457, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5466, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.5230783383262022, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5472, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.7857142857142856, | |
| "grad_norm": 0.5674014749330538, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5464, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.9047619047619047, | |
| "grad_norm": 0.47822121321020516, | |
| "learning_rate": 5e-06, | |
| "loss": 0.543, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.5688267946243286, | |
| "eval_runtime": 8.5104, | |
| "eval_samples_per_second": 265.321, | |
| "eval_steps_per_second": 1.058, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.0238095238095237, | |
| "grad_norm": 0.9393861948324864, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5387, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.142857142857143, | |
| "grad_norm": 0.5880801701644413, | |
| "learning_rate": 5e-06, | |
| "loss": 0.513, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.261904761904762, | |
| "grad_norm": 0.7573868967310241, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5084, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.380952380952381, | |
| "grad_norm": 0.5149453817597104, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5065, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 0.48769983126777056, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5096, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.619047619047619, | |
| "grad_norm": 0.6168117709375799, | |
| "learning_rate": 5e-06, | |
| "loss": 0.51, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.738095238095238, | |
| "grad_norm": 0.6182244851253141, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5075, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.857142857142857, | |
| "grad_norm": 0.5115158649844643, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5129, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.9761904761904763, | |
| "grad_norm": 0.6056742291106577, | |
| "learning_rate": 5e-06, | |
| "loss": 0.51, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.5704450607299805, | |
| "eval_runtime": 8.2035, | |
| "eval_samples_per_second": 275.249, | |
| "eval_steps_per_second": 1.097, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 252, | |
| "total_flos": 422109385850880.0, | |
| "train_loss": 0.5631871157222323, | |
| "train_runtime": 1874.1971, | |
| "train_samples_per_second": 68.655, | |
| "train_steps_per_second": 0.134 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 252, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 422109385850880.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |