| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 6.024096385542169, |
| "global_step": 500, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.12, |
| "learning_rate": 4e-05, |
| "loss": 6.9946, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 8e-05, |
| "loss": 5.455, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 0.00012, |
| "loss": 4.45, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 0.00016, |
| "loss": 3.8135, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 0.0002, |
| "loss": 3.4829, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 0.00019555555555555556, |
| "loss": 3.1942, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 0.00019111111111111114, |
| "loss": 2.9794, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 0.0001866666666666667, |
| "loss": 2.8266, |
| "step": 80 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 0.00018222222222222224, |
| "loss": 2.6337, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 0.00017777777777777779, |
| "loss": 2.5671, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 0.00017333333333333334, |
| "loss": 2.4814, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 0.00016888888888888889, |
| "loss": 2.4243, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 0.00016444444444444444, |
| "loss": 2.3516, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 0.00016, |
| "loss": 2.3141, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 0.00015555555555555556, |
| "loss": 2.2503, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 0.0001511111111111111, |
| "loss": 2.1908, |
| "step": 160 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 0.00014666666666666666, |
| "loss": 2.1537, |
| "step": 170 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 0.00014222222222222224, |
| "loss": 2.0442, |
| "step": 180 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 0.0001377777777777778, |
| "loss": 2.0096, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 0.00013333333333333334, |
| "loss": 1.9876, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 0.00012888888888888892, |
| "loss": 1.9789, |
| "step": 210 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 0.00012444444444444444, |
| "loss": 1.9375, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 0.00012, |
| "loss": 1.9222, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 0.00011555555555555555, |
| "loss": 1.8738, |
| "step": 240 |
| }, |
| { |
| "epoch": 3.01, |
| "learning_rate": 0.00011111111111111112, |
| "loss": 1.8951, |
| "step": 250 |
| }, |
| { |
| "epoch": 3.01, |
| "eval_bleu": 3.9246, |
| "eval_gen_len": 18.6281, |
| "eval_loss": 2.265472412109375, |
| "eval_runtime": 47.3348, |
| "eval_samples_per_second": 88.73, |
| "eval_steps_per_second": 0.697, |
| "step": 250 |
| }, |
| { |
| "epoch": 3.13, |
| "learning_rate": 0.00010666666666666667, |
| "loss": 1.7962, |
| "step": 260 |
| }, |
| { |
| "epoch": 3.25, |
| "learning_rate": 0.00010222222222222222, |
| "loss": 1.7742, |
| "step": 270 |
| }, |
| { |
| "epoch": 3.37, |
| "learning_rate": 9.777777777777778e-05, |
| "loss": 1.7561, |
| "step": 280 |
| }, |
| { |
| "epoch": 3.49, |
| "learning_rate": 9.333333333333334e-05, |
| "loss": 1.7248, |
| "step": 290 |
| }, |
| { |
| "epoch": 3.61, |
| "learning_rate": 8.888888888888889e-05, |
| "loss": 1.7343, |
| "step": 300 |
| }, |
| { |
| "epoch": 3.73, |
| "learning_rate": 8.444444444444444e-05, |
| "loss": 1.7053, |
| "step": 310 |
| }, |
| { |
| "epoch": 3.86, |
| "learning_rate": 8e-05, |
| "loss": 1.6866, |
| "step": 320 |
| }, |
| { |
| "epoch": 3.98, |
| "learning_rate": 7.555555555555556e-05, |
| "loss": 1.665, |
| "step": 330 |
| }, |
| { |
| "epoch": 4.1, |
| "learning_rate": 7.111111111111112e-05, |
| "loss": 1.6531, |
| "step": 340 |
| }, |
| { |
| "epoch": 4.22, |
| "learning_rate": 6.666666666666667e-05, |
| "loss": 1.615, |
| "step": 350 |
| }, |
| { |
| "epoch": 4.34, |
| "learning_rate": 6.222222222222222e-05, |
| "loss": 1.6093, |
| "step": 360 |
| }, |
| { |
| "epoch": 4.46, |
| "learning_rate": 5.7777777777777776e-05, |
| "loss": 1.6087, |
| "step": 370 |
| }, |
| { |
| "epoch": 4.58, |
| "learning_rate": 5.333333333333333e-05, |
| "loss": 1.586, |
| "step": 380 |
| }, |
| { |
| "epoch": 4.7, |
| "learning_rate": 4.888888888888889e-05, |
| "loss": 1.5655, |
| "step": 390 |
| }, |
| { |
| "epoch": 4.82, |
| "learning_rate": 4.4444444444444447e-05, |
| "loss": 1.5485, |
| "step": 400 |
| }, |
| { |
| "epoch": 4.94, |
| "learning_rate": 4e-05, |
| "loss": 1.582, |
| "step": 410 |
| }, |
| { |
| "epoch": 5.06, |
| "learning_rate": 3.555555555555556e-05, |
| "loss": 1.5616, |
| "step": 420 |
| }, |
| { |
| "epoch": 5.18, |
| "learning_rate": 3.111111111111111e-05, |
| "loss": 1.5579, |
| "step": 430 |
| }, |
| { |
| "epoch": 5.3, |
| "learning_rate": 2.6666666666666667e-05, |
| "loss": 1.5261, |
| "step": 440 |
| }, |
| { |
| "epoch": 5.42, |
| "learning_rate": 2.2222222222222223e-05, |
| "loss": 1.5194, |
| "step": 450 |
| }, |
| { |
| "epoch": 5.54, |
| "learning_rate": 1.777777777777778e-05, |
| "loss": 1.5099, |
| "step": 460 |
| }, |
| { |
| "epoch": 5.66, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 1.5227, |
| "step": 470 |
| }, |
| { |
| "epoch": 5.78, |
| "learning_rate": 8.88888888888889e-06, |
| "loss": 1.4967, |
| "step": 480 |
| }, |
| { |
| "epoch": 5.9, |
| "learning_rate": 4.444444444444445e-06, |
| "loss": 1.5215, |
| "step": 490 |
| }, |
| { |
| "epoch": 6.02, |
| "learning_rate": 0.0, |
| "loss": 1.5435, |
| "step": 500 |
| }, |
| { |
| "epoch": 6.02, |
| "eval_bleu": 4.1139, |
| "eval_gen_len": 18.6052, |
| "eval_loss": 2.260429859161377, |
| "eval_runtime": 18.0892, |
| "eval_samples_per_second": 232.183, |
| "eval_steps_per_second": 1.824, |
| "step": 500 |
| }, |
| { |
| "epoch": 6.02, |
| "step": 500, |
| "total_flos": 3.792343111041024e+16, |
| "train_loss": 2.211637247085571, |
| "train_runtime": 368.6992, |
| "train_samples_per_second": 347.166, |
| "train_steps_per_second": 1.356 |
| } |
| ], |
| "max_steps": 500, |
| "num_train_epochs": 7, |
| "total_flos": 3.792343111041024e+16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|