| { |
| "best_metric": 1.1523780822753906, |
| "best_model_checkpoint": "flowers_image_detection/checkpoint-2780", |
| "epoch": 20.0, |
| "eval_steps": 500, |
| "global_step": 2780, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.42882882882882883, |
| "eval_loss": 3.241873025894165, |
| "eval_runtime": 10.7562, |
| "eval_samples_per_second": 103.196, |
| "eval_steps_per_second": 12.923, |
| "step": 139 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.6378378378378379, |
| "eval_loss": 2.929600477218628, |
| "eval_runtime": 10.7704, |
| "eval_samples_per_second": 103.06, |
| "eval_steps_per_second": 12.906, |
| "step": 278 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.7585585585585586, |
| "eval_loss": 2.6281425952911377, |
| "eval_runtime": 10.6539, |
| "eval_samples_per_second": 104.187, |
| "eval_steps_per_second": 13.047, |
| "step": 417 |
| }, |
| { |
| "epoch": 3.6, |
| "learning_rate": 8.351648351648353e-06, |
| "loss": 2.9311, |
| "step": 500 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.7972972972972973, |
| "eval_loss": 2.3711183071136475, |
| "eval_runtime": 10.683, |
| "eval_samples_per_second": 103.903, |
| "eval_steps_per_second": 13.011, |
| "step": 556 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.8315315315315316, |
| "eval_loss": 2.161473274230957, |
| "eval_runtime": 10.8746, |
| "eval_samples_per_second": 102.073, |
| "eval_steps_per_second": 12.782, |
| "step": 695 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_accuracy": 0.8468468468468469, |
| "eval_loss": 1.9815605878829956, |
| "eval_runtime": 10.6654, |
| "eval_samples_per_second": 104.075, |
| "eval_steps_per_second": 13.033, |
| "step": 834 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_accuracy": 0.8657657657657658, |
| "eval_loss": 1.8246235847473145, |
| "eval_runtime": 10.8392, |
| "eval_samples_per_second": 102.406, |
| "eval_steps_per_second": 12.824, |
| "step": 973 |
| }, |
| { |
| "epoch": 7.19, |
| "learning_rate": 6.5201465201465204e-06, |
| "loss": 1.9674, |
| "step": 1000 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_accuracy": 0.8747747747747747, |
| "eval_loss": 1.7018743753433228, |
| "eval_runtime": 10.6972, |
| "eval_samples_per_second": 103.765, |
| "eval_steps_per_second": 12.994, |
| "step": 1112 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_accuracy": 0.8738738738738738, |
| "eval_loss": 1.5962648391723633, |
| "eval_runtime": 10.9948, |
| "eval_samples_per_second": 100.957, |
| "eval_steps_per_second": 12.642, |
| "step": 1251 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_accuracy": 0.881981981981982, |
| "eval_loss": 1.5058796405792236, |
| "eval_runtime": 10.5932, |
| "eval_samples_per_second": 104.784, |
| "eval_steps_per_second": 13.122, |
| "step": 1390 |
| }, |
| { |
| "epoch": 10.79, |
| "learning_rate": 4.688644688644689e-06, |
| "loss": 1.4254, |
| "step": 1500 |
| }, |
| { |
| "epoch": 11.0, |
| "eval_accuracy": 0.8855855855855855, |
| "eval_loss": 1.4283403158187866, |
| "eval_runtime": 10.9479, |
| "eval_samples_per_second": 101.39, |
| "eval_steps_per_second": 12.697, |
| "step": 1529 |
| }, |
| { |
| "epoch": 12.0, |
| "eval_accuracy": 0.8882882882882883, |
| "eval_loss": 1.3638710975646973, |
| "eval_runtime": 10.6646, |
| "eval_samples_per_second": 104.082, |
| "eval_steps_per_second": 13.034, |
| "step": 1668 |
| }, |
| { |
| "epoch": 13.0, |
| "eval_accuracy": 0.8900900900900901, |
| "eval_loss": 1.3107937574386597, |
| "eval_runtime": 10.893, |
| "eval_samples_per_second": 101.9, |
| "eval_steps_per_second": 12.76, |
| "step": 1807 |
| }, |
| { |
| "epoch": 14.0, |
| "eval_accuracy": 0.8927927927927928, |
| "eval_loss": 1.2697221040725708, |
| "eval_runtime": 10.6225, |
| "eval_samples_per_second": 104.495, |
| "eval_steps_per_second": 13.085, |
| "step": 1946 |
| }, |
| { |
| "epoch": 14.39, |
| "learning_rate": 2.8571428571428573e-06, |
| "loss": 1.1236, |
| "step": 2000 |
| }, |
| { |
| "epoch": 15.0, |
| "eval_accuracy": 0.8927927927927928, |
| "eval_loss": 1.2321548461914062, |
| "eval_runtime": 10.9992, |
| "eval_samples_per_second": 100.917, |
| "eval_steps_per_second": 12.637, |
| "step": 2085 |
| }, |
| { |
| "epoch": 16.0, |
| "eval_accuracy": 0.8936936936936937, |
| "eval_loss": 1.2019907236099243, |
| "eval_runtime": 10.6905, |
| "eval_samples_per_second": 103.831, |
| "eval_steps_per_second": 13.002, |
| "step": 2224 |
| }, |
| { |
| "epoch": 17.0, |
| "eval_accuracy": 0.8945945945945946, |
| "eval_loss": 1.1794919967651367, |
| "eval_runtime": 10.9046, |
| "eval_samples_per_second": 101.792, |
| "eval_steps_per_second": 12.747, |
| "step": 2363 |
| }, |
| { |
| "epoch": 17.99, |
| "learning_rate": 1.0256410256410257e-06, |
| "loss": 0.9693, |
| "step": 2500 |
| }, |
| { |
| "epoch": 18.0, |
| "eval_accuracy": 0.9, |
| "eval_loss": 1.1641703844070435, |
| "eval_runtime": 10.5836, |
| "eval_samples_per_second": 104.879, |
| "eval_steps_per_second": 13.134, |
| "step": 2502 |
| }, |
| { |
| "epoch": 19.0, |
| "eval_accuracy": 0.9009009009009009, |
| "eval_loss": 1.155342698097229, |
| "eval_runtime": 10.5817, |
| "eval_samples_per_second": 104.898, |
| "eval_steps_per_second": 13.136, |
| "step": 2641 |
| }, |
| { |
| "epoch": 20.0, |
| "eval_accuracy": 0.9, |
| "eval_loss": 1.1523780822753906, |
| "eval_runtime": 10.6258, |
| "eval_samples_per_second": 104.463, |
| "eval_steps_per_second": 13.081, |
| "step": 2780 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 2780, |
| "num_train_epochs": 20, |
| "save_steps": 500, |
| "total_flos": 6.876937896135229e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|