| { |
| "best_metric": 0.9975817923186344, |
| "best_model_checkpoint": "VitDisease/checkpoint-384", |
| "epoch": 2.9960988296488944, |
| "eval_steps": 500, |
| "global_step": 576, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05, |
| "learning_rate": 8.620689655172414e-06, |
| "loss": 3.6405, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.7241379310344828e-05, |
| "loss": 3.5875, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 2.5862068965517244e-05, |
| "loss": 3.4689, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 3.4482758620689657e-05, |
| "loss": 3.29, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 4.3103448275862066e-05, |
| "loss": 3.0405, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 4.980694980694981e-05, |
| "loss": 2.7252, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 4.884169884169885e-05, |
| "loss": 2.3972, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 4.787644787644788e-05, |
| "loss": 2.1134, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 4.6911196911196914e-05, |
| "loss": 1.8631, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 4.594594594594595e-05, |
| "loss": 1.6843, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 4.498069498069498e-05, |
| "loss": 1.5059, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 4.401544401544402e-05, |
| "loss": 1.3564, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.305019305019305e-05, |
| "loss": 1.2375, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 4.2084942084942086e-05, |
| "loss": 1.1429, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 4.111969111969112e-05, |
| "loss": 1.05, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 4.015444015444015e-05, |
| "loss": 0.9708, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 3.918918918918919e-05, |
| "loss": 0.9041, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 3.822393822393823e-05, |
| "loss": 0.8441, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 3.725868725868726e-05, |
| "loss": 0.7869, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.9927453769559033, |
| "eval_loss": 0.7515802979469299, |
| "eval_runtime": 71.4706, |
| "eval_samples_per_second": 98.362, |
| "eval_steps_per_second": 1.539, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 3.6293436293436295e-05, |
| "loss": 0.7364, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 3.532818532818533e-05, |
| "loss": 0.6957, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 3.436293436293436e-05, |
| "loss": 0.6565, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 3.33976833976834e-05, |
| "loss": 0.6302, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 3.2432432432432436e-05, |
| "loss": 0.5991, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 3.1467181467181466e-05, |
| "loss": 0.5714, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 3.0501930501930504e-05, |
| "loss": 0.5468, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 2.953667953667954e-05, |
| "loss": 0.5258, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 2.857142857142857e-05, |
| "loss": 0.5043, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 2.7606177606177608e-05, |
| "loss": 0.4892, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 2.6640926640926645e-05, |
| "loss": 0.4738, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 2.5675675675675675e-05, |
| "loss": 0.4531, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 2.4710424710424712e-05, |
| "loss": 0.4447, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 2.3745173745173746e-05, |
| "loss": 0.4344, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 2.277992277992278e-05, |
| "loss": 0.422, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 2.1814671814671817e-05, |
| "loss": 0.4091, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 2.084942084942085e-05, |
| "loss": 0.396, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 1.9884169884169884e-05, |
| "loss": 0.3917, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 1.891891891891892e-05, |
| "loss": 0.3828, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.9975817923186344, |
| "eval_loss": 0.3735862076282501, |
| "eval_runtime": 71.8988, |
| "eval_samples_per_second": 97.776, |
| "eval_steps_per_second": 1.53, |
| "step": 384 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 1.7953667953667955e-05, |
| "loss": 0.3697, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 1.698841698841699e-05, |
| "loss": 0.3584, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 1.6023166023166026e-05, |
| "loss": 0.3505, |
| "step": 410 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 1.505791505791506e-05, |
| "loss": 0.3437, |
| "step": 420 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 1.4092664092664093e-05, |
| "loss": 0.3378, |
| "step": 430 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 1.3127413127413127e-05, |
| "loss": 0.3344, |
| "step": 440 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 1.2162162162162164e-05, |
| "loss": 0.3294, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 1.1196911196911197e-05, |
| "loss": 0.3237, |
| "step": 460 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.0231660231660233e-05, |
| "loss": 0.3187, |
| "step": 470 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 9.266409266409266e-06, |
| "loss": 0.3173, |
| "step": 480 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 8.301158301158302e-06, |
| "loss": 0.3168, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 7.335907335907337e-06, |
| "loss": 0.3132, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 6.370656370656371e-06, |
| "loss": 0.3111, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.405405405405406e-06, |
| "loss": 0.3061, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 4.4401544401544405e-06, |
| "loss": 0.3077, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 3.474903474903475e-06, |
| "loss": 0.3033, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 2.5096525096525096e-06, |
| "loss": 0.3021, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 1.5444015444015445e-06, |
| "loss": 0.3015, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 5.791505791505791e-07, |
| "loss": 0.299, |
| "step": 570 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.9975817923186344, |
| "eval_loss": 0.30786001682281494, |
| "eval_runtime": 70.6316, |
| "eval_samples_per_second": 99.531, |
| "eval_steps_per_second": 1.557, |
| "step": 576 |
| }, |
| { |
| "epoch": 3.0, |
| "step": 576, |
| "total_flos": 1.142879536778512e+19, |
| "train_loss": 0.932237885064549, |
| "train_runtime": 3472.3536, |
| "train_samples_per_second": 42.512, |
| "train_steps_per_second": 0.166 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 576, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "total_flos": 1.142879536778512e+19, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|