| { | |
| "best_metric": 0.030118444934487343, | |
| "best_model_checkpoint": "./vit-base-crack-classification/checkpoint-609", | |
| "epoch": 3.0, | |
| "eval_steps": 100, | |
| "global_step": 609, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00019671592775041052, | |
| "loss": 1.4989, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 0.00019343185550082105, | |
| "loss": 1.059, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 0.00019014778325123153, | |
| "loss": 0.8111, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.00018686371100164203, | |
| "loss": 0.6727, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.00018357963875205257, | |
| "loss": 0.7606, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00018029556650246307, | |
| "loss": 0.688, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00017701149425287358, | |
| "loss": 0.6076, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 0.00017372742200328408, | |
| "loss": 0.5333, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.0001704433497536946, | |
| "loss": 0.4982, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0001671592775041051, | |
| "loss": 0.4319, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.0001638752052545156, | |
| "loss": 0.5499, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 0.00016059113300492613, | |
| "loss": 0.3999, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00015730706075533664, | |
| "loss": 0.443, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 0.00015402298850574712, | |
| "loss": 0.417, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 0.00015073891625615765, | |
| "loss": 0.3383, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 0.00014745484400656815, | |
| "loss": 0.4005, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 0.00014417077175697866, | |
| "loss": 0.3608, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 0.00014088669950738917, | |
| "loss": 0.3782, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 0.0001376026272577997, | |
| "loss": 0.3411, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 0.00013431855500821018, | |
| "loss": 0.2138, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.9027777777777778, | |
| "eval_loss": 0.2958821952342987, | |
| "eval_runtime": 15.6222, | |
| "eval_samples_per_second": 46.088, | |
| "eval_steps_per_second": 5.761, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 0.00013103448275862068, | |
| "loss": 0.1832, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 0.00012775041050903122, | |
| "loss": 0.1517, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 0.00012446633825944172, | |
| "loss": 0.1099, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 0.00012118226600985223, | |
| "loss": 0.1762, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 0.00011789819376026272, | |
| "loss": 0.1496, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 0.00011461412151067324, | |
| "loss": 0.1287, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 0.00011133004926108374, | |
| "loss": 0.1389, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 0.00010804597701149426, | |
| "loss": 0.1583, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 0.00010476190476190477, | |
| "loss": 0.1214, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 0.00010147783251231529, | |
| "loss": 0.1566, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.819376026272578e-05, | |
| "loss": 0.1256, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.490968801313629e-05, | |
| "loss": 0.1123, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.16256157635468e-05, | |
| "loss": 0.0853, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.834154351395731e-05, | |
| "loss": 0.067, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.505747126436782e-05, | |
| "loss": 0.1035, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.177339901477834e-05, | |
| "loss": 0.0681, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.848932676518884e-05, | |
| "loss": 0.0382, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.520525451559935e-05, | |
| "loss": 0.0537, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 7.192118226600985e-05, | |
| "loss": 0.0283, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.863711001642037e-05, | |
| "loss": 0.0291, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.9833333333333333, | |
| "eval_loss": 0.06349290162324905, | |
| "eval_runtime": 15.4242, | |
| "eval_samples_per_second": 46.68, | |
| "eval_steps_per_second": 5.835, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.535303776683088e-05, | |
| "loss": 0.0275, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.206896551724138e-05, | |
| "loss": 0.0318, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.878489326765189e-05, | |
| "loss": 0.0312, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.55008210180624e-05, | |
| "loss": 0.0284, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.2216748768472914e-05, | |
| "loss": 0.0217, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.893267651888342e-05, | |
| "loss": 0.014, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.5648604269293925e-05, | |
| "loss": 0.02, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.236453201970443e-05, | |
| "loss": 0.0148, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 3.908045977011495e-05, | |
| "loss": 0.0136, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 3.5796387520525456e-05, | |
| "loss": 0.0129, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 3.251231527093596e-05, | |
| "loss": 0.0375, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 2.9228243021346467e-05, | |
| "loss": 0.0124, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 2.5944170771756983e-05, | |
| "loss": 0.0129, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 2.266009852216749e-05, | |
| "loss": 0.0189, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 1.9376026272577998e-05, | |
| "loss": 0.0142, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.6091954022988507e-05, | |
| "loss": 0.0332, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.2807881773399016e-05, | |
| "loss": 0.0116, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 9.523809523809523e-06, | |
| "loss": 0.0311, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 6.239737274220033e-06, | |
| "loss": 0.012, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 2.955665024630542e-06, | |
| "loss": 0.0128, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.9944444444444445, | |
| "eval_loss": 0.030118444934487343, | |
| "eval_runtime": 15.4638, | |
| "eval_samples_per_second": 46.56, | |
| "eval_steps_per_second": 5.82, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 609, | |
| "total_flos": 1.5064982852419584e+18, | |
| "train_loss": 0.23008910012362627, | |
| "train_runtime": 477.4194, | |
| "train_samples_per_second": 40.719, | |
| "train_steps_per_second": 1.276 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 609, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 1.5064982852419584e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |