| { | |
| "best_metric": 0.7897123694419861, | |
| "best_model_checkpoint": "./vit-base-crack-classification-2/checkpoint-609", | |
| "epoch": 3.0, | |
| "eval_steps": 100, | |
| "global_step": 609, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.835796387520527e-06, | |
| "loss": 1.7722, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.671592775041053e-06, | |
| "loss": 1.7365, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.507389162561577e-06, | |
| "loss": 1.6926, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.343185550082103e-06, | |
| "loss": 1.6732, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.178981937602627e-06, | |
| "loss": 1.6336, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.014778325123153e-06, | |
| "loss": 1.5896, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 8.85057471264368e-06, | |
| "loss": 1.5829, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 8.686371100164205e-06, | |
| "loss": 1.5218, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 8.52216748768473e-06, | |
| "loss": 1.5235, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 8.357963875205255e-06, | |
| "loss": 1.4616, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 8.193760262725781e-06, | |
| "loss": 1.3986, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.029556650246306e-06, | |
| "loss": 1.3888, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 7.865353037766832e-06, | |
| "loss": 1.3618, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 7.701149425287356e-06, | |
| "loss": 1.3174, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 7.536945812807882e-06, | |
| "loss": 1.2934, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 7.372742200328408e-06, | |
| "loss": 1.2575, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.208538587848934e-06, | |
| "loss": 1.158, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 7.044334975369459e-06, | |
| "loss": 1.1727, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 6.880131362889985e-06, | |
| "loss": 1.1927, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 6.715927750410509e-06, | |
| "loss": 1.1735, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.7361111111111112, | |
| "eval_loss": 1.1257396936416626, | |
| "eval_runtime": 13.5112, | |
| "eval_samples_per_second": 53.289, | |
| "eval_steps_per_second": 6.661, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 6.568144499178982e-06, | |
| "loss": 1.1626, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 6.403940886699508e-06, | |
| "loss": 1.0837, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 6.239737274220034e-06, | |
| "loss": 1.0712, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 6.075533661740559e-06, | |
| "loss": 0.9507, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 5.911330049261085e-06, | |
| "loss": 0.9846, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 5.747126436781609e-06, | |
| "loss": 0.9912, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 5.582922824302135e-06, | |
| "loss": 1.0554, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 5.41871921182266e-06, | |
| "loss": 1.0018, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 5.254515599343186e-06, | |
| "loss": 0.9575, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 5.090311986863712e-06, | |
| "loss": 0.9674, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.926108374384237e-06, | |
| "loss": 0.9157, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.761904761904762e-06, | |
| "loss": 0.9381, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.5977011494252875e-06, | |
| "loss": 0.9266, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.4334975369458135e-06, | |
| "loss": 0.9534, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.2692939244663386e-06, | |
| "loss": 0.884, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.105090311986864e-06, | |
| "loss": 0.8147, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 3.94088669950739e-06, | |
| "loss": 0.8547, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 3.7766830870279148e-06, | |
| "loss": 0.8387, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 3.6124794745484403e-06, | |
| "loss": 0.8161, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 3.448275862068966e-06, | |
| "loss": 0.8295, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.8041666666666667, | |
| "eval_loss": 0.8613185286521912, | |
| "eval_runtime": 13.5409, | |
| "eval_samples_per_second": 53.172, | |
| "eval_steps_per_second": 6.647, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 3.284072249589491e-06, | |
| "loss": 0.8765, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 3.119868637110017e-06, | |
| "loss": 0.8456, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 2.9556650246305424e-06, | |
| "loss": 0.8405, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 2.7914614121510676e-06, | |
| "loss": 0.7693, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 2.627257799671593e-06, | |
| "loss": 0.8084, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.4630541871921186e-06, | |
| "loss": 0.782, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.2988505747126437e-06, | |
| "loss": 0.7105, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.1346469622331693e-06, | |
| "loss": 0.8208, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.970443349753695e-06, | |
| "loss": 0.7228, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.8062397372742201e-06, | |
| "loss": 0.7715, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.6420361247947455e-06, | |
| "loss": 0.7872, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.4778325123152712e-06, | |
| "loss": 0.7411, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 1.3136288998357965e-06, | |
| "loss": 0.7804, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 1.1494252873563219e-06, | |
| "loss": 0.7374, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 9.852216748768474e-07, | |
| "loss": 0.7187, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 8.210180623973727e-07, | |
| "loss": 0.6989, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 6.568144499178983e-07, | |
| "loss": 0.7046, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.926108374384237e-07, | |
| "loss": 0.7508, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 3.2840722495894914e-07, | |
| "loss": 0.7759, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.6420361247947457e-07, | |
| "loss": 0.7085, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.8236111111111111, | |
| "eval_loss": 0.7897123694419861, | |
| "eval_runtime": 13.4904, | |
| "eval_samples_per_second": 53.371, | |
| "eval_steps_per_second": 6.671, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 609, | |
| "total_flos": 1.5064982852419584e+18, | |
| "train_loss": 1.0488137838679974, | |
| "train_runtime": 435.1954, | |
| "train_samples_per_second": 44.67, | |
| "train_steps_per_second": 1.399 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 609, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 1.5064982852419584e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |