| { | |
| "best_metric": 0.18068847060203552, | |
| "best_model_checkpoint": "swin-tiny-patch4-window7-224-crack-detector\\checkpoint-585", | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 585, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 8.47457627118644e-06, | |
| "loss": 1.3866, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.694915254237288e-05, | |
| "loss": 1.3138, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 2.5423728813559322e-05, | |
| "loss": 1.2036, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.389830508474576e-05, | |
| "loss": 1.012, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.2372881355932206e-05, | |
| "loss": 0.8611, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.990494296577947e-05, | |
| "loss": 0.7258, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.8954372623574146e-05, | |
| "loss": 0.6027, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.800380228136883e-05, | |
| "loss": 0.5971, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.70532319391635e-05, | |
| "loss": 0.5713, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.610266159695818e-05, | |
| "loss": 0.481, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.5152091254752856e-05, | |
| "loss": 0.5275, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.4201520912547525e-05, | |
| "loss": 0.4933, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.325095057034221e-05, | |
| "loss": 0.4915, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.2300380228136884e-05, | |
| "loss": 0.5065, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.134980988593156e-05, | |
| "loss": 0.4511, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.0399239543726235e-05, | |
| "loss": 0.366, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.944866920152092e-05, | |
| "loss": 0.4178, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.849809885931559e-05, | |
| "loss": 0.4868, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.754752851711027e-05, | |
| "loss": 0.3863, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": { | |
| "accuracy": 0.885576923076923 | |
| }, | |
| "eval_f1": { | |
| "f1": 0.8829318618369404 | |
| }, | |
| "eval_loss": 0.33488142490386963, | |
| "eval_precision": { | |
| "precision": 0.8830357915066687 | |
| }, | |
| "eval_recall": { | |
| "recall": 0.8864842943431257 | |
| }, | |
| "eval_runtime": 19.1649, | |
| "eval_samples_per_second": 162.798, | |
| "eval_steps_per_second": 10.175, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.6596958174904945e-05, | |
| "loss": 0.4016, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.564638783269962e-05, | |
| "loss": 0.4089, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.46958174904943e-05, | |
| "loss": 0.3912, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.374524714828898e-05, | |
| "loss": 0.352, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.2794676806083655e-05, | |
| "loss": 0.3791, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.1844106463878324e-05, | |
| "loss": 0.3799, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.0893536121673007e-05, | |
| "loss": 0.3602, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.994296577946768e-05, | |
| "loss": 0.3451, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.8992395437262358e-05, | |
| "loss": 0.4013, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.8041825095057034e-05, | |
| "loss": 0.3138, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.7091254752851713e-05, | |
| "loss": 0.3488, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.614068441064639e-05, | |
| "loss": 0.3306, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.5190114068441068e-05, | |
| "loss": 0.3477, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.423954372623574e-05, | |
| "loss": 0.3075, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.328897338403042e-05, | |
| "loss": 0.2544, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.2338403041825095e-05, | |
| "loss": 0.2909, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.138783269961977e-05, | |
| "loss": 0.2901, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.043726235741445e-05, | |
| "loss": 0.3154, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.9486692015209126e-05, | |
| "loss": 0.2921, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.8536121673003802e-05, | |
| "loss": 0.2685, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": { | |
| "accuracy": 0.9080128205128205 | |
| }, | |
| "eval_f1": { | |
| "f1": 0.9106277459775055 | |
| }, | |
| "eval_loss": 0.2715485692024231, | |
| "eval_precision": { | |
| "precision": 0.9130231253775549 | |
| }, | |
| "eval_recall": { | |
| "recall": 0.9148104520472664 | |
| }, | |
| "eval_runtime": 19.694, | |
| "eval_samples_per_second": 158.424, | |
| "eval_steps_per_second": 9.901, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.758555133079848e-05, | |
| "loss": 0.3095, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.6634980988593157e-05, | |
| "loss": 0.275, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.5684410646387833e-05, | |
| "loss": 0.2995, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.4733840304182512e-05, | |
| "loss": 0.2558, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.3783269961977188e-05, | |
| "loss": 0.3233, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.2832699619771862e-05, | |
| "loss": 0.2716, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.1882129277566541e-05, | |
| "loss": 0.2387, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.0931558935361218e-05, | |
| "loss": 0.275, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 9.980988593155894e-06, | |
| "loss": 0.2475, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.03041825095057e-06, | |
| "loss": 0.2272, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.079847908745247e-06, | |
| "loss": 0.2475, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.129277566539924e-06, | |
| "loss": 0.2681, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.1787072243346015e-06, | |
| "loss": 0.2519, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.228136882129278e-06, | |
| "loss": 0.244, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.277566539923955e-06, | |
| "loss": 0.243, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.3269961977186314e-06, | |
| "loss": 0.2361, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.376425855513308e-06, | |
| "loss": 0.2095, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.4258555133079848e-06, | |
| "loss": 0.2171, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.752851711026616e-07, | |
| "loss": 0.2235, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": { | |
| "accuracy": 0.9384615384615385 | |
| }, | |
| "eval_f1": { | |
| "f1": 0.9382975252490704 | |
| }, | |
| "eval_loss": 0.18068847060203552, | |
| "eval_precision": { | |
| "precision": 0.9382005688460371 | |
| }, | |
| "eval_recall": { | |
| "recall": 0.9395073274524703 | |
| }, | |
| "eval_runtime": 19.9688, | |
| "eval_samples_per_second": 156.243, | |
| "eval_steps_per_second": 9.765, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 585, | |
| "total_flos": 9.306602345005056e+17, | |
| "train_loss": 0.4218194468408568, | |
| "train_runtime": 526.8698, | |
| "train_samples_per_second": 71.061, | |
| "train_steps_per_second": 1.11 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 585, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 9.306602345005056e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |