| { |
| "best_metric": 0.894109396914446, |
| "best_model_checkpoint": "swin-tiny-patch4-window7-224-classification/checkpoint-1000", |
| "epoch": 9.975062344139651, |
| "eval_steps": 500, |
| "global_step": 1000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.1, |
| "grad_norm": 5.313328742980957, |
| "learning_rate": 1e-05, |
| "loss": 2.0631, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 5.775210857391357, |
| "learning_rate": 2e-05, |
| "loss": 1.7746, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 4.816771984100342, |
| "learning_rate": 3e-05, |
| "loss": 1.5563, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 6.007688999176025, |
| "learning_rate": 4e-05, |
| "loss": 1.2361, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 7.807107925415039, |
| "learning_rate": 5e-05, |
| "loss": 1.0685, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 6.920202732086182, |
| "learning_rate": 6e-05, |
| "loss": 0.9716, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 5.678104877471924, |
| "learning_rate": 7e-05, |
| "loss": 0.9274, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 7.1128010749816895, |
| "learning_rate": 8e-05, |
| "loss": 0.8962, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 8.700326919555664, |
| "learning_rate": 9e-05, |
| "loss": 0.8924, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 7.496997833251953, |
| "learning_rate": 0.0001, |
| "loss": 0.8514, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.705820476858345, |
| "eval_loss": 0.7862311005592346, |
| "eval_runtime": 38.0587, |
| "eval_samples_per_second": 74.937, |
| "eval_steps_per_second": 1.182, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 5.842683792114258, |
| "learning_rate": 9.888888888888889e-05, |
| "loss": 0.7961, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 7.082647323608398, |
| "learning_rate": 9.777777777777778e-05, |
| "loss": 0.8179, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 6.531412601470947, |
| "learning_rate": 9.666666666666667e-05, |
| "loss": 0.772, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 4.981847763061523, |
| "learning_rate": 9.555555555555557e-05, |
| "loss": 0.7298, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 6.9995198249816895, |
| "learning_rate": 9.444444444444444e-05, |
| "loss": 0.6882, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 4.9766693115234375, |
| "learning_rate": 9.333333333333334e-05, |
| "loss": 0.7048, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 5.899988174438477, |
| "learning_rate": 9.222222222222223e-05, |
| "loss": 0.7254, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 4.040374755859375, |
| "learning_rate": 9.111111111111112e-05, |
| "loss": 0.6543, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 6.156548023223877, |
| "learning_rate": 9e-05, |
| "loss": 0.6329, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 6.268504619598389, |
| "learning_rate": 8.888888888888889e-05, |
| "loss": 0.631, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.7885694249649369, |
| "eval_loss": 0.5491923093795776, |
| "eval_runtime": 38.032, |
| "eval_samples_per_second": 74.989, |
| "eval_steps_per_second": 1.183, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 6.425785541534424, |
| "learning_rate": 8.777777777777778e-05, |
| "loss": 0.612, |
| "step": 210 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 5.754735469818115, |
| "learning_rate": 8.666666666666667e-05, |
| "loss": 0.6117, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 5.51466703414917, |
| "learning_rate": 8.555555555555556e-05, |
| "loss": 0.6041, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 5.170792102813721, |
| "learning_rate": 8.444444444444444e-05, |
| "loss": 0.6057, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 5.319341659545898, |
| "learning_rate": 8.333333333333334e-05, |
| "loss": 0.556, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.59, |
| "grad_norm": 7.033453941345215, |
| "learning_rate": 8.222222222222222e-05, |
| "loss": 0.5918, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.69, |
| "grad_norm": 6.37441349029541, |
| "learning_rate": 8.111111111111112e-05, |
| "loss": 0.5603, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.79, |
| "grad_norm": 4.926444053649902, |
| "learning_rate": 8e-05, |
| "loss": 0.5827, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.89, |
| "grad_norm": 4.273934841156006, |
| "learning_rate": 7.88888888888889e-05, |
| "loss": 0.5406, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.99, |
| "grad_norm": 5.528937816619873, |
| "learning_rate": 7.777777777777778e-05, |
| "loss": 0.5516, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.99, |
| "eval_accuracy": 0.8078541374474053, |
| "eval_loss": 0.5194684863090515, |
| "eval_runtime": 38.7637, |
| "eval_samples_per_second": 73.574, |
| "eval_steps_per_second": 1.161, |
| "step": 300 |
| }, |
| { |
| "epoch": 3.09, |
| "grad_norm": 9.39556884765625, |
| "learning_rate": 7.666666666666667e-05, |
| "loss": 0.5039, |
| "step": 310 |
| }, |
| { |
| "epoch": 3.19, |
| "grad_norm": 5.658567428588867, |
| "learning_rate": 7.555555555555556e-05, |
| "loss": 0.497, |
| "step": 320 |
| }, |
| { |
| "epoch": 3.29, |
| "grad_norm": 5.974184513092041, |
| "learning_rate": 7.444444444444444e-05, |
| "loss": 0.5041, |
| "step": 330 |
| }, |
| { |
| "epoch": 3.39, |
| "grad_norm": 5.876557350158691, |
| "learning_rate": 7.333333333333333e-05, |
| "loss": 0.4978, |
| "step": 340 |
| }, |
| { |
| "epoch": 3.49, |
| "grad_norm": 7.005821704864502, |
| "learning_rate": 7.222222222222222e-05, |
| "loss": 0.5145, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.59, |
| "grad_norm": 4.967439651489258, |
| "learning_rate": 7.111111111111112e-05, |
| "loss": 0.4803, |
| "step": 360 |
| }, |
| { |
| "epoch": 3.69, |
| "grad_norm": 4.976831436157227, |
| "learning_rate": 7e-05, |
| "loss": 0.4816, |
| "step": 370 |
| }, |
| { |
| "epoch": 3.79, |
| "grad_norm": 6.872544288635254, |
| "learning_rate": 6.88888888888889e-05, |
| "loss": 0.4811, |
| "step": 380 |
| }, |
| { |
| "epoch": 3.89, |
| "grad_norm": 5.97269868850708, |
| "learning_rate": 6.777777777777778e-05, |
| "loss": 0.4978, |
| "step": 390 |
| }, |
| { |
| "epoch": 3.99, |
| "grad_norm": 5.982359886169434, |
| "learning_rate": 6.666666666666667e-05, |
| "loss": 0.4597, |
| "step": 400 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.8229312762973352, |
| "eval_loss": 0.48398053646087646, |
| "eval_runtime": 38.5586, |
| "eval_samples_per_second": 73.965, |
| "eval_steps_per_second": 1.167, |
| "step": 401 |
| }, |
| { |
| "epoch": 4.09, |
| "grad_norm": 4.751348972320557, |
| "learning_rate": 6.555555555555556e-05, |
| "loss": 0.4443, |
| "step": 410 |
| }, |
| { |
| "epoch": 4.19, |
| "grad_norm": 4.549768447875977, |
| "learning_rate": 6.444444444444446e-05, |
| "loss": 0.423, |
| "step": 420 |
| }, |
| { |
| "epoch": 4.29, |
| "grad_norm": 6.269441604614258, |
| "learning_rate": 6.333333333333333e-05, |
| "loss": 0.449, |
| "step": 430 |
| }, |
| { |
| "epoch": 4.39, |
| "grad_norm": 4.581325531005859, |
| "learning_rate": 6.222222222222222e-05, |
| "loss": 0.4099, |
| "step": 440 |
| }, |
| { |
| "epoch": 4.49, |
| "grad_norm": 5.31665563583374, |
| "learning_rate": 6.111111111111112e-05, |
| "loss": 0.4075, |
| "step": 450 |
| }, |
| { |
| "epoch": 4.59, |
| "grad_norm": 5.4510416984558105, |
| "learning_rate": 6e-05, |
| "loss": 0.4195, |
| "step": 460 |
| }, |
| { |
| "epoch": 4.69, |
| "grad_norm": 5.312283515930176, |
| "learning_rate": 5.8888888888888896e-05, |
| "loss": 0.3829, |
| "step": 470 |
| }, |
| { |
| "epoch": 4.79, |
| "grad_norm": 5.865201950073242, |
| "learning_rate": 5.7777777777777776e-05, |
| "loss": 0.4209, |
| "step": 480 |
| }, |
| { |
| "epoch": 4.89, |
| "grad_norm": 5.123416423797607, |
| "learning_rate": 5.666666666666667e-05, |
| "loss": 0.4094, |
| "step": 490 |
| }, |
| { |
| "epoch": 4.99, |
| "grad_norm": 4.805253982543945, |
| "learning_rate": 5.555555555555556e-05, |
| "loss": 0.4094, |
| "step": 500 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.8597475455820477, |
| "eval_loss": 0.3945947587490082, |
| "eval_runtime": 38.9464, |
| "eval_samples_per_second": 73.229, |
| "eval_steps_per_second": 1.155, |
| "step": 501 |
| }, |
| { |
| "epoch": 5.09, |
| "grad_norm": 4.796308517456055, |
| "learning_rate": 5.4444444444444446e-05, |
| "loss": 0.3555, |
| "step": 510 |
| }, |
| { |
| "epoch": 5.19, |
| "grad_norm": 4.460272789001465, |
| "learning_rate": 5.333333333333333e-05, |
| "loss": 0.3719, |
| "step": 520 |
| }, |
| { |
| "epoch": 5.29, |
| "grad_norm": 5.035310745239258, |
| "learning_rate": 5.222222222222223e-05, |
| "loss": 0.3785, |
| "step": 530 |
| }, |
| { |
| "epoch": 5.39, |
| "grad_norm": 5.145148754119873, |
| "learning_rate": 5.111111111111111e-05, |
| "loss": 0.3881, |
| "step": 540 |
| }, |
| { |
| "epoch": 5.49, |
| "grad_norm": 6.646022319793701, |
| "learning_rate": 5e-05, |
| "loss": 0.3776, |
| "step": 550 |
| }, |
| { |
| "epoch": 5.59, |
| "grad_norm": 6.039920330047607, |
| "learning_rate": 4.888888888888889e-05, |
| "loss": 0.3891, |
| "step": 560 |
| }, |
| { |
| "epoch": 5.69, |
| "grad_norm": 4.1056623458862305, |
| "learning_rate": 4.7777777777777784e-05, |
| "loss": 0.3369, |
| "step": 570 |
| }, |
| { |
| "epoch": 5.79, |
| "grad_norm": 5.633683204650879, |
| "learning_rate": 4.666666666666667e-05, |
| "loss": 0.3606, |
| "step": 580 |
| }, |
| { |
| "epoch": 5.89, |
| "grad_norm": 5.873307704925537, |
| "learning_rate": 4.555555555555556e-05, |
| "loss": 0.3386, |
| "step": 590 |
| }, |
| { |
| "epoch": 5.99, |
| "grad_norm": 5.3684983253479, |
| "learning_rate": 4.4444444444444447e-05, |
| "loss": 0.3289, |
| "step": 600 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_accuracy": 0.8572931276297335, |
| "eval_loss": 0.3786664605140686, |
| "eval_runtime": 40.3623, |
| "eval_samples_per_second": 70.66, |
| "eval_steps_per_second": 1.115, |
| "step": 601 |
| }, |
| { |
| "epoch": 6.08, |
| "grad_norm": 4.093196392059326, |
| "learning_rate": 4.3333333333333334e-05, |
| "loss": 0.3251, |
| "step": 610 |
| }, |
| { |
| "epoch": 6.18, |
| "grad_norm": 4.739712715148926, |
| "learning_rate": 4.222222222222222e-05, |
| "loss": 0.3029, |
| "step": 620 |
| }, |
| { |
| "epoch": 6.28, |
| "grad_norm": 3.9042539596557617, |
| "learning_rate": 4.111111111111111e-05, |
| "loss": 0.3269, |
| "step": 630 |
| }, |
| { |
| "epoch": 6.38, |
| "grad_norm": 5.695065498352051, |
| "learning_rate": 4e-05, |
| "loss": 0.3318, |
| "step": 640 |
| }, |
| { |
| "epoch": 6.48, |
| "grad_norm": 4.814879894256592, |
| "learning_rate": 3.888888888888889e-05, |
| "loss": 0.3087, |
| "step": 650 |
| }, |
| { |
| "epoch": 6.58, |
| "grad_norm": 4.633285999298096, |
| "learning_rate": 3.777777777777778e-05, |
| "loss": 0.3416, |
| "step": 660 |
| }, |
| { |
| "epoch": 6.68, |
| "grad_norm": 4.862430095672607, |
| "learning_rate": 3.6666666666666666e-05, |
| "loss": 0.3063, |
| "step": 670 |
| }, |
| { |
| "epoch": 6.78, |
| "grad_norm": 4.930345058441162, |
| "learning_rate": 3.555555555555556e-05, |
| "loss": 0.3225, |
| "step": 680 |
| }, |
| { |
| "epoch": 6.88, |
| "grad_norm": 4.587200164794922, |
| "learning_rate": 3.444444444444445e-05, |
| "loss": 0.3139, |
| "step": 690 |
| }, |
| { |
| "epoch": 6.98, |
| "grad_norm": 5.570033550262451, |
| "learning_rate": 3.3333333333333335e-05, |
| "loss": 0.3145, |
| "step": 700 |
| }, |
| { |
| "epoch": 6.99, |
| "eval_accuracy": 0.8772791023842917, |
| "eval_loss": 0.3536257743835449, |
| "eval_runtime": 42.7073, |
| "eval_samples_per_second": 66.78, |
| "eval_steps_per_second": 1.054, |
| "step": 701 |
| }, |
| { |
| "epoch": 7.08, |
| "grad_norm": 5.562831401824951, |
| "learning_rate": 3.222222222222223e-05, |
| "loss": 0.2998, |
| "step": 710 |
| }, |
| { |
| "epoch": 7.18, |
| "grad_norm": 5.306057453155518, |
| "learning_rate": 3.111111111111111e-05, |
| "loss": 0.2807, |
| "step": 720 |
| }, |
| { |
| "epoch": 7.28, |
| "grad_norm": 5.671096324920654, |
| "learning_rate": 3e-05, |
| "loss": 0.2794, |
| "step": 730 |
| }, |
| { |
| "epoch": 7.38, |
| "grad_norm": 5.412108421325684, |
| "learning_rate": 2.8888888888888888e-05, |
| "loss": 0.2814, |
| "step": 740 |
| }, |
| { |
| "epoch": 7.48, |
| "grad_norm": 4.249139308929443, |
| "learning_rate": 2.777777777777778e-05, |
| "loss": 0.2531, |
| "step": 750 |
| }, |
| { |
| "epoch": 7.58, |
| "grad_norm": 5.714820861816406, |
| "learning_rate": 2.6666666666666667e-05, |
| "loss": 0.2712, |
| "step": 760 |
| }, |
| { |
| "epoch": 7.68, |
| "grad_norm": 5.9426469802856445, |
| "learning_rate": 2.5555555555555554e-05, |
| "loss": 0.2589, |
| "step": 770 |
| }, |
| { |
| "epoch": 7.78, |
| "grad_norm": 4.787610054016113, |
| "learning_rate": 2.4444444444444445e-05, |
| "loss": 0.2745, |
| "step": 780 |
| }, |
| { |
| "epoch": 7.88, |
| "grad_norm": 3.9021761417388916, |
| "learning_rate": 2.3333333333333336e-05, |
| "loss": 0.2728, |
| "step": 790 |
| }, |
| { |
| "epoch": 7.98, |
| "grad_norm": 5.153603553771973, |
| "learning_rate": 2.2222222222222223e-05, |
| "loss": 0.2911, |
| "step": 800 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_accuracy": 0.885343618513324, |
| "eval_loss": 0.32486388087272644, |
| "eval_runtime": 42.9219, |
| "eval_samples_per_second": 66.446, |
| "eval_steps_per_second": 1.048, |
| "step": 802 |
| }, |
| { |
| "epoch": 8.08, |
| "grad_norm": 5.644151210784912, |
| "learning_rate": 2.111111111111111e-05, |
| "loss": 0.2559, |
| "step": 810 |
| }, |
| { |
| "epoch": 8.18, |
| "grad_norm": 4.023806095123291, |
| "learning_rate": 2e-05, |
| "loss": 0.2415, |
| "step": 820 |
| }, |
| { |
| "epoch": 8.28, |
| "grad_norm": 6.214153289794922, |
| "learning_rate": 1.888888888888889e-05, |
| "loss": 0.2447, |
| "step": 830 |
| }, |
| { |
| "epoch": 8.38, |
| "grad_norm": 3.903085231781006, |
| "learning_rate": 1.777777777777778e-05, |
| "loss": 0.2602, |
| "step": 840 |
| }, |
| { |
| "epoch": 8.48, |
| "grad_norm": 5.1828131675720215, |
| "learning_rate": 1.6666666666666667e-05, |
| "loss": 0.2529, |
| "step": 850 |
| }, |
| { |
| "epoch": 8.58, |
| "grad_norm": 4.840915203094482, |
| "learning_rate": 1.5555555555555555e-05, |
| "loss": 0.2302, |
| "step": 860 |
| }, |
| { |
| "epoch": 8.68, |
| "grad_norm": 5.424919128417969, |
| "learning_rate": 1.4444444444444444e-05, |
| "loss": 0.246, |
| "step": 870 |
| }, |
| { |
| "epoch": 8.78, |
| "grad_norm": 6.8232316970825195, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 0.2435, |
| "step": 880 |
| }, |
| { |
| "epoch": 8.88, |
| "grad_norm": 4.856207847595215, |
| "learning_rate": 1.2222222222222222e-05, |
| "loss": 0.2376, |
| "step": 890 |
| }, |
| { |
| "epoch": 8.98, |
| "grad_norm": 5.239475250244141, |
| "learning_rate": 1.1111111111111112e-05, |
| "loss": 0.2487, |
| "step": 900 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_accuracy": 0.8937587657784011, |
| "eval_loss": 0.30939263105392456, |
| "eval_runtime": 39.5102, |
| "eval_samples_per_second": 72.184, |
| "eval_steps_per_second": 1.139, |
| "step": 902 |
| }, |
| { |
| "epoch": 9.08, |
| "grad_norm": 5.607595443725586, |
| "learning_rate": 1e-05, |
| "loss": 0.2253, |
| "step": 910 |
| }, |
| { |
| "epoch": 9.18, |
| "grad_norm": 3.9586544036865234, |
| "learning_rate": 8.88888888888889e-06, |
| "loss": 0.2478, |
| "step": 920 |
| }, |
| { |
| "epoch": 9.28, |
| "grad_norm": 3.8814537525177, |
| "learning_rate": 7.777777777777777e-06, |
| "loss": 0.2173, |
| "step": 930 |
| }, |
| { |
| "epoch": 9.38, |
| "grad_norm": 7.456895351409912, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 0.2315, |
| "step": 940 |
| }, |
| { |
| "epoch": 9.48, |
| "grad_norm": 4.543576717376709, |
| "learning_rate": 5.555555555555556e-06, |
| "loss": 0.2281, |
| "step": 950 |
| }, |
| { |
| "epoch": 9.58, |
| "grad_norm": 4.659777641296387, |
| "learning_rate": 4.444444444444445e-06, |
| "loss": 0.2324, |
| "step": 960 |
| }, |
| { |
| "epoch": 9.68, |
| "grad_norm": 4.048426151275635, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 0.2132, |
| "step": 970 |
| }, |
| { |
| "epoch": 9.78, |
| "grad_norm": 3.969958543777466, |
| "learning_rate": 2.2222222222222225e-06, |
| "loss": 0.2076, |
| "step": 980 |
| }, |
| { |
| "epoch": 9.88, |
| "grad_norm": 5.621995449066162, |
| "learning_rate": 1.1111111111111112e-06, |
| "loss": 0.2262, |
| "step": 990 |
| }, |
| { |
| "epoch": 9.98, |
| "grad_norm": 4.668450355529785, |
| "learning_rate": 0.0, |
| "loss": 0.2291, |
| "step": 1000 |
| }, |
| { |
| "epoch": 9.98, |
| "eval_accuracy": 0.894109396914446, |
| "eval_loss": 0.303189218044281, |
| "eval_runtime": 39.0907, |
| "eval_samples_per_second": 72.958, |
| "eval_steps_per_second": 1.151, |
| "step": 1000 |
| }, |
| { |
| "epoch": 9.98, |
| "step": 1000, |
| "total_flos": 6.364199987970048e+18, |
| "train_loss": 0.48602974390983583, |
| "train_runtime": 5351.8475, |
| "train_samples_per_second": 47.954, |
| "train_steps_per_second": 0.187 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 1000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 500, |
| "total_flos": 6.364199987970048e+18, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|