| { |
| "best_metric": 0.7292817679558011, |
| "best_model_checkpoint": "musical_instrument_detection/checkpoint-680", |
| "epoch": 6.0, |
| "global_step": 816, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.07, |
| "learning_rate": 2.2058823529411767e-06, |
| "loss": 1.3866, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 4.65686274509804e-06, |
| "loss": 1.383, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 7.107843137254902e-06, |
| "loss": 1.3851, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 9.31372549019608e-06, |
| "loss": 1.3729, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.1764705882352942e-05, |
| "loss": 1.3689, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.4215686274509804e-05, |
| "loss": 1.3561, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.6666666666666667e-05, |
| "loss": 1.3463, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.9117647058823528e-05, |
| "loss": 1.3176, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 2.1568627450980395e-05, |
| "loss": 1.3063, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 2.3774509803921568e-05, |
| "loss": 1.2662, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 2.6225490196078435e-05, |
| "loss": 1.2082, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 2.8676470588235295e-05, |
| "loss": 1.205, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 3.112745098039216e-05, |
| "loss": 1.1802, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.5322283609576427, |
| "eval_loss": 1.038236141204834, |
| "eval_runtime": 5.4332, |
| "eval_samples_per_second": 99.942, |
| "eval_steps_per_second": 6.258, |
| "step": 136 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 3.357843137254902e-05, |
| "loss": 1.1219, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 3.5784313725490195e-05, |
| "loss": 1.0772, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 3.8235294117647055e-05, |
| "loss": 1.016, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 4.068627450980392e-05, |
| "loss": 0.9292, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 4.313725490196079e-05, |
| "loss": 0.8552, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 4.558823529411765e-05, |
| "loss": 0.7906, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 4.803921568627452e-05, |
| "loss": 0.7561, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 4.994553376906319e-05, |
| "loss": 0.7469, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 4.967320261437909e-05, |
| "loss": 0.7354, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 4.940087145969499e-05, |
| "loss": 0.7026, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 4.91285403050109e-05, |
| "loss": 0.6496, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 4.88562091503268e-05, |
| "loss": 0.5919, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 4.85838779956427e-05, |
| "loss": 0.6196, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 4.8338779956427017e-05, |
| "loss": 0.5676, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.7090239410681399, |
| "eval_loss": 0.5477131009101868, |
| "eval_runtime": 5.4026, |
| "eval_samples_per_second": 100.508, |
| "eval_steps_per_second": 6.293, |
| "step": 272 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.809368191721133e-05, |
| "loss": 0.5971, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.7821350762527234e-05, |
| "loss": 0.6122, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 4.7549019607843135e-05, |
| "loss": 0.4923, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 4.7276688453159044e-05, |
| "loss": 0.528, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 4.700435729847495e-05, |
| "loss": 0.5466, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 4.673202614379085e-05, |
| "loss": 0.5894, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 4.6459694989106755e-05, |
| "loss": 0.5512, |
| "step": 340 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 4.6187363834422656e-05, |
| "loss": 0.4994, |
| "step": 350 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 4.5915032679738564e-05, |
| "loss": 0.5158, |
| "step": 360 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.564270152505447e-05, |
| "loss": 0.5905, |
| "step": 370 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 4.5370370370370374e-05, |
| "loss": 0.5186, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 4.5098039215686275e-05, |
| "loss": 0.729, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 4.482570806100218e-05, |
| "loss": 0.5502, |
| "step": 400 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.7255985267034991, |
| "eval_loss": 0.5312222242355347, |
| "eval_runtime": 5.3808, |
| "eval_samples_per_second": 100.915, |
| "eval_steps_per_second": 6.319, |
| "step": 408 |
| }, |
| { |
| "epoch": 3.01, |
| "learning_rate": 4.4553376906318085e-05, |
| "loss": 0.5265, |
| "step": 410 |
| }, |
| { |
| "epoch": 3.09, |
| "learning_rate": 4.4281045751633986e-05, |
| "loss": 0.5098, |
| "step": 420 |
| }, |
| { |
| "epoch": 3.16, |
| "learning_rate": 4.400871459694989e-05, |
| "loss": 0.5669, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.24, |
| "learning_rate": 4.37363834422658e-05, |
| "loss": 0.474, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.31, |
| "learning_rate": 4.3464052287581704e-05, |
| "loss": 0.4687, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.38, |
| "learning_rate": 4.3191721132897605e-05, |
| "loss": 0.5691, |
| "step": 460 |
| }, |
| { |
| "epoch": 3.46, |
| "learning_rate": 4.291938997821351e-05, |
| "loss": 0.7491, |
| "step": 470 |
| }, |
| { |
| "epoch": 3.53, |
| "learning_rate": 4.2647058823529415e-05, |
| "loss": 0.4712, |
| "step": 480 |
| }, |
| { |
| "epoch": 3.6, |
| "learning_rate": 4.2374727668845316e-05, |
| "loss": 0.489, |
| "step": 490 |
| }, |
| { |
| "epoch": 3.68, |
| "learning_rate": 4.2102396514161224e-05, |
| "loss": 0.5199, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.75, |
| "learning_rate": 4.1830065359477126e-05, |
| "loss": 0.6991, |
| "step": 510 |
| }, |
| { |
| "epoch": 3.82, |
| "learning_rate": 4.1557734204793034e-05, |
| "loss": 0.5455, |
| "step": 520 |
| }, |
| { |
| "epoch": 3.9, |
| "learning_rate": 4.1285403050108935e-05, |
| "loss": 0.4854, |
| "step": 530 |
| }, |
| { |
| "epoch": 3.97, |
| "learning_rate": 4.101307189542484e-05, |
| "loss": 0.5191, |
| "step": 540 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.7182320441988951, |
| "eval_loss": 0.49646708369255066, |
| "eval_runtime": 5.3289, |
| "eval_samples_per_second": 101.896, |
| "eval_steps_per_second": 6.38, |
| "step": 544 |
| }, |
| { |
| "epoch": 4.04, |
| "learning_rate": 4.074074074074074e-05, |
| "loss": 0.4407, |
| "step": 550 |
| }, |
| { |
| "epoch": 4.12, |
| "learning_rate": 4.0468409586056646e-05, |
| "loss": 0.6965, |
| "step": 560 |
| }, |
| { |
| "epoch": 4.19, |
| "learning_rate": 4.0196078431372555e-05, |
| "loss": 0.6421, |
| "step": 570 |
| }, |
| { |
| "epoch": 4.26, |
| "learning_rate": 3.9923747276688456e-05, |
| "loss": 0.4867, |
| "step": 580 |
| }, |
| { |
| "epoch": 4.34, |
| "learning_rate": 3.965141612200436e-05, |
| "loss": 0.4978, |
| "step": 590 |
| }, |
| { |
| "epoch": 4.41, |
| "learning_rate": 3.9379084967320266e-05, |
| "loss": 0.467, |
| "step": 600 |
| }, |
| { |
| "epoch": 4.49, |
| "learning_rate": 3.910675381263617e-05, |
| "loss": 0.4662, |
| "step": 610 |
| }, |
| { |
| "epoch": 4.56, |
| "learning_rate": 3.8834422657952075e-05, |
| "loss": 0.4233, |
| "step": 620 |
| }, |
| { |
| "epoch": 4.63, |
| "learning_rate": 3.8562091503267977e-05, |
| "loss": 0.4864, |
| "step": 630 |
| }, |
| { |
| "epoch": 4.71, |
| "learning_rate": 3.828976034858388e-05, |
| "loss": 0.5433, |
| "step": 640 |
| }, |
| { |
| "epoch": 4.78, |
| "learning_rate": 3.8017429193899786e-05, |
| "loss": 0.4851, |
| "step": 650 |
| }, |
| { |
| "epoch": 4.85, |
| "learning_rate": 3.774509803921569e-05, |
| "loss": 0.4162, |
| "step": 660 |
| }, |
| { |
| "epoch": 4.93, |
| "learning_rate": 3.747276688453159e-05, |
| "loss": 0.4304, |
| "step": 670 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 3.72004357298475e-05, |
| "loss": 0.4473, |
| "step": 680 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.7292817679558011, |
| "eval_loss": 0.437788724899292, |
| "eval_runtime": 5.426, |
| "eval_samples_per_second": 100.073, |
| "eval_steps_per_second": 6.266, |
| "step": 680 |
| }, |
| { |
| "epoch": 5.07, |
| "learning_rate": 3.6928104575163405e-05, |
| "loss": 0.4891, |
| "step": 690 |
| }, |
| { |
| "epoch": 5.15, |
| "learning_rate": 3.665577342047931e-05, |
| "loss": 0.5293, |
| "step": 700 |
| }, |
| { |
| "epoch": 5.22, |
| "learning_rate": 3.638344226579521e-05, |
| "loss": 0.4155, |
| "step": 710 |
| }, |
| { |
| "epoch": 5.29, |
| "learning_rate": 3.611111111111111e-05, |
| "loss": 0.4949, |
| "step": 720 |
| }, |
| { |
| "epoch": 5.37, |
| "learning_rate": 3.583877995642702e-05, |
| "loss": 0.4764, |
| "step": 730 |
| }, |
| { |
| "epoch": 5.44, |
| "learning_rate": 3.556644880174292e-05, |
| "loss": 0.4032, |
| "step": 740 |
| }, |
| { |
| "epoch": 5.51, |
| "learning_rate": 3.529411764705883e-05, |
| "loss": 0.4569, |
| "step": 750 |
| }, |
| { |
| "epoch": 5.59, |
| "learning_rate": 3.502178649237473e-05, |
| "loss": 0.4646, |
| "step": 760 |
| }, |
| { |
| "epoch": 5.66, |
| "learning_rate": 3.474945533769064e-05, |
| "loss": 0.4825, |
| "step": 770 |
| }, |
| { |
| "epoch": 5.74, |
| "learning_rate": 3.447712418300654e-05, |
| "loss": 0.3938, |
| "step": 780 |
| }, |
| { |
| "epoch": 5.81, |
| "learning_rate": 3.420479302832244e-05, |
| "loss": 0.4213, |
| "step": 790 |
| }, |
| { |
| "epoch": 5.88, |
| "learning_rate": 3.393246187363834e-05, |
| "loss": 0.4944, |
| "step": 800 |
| }, |
| { |
| "epoch": 5.96, |
| "learning_rate": 3.366013071895425e-05, |
| "loss": 0.466, |
| "step": 810 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_accuracy": 0.714548802946593, |
| "eval_loss": 0.4857429563999176, |
| "eval_runtime": 5.4083, |
| "eval_samples_per_second": 100.401, |
| "eval_steps_per_second": 6.287, |
| "step": 816 |
| } |
| ], |
| "max_steps": 2040, |
| "num_train_epochs": 15, |
| "total_flos": 3.94420584390912e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|