| { |
| "best_metric": 108.29063415527344, |
| "best_model_checkpoint": "./ViT_ordinary/checkpoint-2500", |
| "epoch": 2.2893772893772892, |
| "global_step": 2500, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.1452785890167783e-06, |
| "loss": 1143.4345, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 2.2905571780335565e-06, |
| "loss": 1227.331, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 3.435835767050335e-06, |
| "loss": 1257.7006, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 4.581114356067113e-06, |
| "loss": 1220.7573, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 5.726392945083892e-06, |
| "loss": 1314.9278, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 6.87167153410067e-06, |
| "loss": 1161.0052, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 8.016950123117448e-06, |
| "loss": 1195.8272, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 9.162228712134226e-06, |
| "loss": 1135.2035, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.0307507301151004e-05, |
| "loss": 1105.4266, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.1452785890167784e-05, |
| "loss": 1138.5144, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.09, |
| "eval_RMSE": 45.47388458251953, |
| "eval_loss": 2676.440673828125, |
| "eval_runtime": 88.8721, |
| "eval_samples_per_second": 91.941, |
| "eval_steps_per_second": 2.881, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.259806447918456e-05, |
| "loss": 936.6518, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.374334306820134e-05, |
| "loss": 1003.3095, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.4888621657218118e-05, |
| "loss": 994.7256, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.6033900246234896e-05, |
| "loss": 962.0123, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.7179178835251674e-05, |
| "loss": 1012.3558, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.8324457424268452e-05, |
| "loss": 923.0986, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.946973601328523e-05, |
| "loss": 1055.8317, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 2.0615014602302008e-05, |
| "loss": 1189.4732, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 2.1760293191318786e-05, |
| "loss": 998.1689, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 2.2905571780335568e-05, |
| "loss": 940.4191, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.18, |
| "eval_RMSE": 44.35804748535156, |
| "eval_loss": 2574.700927734375, |
| "eval_runtime": 88.6183, |
| "eval_samples_per_second": 92.204, |
| "eval_steps_per_second": 2.889, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 2.4050850369352342e-05, |
| "loss": 1081.1541, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 2.519612895836912e-05, |
| "loss": 953.7207, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 2.63414075473859e-05, |
| "loss": 920.4622, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 2.748668613640268e-05, |
| "loss": 1037.0487, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 2.8631964725419458e-05, |
| "loss": 991.3824, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 2.9777243314436236e-05, |
| "loss": 1079.5839, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 3.092252190345301e-05, |
| "loss": 989.4474, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 3.206780049246979e-05, |
| "loss": 987.4904, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 3.321307908148657e-05, |
| "loss": 1038.7738, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 3.435835767050335e-05, |
| "loss": 921.6753, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.27, |
| "eval_RMSE": 43.82501220703125, |
| "eval_loss": 2526.614501953125, |
| "eval_runtime": 89.1648, |
| "eval_samples_per_second": 91.639, |
| "eval_steps_per_second": 2.871, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 3.550363625952013e-05, |
| "loss": 909.3841, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 3.6648914848536904e-05, |
| "loss": 931.0681, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 3.7794193437553686e-05, |
| "loss": 969.0332, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 3.893947202657046e-05, |
| "loss": 1054.9324, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 4.008475061558724e-05, |
| "loss": 989.6485, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 4.1230029204604016e-05, |
| "loss": 940.8793, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 4.237530779362079e-05, |
| "loss": 935.6209, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 4.352058638263757e-05, |
| "loss": 1083.344, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 4.4665864971654354e-05, |
| "loss": 918.0279, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 4.5811143560671135e-05, |
| "loss": 938.6592, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.37, |
| "eval_RMSE": 43.11045837402344, |
| "eval_loss": 2458.26171875, |
| "eval_runtime": 89.3306, |
| "eval_samples_per_second": 91.469, |
| "eval_steps_per_second": 2.866, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 4.69564221496879e-05, |
| "loss": 980.8854, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 4.8101700738704685e-05, |
| "loss": 1039.0189, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 4.9246979327721466e-05, |
| "loss": 876.2118, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 5.039225791673824e-05, |
| "loss": 922.5582, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 5.153753650575502e-05, |
| "loss": 918.9763, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 5.26828150947718e-05, |
| "loss": 967.343, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 5.3828093683788585e-05, |
| "loss": 866.294, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 5.497337227280536e-05, |
| "loss": 866.6839, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 5.611865086182213e-05, |
| "loss": 839.3744, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 5.7263929450838916e-05, |
| "loss": 904.052, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.46, |
| "eval_RMSE": 42.12101364135742, |
| "eval_loss": 2360.75439453125, |
| "eval_runtime": 82.5537, |
| "eval_samples_per_second": 98.978, |
| "eval_steps_per_second": 3.101, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 5.840920803985569e-05, |
| "loss": 770.1921, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 5.955448662887247e-05, |
| "loss": 867.2703, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 6.0699765217889246e-05, |
| "loss": 803.1129, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 6.184504380690602e-05, |
| "loss": 838.8234, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 6.29903223959228e-05, |
| "loss": 854.0321, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 6.413560098493958e-05, |
| "loss": 843.9535, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 6.528087957395637e-05, |
| "loss": 819.4999, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 6.642615816297315e-05, |
| "loss": 833.5234, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 6.757143675198991e-05, |
| "loss": 919.277, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 6.87167153410067e-05, |
| "loss": 852.848, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.55, |
| "eval_RMSE": 40.79917526245117, |
| "eval_loss": 2225.54150390625, |
| "eval_runtime": 90.1848, |
| "eval_samples_per_second": 90.603, |
| "eval_steps_per_second": 2.839, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 6.986199393002348e-05, |
| "loss": 750.3481, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 7.100727251904026e-05, |
| "loss": 879.5618, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.215255110805703e-05, |
| "loss": 797.6468, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.329782969707381e-05, |
| "loss": 835.0447, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.444310828609059e-05, |
| "loss": 755.9302, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.558838687510737e-05, |
| "loss": 749.4049, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.673366546412415e-05, |
| "loss": 727.4822, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 7.787894405314092e-05, |
| "loss": 664.3502, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 7.90242226421577e-05, |
| "loss": 737.5342, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 8.016950123117448e-05, |
| "loss": 685.9979, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.64, |
| "eval_RMSE": 39.182891845703125, |
| "eval_loss": 2045.955322265625, |
| "eval_runtime": 86.8947, |
| "eval_samples_per_second": 94.033, |
| "eval_steps_per_second": 2.946, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 8.131477982019125e-05, |
| "loss": 818.1389, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 8.246005840920803e-05, |
| "loss": 679.9854, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 8.360533699822481e-05, |
| "loss": 696.1048, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 8.475061558724158e-05, |
| "loss": 694.8774, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 8.589589417625838e-05, |
| "loss": 673.0095, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 8.704117276527514e-05, |
| "loss": 638.9566, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 8.818645135429193e-05, |
| "loss": 660.6367, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 8.933172994330871e-05, |
| "loss": 622.8923, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 9.047700853232548e-05, |
| "loss": 570.6501, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 9.162228712134227e-05, |
| "loss": 612.6461, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.73, |
| "eval_RMSE": 37.14878845214844, |
| "eval_loss": 1821.157958984375, |
| "eval_runtime": 90.6758, |
| "eval_samples_per_second": 90.112, |
| "eval_steps_per_second": 2.823, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 9.276756571035904e-05, |
| "loss": 632.5693, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 9.39128442993758e-05, |
| "loss": 633.8072, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 9.50581228883926e-05, |
| "loss": 625.9993, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 9.620340147740937e-05, |
| "loss": 616.1023, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 9.734868006642615e-05, |
| "loss": 542.3134, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 9.849395865544293e-05, |
| "loss": 510.9979, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 9.96392372444597e-05, |
| "loss": 527.1517, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 0.00010078451583347648, |
| "loss": 595.9219, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 0.00010192979442249326, |
| "loss": 480.4347, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 0.00010307507301151004, |
| "loss": 487.5384, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.82, |
| "eval_RMSE": 34.58900833129883, |
| "eval_loss": 1558.3662109375, |
| "eval_runtime": 89.9793, |
| "eval_samples_per_second": 90.81, |
| "eval_steps_per_second": 2.845, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 0.00010422035160052684, |
| "loss": 538.8044, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 0.0001053656301895436, |
| "loss": 438.5807, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 0.00010651090877856037, |
| "loss": 492.5742, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 0.00010765618736757717, |
| "loss": 368.3077, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 0.00010880146595659392, |
| "loss": 374.0631, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 0.00010994674454561072, |
| "loss": 505.5752, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 0.0001110920231346275, |
| "loss": 392.0846, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 0.00011223730172364425, |
| "loss": 348.8105, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 0.00011338258031266105, |
| "loss": 351.132, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 0.00011452785890167783, |
| "loss": 363.1367, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.92, |
| "eval_RMSE": 31.56056785583496, |
| "eval_loss": 1274.2431640625, |
| "eval_runtime": 89.8821, |
| "eval_samples_per_second": 90.908, |
| "eval_steps_per_second": 2.848, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 0.0001156731374906946, |
| "loss": 296.5691, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 0.00011681841607971138, |
| "loss": 357.6329, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 0.00011796369466872816, |
| "loss": 361.3535, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 0.00011910897325774494, |
| "loss": 322.6138, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 0.00012025425184676171, |
| "loss": 405.8568, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 0.00012139953043577849, |
| "loss": 352.4393, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 0.00012254480902479527, |
| "loss": 367.3307, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 0.00012369008761381204, |
| "loss": 300.1264, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 0.00012483536620282884, |
| "loss": 344.8634, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 0.0001259806447918456, |
| "loss": 275.3597, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.01, |
| "eval_RMSE": 28.991735458374023, |
| "eval_loss": 1044.326416015625, |
| "eval_runtime": 90.499, |
| "eval_samples_per_second": 90.288, |
| "eval_steps_per_second": 2.829, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 0.00012712592338086237, |
| "loss": 354.2523, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 0.00012827120196987917, |
| "loss": 283.569, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 0.00012941648055889594, |
| "loss": 248.2783, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 0.00013056175914791273, |
| "loss": 262.5986, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 0.0001317070377369295, |
| "loss": 294.735, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 0.0001328523163259463, |
| "loss": 301.2489, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 0.00013399759491496306, |
| "loss": 252.3795, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 0.00013514287350397983, |
| "loss": 247.8596, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 0.00013628815209299662, |
| "loss": 253.8303, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 0.0001374334306820134, |
| "loss": 251.5496, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.1, |
| "eval_RMSE": 27.961669921875, |
| "eval_loss": 858.2109985351562, |
| "eval_runtime": 90.7351, |
| "eval_samples_per_second": 90.053, |
| "eval_steps_per_second": 2.821, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 0.00013857870927103016, |
| "loss": 207.0559, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 0.00013972398786004695, |
| "loss": 231.4458, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 0.00014086926644906372, |
| "loss": 252.8603, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 0.00014201454503808052, |
| "loss": 204.8269, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 0.00014315982362709729, |
| "loss": 230.3754, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 0.00014430510221611405, |
| "loss": 170.1902, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 0.00014545038080513085, |
| "loss": 207.8812, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 0.00014659565939414762, |
| "loss": 250.5713, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 0.00014774093798316438, |
| "loss": 210.8263, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 0.00014888621657218118, |
| "loss": 232.956, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.19, |
| "eval_RMSE": 28.184946060180664, |
| "eval_loss": 733.7390747070312, |
| "eval_runtime": 89.6232, |
| "eval_samples_per_second": 91.171, |
| "eval_steps_per_second": 2.856, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 0.00015003149516119795, |
| "loss": 209.1999, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 0.00015117677375021474, |
| "loss": 219.4115, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 0.0001523220523392315, |
| "loss": 142.051, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 0.0001534673309282483, |
| "loss": 182.2509, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 0.00015461260951726507, |
| "loss": 168.6032, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 0.00015575788810628184, |
| "loss": 171.8442, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 0.00015690316669529864, |
| "loss": 171.8282, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 0.0001580484452843154, |
| "loss": 186.681, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 0.00015919372387333217, |
| "loss": 173.0698, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 0.00016033900246234897, |
| "loss": 148.5199, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.28, |
| "eval_RMSE": 27.53377914428711, |
| "eval_loss": 608.1200561523438, |
| "eval_runtime": 87.4424, |
| "eval_samples_per_second": 93.444, |
| "eval_steps_per_second": 2.928, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 0.00016148428105136573, |
| "loss": 195.879, |
| "step": 1410 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 0.0001626295596403825, |
| "loss": 161.7053, |
| "step": 1420 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 0.0001637748382293993, |
| "loss": 144.2296, |
| "step": 1430 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 0.00016492011681841607, |
| "loss": 146.4364, |
| "step": 1440 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 0.00016606539540743283, |
| "loss": 144.3109, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 0.00016721067399644963, |
| "loss": 184.0622, |
| "step": 1460 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 0.00016835595258546642, |
| "loss": 127.7919, |
| "step": 1470 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 0.00016950123117448316, |
| "loss": 133.7518, |
| "step": 1480 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 0.00017064650976349996, |
| "loss": 112.2735, |
| "step": 1490 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 0.00017179178835251675, |
| "loss": 126.9139, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.37, |
| "eval_RMSE": 26.30722999572754, |
| "eval_loss": 453.3481750488281, |
| "eval_runtime": 88.0738, |
| "eval_samples_per_second": 92.774, |
| "eval_steps_per_second": 2.907, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 0.00017293706694153352, |
| "loss": 134.5855, |
| "step": 1510 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 0.0001740823455305503, |
| "loss": 124.9829, |
| "step": 1520 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 0.00017522762411956708, |
| "loss": 111.1943, |
| "step": 1530 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 0.00017637290270858385, |
| "loss": 124.6046, |
| "step": 1540 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 0.00017751818129760062, |
| "loss": 97.5316, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 0.00017866345988661742, |
| "loss": 130.0623, |
| "step": 1560 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 0.0001798087384756342, |
| "loss": 140.2047, |
| "step": 1570 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 0.00018095401706465095, |
| "loss": 111.0409, |
| "step": 1580 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 0.00018209929565366775, |
| "loss": 112.2171, |
| "step": 1590 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 0.00018324457424268454, |
| "loss": 113.2355, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.47, |
| "eval_RMSE": 26.593936920166016, |
| "eval_loss": 397.5113220214844, |
| "eval_runtime": 89.5143, |
| "eval_samples_per_second": 91.282, |
| "eval_steps_per_second": 2.86, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 0.00018438985283170128, |
| "loss": 121.8388, |
| "step": 1610 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 0.00018553513142071808, |
| "loss": 116.3681, |
| "step": 1620 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 0.00018668041000973487, |
| "loss": 113.7981, |
| "step": 1630 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 0.0001878256885987516, |
| "loss": 119.2628, |
| "step": 1640 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 0.0001889709671877684, |
| "loss": 93.8756, |
| "step": 1650 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 0.0001901162457767852, |
| "loss": 118.5608, |
| "step": 1660 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 0.00019126152436580197, |
| "loss": 101.4624, |
| "step": 1670 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 0.00019240680295481874, |
| "loss": 96.8405, |
| "step": 1680 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 0.00019355208154383553, |
| "loss": 93.3038, |
| "step": 1690 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 0.0001946973601328523, |
| "loss": 99.5798, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.56, |
| "eval_RMSE": 25.650577545166016, |
| "eval_loss": 304.5827331542969, |
| "eval_runtime": 88.1891, |
| "eval_samples_per_second": 92.653, |
| "eval_steps_per_second": 2.903, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 0.00019584263872186907, |
| "loss": 116.748, |
| "step": 1710 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 0.00019698791731088586, |
| "loss": 87.1141, |
| "step": 1720 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 0.00019813319589990266, |
| "loss": 97.8321, |
| "step": 1730 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 0.0001992784744889194, |
| "loss": 79.7827, |
| "step": 1740 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 0.00019995291632467376, |
| "loss": 98.2981, |
| "step": 1750 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 0.00019982566314811633, |
| "loss": 68.1837, |
| "step": 1760 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 0.0001996984099715589, |
| "loss": 88.1333, |
| "step": 1770 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 0.0001995711567950015, |
| "loss": 84.4777, |
| "step": 1780 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 0.00019944390361844406, |
| "loss": 75.6494, |
| "step": 1790 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 0.00019931665044188666, |
| "loss": 101.6581, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.65, |
| "eval_RMSE": 24.347003936767578, |
| "eval_loss": 246.870849609375, |
| "eval_runtime": 91.7415, |
| "eval_samples_per_second": 89.066, |
| "eval_steps_per_second": 2.79, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 0.00019918939726532923, |
| "loss": 79.9642, |
| "step": 1810 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 0.00019906214408877182, |
| "loss": 67.1507, |
| "step": 1820 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 0.0001989348909122144, |
| "loss": 78.1831, |
| "step": 1830 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 0.000198807637735657, |
| "loss": 72.3693, |
| "step": 1840 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 0.00019868038455909956, |
| "loss": 73.2772, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 0.00019855313138254216, |
| "loss": 73.4863, |
| "step": 1860 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 0.00019842587820598472, |
| "loss": 68.861, |
| "step": 1870 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 0.0001982986250294273, |
| "loss": 70.2734, |
| "step": 1880 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 0.00019817137185286986, |
| "loss": 80.5998, |
| "step": 1890 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 0.00019804411867631246, |
| "loss": 74.3472, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.74, |
| "eval_RMSE": 25.971708297729492, |
| "eval_loss": 242.7896270751953, |
| "eval_runtime": 89.3685, |
| "eval_samples_per_second": 91.43, |
| "eval_steps_per_second": 2.865, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 0.00019791686549975503, |
| "loss": 88.6437, |
| "step": 1910 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 0.00019778961232319763, |
| "loss": 95.2708, |
| "step": 1920 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 0.0001976623591466402, |
| "loss": 114.3341, |
| "step": 1930 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 0.0001975351059700828, |
| "loss": 67.5784, |
| "step": 1940 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 0.00019740785279352536, |
| "loss": 73.2939, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 0.00019728059961696793, |
| "loss": 68.14, |
| "step": 1960 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 0.00019715334644041053, |
| "loss": 69.5453, |
| "step": 1970 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 0.0001970260932638531, |
| "loss": 81.5987, |
| "step": 1980 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 0.0001968988400872957, |
| "loss": 71.576, |
| "step": 1990 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 0.00019677158691073826, |
| "loss": 77.3393, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.83, |
| "eval_RMSE": 25.02290153503418, |
| "eval_loss": 158.2856903076172, |
| "eval_runtime": 91.421, |
| "eval_samples_per_second": 89.378, |
| "eval_steps_per_second": 2.8, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 0.00019664433373418083, |
| "loss": 71.151, |
| "step": 2010 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 0.0001965170805576234, |
| "loss": 57.0214, |
| "step": 2020 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 0.000196389827381066, |
| "loss": 60.4723, |
| "step": 2030 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 0.00019626257420450857, |
| "loss": 61.1471, |
| "step": 2040 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 0.00019613532102795116, |
| "loss": 64.6843, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 0.00019600806785139373, |
| "loss": 80.3143, |
| "step": 2060 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 0.00019588081467483633, |
| "loss": 61.2826, |
| "step": 2070 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 0.0001957535614982789, |
| "loss": 87.4192, |
| "step": 2080 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 0.0001956263083217215, |
| "loss": 80.7123, |
| "step": 2090 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 0.00019549905514516406, |
| "loss": 69.8205, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.92, |
| "eval_RMSE": 26.108875274658203, |
| "eval_loss": 163.57669067382812, |
| "eval_runtime": 88.3567, |
| "eval_samples_per_second": 92.477, |
| "eval_steps_per_second": 2.897, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 0.00019537180196860666, |
| "loss": 68.615, |
| "step": 2110 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 0.00019524454879204923, |
| "loss": 66.9629, |
| "step": 2120 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 0.00019511729561549182, |
| "loss": 66.9707, |
| "step": 2130 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 0.00019499004243893437, |
| "loss": 61.8375, |
| "step": 2140 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 0.00019486278926237696, |
| "loss": 86.5031, |
| "step": 2150 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 0.00019473553608581953, |
| "loss": 59.317, |
| "step": 2160 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 0.00019460828290926213, |
| "loss": 46.4391, |
| "step": 2170 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 0.0001944810297327047, |
| "loss": 66.9274, |
| "step": 2180 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 0.0001943537765561473, |
| "loss": 60.0429, |
| "step": 2190 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 0.00019422652337958986, |
| "loss": 49.6239, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.01, |
| "eval_RMSE": 25.317358016967773, |
| "eval_loss": 155.91978454589844, |
| "eval_runtime": 89.937, |
| "eval_samples_per_second": 90.852, |
| "eval_steps_per_second": 2.846, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 0.00019409927020303243, |
| "loss": 48.8787, |
| "step": 2210 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 0.00019397201702647503, |
| "loss": 54.4583, |
| "step": 2220 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 0.0001938447638499176, |
| "loss": 51.7232, |
| "step": 2230 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 0.0001937175106733602, |
| "loss": 55.3078, |
| "step": 2240 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 0.00019359025749680276, |
| "loss": 52.4878, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 0.00019346300432024533, |
| "loss": 51.498, |
| "step": 2260 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 0.0001933357511436879, |
| "loss": 55.693, |
| "step": 2270 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 0.0001932084979671305, |
| "loss": 60.7052, |
| "step": 2280 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 0.00019308124479057307, |
| "loss": 57.5329, |
| "step": 2290 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 0.00019295399161401567, |
| "loss": 49.7357, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.11, |
| "eval_RMSE": 25.681665420532227, |
| "eval_loss": 113.43367004394531, |
| "eval_runtime": 88.8092, |
| "eval_samples_per_second": 92.006, |
| "eval_steps_per_second": 2.883, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 0.00019282673843745823, |
| "loss": 40.8768, |
| "step": 2310 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 0.00019269948526090083, |
| "loss": 51.1638, |
| "step": 2320 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 0.0001925722320843434, |
| "loss": 53.8347, |
| "step": 2330 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 0.000192444978907786, |
| "loss": 61.9818, |
| "step": 2340 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 0.00019231772573122857, |
| "loss": 65.0766, |
| "step": 2350 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 0.00019219047255467116, |
| "loss": 54.4025, |
| "step": 2360 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 0.00019206321937811373, |
| "loss": 43.9705, |
| "step": 2370 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 0.00019193596620155633, |
| "loss": 58.1021, |
| "step": 2380 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 0.00019180871302499887, |
| "loss": 41.1698, |
| "step": 2390 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 0.00019168145984844147, |
| "loss": 60.9505, |
| "step": 2400 |
| }, |
| { |
| "epoch": 2.2, |
| "eval_RMSE": 25.213809967041016, |
| "eval_loss": 128.33050537109375, |
| "eval_runtime": 89.5513, |
| "eval_samples_per_second": 91.244, |
| "eval_steps_per_second": 2.859, |
| "step": 2400 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 0.00019155420667188404, |
| "loss": 50.3419, |
| "step": 2410 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 0.00019142695349532663, |
| "loss": 34.2566, |
| "step": 2420 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 0.0001912997003187692, |
| "loss": 45.824, |
| "step": 2430 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 0.0001911724471422118, |
| "loss": 51.5597, |
| "step": 2440 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 0.00019104519396565437, |
| "loss": 63.5468, |
| "step": 2450 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 0.00019091794078909694, |
| "loss": 47.8715, |
| "step": 2460 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 0.00019079068761253953, |
| "loss": 56.6548, |
| "step": 2470 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 0.0001906634344359821, |
| "loss": 46.0047, |
| "step": 2480 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 0.0001905361812594247, |
| "loss": 55.3212, |
| "step": 2490 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 0.00019040892808286727, |
| "loss": 49.0666, |
| "step": 2500 |
| }, |
| { |
| "epoch": 2.29, |
| "eval_RMSE": 26.342649459838867, |
| "eval_loss": 108.29063415527344, |
| "eval_runtime": 89.6891, |
| "eval_samples_per_second": 91.104, |
| "eval_steps_per_second": 2.854, |
| "step": 2500 |
| } |
| ], |
| "max_steps": 4368, |
| "num_train_epochs": 4, |
| "total_flos": 9.294770879473766e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|