| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 19.97563946406821, | |
| "global_step": 8200, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.999981652304564e-05, | |
| "loss": 3.5337, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.999933765030933e-05, | |
| "loss": 2.4344, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.999845697279959e-05, | |
| "loss": 1.7436, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.999720936402992e-05, | |
| "loss": 1.9107, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.999559484231293e-05, | |
| "loss": 1.5355, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.9993613431346816e-05, | |
| "loss": 1.3407, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.999126516021505e-05, | |
| "loss": 1.2717, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.998855006338591e-05, | |
| "loss": 1.294, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.9985468180712016e-05, | |
| "loss": 1.0712, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.998201955742973e-05, | |
| "loss": 1.1961, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.997820424415846e-05, | |
| "loss": 1.3744, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.99740222969e-05, | |
| "loss": 1.1906, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.996947377703761e-05, | |
| "loss": 1.026, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.9964558751335176e-05, | |
| "loss": 0.9262, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.995982192522523e-05, | |
| "loss": 1.1451, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.9954210741627514e-05, | |
| "loss": 0.9061, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.994823327622299e-05, | |
| "loss": 0.8371, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.99418896167498e-05, | |
| "loss": 1.1765, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.9935179856321204e-05, | |
| "loss": 1.0249, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.992810409342409e-05, | |
| "loss": 1.2131, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.992066243191761e-05, | |
| "loss": 0.9495, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.991285498103165e-05, | |
| "loss": 1.1412, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.99046818553652e-05, | |
| "loss": 1.0495, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.989614317488466e-05, | |
| "loss": 1.162, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.988723906492211e-05, | |
| "loss": 1.0261, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.98789130316581e-05, | |
| "loss": 0.9507, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.986931497017868e-05, | |
| "loss": 0.9162, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.985935187300584e-05, | |
| "loss": 0.9901, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.98490238863795e-05, | |
| "loss": 0.9501, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.983833116189544e-05, | |
| "loss": 0.9202, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.982727385650313e-05, | |
| "loss": 0.9175, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.9815852132503465e-05, | |
| "loss": 0.8588, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.980406615754627e-05, | |
| "loss": 1.1674, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.979191610462793e-05, | |
| "loss": 0.9042, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.977940215208884e-05, | |
| "loss": 0.8963, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.976652448361072e-05, | |
| "loss": 0.9845, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.9753283288214036e-05, | |
| "loss": 0.9651, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.973967876025509e-05, | |
| "loss": 1.0159, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.97257110994233e-05, | |
| "loss": 0.8109, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.971138051073816e-05, | |
| "loss": 1.0778, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.9696687204546286e-05, | |
| "loss": 0.9938, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.968163139651835e-05, | |
| "loss": 0.9635, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.966621330764582e-05, | |
| "loss": 0.9848, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.9650433164237844e-05, | |
| "loss": 0.9077, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.9634291197917816e-05, | |
| "loss": 0.8833, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.961778764562005e-05, | |
| "loss": 0.8134, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.960092274958627e-05, | |
| "loss": 0.7578, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.958369675736205e-05, | |
| "loss": 0.8654, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.95661099217932e-05, | |
| "loss": 0.8552, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.954816250102204e-05, | |
| "loss": 0.6767, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.952985475848362e-05, | |
| "loss": 0.8525, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.9511186962901846e-05, | |
| "loss": 0.7788, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.949215938828554e-05, | |
| "loss": 0.9851, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.9472772313924425e-05, | |
| "loss": 0.8086, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.9453026024385e-05, | |
| "loss": 0.7749, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.943292080950639e-05, | |
| "loss": 0.9572, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 4.941245696439609e-05, | |
| "loss": 0.7659, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.9391634789425616e-05, | |
| "loss": 0.7514, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 4.93704545902261e-05, | |
| "loss": 0.6983, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.934891667768383e-05, | |
| "loss": 0.8481, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.9327021367935643e-05, | |
| "loss": 0.8893, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.930476898236432e-05, | |
| "loss": 0.606, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.9282159847593854e-05, | |
| "loss": 0.9614, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.925919429548467e-05, | |
| "loss": 0.8691, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.923587266312873e-05, | |
| "loss": 0.9466, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.921219529284459e-05, | |
| "loss": 0.8841, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.91881625321724e-05, | |
| "loss": 0.9382, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.9163774733868784e-05, | |
| "loss": 0.7369, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 4.913903225590164e-05, | |
| "loss": 0.8588, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.9113935461444955e-05, | |
| "loss": 0.9441, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.9088484718873385e-05, | |
| "loss": 0.8136, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.90626804017569e-05, | |
| "loss": 0.838, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.9036522888855325e-05, | |
| "loss": 0.9653, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.901001256411271e-05, | |
| "loss": 0.797, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.898314981665174e-05, | |
| "loss": 0.8807, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.8955935040768034e-05, | |
| "loss": 0.9804, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.892836863592432e-05, | |
| "loss": 0.8705, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.890045100674461e-05, | |
| "loss": 0.8691, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 4.887218256300821e-05, | |
| "loss": 0.8105, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.884356371964377e-05, | |
| "loss": 0.8711, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.8814594896723155e-05, | |
| "loss": 0.856, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.8785276519455264e-05, | |
| "loss": 0.7324, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.875560901817982e-05, | |
| "loss": 0.7865, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.8725592828361066e-05, | |
| "loss": 0.7155, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.86952283905813e-05, | |
| "loss": 0.6735, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.866451615053451e-05, | |
| "loss": 0.5544, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.863345655901975e-05, | |
| "loss": 0.816, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.8602050071934554e-05, | |
| "loss": 0.5772, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.857029715026826e-05, | |
| "loss": 0.6604, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.853819826009521e-05, | |
| "loss": 0.602, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 4.850575387256794e-05, | |
| "loss": 0.7244, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 4.8472964463910224e-05, | |
| "loss": 0.6775, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.8439830515410155e-05, | |
| "loss": 0.6261, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.840635251341299e-05, | |
| "loss": 0.8245, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.837253094931409e-05, | |
| "loss": 0.6961, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.833836631955164e-05, | |
| "loss": 0.9983, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.8303859125599446e-05, | |
| "loss": 0.6283, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.8269009873959474e-05, | |
| "loss": 0.9679, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.8233819076154494e-05, | |
| "loss": 0.7606, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.8198287248720546e-05, | |
| "loss": 0.6978, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 4.8162414913199336e-05, | |
| "loss": 0.5871, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.812620259613063e-05, | |
| "loss": 0.7016, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.808965082904446e-05, | |
| "loss": 0.8239, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.805276014845339e-05, | |
| "loss": 0.8266, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.801553109584459e-05, | |
| "loss": 0.7679, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.7977964217671925e-05, | |
| "loss": 0.6345, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.79400600653479e-05, | |
| "loss": 0.77, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.790181919523559e-05, | |
| "loss": 0.753, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.7863242168640466e-05, | |
| "loss": 0.6868, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.782432955180216e-05, | |
| "loss": 0.7269, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.7785081915886134e-05, | |
| "loss": 0.7442, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.7745499836975335e-05, | |
| "loss": 0.837, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.77055838960617e-05, | |
| "loss": 0.563, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.7665334679037656e-05, | |
| "loss": 0.6673, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.76247527766875e-05, | |
| "loss": 0.5567, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.758383878467873e-05, | |
| "loss": 0.6742, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.754259330355334e-05, | |
| "loss": 0.6214, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.750101693871893e-05, | |
| "loss": 0.7883, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.7459110300439905e-05, | |
| "loss": 0.8314, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.7416874003828425e-05, | |
| "loss": 0.6504, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.7374308668835484e-05, | |
| "loss": 0.8773, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.7331414920241704e-05, | |
| "loss": 0.7314, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.728819338764824e-05, | |
| "loss": 0.7255, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 4.72446447054675e-05, | |
| "loss": 0.6126, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.7200769512913855e-05, | |
| "loss": 0.6329, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.715656845399425e-05, | |
| "loss": 0.6519, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.711204217749872e-05, | |
| "loss": 0.7236, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.706719133699093e-05, | |
| "loss": 0.5436, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.702201659079851e-05, | |
| "loss": 0.5856, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.697651860200347e-05, | |
| "loss": 0.6632, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 4.6930698038432394e-05, | |
| "loss": 0.4646, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.688455557264667e-05, | |
| "loss": 0.6182, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 4.683809188193263e-05, | |
| "loss": 0.4338, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 4.679130764829161e-05, | |
| "loss": 0.757, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.674420355842987e-05, | |
| "loss": 0.4597, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.6696780303748626e-05, | |
| "loss": 0.6495, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 4.6649038580333827e-05, | |
| "loss": 0.685, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.660097908894595e-05, | |
| "loss": 0.5783, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 4.655260253500972e-05, | |
| "loss": 0.5711, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.650390962860377e-05, | |
| "loss": 0.5597, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 4.645490108445018e-05, | |
| "loss": 0.6217, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.640557762190403e-05, | |
| "loss": 0.5919, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.635593996494281e-05, | |
| "loss": 0.6931, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 4.630598884215582e-05, | |
| "loss": 0.5832, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.625572498673344e-05, | |
| "loss": 0.5968, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 4.620514913645641e-05, | |
| "loss": 0.7203, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 4.6154262033684944e-05, | |
| "loss": 0.5169, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.610306442534792e-05, | |
| "loss": 0.6657, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.6051557062931824e-05, | |
| "loss": 0.5104, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.599974070246978e-05, | |
| "loss": 0.5557, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.594761610453043e-05, | |
| "loss": 0.6474, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.5895184034206765e-05, | |
| "loss": 0.5972, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.58424452611049e-05, | |
| "loss": 0.5479, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 4.578940055933282e-05, | |
| "loss": 0.4461, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 4.573605070748894e-05, | |
| "loss": 0.7081, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 4.5682396488650716e-05, | |
| "loss": 0.7244, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 4.562843869036317e-05, | |
| "loss": 0.656, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.557417810462729e-05, | |
| "loss": 0.5638, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.551961552788847e-05, | |
| "loss": 0.484, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.54647517610247e-05, | |
| "loss": 0.4737, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.5409587609334954e-05, | |
| "loss": 0.7067, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 4.535412388252727e-05, | |
| "loss": 0.5712, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 4.529836139470689e-05, | |
| "loss": 0.6356, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.524230096436435e-05, | |
| "loss": 0.6097, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.5185943414363396e-05, | |
| "loss": 0.4282, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 4.512928957192896e-05, | |
| "loss": 0.3885, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 4.5072340268635e-05, | |
| "loss": 0.2796, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 4.5015096340392304e-05, | |
| "loss": 0.5265, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 4.495755862743619e-05, | |
| "loss": 0.6607, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 4.48997279743142e-05, | |
| "loss": 0.3981, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 4.484160522987372e-05, | |
| "loss": 0.4307, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 4.4783191247249457e-05, | |
| "loss": 0.56, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 4.4724486883851e-05, | |
| "loss": 0.4518, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 4.466549300135016e-05, | |
| "loss": 0.5755, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 4.4606210465668384e-05, | |
| "loss": 0.5125, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 4.454664014696398e-05, | |
| "loss": 0.5609, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 4.448678291961942e-05, | |
| "loss": 0.4354, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 4.442663966222842e-05, | |
| "loss": 0.6086, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 4.4366211257583145e-05, | |
| "loss": 0.5637, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 4.4305498592661146e-05, | |
| "loss": 0.4855, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 4.424450255861242e-05, | |
| "loss": 0.361, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 4.4183224050746286e-05, | |
| "loss": 0.6417, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 4.412166396851827e-05, | |
| "loss": 0.473, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 4.4059823215516865e-05, | |
| "loss": 0.587, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 4.399770269945034e-05, | |
| "loss": 0.4707, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 4.393530333213332e-05, | |
| "loss": 0.4057, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 4.387262602947349e-05, | |
| "loss": 0.404, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 4.380967171145809e-05, | |
| "loss": 0.478, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 4.374644130214044e-05, | |
| "loss": 0.4948, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 4.3682935729626385e-05, | |
| "loss": 0.5436, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 4.361915592606063e-05, | |
| "loss": 0.4618, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 4.355510282761312e-05, | |
| "loss": 0.475, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 4.349077737446525e-05, | |
| "loss": 0.4473, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 4.342618051079606e-05, | |
| "loss": 0.4606, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 4.336131318476842e-05, | |
| "loss": 0.4914, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.32961763485151e-05, | |
| "loss": 0.5147, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 4.323077095812476e-05, | |
| "loss": 0.5582, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 4.316509797362793e-05, | |
| "loss": 0.5085, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 4.309915835898296e-05, | |
| "loss": 0.5099, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 4.3032953082061823e-05, | |
| "loss": 0.4276, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 4.296648311463591e-05, | |
| "loss": 0.4463, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 4.289974943236182e-05, | |
| "loss": 0.5247, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 4.2832753014766955e-05, | |
| "loss": 0.5736, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 4.276549484523521e-05, | |
| "loss": 0.4952, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 4.269797591099253e-05, | |
| "loss": 0.4736, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 4.263019720309236e-05, | |
| "loss": 0.4767, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 4.2562159716401193e-05, | |
| "loss": 0.4043, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 5.07, | |
| "learning_rate": 4.2493864449583883e-05, | |
| "loss": 0.4051, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 4.2425312405089056e-05, | |
| "loss": 0.343, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 5.12, | |
| "learning_rate": 4.2356504589134316e-05, | |
| "loss": 0.3777, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 5.14, | |
| "learning_rate": 4.228744201169156e-05, | |
| "loss": 0.3874, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 4.2218125686472075e-05, | |
| "loss": 0.5224, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 5.19, | |
| "learning_rate": 4.2148556630911736e-05, | |
| "loss": 0.4146, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 5.21, | |
| "learning_rate": 4.207873586615603e-05, | |
| "loss": 0.4122, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 4.200866441704504e-05, | |
| "loss": 0.3635, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 4.1938343312098446e-05, | |
| "loss": 0.2448, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 4.186777358350044e-05, | |
| "loss": 0.4659, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 4.179695626708452e-05, | |
| "loss": 0.3774, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "learning_rate": 4.1725892402318334e-05, | |
| "loss": 0.2299, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 4.1654583032288394e-05, | |
| "loss": 0.371, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 4.1583029203684784e-05, | |
| "loss": 0.33, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 4.1511231966785796e-05, | |
| "loss": 0.4362, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 4.1439192375442495e-05, | |
| "loss": 0.2833, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 5.46, | |
| "learning_rate": 4.136691148706325e-05, | |
| "loss": 0.3126, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 5.48, | |
| "learning_rate": 4.129439036259826e-05, | |
| "loss": 0.3969, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 4.122163006652393e-05, | |
| "loss": 0.3791, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 4.114863166682725e-05, | |
| "loss": 0.4307, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 4.107539623499014e-05, | |
| "loss": 0.3985, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 5.58, | |
| "learning_rate": 4.100192484597374e-05, | |
| "loss": 0.3489, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 4.0928218578202574e-05, | |
| "loss": 0.3095, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 4.0854278513548764e-05, | |
| "loss": 0.3725, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 5.65, | |
| "learning_rate": 4.078010573731612e-05, | |
| "loss": 0.428, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 5.68, | |
| "learning_rate": 4.0705701338224276e-05, | |
| "loss": 0.4104, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 5.7, | |
| "learning_rate": 4.063106640839264e-05, | |
| "loss": 0.4075, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 5.72, | |
| "learning_rate": 4.055620204332435e-05, | |
| "loss": 0.3711, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 5.75, | |
| "learning_rate": 4.048110934189028e-05, | |
| "loss": 0.4147, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 4.040578940631284e-05, | |
| "loss": 0.4086, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 4.0330243342149824e-05, | |
| "loss": 0.5578, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 4.0254472258278176e-05, | |
| "loss": 0.3451, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 5.85, | |
| "learning_rate": 4.01784772668777e-05, | |
| "loss": 0.325, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 4.010225948341476e-05, | |
| "loss": 0.3898, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 4.002582002662592e-05, | |
| "loss": 0.3024, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 5.92, | |
| "learning_rate": 3.9949160018501454e-05, | |
| "loss": 0.4362, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 5.94, | |
| "learning_rate": 3.987228058426896e-05, | |
| "loss": 0.4197, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 3.979518285237679e-05, | |
| "loss": 0.4507, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 5.99, | |
| "learning_rate": 3.971786795447751e-05, | |
| "loss": 0.4414, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 6.02, | |
| "learning_rate": 3.964033702541127e-05, | |
| "loss": 0.3511, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 6.04, | |
| "learning_rate": 3.956259120318918e-05, | |
| "loss": 0.2937, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 3.948463162897656e-05, | |
| "loss": 0.2735, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 3.940645944707625e-05, | |
| "loss": 0.2736, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 3.9328075804911746e-05, | |
| "loss": 0.33, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 6.14, | |
| "learning_rate": 3.9249481853010403e-05, | |
| "loss": 0.2393, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 6.16, | |
| "learning_rate": 3.917067874498653e-05, | |
| "loss": 0.3025, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 6.19, | |
| "learning_rate": 3.909166763752448e-05, | |
| "loss": 0.2683, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 3.901244969036163e-05, | |
| "loss": 0.2801, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 6.24, | |
| "learning_rate": 3.8933026066271396e-05, | |
| "loss": 0.3711, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "learning_rate": 3.8853397931046156e-05, | |
| "loss": 0.4418, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 3.877356645348011e-05, | |
| "loss": 0.2819, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 3.86935328053522e-05, | |
| "loss": 0.3661, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 6.33, | |
| "learning_rate": 3.861329816140881e-05, | |
| "loss": 0.2374, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 6.36, | |
| "learning_rate": 3.853286369934658e-05, | |
| "loss": 0.2686, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 6.38, | |
| "learning_rate": 3.845223059979512e-05, | |
| "loss": 0.3097, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 6.41, | |
| "learning_rate": 3.83714000462997e-05, | |
| "loss": 0.2125, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 3.8290373225303786e-05, | |
| "loss": 0.2657, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 6.46, | |
| "learning_rate": 3.820915132613174e-05, | |
| "loss": 0.284, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 6.48, | |
| "learning_rate": 3.81277355409713e-05, | |
| "loss": 0.3823, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 6.5, | |
| "learning_rate": 3.804612706485609e-05, | |
| "loss": 0.2053, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 3.796432709564807e-05, | |
| "loss": 0.3304, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 6.55, | |
| "learning_rate": 3.7882336834019994e-05, | |
| "loss": 0.2427, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 6.58, | |
| "learning_rate": 3.780015748343773e-05, | |
| "loss": 0.3177, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 3.771779025014264e-05, | |
| "loss": 0.2807, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 6.63, | |
| "learning_rate": 3.7635236343133854e-05, | |
| "loss": 0.3766, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 3.755249697415051e-05, | |
| "loss": 0.356, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 3.746957335765402e-05, | |
| "loss": 0.2231, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 6.7, | |
| "learning_rate": 3.7386466710810194e-05, | |
| "loss": 0.3641, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 6.72, | |
| "learning_rate": 3.730317825347137e-05, | |
| "loss": 0.2509, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 6.75, | |
| "learning_rate": 3.721970920815856e-05, | |
| "loss": 0.3387, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 6.77, | |
| "learning_rate": 3.713606080004346e-05, | |
| "loss": 0.3275, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 6.8, | |
| "learning_rate": 3.705223425693049e-05, | |
| "loss": 0.2544, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 3.696823080923873e-05, | |
| "loss": 0.2391, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 3.688405168998394e-05, | |
| "loss": 0.3147, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 3.6799698134760395e-05, | |
| "loss": 0.2503, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 3.671517138172277e-05, | |
| "loss": 0.3549, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 6.92, | |
| "learning_rate": 3.663047267156794e-05, | |
| "loss": 0.3271, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 6.94, | |
| "learning_rate": 3.654560324751684e-05, | |
| "loss": 0.2629, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 6.97, | |
| "learning_rate": 3.6460564355296124e-05, | |
| "loss": 0.2522, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 6.99, | |
| "learning_rate": 3.6375357243119963e-05, | |
| "loss": 0.3017, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 7.02, | |
| "learning_rate": 3.628998316167167e-05, | |
| "loss": 0.2657, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 3.620444336408535e-05, | |
| "loss": 0.1793, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 3.611873910592753e-05, | |
| "loss": 0.1965, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 3.603287164517872e-05, | |
| "loss": 0.1778, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 3.5946842242214934e-05, | |
| "loss": 0.1364, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 3.586065215978919e-05, | |
| "loss": 0.1961, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 7.16, | |
| "learning_rate": 3.577430266301299e-05, | |
| "loss": 0.1968, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 7.19, | |
| "learning_rate": 3.568779501933777e-05, | |
| "loss": 0.1688, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 3.5601130498536226e-05, | |
| "loss": 0.1694, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 7.24, | |
| "learning_rate": 3.551431037268374e-05, | |
| "loss": 0.2817, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 3.5427335916139724e-05, | |
| "loss": 0.2182, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 7.28, | |
| "learning_rate": 3.534020840552883e-05, | |
| "loss": 0.1984, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 3.5252929119722275e-05, | |
| "loss": 0.1884, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 7.33, | |
| "learning_rate": 3.516549933981907e-05, | |
| "loss": 0.1926, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 3.50779203491272e-05, | |
| "loss": 0.2422, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 7.38, | |
| "learning_rate": 3.499019343314478e-05, | |
| "loss": 0.2743, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 7.41, | |
| "learning_rate": 3.490231987954119e-05, | |
| "loss": 0.1692, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 3.4814300978138194e-05, | |
| "loss": 0.1952, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 7.45, | |
| "learning_rate": 3.4726138020890994e-05, | |
| "loss": 0.2851, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 7.48, | |
| "learning_rate": 3.463783230186925e-05, | |
| "loss": 0.2437, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "learning_rate": 3.454938511723812e-05, | |
| "loss": 0.2101, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 3.44607977652392e-05, | |
| "loss": 0.212, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 3.4372071546171516e-05, | |
| "loss": 0.2584, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 7.58, | |
| "learning_rate": 3.428320776237235e-05, | |
| "loss": 0.2939, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 3.419420771819824e-05, | |
| "loss": 0.2372, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 3.4105072720005744e-05, | |
| "loss": 0.2033, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 7.65, | |
| "learning_rate": 3.401580407613229e-05, | |
| "loss": 0.2327, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 3.392640309687701e-05, | |
| "loss": 0.2603, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 3.383687109448143e-05, | |
| "loss": 0.2729, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 3.37472093831103e-05, | |
| "loss": 0.2348, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 7.75, | |
| "learning_rate": 3.365741927883223e-05, | |
| "loss": 0.1297, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 7.77, | |
| "learning_rate": 3.3567502099600414e-05, | |
| "loss": 0.2293, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 3.347745916523326e-05, | |
| "loss": 0.212, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "learning_rate": 3.3387291797395034e-05, | |
| "loss": 0.2167, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 3.329700131957648e-05, | |
| "loss": 0.267, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 3.3206589057075314e-05, | |
| "loss": 0.212, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 3.3116056336976885e-05, | |
| "loss": 0.1898, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 7.92, | |
| "learning_rate": 3.302540448813462e-05, | |
| "loss": 0.1769, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 3.293463484115051e-05, | |
| "loss": 0.2688, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 7.97, | |
| "learning_rate": 3.284374872835563e-05, | |
| "loss": 0.2554, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 7.99, | |
| "learning_rate": 3.2752747483790556e-05, | |
| "loss": 0.1952, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 8.01, | |
| "learning_rate": 3.26616324431858e-05, | |
| "loss": 0.1999, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "learning_rate": 3.2570404943942136e-05, | |
| "loss": 0.1176, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 3.247906632511109e-05, | |
| "loss": 0.1442, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 8.09, | |
| "learning_rate": 3.238761792737518e-05, | |
| "loss": 0.187, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 8.11, | |
| "learning_rate": 3.22960610930283e-05, | |
| "loss": 0.1203, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 8.14, | |
| "learning_rate": 3.220439716595596e-05, | |
| "loss": 0.1056, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 3.211262749161562e-05, | |
| "loss": 0.1151, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 8.19, | |
| "learning_rate": 3.2020753417016914e-05, | |
| "loss": 0.1765, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 8.21, | |
| "learning_rate": 3.192877629070187e-05, | |
| "loss": 0.1612, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 8.23, | |
| "learning_rate": 3.183669746272511e-05, | |
| "loss": 0.1169, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 3.17445182846341e-05, | |
| "loss": 0.1377, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 8.28, | |
| "learning_rate": 3.1652240109449204e-05, | |
| "loss": 0.2031, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 8.31, | |
| "learning_rate": 3.155986429164392e-05, | |
| "loss": 0.1687, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 8.33, | |
| "learning_rate": 3.1467392187124934e-05, | |
| "loss": 0.1099, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 3.137482515321226e-05, | |
| "loss": 0.1242, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 3.1282164548619294e-05, | |
| "loss": 0.1616, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 3.118941173343287e-05, | |
| "loss": 0.1509, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 8.43, | |
| "learning_rate": 3.109656806909333e-05, | |
| "loss": 0.1044, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 8.45, | |
| "learning_rate": 3.100363491837446e-05, | |
| "loss": 0.1792, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 8.48, | |
| "learning_rate": 3.091061364536361e-05, | |
| "loss": 0.1662, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 8.5, | |
| "learning_rate": 3.0817505615441564e-05, | |
| "loss": 0.1553, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 8.53, | |
| "learning_rate": 3.072431219526252e-05, | |
| "loss": 0.1232, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 8.55, | |
| "learning_rate": 3.06310347527341e-05, | |
| "loss": 0.1721, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 8.57, | |
| "learning_rate": 3.053767465699717e-05, | |
| "loss": 0.1511, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 3.044423327840582e-05, | |
| "loss": 0.1952, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 3.0350711988507206e-05, | |
| "loss": 0.1502, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 8.65, | |
| "learning_rate": 3.025711216002145e-05, | |
| "loss": 0.1683, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 8.67, | |
| "learning_rate": 3.0163435166821474e-05, | |
| "loss": 0.1556, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 3.0069682383912813e-05, | |
| "loss": 0.1446, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 8.72, | |
| "learning_rate": 2.9975855187413492e-05, | |
| "loss": 0.132, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 8.75, | |
| "learning_rate": 2.9881954954533758e-05, | |
| "loss": 0.1375, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 8.77, | |
| "learning_rate": 2.9787983063555918e-05, | |
| "loss": 0.1546, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 8.79, | |
| "learning_rate": 2.9693940893814072e-05, | |
| "loss": 0.2073, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 8.82, | |
| "learning_rate": 2.960924399355443e-05, | |
| "loss": 0.1486, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 8.84, | |
| "learning_rate": 2.9515072097904667e-05, | |
| "loss": 0.14, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 8.87, | |
| "learning_rate": 2.9420833929320722e-05, | |
| "loss": 0.214, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 2.9326530871045176e-05, | |
| "loss": 0.1181, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 8.92, | |
| "learning_rate": 2.923216430727306e-05, | |
| "loss": 0.114, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 8.94, | |
| "learning_rate": 2.9137735623131544e-05, | |
| "loss": 0.1473, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 2.9043246204659626e-05, | |
| "loss": 0.1988, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 8.99, | |
| "learning_rate": 2.8948697438787754e-05, | |
| "loss": 0.1839, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 9.01, | |
| "learning_rate": 2.8854090713317512e-05, | |
| "loss": 0.1057, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 2.8759427416901192e-05, | |
| "loss": 0.0707, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 9.06, | |
| "learning_rate": 2.8664708939021466e-05, | |
| "loss": 0.0814, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 2.8569936669970966e-05, | |
| "loss": 0.0954, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 9.11, | |
| "learning_rate": 2.8475112000831865e-05, | |
| "loss": 0.0707, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "learning_rate": 2.838023632345549e-05, | |
| "loss": 0.1007, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 9.16, | |
| "learning_rate": 2.8285311030441858e-05, | |
| "loss": 0.0725, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 2.819033751511927e-05, | |
| "loss": 0.0866, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 2.809531717152383e-05, | |
| "loss": 0.1054, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 9.23, | |
| "learning_rate": 2.8000251394379007e-05, | |
| "loss": 0.0937, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 2.790514157907512e-05, | |
| "loss": 0.1217, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 9.28, | |
| "learning_rate": 2.7809989121648927e-05, | |
| "loss": 0.173, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 2.7714795418763068e-05, | |
| "loss": 0.1031, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 9.33, | |
| "learning_rate": 2.7619561867685595e-05, | |
| "loss": 0.1018, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 9.35, | |
| "learning_rate": 2.7524289866269465e-05, | |
| "loss": 0.1389, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 9.38, | |
| "learning_rate": 2.7428980812932014e-05, | |
| "loss": 0.1034, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 2.7333636106634415e-05, | |
| "loss": 0.0928, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 2.723825714686119e-05, | |
| "loss": 0.0861, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 2.714284533359961e-05, | |
| "loss": 0.1036, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 9.48, | |
| "learning_rate": 2.7047402067319194e-05, | |
| "loss": 0.113, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 2.695192874895112e-05, | |
| "loss": 0.1158, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 9.52, | |
| "learning_rate": 2.68564267798677e-05, | |
| "loss": 0.1194, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 9.55, | |
| "learning_rate": 2.6760897561861743e-05, | |
| "loss": 0.1169, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 2.666534249712606e-05, | |
| "loss": 0.0836, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 9.6, | |
| "learning_rate": 2.656976298823284e-05, | |
| "loss": 0.0845, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 9.62, | |
| "learning_rate": 2.647416043811304e-05, | |
| "loss": 0.0784, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "learning_rate": 2.6378536250035852e-05, | |
| "loss": 0.1129, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 9.67, | |
| "learning_rate": 2.6282891827588046e-05, | |
| "loss": 0.1254, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 2.618722857465343e-05, | |
| "loss": 0.0714, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 9.72, | |
| "learning_rate": 2.609154789539216e-05, | |
| "loss": 0.0942, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 9.74, | |
| "learning_rate": 2.5995851194220222e-05, | |
| "loss": 0.0789, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 9.77, | |
| "learning_rate": 2.590013987578876e-05, | |
| "loss": 0.1355, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 9.79, | |
| "learning_rate": 2.5804415344963462e-05, | |
| "loss": 0.1063, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 9.82, | |
| "learning_rate": 2.5708679006803964e-05, | |
| "loss": 0.1322, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 9.84, | |
| "learning_rate": 2.56129322665432e-05, | |
| "loss": 0.1199, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 2.5517176529566806e-05, | |
| "loss": 0.0755, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 2.542141320139244e-05, | |
| "loss": 0.1086, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 2.532564368764922e-05, | |
| "loss": 0.1059, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 2.5229869394057036e-05, | |
| "loss": 0.1158, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 9.96, | |
| "learning_rate": 2.513409172640595e-05, | |
| "loss": 0.1189, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 9.99, | |
| "learning_rate": 2.5038312090535543e-05, | |
| "loss": 0.1343, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 10.01, | |
| "learning_rate": 2.4942531892314273e-05, | |
| "loss": 0.0777, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 10.04, | |
| "learning_rate": 2.4846752537618875e-05, | |
| "loss": 0.1004, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 10.06, | |
| "learning_rate": 2.4750975432313696e-05, | |
| "loss": 0.0669, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 2.4655201982230044e-05, | |
| "loss": 0.0538, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 2.4559433593145617e-05, | |
| "loss": 0.0533, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 10.13, | |
| "learning_rate": 2.4463671670763787e-05, | |
| "loss": 0.0622, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "learning_rate": 2.436791762069303e-05, | |
| "loss": 0.071, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 10.18, | |
| "learning_rate": 2.4272172848426257e-05, | |
| "loss": 0.0628, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 10.21, | |
| "learning_rate": 2.417643875932021e-05, | |
| "loss": 0.0505, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 2.408071675857482e-05, | |
| "loss": 0.0769, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 10.26, | |
| "learning_rate": 2.3985008251212574e-05, | |
| "loss": 0.0689, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 10.28, | |
| "learning_rate": 2.3889314642057913e-05, | |
| "loss": 0.0869, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 10.3, | |
| "learning_rate": 2.3793637335716586e-05, | |
| "loss": 0.0823, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 2.369797773655506e-05, | |
| "loss": 0.0665, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 10.35, | |
| "learning_rate": 2.3602337248679885e-05, | |
| "loss": 0.0624, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 10.38, | |
| "learning_rate": 2.3506717275917093e-05, | |
| "loss": 0.0801, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 2.3411119221791605e-05, | |
| "loss": 0.0624, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 10.43, | |
| "learning_rate": 2.3315544489506592e-05, | |
| "loss": 0.0494, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "learning_rate": 2.3219994481922934e-05, | |
| "loss": 0.0701, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 10.48, | |
| "learning_rate": 2.312447060153856e-05, | |
| "loss": 0.0666, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 10.5, | |
| "learning_rate": 2.3028974250467946e-05, | |
| "loss": 0.0934, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 10.52, | |
| "learning_rate": 2.2933506830421435e-05, | |
| "loss": 0.0868, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 2.2838069742684766e-05, | |
| "loss": 0.0709, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 10.57, | |
| "learning_rate": 2.2742664388098438e-05, | |
| "loss": 0.0514, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 10.6, | |
| "learning_rate": 2.2647292167037144e-05, | |
| "loss": 0.0739, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 10.62, | |
| "learning_rate": 2.255195447938927e-05, | |
| "loss": 0.052, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 10.65, | |
| "learning_rate": 2.2456652724536286e-05, | |
| "loss": 0.0828, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "learning_rate": 2.2361388301332265e-05, | |
| "loss": 0.0812, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 10.69, | |
| "learning_rate": 2.2266162608083306e-05, | |
| "loss": 0.0832, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 10.72, | |
| "learning_rate": 2.217097704252701e-05, | |
| "loss": 0.0488, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 2.2075833001812003e-05, | |
| "loss": 0.0744, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 2.198073188247738e-05, | |
| "loss": 0.0664, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 10.79, | |
| "learning_rate": 2.1885675080432247e-05, | |
| "loss": 0.0743, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 10.82, | |
| "learning_rate": 2.1790663990935202e-05, | |
| "loss": 0.0809, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 10.84, | |
| "learning_rate": 2.169570000857387e-05, | |
| "loss": 0.0585, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 10.86, | |
| "learning_rate": 2.1600784527244445e-05, | |
| "loss": 0.0781, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 2.150591894013118e-05, | |
| "loss": 0.0991, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 2.1411104639686014e-05, | |
| "loss": 0.0574, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 10.94, | |
| "learning_rate": 2.1316343017608062e-05, | |
| "loss": 0.05, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 2.1221635464823237e-05, | |
| "loss": 0.0812, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 10.99, | |
| "learning_rate": 2.1126983371463795e-05, | |
| "loss": 0.0496, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 11.01, | |
| "learning_rate": 2.1032388126847967e-05, | |
| "loss": 0.0417, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 11.04, | |
| "learning_rate": 2.0937851119459548e-05, | |
| "loss": 0.0426, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 11.06, | |
| "learning_rate": 2.0843373736927504e-05, | |
| "loss": 0.0482, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 11.08, | |
| "learning_rate": 2.074895736600564e-05, | |
| "loss": 0.0399, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 2.0654603392552193e-05, | |
| "loss": 0.0334, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 11.13, | |
| "learning_rate": 2.0560313201509554e-05, | |
| "loss": 0.0506, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 11.16, | |
| "learning_rate": 2.0466088176883875e-05, | |
| "loss": 0.042, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 11.18, | |
| "learning_rate": 2.037192970172481e-05, | |
| "loss": 0.0464, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 11.21, | |
| "learning_rate": 2.027783915810518e-05, | |
| "loss": 0.0618, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 11.23, | |
| "learning_rate": 2.0183817927100683e-05, | |
| "loss": 0.0543, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 11.25, | |
| "learning_rate": 2.0089867388769666e-05, | |
| "loss": 0.046, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 11.28, | |
| "learning_rate": 1.9995988922132803e-05, | |
| "loss": 0.0532, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 11.3, | |
| "learning_rate": 1.9902183905152907e-05, | |
| "loss": 0.0452, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 11.33, | |
| "learning_rate": 1.980845371471469e-05, | |
| "loss": 0.0493, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 11.35, | |
| "learning_rate": 1.9714799726604537e-05, | |
| "loss": 0.0422, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 11.38, | |
| "learning_rate": 1.962122331549033e-05, | |
| "loss": 0.0492, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 1.9527725854901268e-05, | |
| "loss": 0.0578, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 11.43, | |
| "learning_rate": 1.9434308717207682e-05, | |
| "loss": 0.0574, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 11.45, | |
| "learning_rate": 1.9340973273600942e-05, | |
| "loss": 0.0348, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 11.47, | |
| "learning_rate": 1.9247720894073264e-05, | |
| "loss": 0.0606, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 11.5, | |
| "learning_rate": 1.9154552947397668e-05, | |
| "loss": 0.044, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 11.52, | |
| "learning_rate": 1.906147080110784e-05, | |
| "loss": 0.0497, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 11.55, | |
| "learning_rate": 1.8968475821478066e-05, | |
| "loss": 0.0604, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 11.57, | |
| "learning_rate": 1.8875569373503215e-05, | |
| "loss": 0.0497, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 1.8782752820878634e-05, | |
| "loss": 0.0315, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 11.62, | |
| "learning_rate": 1.8690027525980212e-05, | |
| "loss": 0.0475, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 11.64, | |
| "learning_rate": 1.8597394849844317e-05, | |
| "loss": 0.0521, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 11.67, | |
| "learning_rate": 1.8504856152147855e-05, | |
| "loss": 0.0616, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 11.69, | |
| "learning_rate": 1.8412412791188305e-05, | |
| "loss": 0.0409, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 11.72, | |
| "learning_rate": 1.8320066123863754e-05, | |
| "loss": 0.0435, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 11.74, | |
| "learning_rate": 1.8227817505653043e-05, | |
| "loss": 0.0452, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 11.77, | |
| "learning_rate": 1.8135668290595804e-05, | |
| "loss": 0.0313, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 11.79, | |
| "learning_rate": 1.8043619831272625e-05, | |
| "loss": 0.0592, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 11.81, | |
| "learning_rate": 1.795167347878518e-05, | |
| "loss": 0.0313, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 11.84, | |
| "learning_rate": 1.7859830582736404e-05, | |
| "loss": 0.0377, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 11.86, | |
| "learning_rate": 1.77680924912107e-05, | |
| "loss": 0.0661, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 1.7676460550754104e-05, | |
| "loss": 0.0356, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 11.91, | |
| "learning_rate": 1.758493610635457e-05, | |
| "loss": 0.0513, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "learning_rate": 1.74935205014222e-05, | |
| "loss": 0.0421, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 11.96, | |
| "learning_rate": 1.740221507776954e-05, | |
| "loss": 0.054, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 11.99, | |
| "learning_rate": 1.731102117559187e-05, | |
| "loss": 0.0571, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 12.01, | |
| "learning_rate": 1.7219940133447544e-05, | |
| "loss": 0.0428, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 12.03, | |
| "learning_rate": 1.7128973288238344e-05, | |
| "loss": 0.0234, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "learning_rate": 1.7047201869528476e-05, | |
| "loss": 0.0493, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 12.08, | |
| "learning_rate": 1.69564556756413e-05, | |
| "loss": 0.0273, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 12.11, | |
| "learning_rate": 1.6865827546155305e-05, | |
| "loss": 0.0274, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 12.13, | |
| "learning_rate": 1.6775318811324364e-05, | |
| "loss": 0.0227, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 12.16, | |
| "learning_rate": 1.668493079964982e-05, | |
| "loss": 0.0322, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 12.18, | |
| "learning_rate": 1.6594664837861045e-05, | |
| "loss": 0.0249, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 12.2, | |
| "learning_rate": 1.6504522250895954e-05, | |
| "loss": 0.0484, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 12.23, | |
| "learning_rate": 1.6414504361881512e-05, | |
| "loss": 0.0411, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 12.25, | |
| "learning_rate": 1.6324612492114378e-05, | |
| "loss": 0.0371, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 12.28, | |
| "learning_rate": 1.6234847961041465e-05, | |
| "loss": 0.03, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 1.6145212086240607e-05, | |
| "loss": 0.0254, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 12.33, | |
| "learning_rate": 1.6055706183401176e-05, | |
| "loss": 0.0338, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 12.35, | |
| "learning_rate": 1.596633156630481e-05, | |
| "loss": 0.0277, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 12.38, | |
| "learning_rate": 1.5877089546806125e-05, | |
| "loss": 0.0356, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 1.5787981434813416e-05, | |
| "loss": 0.0268, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 12.42, | |
| "learning_rate": 1.5699008538269494e-05, | |
| "loss": 0.0359, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 12.45, | |
| "learning_rate": 1.5610172163132445e-05, | |
| "loss": 0.0246, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 12.47, | |
| "learning_rate": 1.5521473613356476e-05, | |
| "loss": 0.0238, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "learning_rate": 1.5432914190872757e-05, | |
| "loss": 0.028, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 12.52, | |
| "learning_rate": 1.5344495195570348e-05, | |
| "loss": 0.0356, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 12.55, | |
| "learning_rate": 1.5256217925277087e-05, | |
| "loss": 0.0509, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 12.57, | |
| "learning_rate": 1.5168083675740539e-05, | |
| "loss": 0.036, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 12.59, | |
| "learning_rate": 1.5080093740609008e-05, | |
| "loss": 0.0336, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 1.4992249411412513e-05, | |
| "loss": 0.0289, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 12.64, | |
| "learning_rate": 1.4904551977543859e-05, | |
| "loss": 0.0364, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 12.67, | |
| "learning_rate": 1.4817002726239682e-05, | |
| "loss": 0.0348, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 12.69, | |
| "learning_rate": 1.4729602942561588e-05, | |
| "loss": 0.0381, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 12.72, | |
| "learning_rate": 1.4642353909377268e-05, | |
| "loss": 0.039, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 12.74, | |
| "learning_rate": 1.4555256907341667e-05, | |
| "loss": 0.027, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 12.76, | |
| "learning_rate": 1.4468313214878199e-05, | |
| "loss": 0.029, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 12.79, | |
| "learning_rate": 1.4381524108159989e-05, | |
| "loss": 0.0477, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 12.81, | |
| "learning_rate": 1.4294890861091087e-05, | |
| "loss": 0.0416, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 12.84, | |
| "learning_rate": 1.4208414745287862e-05, | |
| "loss": 0.048, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 12.86, | |
| "learning_rate": 1.4122097030060249e-05, | |
| "loss": 0.03, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 1.4035938982393176e-05, | |
| "loss": 0.0434, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 1.3949941866927946e-05, | |
| "loss": 0.0295, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 12.94, | |
| "learning_rate": 1.3864106945943642e-05, | |
| "loss": 0.0311, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 12.96, | |
| "learning_rate": 1.3778435479338666e-05, | |
| "loss": 0.0353, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 12.98, | |
| "learning_rate": 1.3692928724612203e-05, | |
| "loss": 0.028, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 13.01, | |
| "learning_rate": 1.3607587936845761e-05, | |
| "loss": 0.0362, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 13.03, | |
| "learning_rate": 1.3522414368684783e-05, | |
| "loss": 0.0285, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 13.06, | |
| "learning_rate": 1.3445902163436381e-05, | |
| "loss": 0.0253, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 13.08, | |
| "learning_rate": 1.3361049754758403e-05, | |
| "loss": 0.0245, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 1.3276368184411857e-05, | |
| "loss": 0.0239, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 13.13, | |
| "learning_rate": 1.3191858695366082e-05, | |
| "loss": 0.0212, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 13.15, | |
| "learning_rate": 1.3107522528064552e-05, | |
| "loss": 0.0249, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 13.18, | |
| "learning_rate": 1.302336092040673e-05, | |
| "loss": 0.027, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 13.2, | |
| "learning_rate": 1.2939375107729852e-05, | |
| "loss": 0.0196, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 13.23, | |
| "learning_rate": 1.2855566322790797e-05, | |
| "loss": 0.019, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 13.25, | |
| "learning_rate": 1.2771935795748036e-05, | |
| "loss": 0.024, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 13.28, | |
| "learning_rate": 1.2688484754143493e-05, | |
| "loss": 0.0243, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 13.3, | |
| "learning_rate": 1.2605214422884615e-05, | |
| "loss": 0.0232, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 13.33, | |
| "learning_rate": 1.2522126024226346e-05, | |
| "loss": 0.0226, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 13.35, | |
| "learning_rate": 1.2439220777753193e-05, | |
| "loss": 0.0311, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 13.37, | |
| "learning_rate": 1.2356499900361332e-05, | |
| "loss": 0.0239, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 13.4, | |
| "learning_rate": 1.2273964606240718e-05, | |
| "loss": 0.0221, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 13.42, | |
| "learning_rate": 1.2191616106857312e-05, | |
| "loss": 0.0203, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 13.45, | |
| "learning_rate": 1.2109455610935261e-05, | |
| "loss": 0.0257, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 13.47, | |
| "learning_rate": 1.2027484324439178e-05, | |
| "loss": 0.0298, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 13.5, | |
| "learning_rate": 1.194570345055641e-05, | |
| "loss": 0.0242, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 13.52, | |
| "learning_rate": 1.1864114189679412e-05, | |
| "loss": 0.0285, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 13.54, | |
| "learning_rate": 1.1782717739388116e-05, | |
| "loss": 0.025, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 13.57, | |
| "learning_rate": 1.1701515294432346e-05, | |
| "loss": 0.02, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 13.59, | |
| "learning_rate": 1.162050804671429e-05, | |
| "loss": 0.0232, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 13.62, | |
| "learning_rate": 1.1539697185270982e-05, | |
| "loss": 0.0245, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 13.64, | |
| "learning_rate": 1.1459083896256887e-05, | |
| "loss": 0.0226, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 13.67, | |
| "learning_rate": 1.1378669362926467e-05, | |
| "loss": 0.0344, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 13.69, | |
| "learning_rate": 1.1298454765616812e-05, | |
| "loss": 0.031, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "learning_rate": 1.1218441281730332e-05, | |
| "loss": 0.029, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 13.74, | |
| "learning_rate": 1.1138630085717438e-05, | |
| "loss": 0.0288, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 13.76, | |
| "learning_rate": 1.1059022349059361e-05, | |
| "loss": 0.0358, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 13.79, | |
| "learning_rate": 1.0979619240250888e-05, | |
| "loss": 0.0236, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 13.81, | |
| "learning_rate": 1.0900421924783271e-05, | |
| "loss": 0.0383, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 13.84, | |
| "learning_rate": 1.0821431565127102e-05, | |
| "loss": 0.0238, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 13.86, | |
| "learning_rate": 1.0742649320715209e-05, | |
| "loss": 0.0224, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 13.89, | |
| "learning_rate": 1.0664076347925703e-05, | |
| "loss": 0.0243, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 13.91, | |
| "learning_rate": 1.0585713800064963e-05, | |
| "loss": 0.0267, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 13.93, | |
| "learning_rate": 1.0507562827350719e-05, | |
| "loss": 0.0278, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 1.0429624576895179e-05, | |
| "loss": 0.0252, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 13.98, | |
| "learning_rate": 1.0351900192688144e-05, | |
| "loss": 0.0214, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 14.01, | |
| "learning_rate": 1.027439081558029e-05, | |
| "loss": 0.0271, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 14.03, | |
| "learning_rate": 1.0197097583266374e-05, | |
| "loss": 0.0194, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 14.06, | |
| "learning_rate": 1.0120021630268541e-05, | |
| "loss": 0.0247, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 14.08, | |
| "learning_rate": 1.0043164087919696e-05, | |
| "loss": 0.0166, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 14.1, | |
| "learning_rate": 9.966526084346836e-06, | |
| "loss": 0.025, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 14.13, | |
| "learning_rate": 9.890108744454573e-06, | |
| "loss": 0.0262, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 14.15, | |
| "learning_rate": 9.813913189908571e-06, | |
| "loss": 0.0163, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 14.18, | |
| "learning_rate": 9.73794053911909e-06, | |
| "loss": 0.0217, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 9.662191907224583e-06, | |
| "loss": 0.0171, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 14.23, | |
| "learning_rate": 9.5866684060753e-06, | |
| "loss": 0.0161, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 14.25, | |
| "learning_rate": 9.511371144217004e-06, | |
| "loss": 0.0265, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 14.28, | |
| "learning_rate": 9.436301226874677e-06, | |
| "loss": 0.0172, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 14.3, | |
| "learning_rate": 9.361459755936317e-06, | |
| "loss": 0.0179, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 14.32, | |
| "learning_rate": 9.28684782993671e-06, | |
| "loss": 0.0211, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 14.35, | |
| "learning_rate": 9.212466544041384e-06, | |
| "loss": 0.0221, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 14.37, | |
| "learning_rate": 9.138316990030485e-06, | |
| "loss": 0.0175, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 14.4, | |
| "learning_rate": 9.064400256282757e-06, | |
| "loss": 0.0177, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 14.42, | |
| "learning_rate": 8.990717427759585e-06, | |
| "loss": 0.0168, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 14.45, | |
| "learning_rate": 8.917269585989026e-06, | |
| "loss": 0.0242, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 14.47, | |
| "learning_rate": 8.844057809049999e-06, | |
| "loss": 0.0249, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 14.49, | |
| "learning_rate": 8.771083171556408e-06, | |
| "loss": 0.0158, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 14.52, | |
| "learning_rate": 8.698346744641392e-06, | |
| "loss": 0.0204, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 14.54, | |
| "learning_rate": 8.625849595941608e-06, | |
| "loss": 0.0225, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 14.57, | |
| "learning_rate": 8.553592789581518e-06, | |
| "loss": 0.0158, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 14.59, | |
| "learning_rate": 8.48157738615784e-06, | |
| "loss": 0.029, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 14.62, | |
| "learning_rate": 8.409804442723918e-06, | |
| "loss": 0.0187, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 14.64, | |
| "learning_rate": 8.338275012774247e-06, | |
| "loss": 0.0211, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 14.67, | |
| "learning_rate": 8.266990146228964e-06, | |
| "loss": 0.0184, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 14.69, | |
| "learning_rate": 8.195950889418502e-06, | |
| "loss": 0.0167, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 14.71, | |
| "learning_rate": 8.12515828506817e-06, | |
| "loss": 0.0199, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 14.74, | |
| "learning_rate": 8.054613372282891e-06, | |
| "loss": 0.0184, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 14.76, | |
| "learning_rate": 7.984317186531928e-06, | |
| "loss": 0.0196, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 14.79, | |
| "learning_rate": 7.914270759633668e-06, | |
| "loss": 0.0185, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 14.81, | |
| "learning_rate": 7.84447511974053e-06, | |
| "loss": 0.0178, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 14.84, | |
| "learning_rate": 7.774931291323825e-06, | |
| "loss": 0.022, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 14.86, | |
| "learning_rate": 7.705640295158741e-06, | |
| "loss": 0.0201, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 14.88, | |
| "learning_rate": 7.636603148309363e-06, | |
| "loss": 0.0221, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 14.91, | |
| "learning_rate": 7.567820864113706e-06, | |
| "loss": 0.0212, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 14.93, | |
| "learning_rate": 7.499294452168904e-06, | |
| "loss": 0.0174, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 14.96, | |
| "learning_rate": 7.431024918316343e-06, | |
| "loss": 0.0275, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "learning_rate": 7.363013264626914e-06, | |
| "loss": 0.022, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 15.01, | |
| "learning_rate": 7.295260489386313e-06, | |
| "loss": 0.0215, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 15.03, | |
| "learning_rate": 7.22776758708035e-06, | |
| "loss": 0.0165, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 15.05, | |
| "learning_rate": 7.160535548380409e-06, | |
| "loss": 0.0145, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 15.08, | |
| "learning_rate": 7.093565360128862e-06, | |
| "loss": 0.0144, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 15.1, | |
| "learning_rate": 7.026858005324616e-06, | |
| "loss": 0.0173, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 15.13, | |
| "learning_rate": 6.96041446310863e-06, | |
| "loss": 0.0141, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 15.15, | |
| "learning_rate": 6.894235708749622e-06, | |
| "loss": 0.013, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 15.18, | |
| "learning_rate": 6.828322713629689e-06, | |
| "loss": 0.0174, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 15.2, | |
| "learning_rate": 6.762676445230084e-06, | |
| "loss": 0.0159, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 15.23, | |
| "learning_rate": 6.697297867117e-06, | |
| "loss": 0.016, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 15.25, | |
| "learning_rate": 6.63218793892742e-06, | |
| "loss": 0.0182, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 15.27, | |
| "learning_rate": 6.567347616355049e-06, | |
| "loss": 0.0157, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 15.3, | |
| "learning_rate": 6.502777851136288e-06, | |
| "loss": 0.0158, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 15.32, | |
| "learning_rate": 6.438479591036239e-06, | |
| "loss": 0.0147, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 15.35, | |
| "learning_rate": 6.37445377983483e-06, | |
| "loss": 0.0171, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 15.37, | |
| "learning_rate": 6.310701357312909e-06, | |
| "loss": 0.0189, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 15.4, | |
| "learning_rate": 6.247223259238511e-06, | |
| "loss": 0.0153, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 15.42, | |
| "learning_rate": 6.184020417353084e-06, | |
| "loss": 0.0179, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 15.44, | |
| "learning_rate": 6.121093759357824e-06, | |
| "loss": 0.0166, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 15.47, | |
| "learning_rate": 6.05844420890006e-06, | |
| "loss": 0.0226, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 15.49, | |
| "learning_rate": 5.996072685559675e-06, | |
| "loss": 0.0184, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 15.52, | |
| "learning_rate": 5.9339801048356515e-06, | |
| "loss": 0.0162, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 15.54, | |
| "learning_rate": 5.87216737813259e-06, | |
| "loss": 0.0177, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 15.57, | |
| "learning_rate": 5.810635412747373e-06, | |
| "loss": 0.0162, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 15.59, | |
| "learning_rate": 5.749385111855787e-06, | |
| "loss": 0.017, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 15.62, | |
| "learning_rate": 5.688417374499335e-06, | |
| "loss": 0.0165, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 15.64, | |
| "learning_rate": 5.627733095572002e-06, | |
| "loss": 0.0225, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 15.66, | |
| "learning_rate": 5.567333165807115e-06, | |
| "loss": 0.018, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 15.69, | |
| "learning_rate": 5.5072184717643e-06, | |
| "loss": 0.0156, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 15.71, | |
| "learning_rate": 5.447389895816415e-06, | |
| "loss": 0.017, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 15.74, | |
| "learning_rate": 5.387848316136665e-06, | |
| "loss": 0.0167, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 15.76, | |
| "learning_rate": 5.328594606685661e-06, | |
| "loss": 0.0149, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 15.79, | |
| "learning_rate": 5.269629637198617e-06, | |
| "loss": 0.0172, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 15.81, | |
| "learning_rate": 5.210954273172577e-06, | |
| "loss": 0.025, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 15.83, | |
| "learning_rate": 5.1525693758537025e-06, | |
| "loss": 0.0209, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 15.86, | |
| "learning_rate": 5.0944758022246435e-06, | |
| "loss": 0.0237, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 15.88, | |
| "learning_rate": 5.0366744049919614e-06, | |
| "loss": 0.0244, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 15.91, | |
| "learning_rate": 4.979166032573607e-06, | |
| "loss": 0.0161, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 15.93, | |
| "learning_rate": 4.921951529086447e-06, | |
| "loss": 0.0197, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 15.96, | |
| "learning_rate": 4.865031734333919e-06, | |
| "loss": 0.0206, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 15.98, | |
| "learning_rate": 4.808407483793667e-06, | |
| "loss": 0.0169, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 4.752079608605295e-06, | |
| "loss": 0.0167, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 16.03, | |
| "learning_rate": 4.696048935558167e-06, | |
| "loss": 0.0167, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 16.05, | |
| "learning_rate": 4.640316287079252e-06, | |
| "loss": 0.0116, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 16.08, | |
| "learning_rate": 4.584882481221081e-06, | |
| "loss": 0.0141, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 16.1, | |
| "learning_rate": 4.529748331649728e-06, | |
| "loss": 0.014, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 16.13, | |
| "learning_rate": 4.474914647632855e-06, | |
| "loss": 0.0145, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 16.15, | |
| "learning_rate": 4.420382234027859e-06, | |
| "loss": 0.0246, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 16.18, | |
| "learning_rate": 4.366151891270018e-06, | |
| "loss": 0.0137, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 16.2, | |
| "learning_rate": 4.312224415360791e-06, | |
| "loss": 0.0146, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 16.22, | |
| "learning_rate": 4.258600597856099e-06, | |
| "loss": 0.0143, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 16.25, | |
| "learning_rate": 4.205281225854727e-06, | |
| "loss": 0.0139, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 16.27, | |
| "learning_rate": 4.152267081986741e-06, | |
| "loss": 0.014, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 16.3, | |
| "learning_rate": 4.099558944402043e-06, | |
| "loss": 0.0135, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 16.32, | |
| "learning_rate": 4.047157586758907e-06, | |
| "loss": 0.0153, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 16.35, | |
| "learning_rate": 3.99506377821266e-06, | |
| "loss": 0.0145, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 16.37, | |
| "learning_rate": 3.943278283404369e-06, | |
| "loss": 0.0184, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 16.39, | |
| "learning_rate": 3.891801862449629e-06, | |
| "loss": 0.0153, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 16.42, | |
| "learning_rate": 3.840635270927381e-06, | |
| "loss": 0.0151, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 16.44, | |
| "learning_rate": 3.7897792598688643e-06, | |
| "loss": 0.0127, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 16.47, | |
| "learning_rate": 3.7392345757465586e-06, | |
| "loss": 0.0199, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 16.49, | |
| "learning_rate": 3.6890019604632424e-06, | |
| "loss": 0.0148, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 16.52, | |
| "learning_rate": 3.6390821513410974e-06, | |
| "loss": 0.0133, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 16.54, | |
| "learning_rate": 3.5894758811108793e-06, | |
| "loss": 0.014, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 16.57, | |
| "learning_rate": 3.540183877901182e-06, | |
| "loss": 0.0166, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 16.59, | |
| "learning_rate": 3.491206865227739e-06, | |
| "loss": 0.0166, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 16.61, | |
| "learning_rate": 3.4425455619827963e-06, | |
| "loss": 0.0163, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 16.64, | |
| "learning_rate": 3.394200682424578e-06, | |
| "loss": 0.0185, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 16.66, | |
| "learning_rate": 3.3461729361667722e-06, | |
| "loss": 0.015, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 16.69, | |
| "learning_rate": 3.2984630281681555e-06, | |
| "loss": 0.0171, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 16.71, | |
| "learning_rate": 3.251071658722213e-06, | |
| "loss": 0.016, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 16.74, | |
| "learning_rate": 3.203999523446885e-06, | |
| "loss": 0.02, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "learning_rate": 3.1572473132743222e-06, | |
| "loss": 0.0132, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 16.78, | |
| "learning_rate": 3.1108157144407764e-06, | |
| "loss": 0.0157, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 16.81, | |
| "learning_rate": 3.0647054084765176e-06, | |
| "loss": 0.0149, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 16.83, | |
| "learning_rate": 3.018917072195823e-06, | |
| "loss": 0.0138, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 16.86, | |
| "learning_rate": 2.9734513776870616e-06, | |
| "loss": 0.0146, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 16.88, | |
| "learning_rate": 2.928308992302792e-06, | |
| "loss": 0.0188, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 16.91, | |
| "learning_rate": 2.8834905786500083e-06, | |
| "loss": 0.023, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 16.93, | |
| "learning_rate": 2.838996794580398e-06, | |
| "loss": 0.0158, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 16.95, | |
| "learning_rate": 2.7948282931806793e-06, | |
| "loss": 0.0142, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 16.98, | |
| "learning_rate": 2.750985722763022e-06, | |
| "loss": 0.0168, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 2.7074697268555244e-06, | |
| "loss": 0.0161, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 17.03, | |
| "learning_rate": 2.6642809441927817e-06, | |
| "loss": 0.0132, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 17.05, | |
| "learning_rate": 2.621420008706499e-06, | |
| "loss": 0.0131, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 17.08, | |
| "learning_rate": 2.5788875495161845e-06, | |
| "loss": 0.0122, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 17.1, | |
| "learning_rate": 2.5366841909199356e-06, | |
| "loss": 0.0142, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 17.13, | |
| "learning_rate": 2.4948105523852323e-06, | |
| "loss": 0.0234, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 17.15, | |
| "learning_rate": 2.4532672485398976e-06, | |
| "loss": 0.014, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 17.17, | |
| "learning_rate": 2.4120548891630347e-06, | |
| "loss": 0.0154, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 17.2, | |
| "learning_rate": 2.3711740791761083e-06, | |
| "loss": 0.0132, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 17.22, | |
| "learning_rate": 2.3306254186340303e-06, | |
| "loss": 0.013, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 17.25, | |
| "learning_rate": 2.2904095027163786e-06, | |
| "loss": 0.0242, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 17.27, | |
| "learning_rate": 2.2505269217186596e-06, | |
| "loss": 0.0143, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 17.3, | |
| "learning_rate": 2.210978261043631e-06, | |
| "loss": 0.0121, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 17.32, | |
| "learning_rate": 2.171764101192722e-06, | |
| "loss": 0.0138, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 17.34, | |
| "learning_rate": 2.1328850177574936e-06, | |
| "loss": 0.0133, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 17.37, | |
| "learning_rate": 2.0943415814112162e-06, | |
| "loss": 0.0125, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 17.39, | |
| "learning_rate": 2.0561343579004715e-06, | |
| "loss": 0.0145, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 17.42, | |
| "learning_rate": 2.018263908036863e-06, | |
| "loss": 0.0148, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 17.44, | |
| "learning_rate": 1.980730787688781e-06, | |
| "loss": 0.0133, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 17.47, | |
| "learning_rate": 1.9435355477732204e-06, | |
| "loss": 0.0129, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 17.49, | |
| "learning_rate": 1.9066787342477354e-06, | |
| "loss": 0.0147, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 17.52, | |
| "learning_rate": 1.8701608881023958e-06, | |
| "loss": 0.0154, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 17.54, | |
| "learning_rate": 1.8339825453518645e-06, | |
| "loss": 0.0134, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 17.56, | |
| "learning_rate": 1.7981442370274992e-06, | |
| "loss": 0.013, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 17.59, | |
| "learning_rate": 1.7626464891696031e-06, | |
| "loss": 0.0166, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 17.61, | |
| "learning_rate": 1.727489822819664e-06, | |
| "loss": 0.0122, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 17.64, | |
| "learning_rate": 1.6926747540127297e-06, | |
| "loss": 0.0139, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 17.66, | |
| "learning_rate": 1.6582017937698285e-06, | |
| "loss": 0.0141, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 17.69, | |
| "learning_rate": 1.62407144809045e-06, | |
| "loss": 0.0141, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 17.71, | |
| "learning_rate": 1.5902842179451482e-06, | |
| "loss": 0.0133, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 17.73, | |
| "learning_rate": 1.5568405992681722e-06, | |
| "loss": 0.015, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 17.76, | |
| "learning_rate": 1.5237410829501864e-06, | |
| "loss": 0.015, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "learning_rate": 1.4909861548310672e-06, | |
| "loss": 0.0195, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 17.81, | |
| "learning_rate": 1.4585762956927624e-06, | |
| "loss": 0.0165, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 17.83, | |
| "learning_rate": 1.4265119812522597e-06, | |
| "loss": 0.0145, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 17.86, | |
| "learning_rate": 1.394793682154577e-06, | |
| "loss": 0.0125, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 17.88, | |
| "learning_rate": 1.363421863965869e-06, | |
| "loss": 0.0139, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 17.9, | |
| "learning_rate": 1.3323969871665897e-06, | |
| "loss": 0.015, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 17.93, | |
| "learning_rate": 1.3017195071447246e-06, | |
| "loss": 0.0149, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 17.95, | |
| "learning_rate": 1.2713898741891244e-06, | |
| "loss": 0.0184, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 17.98, | |
| "learning_rate": 1.2414085334828878e-06, | |
| "loss": 0.0132, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "learning_rate": 1.2117759250968224e-06, | |
| "loss": 0.0149, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 18.03, | |
| "learning_rate": 1.1824924839829776e-06, | |
| "loss": 0.0134, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 18.05, | |
| "learning_rate": 1.1535586399682885e-06, | |
| "loss": 0.0126, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 18.08, | |
| "learning_rate": 1.1249748177482366e-06, | |
| "loss": 0.0137, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 18.1, | |
| "learning_rate": 1.0967414368806383e-06, | |
| "loss": 0.0134, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 18.12, | |
| "learning_rate": 1.0688589117794717e-06, | |
| "loss": 0.0129, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 18.15, | |
| "learning_rate": 1.0413276517087956e-06, | |
| "loss": 0.0121, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 18.17, | |
| "learning_rate": 1.0141480607767513e-06, | |
| "loss": 0.0159, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 18.2, | |
| "learning_rate": 9.87320537929623e-07, | |
| "loss": 0.0141, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 18.22, | |
| "learning_rate": 9.6084547694599e-07, | |
| "loss": 0.0122, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 18.25, | |
| "learning_rate": 9.347232664309368e-07, | |
| "loss": 0.013, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 18.27, | |
| "learning_rate": 9.089542898103459e-07, | |
| "loss": 0.0117, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 18.29, | |
| "learning_rate": 8.835389253252918e-07, | |
| "loss": 0.0143, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 18.32, | |
| "learning_rate": 8.584775460264621e-07, | |
| "loss": 0.0122, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 18.34, | |
| "learning_rate": 8.33770519768709e-07, | |
| "loss": 0.0122, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 18.37, | |
| "learning_rate": 8.09418209205623e-07, | |
| "loss": 0.0135, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 18.39, | |
| "learning_rate": 7.854209717842231e-07, | |
| "loss": 0.0141, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 18.42, | |
| "learning_rate": 7.617791597397117e-07, | |
| "loss": 0.0109, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 18.44, | |
| "learning_rate": 7.384931200903084e-07, | |
| "loss": 0.0124, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 18.47, | |
| "learning_rate": 7.155631946321406e-07, | |
| "loss": 0.012, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 18.49, | |
| "learning_rate": 6.929897199342395e-07, | |
| "loss": 0.0136, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 18.51, | |
| "learning_rate": 6.707730273335932e-07, | |
| "loss": 0.0159, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 18.54, | |
| "learning_rate": 6.489134429302907e-07, | |
| "loss": 0.016, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 18.56, | |
| "learning_rate": 6.274112875827326e-07, | |
| "loss": 0.012, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 18.59, | |
| "learning_rate": 6.062668769029167e-07, | |
| "loss": 0.0124, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 18.61, | |
| "learning_rate": 5.854805212518022e-07, | |
| "loss": 0.0205, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 18.64, | |
| "learning_rate": 5.650525257347744e-07, | |
| "loss": 0.0136, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 18.66, | |
| "learning_rate": 5.449831901971431e-07, | |
| "loss": 0.0133, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 18.68, | |
| "learning_rate": 5.252728092197539e-07, | |
| "loss": 0.015, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 18.71, | |
| "learning_rate": 5.059216721146592e-07, | |
| "loss": 0.0134, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 18.73, | |
| "learning_rate": 4.869300629208762e-07, | |
| "loss": 0.0125, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 18.76, | |
| "learning_rate": 4.6829826040021287e-07, | |
| "loss": 0.0141, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 18.78, | |
| "learning_rate": 4.500265380331797e-07, | |
| "loss": 0.0158, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 18.81, | |
| "learning_rate": 4.3211516401497344e-07, | |
| "loss": 0.013, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 18.83, | |
| "learning_rate": 4.145644012515465e-07, | |
| "loss": 0.0131, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 18.86, | |
| "learning_rate": 3.973745073557328e-07, | |
| "loss": 0.0152, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 18.88, | |
| "learning_rate": 3.805457346434865e-07, | |
| "loss": 0.0138, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 18.9, | |
| "learning_rate": 3.640783301301631e-07, | |
| "loss": 0.0147, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 18.93, | |
| "learning_rate": 3.479725355268998e-07, | |
| "loss": 0.0154, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 18.95, | |
| "learning_rate": 3.3222858723707395e-07, | |
| "loss": 0.0125, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 18.98, | |
| "learning_rate": 3.168467163528116e-07, | |
| "loss": 0.0236, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "learning_rate": 3.018271486516233e-07, | |
| "loss": 0.0128, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 19.03, | |
| "learning_rate": 2.871701045930708e-07, | |
| "loss": 0.0168, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 19.05, | |
| "learning_rate": 2.72875799315539e-07, | |
| "loss": 0.0201, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 19.07, | |
| "learning_rate": 2.589444426330773e-07, | |
| "loss": 0.0121, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 19.1, | |
| "learning_rate": 2.453762390323133e-07, | |
| "loss": 0.0128, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 19.12, | |
| "learning_rate": 2.3217138766946366e-07, | |
| "loss": 0.0146, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 19.15, | |
| "learning_rate": 2.1933008236739993e-07, | |
| "loss": 0.0132, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 19.17, | |
| "learning_rate": 2.068525116128095e-07, | |
| "loss": 0.012, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 19.2, | |
| "learning_rate": 1.9473885855342822e-07, | |
| "loss": 0.012, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 19.22, | |
| "learning_rate": 1.8298930099534818e-07, | |
| "loss": 0.0162, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 19.24, | |
| "learning_rate": 1.716040114004114e-07, | |
| "loss": 0.0117, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 19.27, | |
| "learning_rate": 1.6058315688367854e-07, | |
| "loss": 0.0135, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 19.29, | |
| "learning_rate": 1.4992689921097814e-07, | |
| "loss": 0.0135, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 19.32, | |
| "learning_rate": 1.396353947965251e-07, | |
| "loss": 0.0144, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 19.34, | |
| "learning_rate": 1.2970879470062813e-07, | |
| "loss": 0.0128, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 19.37, | |
| "learning_rate": 1.2014724462747762e-07, | |
| "loss": 0.0121, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 19.39, | |
| "learning_rate": 1.109508849230001e-07, | |
| "loss": 0.0132, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 19.42, | |
| "learning_rate": 1.021198505728016e-07, | |
| "loss": 0.0124, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 19.44, | |
| "learning_rate": 9.365427120018311e-08, | |
| "loss": 0.0141, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 19.46, | |
| "learning_rate": 8.555427106424485e-08, | |
| "loss": 0.0154, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 19.49, | |
| "learning_rate": 7.781996905805167e-08, | |
| "loss": 0.0124, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 19.51, | |
| "learning_rate": 7.045147870690105e-08, | |
| "loss": 0.0118, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 19.54, | |
| "learning_rate": 6.344890816664673e-08, | |
| "loss": 0.0137, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 19.56, | |
| "learning_rate": 5.681236022211378e-08, | |
| "loss": 0.0134, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 19.59, | |
| "learning_rate": 5.0541932285586014e-08, | |
| "loss": 0.0123, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 19.61, | |
| "learning_rate": 4.463771639539038e-08, | |
| "loss": 0.0129, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 19.63, | |
| "learning_rate": 3.909979921452867e-08, | |
| "loss": 0.0132, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 19.66, | |
| "learning_rate": 3.392826202941179e-08, | |
| "loss": 0.0116, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 19.68, | |
| "learning_rate": 2.912318074867748e-08, | |
| "loss": 0.0134, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 19.71, | |
| "learning_rate": 2.4684625902057778e-08, | |
| "loss": 0.0115, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 19.73, | |
| "learning_rate": 2.061266263935213e-08, | |
| "loss": 0.014, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 19.76, | |
| "learning_rate": 1.6907350729478134e-08, | |
| "loss": 0.0131, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 19.78, | |
| "learning_rate": 1.356874455958057e-08, | |
| "loss": 0.0145, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 19.81, | |
| "learning_rate": 1.0596893134240393e-08, | |
| "loss": 0.0145, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 19.83, | |
| "learning_rate": 7.991840074764167e-09, | |
| "loss": 0.0131, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 19.85, | |
| "learning_rate": 5.75362361852072e-09, | |
| "loss": 0.0126, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 19.88, | |
| "learning_rate": 3.882276618405456e-09, | |
| "loss": 0.0131, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 19.9, | |
| "learning_rate": 2.377826542343531e-09, | |
| "loss": 0.0149, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 19.93, | |
| "learning_rate": 1.2402954728846228e-09, | |
| "loss": 0.013, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 19.95, | |
| "learning_rate": 4.697001068892925e-10, | |
| "loss": 0.0127, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 19.98, | |
| "learning_rate": 6.605175527640839e-11, | |
| "loss": 0.0116, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 19.98, | |
| "step": 8200, | |
| "total_flos": 3.0532626578313216e+17, | |
| "train_loss": 0.2642190715797791, | |
| "train_runtime": 4480.4571, | |
| "train_samples_per_second": 29.305, | |
| "train_steps_per_second": 1.83 | |
| } | |
| ], | |
| "max_steps": 8200, | |
| "num_train_epochs": 20, | |
| "total_flos": 3.0532626578313216e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |