| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "global_step": 44361, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6934, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1e-05, | |
| "loss": 0.5953, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.5041, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2e-05, | |
| "loss": 0.5038, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.4902, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3e-05, | |
| "loss": 0.4758, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.5029, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4e-05, | |
| "loss": 0.4621, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.477, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5e-05, | |
| "loss": 0.4989, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.994300175554593e-05, | |
| "loss": 0.4836, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.988600351109186e-05, | |
| "loss": 0.4955, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.982900526663779e-05, | |
| "loss": 0.4802, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.977200702218372e-05, | |
| "loss": 0.4741, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.971500877772965e-05, | |
| "loss": 0.4728, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.965801053327558e-05, | |
| "loss": 0.4439, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9601012288821506e-05, | |
| "loss": 0.4795, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.954401404436744e-05, | |
| "loss": 0.4612, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.948701579991337e-05, | |
| "loss": 0.4623, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9430017555459296e-05, | |
| "loss": 0.4743, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9373019311005225e-05, | |
| "loss": 0.434, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.931602106655115e-05, | |
| "loss": 0.461, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.925902282209708e-05, | |
| "loss": 0.4438, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9202024577643015e-05, | |
| "loss": 0.4663, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9145026333188944e-05, | |
| "loss": 0.4321, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.908802808873487e-05, | |
| "loss": 0.4614, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.903102984428079e-05, | |
| "loss": 0.4656, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.897403159982673e-05, | |
| "loss": 0.4643, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8917033355372656e-05, | |
| "loss": 0.4566, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8860035110918584e-05, | |
| "loss": 0.4545, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.880303686646451e-05, | |
| "loss": 0.4477, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.874603862201044e-05, | |
| "loss": 0.4497, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.8689040377556375e-05, | |
| "loss": 0.4629, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.86320421331023e-05, | |
| "loss": 0.4539, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.857504388864823e-05, | |
| "loss": 0.4724, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.851804564419416e-05, | |
| "loss": 0.4619, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.846104739974009e-05, | |
| "loss": 0.444, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.840404915528602e-05, | |
| "loss": 0.4532, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.834705091083195e-05, | |
| "loss": 0.4819, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.829005266637788e-05, | |
| "loss": 0.4672, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.8233054421923806e-05, | |
| "loss": 0.4373, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.8176056177469734e-05, | |
| "loss": 0.4425, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.811905793301567e-05, | |
| "loss": 0.442, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.80620596885616e-05, | |
| "loss": 0.4365, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.8005061444107525e-05, | |
| "loss": 0.4481, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.794806319965345e-05, | |
| "loss": 0.4639, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.789106495519938e-05, | |
| "loss": 0.4527, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.783406671074531e-05, | |
| "loss": 0.4575, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.7777068466291244e-05, | |
| "loss": 0.4626, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.772007022183717e-05, | |
| "loss": 0.4235, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.76630719773831e-05, | |
| "loss": 0.4496, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.760607373292903e-05, | |
| "loss": 0.4355, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.7549075488474956e-05, | |
| "loss": 0.4361, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.749207724402089e-05, | |
| "loss": 0.4655, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.743507899956682e-05, | |
| "loss": 0.4295, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.737808075511275e-05, | |
| "loss": 0.4567, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.7321082510658675e-05, | |
| "loss": 0.4501, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.72640842662046e-05, | |
| "loss": 0.4518, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.720708602175054e-05, | |
| "loss": 0.4625, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.715008777729646e-05, | |
| "loss": 0.4563, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.709308953284239e-05, | |
| "loss": 0.4593, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.7036091288388315e-05, | |
| "loss": 0.4316, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.697909304393425e-05, | |
| "loss": 0.4465, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.692209479948018e-05, | |
| "loss": 0.4427, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6865096555026106e-05, | |
| "loss": 0.4623, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.6808098310572034e-05, | |
| "loss": 0.4505, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.675110006611796e-05, | |
| "loss": 0.446, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.669410182166389e-05, | |
| "loss": 0.4549, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.6637103577209825e-05, | |
| "loss": 0.4467, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.658010533275575e-05, | |
| "loss": 0.4508, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.652310708830168e-05, | |
| "loss": 0.4378, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.646610884384761e-05, | |
| "loss": 0.4326, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.640911059939354e-05, | |
| "loss": 0.4312, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.635211235493947e-05, | |
| "loss": 0.4426, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.62951141104854e-05, | |
| "loss": 0.4246, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.623811586603133e-05, | |
| "loss": 0.4327, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.6181117621577257e-05, | |
| "loss": 0.4526, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.6124119377123185e-05, | |
| "loss": 0.4427, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.606712113266912e-05, | |
| "loss": 0.45, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.601012288821505e-05, | |
| "loss": 0.4537, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.5953124643760975e-05, | |
| "loss": 0.4487, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.5896126399306904e-05, | |
| "loss": 0.4305, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.583912815485283e-05, | |
| "loss": 0.4434, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.5782129910398766e-05, | |
| "loss": 0.4619, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.5725131665944694e-05, | |
| "loss": 0.4481, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.566813342149062e-05, | |
| "loss": 0.4493, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.561113517703655e-05, | |
| "loss": 0.4438, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.555413693258248e-05, | |
| "loss": 0.4334, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.549713868812841e-05, | |
| "loss": 0.433, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.544014044367434e-05, | |
| "loss": 0.4493, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.538314219922027e-05, | |
| "loss": 0.449, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.53261439547662e-05, | |
| "loss": 0.4464, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.526914571031212e-05, | |
| "loss": 0.4241, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.5212147465858054e-05, | |
| "loss": 0.4091, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.515514922140398e-05, | |
| "loss": 0.4402, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.509815097694991e-05, | |
| "loss": 0.4245, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.504115273249584e-05, | |
| "loss": 0.4507, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.4984154488041766e-05, | |
| "loss": 0.4107, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.49271562435877e-05, | |
| "loss": 0.4454, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.487015799913363e-05, | |
| "loss": 0.4451, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.481315975467956e-05, | |
| "loss": 0.4273, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.4756161510225485e-05, | |
| "loss": 0.4523, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.469916326577141e-05, | |
| "loss": 0.4469, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.464216502131735e-05, | |
| "loss": 0.4264, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.4585166776863276e-05, | |
| "loss": 0.4442, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.4528168532409204e-05, | |
| "loss": 0.4361, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.447117028795513e-05, | |
| "loss": 0.4309, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.441417204350106e-05, | |
| "loss": 0.4252, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.4357173799046995e-05, | |
| "loss": 0.4442, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.430017555459292e-05, | |
| "loss": 0.4425, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.424317731013885e-05, | |
| "loss": 0.4469, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.418617906568478e-05, | |
| "loss": 0.4222, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.412918082123071e-05, | |
| "loss": 0.4431, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.4072182576776635e-05, | |
| "loss": 0.4301, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.401518433232257e-05, | |
| "loss": 0.4565, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.39581860878685e-05, | |
| "loss": 0.4332, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.3901187843414426e-05, | |
| "loss": 0.4352, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.3844189598960354e-05, | |
| "loss": 0.4307, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.378719135450628e-05, | |
| "loss": 0.4603, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.373019311005222e-05, | |
| "loss": 0.4314, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.3673194865598145e-05, | |
| "loss": 0.4299, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.361619662114407e-05, | |
| "loss": 0.4427, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.355919837669e-05, | |
| "loss": 0.4358, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.350220013223593e-05, | |
| "loss": 0.4483, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.344520188778186e-05, | |
| "loss": 0.4326, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.3388203643327785e-05, | |
| "loss": 0.4367, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.333120539887371e-05, | |
| "loss": 0.4305, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.327420715441964e-05, | |
| "loss": 0.4274, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.3217208909965576e-05, | |
| "loss": 0.4265, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.3160210665511504e-05, | |
| "loss": 0.4519, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.310321242105743e-05, | |
| "loss": 0.4335, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.304621417660336e-05, | |
| "loss": 0.448, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.298921593214929e-05, | |
| "loss": 0.4098, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.2932217687695217e-05, | |
| "loss": 0.45, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.287521944324115e-05, | |
| "loss": 0.4218, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.281822119878708e-05, | |
| "loss": 0.446, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.276122295433301e-05, | |
| "loss": 0.4319, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.2704224709878936e-05, | |
| "loss": 0.4232, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.2647226465424864e-05, | |
| "loss": 0.4258, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.25902282209708e-05, | |
| "loss": 0.4277, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.2533229976516726e-05, | |
| "loss": 0.4264, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.2476231732062655e-05, | |
| "loss": 0.4419, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.241923348760858e-05, | |
| "loss": 0.4369, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.236223524315451e-05, | |
| "loss": 0.4532, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.2305236998700445e-05, | |
| "loss": 0.4205, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.2248238754246373e-05, | |
| "loss": 0.406, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.21912405097923e-05, | |
| "loss": 0.4436, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.213424226533823e-05, | |
| "loss": 0.4443, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.207724402088416e-05, | |
| "loss": 0.4353, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.202024577643009e-05, | |
| "loss": 0.4181, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.196324753197602e-05, | |
| "loss": 0.4202, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.190624928752195e-05, | |
| "loss": 0.4327, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.184925104306788e-05, | |
| "loss": 0.4169, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.1792252798613805e-05, | |
| "loss": 0.4559, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.173525455415974e-05, | |
| "loss": 0.4114, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.167825630970567e-05, | |
| "loss": 0.4371, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.1621258065251596e-05, | |
| "loss": 0.4198, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.156425982079752e-05, | |
| "loss": 0.4338, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.1507261576343445e-05, | |
| "loss": 0.4651, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.145026333188938e-05, | |
| "loss": 0.4341, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.139326508743531e-05, | |
| "loss": 0.4449, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.1336266842981236e-05, | |
| "loss": 0.4415, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.1279268598527164e-05, | |
| "loss": 0.4214, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.122227035407309e-05, | |
| "loss": 0.4582, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.116527210961903e-05, | |
| "loss": 0.4339, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.1108273865164955e-05, | |
| "loss": 0.4519, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.105127562071088e-05, | |
| "loss": 0.4483, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.099427737625681e-05, | |
| "loss": 0.4335, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.093727913180274e-05, | |
| "loss": 0.4199, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.0880280887348674e-05, | |
| "loss": 0.4409, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.08232826428946e-05, | |
| "loss": 0.4459, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.076628439844053e-05, | |
| "loss": 0.4306, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.070928615398646e-05, | |
| "loss": 0.4147, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.0652287909532386e-05, | |
| "loss": 0.4381, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.059528966507832e-05, | |
| "loss": 0.4233, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.053829142062425e-05, | |
| "loss": 0.4353, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.048129317617018e-05, | |
| "loss": 0.4322, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.0424294931716105e-05, | |
| "loss": 0.4247, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.036729668726203e-05, | |
| "loss": 0.4309, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.031029844280796e-05, | |
| "loss": 0.4046, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.0253300198353896e-05, | |
| "loss": 0.4704, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.0196301953899824e-05, | |
| "loss": 0.4778, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.013930370944575e-05, | |
| "loss": 0.4714, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.008230546499168e-05, | |
| "loss": 0.4971, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.002530722053761e-05, | |
| "loss": 0.4713, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.996830897608354e-05, | |
| "loss": 0.4536, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.991131073162947e-05, | |
| "loss": 0.4443, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.98543124871754e-05, | |
| "loss": 0.4303, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.979731424272133e-05, | |
| "loss": 0.4285, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.9740315998267255e-05, | |
| "loss": 0.4376, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.968331775381318e-05, | |
| "loss": 0.4458, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.962631950935911e-05, | |
| "loss": 0.4414, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.956932126490504e-05, | |
| "loss": 0.4474, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.951232302045097e-05, | |
| "loss": 0.4276, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.94553247759969e-05, | |
| "loss": 0.4529, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.939832653154283e-05, | |
| "loss": 0.4629, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.934132828708876e-05, | |
| "loss": 0.4438, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.9284330042634686e-05, | |
| "loss": 0.4331, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.9227331798180615e-05, | |
| "loss": 0.4478, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.917033355372654e-05, | |
| "loss": 0.4499, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.911333530927248e-05, | |
| "loss": 0.4589, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.9056337064818405e-05, | |
| "loss": 0.4636, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.8999338820364334e-05, | |
| "loss": 0.4601, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.894234057591026e-05, | |
| "loss": 0.4391, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.888534233145619e-05, | |
| "loss": 0.4452, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.8828344087002124e-05, | |
| "loss": 0.4603, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.877134584254805e-05, | |
| "loss": 0.4416, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.871434759809398e-05, | |
| "loss": 0.4419, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.865734935363991e-05, | |
| "loss": 0.4367, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.860035110918584e-05, | |
| "loss": 0.4265, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.854335286473177e-05, | |
| "loss": 0.4387, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.84863546202777e-05, | |
| "loss": 0.4327, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.842935637582363e-05, | |
| "loss": 0.4645, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.8372358131369556e-05, | |
| "loss": 0.439, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.8315359886915484e-05, | |
| "loss": 0.4391, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.825836164246142e-05, | |
| "loss": 0.4795, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.8201363398007347e-05, | |
| "loss": 0.485, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.8144365153553275e-05, | |
| "loss": 0.4669, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.80873669090992e-05, | |
| "loss": 0.4674, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.803036866464513e-05, | |
| "loss": 0.4926, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.7973370420191066e-05, | |
| "loss": 0.4585, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.7916372175736994e-05, | |
| "loss": 0.4782, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.785937393128292e-05, | |
| "loss": 0.4851, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.780237568682884e-05, | |
| "loss": 0.4782, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.774537744237477e-05, | |
| "loss": 0.4744, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.7688379197920706e-05, | |
| "loss": 0.4659, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.7631380953466634e-05, | |
| "loss": 0.471, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.757438270901256e-05, | |
| "loss": 0.4871, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.751738446455849e-05, | |
| "loss": 0.4729, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.746038622010442e-05, | |
| "loss": 0.4772, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.740338797565035e-05, | |
| "loss": 0.4957, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.734638973119628e-05, | |
| "loss": 0.4931, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.728939148674221e-05, | |
| "loss": 0.4642, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.723239324228814e-05, | |
| "loss": 0.4725, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.7175394997834065e-05, | |
| "loss": 0.4871, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.711839675338e-05, | |
| "loss": 0.4686, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.706139850892593e-05, | |
| "loss": 0.4776, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.7004400264471856e-05, | |
| "loss": 0.4852, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.6947402020017784e-05, | |
| "loss": 0.4928, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.689040377556371e-05, | |
| "loss": 0.4728, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.683340553110965e-05, | |
| "loss": 0.462, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.6776407286655575e-05, | |
| "loss": 0.4577, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.67194090422015e-05, | |
| "loss": 0.4697, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.666241079774743e-05, | |
| "loss": 0.4914, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.660541255329336e-05, | |
| "loss": 0.4493, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.654841430883929e-05, | |
| "loss": 0.4601, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.649141606438522e-05, | |
| "loss": 0.4577, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.643441781993115e-05, | |
| "loss": 0.4645, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.637741957547708e-05, | |
| "loss": 0.4722, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.6320421331023006e-05, | |
| "loss": 0.4707, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.6263423086568934e-05, | |
| "loss": 0.4611, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.620642484211487e-05, | |
| "loss": 0.4907, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.61494265976608e-05, | |
| "loss": 0.4848, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.6092428353206725e-05, | |
| "loss": 0.5054, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.603543010875265e-05, | |
| "loss": 0.4926, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.597843186429858e-05, | |
| "loss": 0.4608, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.592143361984451e-05, | |
| "loss": 0.4791, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.586443537539044e-05, | |
| "loss": 0.4826, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.5807437130936365e-05, | |
| "loss": 0.4558, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.5750438886482294e-05, | |
| "loss": 0.4822, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.569344064202823e-05, | |
| "loss": 0.4668, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.5636442397574156e-05, | |
| "loss": 0.4707, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.5579444153120084e-05, | |
| "loss": 0.4704, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.552244590866601e-05, | |
| "loss": 0.4502, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.546544766421194e-05, | |
| "loss": 0.4552, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.540844941975787e-05, | |
| "loss": 0.4639, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.5351451175303803e-05, | |
| "loss": 0.4518, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.529445293084973e-05, | |
| "loss": 0.4547, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.523745468639566e-05, | |
| "loss": 0.4349, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.518045644194159e-05, | |
| "loss": 0.4243, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.5123458197487516e-05, | |
| "loss": 0.4471, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.506645995303345e-05, | |
| "loss": 0.4511, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.500946170857938e-05, | |
| "loss": 0.4786, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.495246346412531e-05, | |
| "loss": 0.4378, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.4895465219671235e-05, | |
| "loss": 0.4542, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.483846697521716e-05, | |
| "loss": 0.4533, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.47814687307631e-05, | |
| "loss": 0.453, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.4724470486309026e-05, | |
| "loss": 0.4585, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.4667472241854954e-05, | |
| "loss": 0.4277, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.461047399740088e-05, | |
| "loss": 0.4447, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.455347575294681e-05, | |
| "loss": 0.4586, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.4496477508492745e-05, | |
| "loss": 0.4695, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.443947926403867e-05, | |
| "loss": 0.4866, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.43824810195846e-05, | |
| "loss": 0.4645, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.432548277513053e-05, | |
| "loss": 0.468, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.426848453067646e-05, | |
| "loss": 0.4599, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.421148628622239e-05, | |
| "loss": 0.4653, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.415448804176832e-05, | |
| "loss": 0.4466, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.409748979731425e-05, | |
| "loss": 0.4781, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.404049155286017e-05, | |
| "loss": 0.4575, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.39834933084061e-05, | |
| "loss": 0.4547, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.392649506395203e-05, | |
| "loss": 0.436, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.386949681949796e-05, | |
| "loss": 0.432, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.381249857504389e-05, | |
| "loss": 0.4602, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.3755500330589816e-05, | |
| "loss": 0.4339, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.3698502086135744e-05, | |
| "loss": 0.4424, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.364150384168168e-05, | |
| "loss": 0.4588, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.358450559722761e-05, | |
| "loss": 0.4564, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.3527507352773535e-05, | |
| "loss": 0.4571, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.347050910831946e-05, | |
| "loss": 0.4501, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.341351086386539e-05, | |
| "loss": 0.4238, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.3356512619411326e-05, | |
| "loss": 0.4521, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.3299514374957254e-05, | |
| "loss": 0.4545, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.324251613050318e-05, | |
| "loss": 0.4586, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.318551788604911e-05, | |
| "loss": 0.4197, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.312851964159504e-05, | |
| "loss": 0.4576, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.307152139714097e-05, | |
| "loss": 0.4315, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.30145231526869e-05, | |
| "loss": 0.4138, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.295752490823283e-05, | |
| "loss": 0.4436, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.290052666377876e-05, | |
| "loss": 0.4423, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.2843528419324685e-05, | |
| "loss": 0.4696, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.278653017487061e-05, | |
| "loss": 0.434, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.272953193041655e-05, | |
| "loss": 0.4509, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.2672533685962476e-05, | |
| "loss": 0.4651, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.2615535441508404e-05, | |
| "loss": 0.5966, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.255853719705433e-05, | |
| "loss": 0.4684, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.250153895260026e-05, | |
| "loss": 0.4602, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.2444540708146195e-05, | |
| "loss": 0.4792, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.238754246369212e-05, | |
| "loss": 0.489, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.233054421923805e-05, | |
| "loss": 0.4577, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.227354597478398e-05, | |
| "loss": 0.474, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.221654773032991e-05, | |
| "loss": 0.4908, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.2159549485875835e-05, | |
| "loss": 0.4544, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.2102551241421763e-05, | |
| "loss": 0.4703, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.204555299696769e-05, | |
| "loss": 0.4626, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.198855475251362e-05, | |
| "loss": 0.452, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.1931556508059554e-05, | |
| "loss": 0.463, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.187455826360548e-05, | |
| "loss": 0.4837, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.181756001915141e-05, | |
| "loss": 0.4514, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.176056177469734e-05, | |
| "loss": 0.4509, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.170356353024327e-05, | |
| "loss": 0.4866, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.16465652857892e-05, | |
| "loss": 0.4715, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.158956704133513e-05, | |
| "loss": 0.4567, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.153256879688106e-05, | |
| "loss": 0.4849, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.1475570552426986e-05, | |
| "loss": 0.4728, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.1418572307972914e-05, | |
| "loss": 0.4786, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.136157406351884e-05, | |
| "loss": 0.4524, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.1304575819064777e-05, | |
| "loss": 0.4636, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.1247577574610705e-05, | |
| "loss": 0.4702, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.119057933015663e-05, | |
| "loss": 0.4745, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.113358108570256e-05, | |
| "loss": 0.4628, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.107658284124849e-05, | |
| "loss": 0.4564, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.1019584596794424e-05, | |
| "loss": 0.441, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.096258635234035e-05, | |
| "loss": 0.4533, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.090558810788628e-05, | |
| "loss": 0.4554, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.084858986343221e-05, | |
| "loss": 0.4478, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.0791591618978136e-05, | |
| "loss": 0.4402, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.073459337452407e-05, | |
| "loss": 0.4424, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.067759513007e-05, | |
| "loss": 0.442, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.062059688561593e-05, | |
| "loss": 0.4467, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.0563598641161855e-05, | |
| "loss": 0.4426, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.0506600396707786e-05, | |
| "loss": 0.4461, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.0449602152253714e-05, | |
| "loss": 0.4512, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.0392603907799642e-05, | |
| "loss": 0.4496, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.0335605663345567e-05, | |
| "loss": 0.4622, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.02786074188915e-05, | |
| "loss": 0.4519, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0221609174437427e-05, | |
| "loss": 0.4418, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0164610929983355e-05, | |
| "loss": 0.4574, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0107612685529286e-05, | |
| "loss": 0.4453, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.0050614441075214e-05, | |
| "loss": 0.4571, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.9993616196621142e-05, | |
| "loss": 0.4594, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.9936617952167074e-05, | |
| "loss": 0.4406, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.9879619707713e-05, | |
| "loss": 0.4556, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.9822621463258933e-05, | |
| "loss": 0.4724, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.976562321880486e-05, | |
| "loss": 0.4933, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.970862497435079e-05, | |
| "loss": 0.469, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.965162672989672e-05, | |
| "loss": 0.4717, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.959462848544265e-05, | |
| "loss": 0.4661, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.953763024098858e-05, | |
| "loss": 0.4688, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.9480631996534508e-05, | |
| "loss": 0.4663, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.9423633752080436e-05, | |
| "loss": 0.4566, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.9366635507626368e-05, | |
| "loss": 0.4506, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.9309637263172296e-05, | |
| "loss": 0.474, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.9252639018718224e-05, | |
| "loss": 0.4725, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9195640774264155e-05, | |
| "loss": 0.4634, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9138642529810083e-05, | |
| "loss": 0.4481, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9081644285356015e-05, | |
| "loss": 0.4626, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.9024646040901943e-05, | |
| "loss": 0.481, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.896764779644787e-05, | |
| "loss": 0.4722, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.8910649551993802e-05, | |
| "loss": 0.442, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.885365130753973e-05, | |
| "loss": 0.4898, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.8796653063085662e-05, | |
| "loss": 0.4528, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.873965481863159e-05, | |
| "loss": 0.4724, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.8682656574177518e-05, | |
| "loss": 0.4491, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.862565832972345e-05, | |
| "loss": 0.4485, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.8568660085269377e-05, | |
| "loss": 0.4482, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.8511661840815305e-05, | |
| "loss": 0.4829, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.845466359636123e-05, | |
| "loss": 0.4445, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.839766535190716e-05, | |
| "loss": 0.4405, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.834066710745309e-05, | |
| "loss": 0.4384, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.8283668862999018e-05, | |
| "loss": 0.4467, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.822667061854495e-05, | |
| "loss": 0.4562, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.8169672374090877e-05, | |
| "loss": 0.4699, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.8112674129636805e-05, | |
| "loss": 0.4678, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.8055675885182737e-05, | |
| "loss": 0.4582, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.7998677640728665e-05, | |
| "loss": 0.4573, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.7941679396274596e-05, | |
| "loss": 0.4693, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.7884681151820524e-05, | |
| "loss": 0.4623, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.7827682907366452e-05, | |
| "loss": 0.4467, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.7770684662912384e-05, | |
| "loss": 0.4454, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.771368641845831e-05, | |
| "loss": 0.4502, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.7656688174004243e-05, | |
| "loss": 0.4671, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.759968992955017e-05, | |
| "loss": 0.4675, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.75426916850961e-05, | |
| "loss": 0.4585, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.748569344064203e-05, | |
| "loss": 0.4725, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.742869519618796e-05, | |
| "loss": 0.4883, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7371696951733887e-05, | |
| "loss": 0.4526, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7314698707279818e-05, | |
| "loss": 0.432, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7257700462825746e-05, | |
| "loss": 0.4647, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7200702218371678e-05, | |
| "loss": 0.4803, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7143703973917606e-05, | |
| "loss": 0.4536, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7086705729463534e-05, | |
| "loss": 0.4601, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.7029707485009465e-05, | |
| "loss": 0.4664, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.6972709240555393e-05, | |
| "loss": 0.4715, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.6915710996101325e-05, | |
| "loss": 0.451, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.6858712751647253e-05, | |
| "loss": 0.4567, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.680171450719318e-05, | |
| "loss": 0.4773, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.6744716262739112e-05, | |
| "loss": 0.4746, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.668771801828504e-05, | |
| "loss": 0.4636, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.663071977383097e-05, | |
| "loss": 0.4715, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.6573721529376893e-05, | |
| "loss": 0.4631, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.6516723284922825e-05, | |
| "loss": 0.4626, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.6459725040468753e-05, | |
| "loss": 0.458, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.640272679601468e-05, | |
| "loss": 0.4568, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.6345728551560612e-05, | |
| "loss": 0.4642, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.628873030710654e-05, | |
| "loss": 0.4558, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.6231732062652468e-05, | |
| "loss": 0.4539, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.61747338181984e-05, | |
| "loss": 0.4756, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6117735573744328e-05, | |
| "loss": 0.4646, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.606073732929026e-05, | |
| "loss": 0.4461, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.6003739084836187e-05, | |
| "loss": 0.4643, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.5946740840382115e-05, | |
| "loss": 0.4728, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.5889742595928047e-05, | |
| "loss": 0.4643, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.5832744351473975e-05, | |
| "loss": 0.4754, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.5775746107019906e-05, | |
| "loss": 0.5094, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.5718747862565834e-05, | |
| "loss": 0.548, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.5661749618111762e-05, | |
| "loss": 0.5629, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.5604751373657694e-05, | |
| "loss": 0.5656, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.5547753129203622e-05, | |
| "loss": 0.5313, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.549075488474955e-05, | |
| "loss": 0.5195, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.543375664029548e-05, | |
| "loss": 0.5275, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.537675839584141e-05, | |
| "loss": 0.5084, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.531976015138734e-05, | |
| "loss": 0.4969, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.526276190693327e-05, | |
| "loss": 0.4924, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.5205763662479197e-05, | |
| "loss": 0.5039, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5148765418025128e-05, | |
| "loss": 0.4906, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5091767173571056e-05, | |
| "loss": 0.4634, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5034768929116988e-05, | |
| "loss": 0.4807, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.4977770684662912e-05, | |
| "loss": 0.4635, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.492077244020884e-05, | |
| "loss": 0.4851, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.4863774195754772e-05, | |
| "loss": 0.4647, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.48067759513007e-05, | |
| "loss": 0.4764, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.474977770684663e-05, | |
| "loss": 0.4688, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.469277946239256e-05, | |
| "loss": 0.4592, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.4635781217938488e-05, | |
| "loss": 0.4809, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.457878297348442e-05, | |
| "loss": 0.4754, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.4521784729030347e-05, | |
| "loss": 0.4727, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.446478648457628e-05, | |
| "loss": 0.4643, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.4407788240122207e-05, | |
| "loss": 0.4628, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.4350789995668135e-05, | |
| "loss": 0.4587, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.4293791751214066e-05, | |
| "loss": 0.4665, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.423679350675999e-05, | |
| "loss": 0.4624, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4179795262305922e-05, | |
| "loss": 0.4667, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.412279701785185e-05, | |
| "loss": 0.4898, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4065798773397778e-05, | |
| "loss": 0.4808, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.400880052894371e-05, | |
| "loss": 0.4781, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.3951802284489638e-05, | |
| "loss": 0.4926, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.389480404003557e-05, | |
| "loss": 0.4804, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.3837805795581497e-05, | |
| "loss": 0.4645, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.3780807551127425e-05, | |
| "loss": 0.4867, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.3723809306673357e-05, | |
| "loss": 0.4763, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.3666811062219285e-05, | |
| "loss": 0.4924, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.3609812817765213e-05, | |
| "loss": 0.4916, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.3552814573311144e-05, | |
| "loss": 0.4834, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.3495816328857072e-05, | |
| "loss": 0.4842, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.3438818084403004e-05, | |
| "loss": 0.4999, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.3381819839948932e-05, | |
| "loss": 0.4782, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.332482159549486e-05, | |
| "loss": 0.4867, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3267823351040788e-05, | |
| "loss": 0.473, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3210825106586716e-05, | |
| "loss": 0.4925, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3153826862132647e-05, | |
| "loss": 0.4853, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3096828617678575e-05, | |
| "loss": 0.4928, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3039830373224504e-05, | |
| "loss": 0.4842, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.2982832128770435e-05, | |
| "loss": 0.4831, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.2925833884316363e-05, | |
| "loss": 0.4822, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.2868835639862294e-05, | |
| "loss": 0.4623, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.2811837395408223e-05, | |
| "loss": 0.4582, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.275483915095415e-05, | |
| "loss": 0.4681, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.2697840906500082e-05, | |
| "loss": 0.467, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.264084266204601e-05, | |
| "loss": 0.4718, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.258384441759194e-05, | |
| "loss": 0.4712, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.252684617313787e-05, | |
| "loss": 0.472, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.2469847928683798e-05, | |
| "loss": 0.4705, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.241284968422973e-05, | |
| "loss": 0.4553, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.2355851439775654e-05, | |
| "loss": 0.4804, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2298853195321585e-05, | |
| "loss": 0.4814, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2241854950867513e-05, | |
| "loss": 0.4678, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.218485670641344e-05, | |
| "loss": 0.4675, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2127858461959373e-05, | |
| "loss": 0.4655, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.20708602175053e-05, | |
| "loss": 0.4687, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2013861973051232e-05, | |
| "loss": 0.4589, | |
| "step": 25050 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.195686372859716e-05, | |
| "loss": 0.4671, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.189986548414309e-05, | |
| "loss": 0.4679, | |
| "step": 25150 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.184286723968902e-05, | |
| "loss": 0.4652, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.1785868995234948e-05, | |
| "loss": 0.4703, | |
| "step": 25250 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.1728870750780876e-05, | |
| "loss": 0.4699, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.1671872506326807e-05, | |
| "loss": 0.4767, | |
| "step": 25350 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.1614874261872735e-05, | |
| "loss": 0.4657, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.1557876017418667e-05, | |
| "loss": 0.4697, | |
| "step": 25450 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.1500877772964595e-05, | |
| "loss": 0.4573, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.1443879528510523e-05, | |
| "loss": 0.4732, | |
| "step": 25550 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.138688128405645e-05, | |
| "loss": 0.4723, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.132988303960238e-05, | |
| "loss": 0.483, | |
| "step": 25650 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.127288479514831e-05, | |
| "loss": 0.4727, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.121588655069424e-05, | |
| "loss": 0.4577, | |
| "step": 25750 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.115888830624017e-05, | |
| "loss": 0.4557, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.1101890061786098e-05, | |
| "loss": 0.46, | |
| "step": 25850 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.1044891817332026e-05, | |
| "loss": 0.4819, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.0987893572877958e-05, | |
| "loss": 0.4792, | |
| "step": 25950 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.0930895328423886e-05, | |
| "loss": 0.4655, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.0873897083969814e-05, | |
| "loss": 0.4756, | |
| "step": 26050 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.0816898839515745e-05, | |
| "loss": 0.4788, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.0759900595061673e-05, | |
| "loss": 0.4684, | |
| "step": 26150 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.0702902350607605e-05, | |
| "loss": 0.4458, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.0645904106153533e-05, | |
| "loss": 0.4838, | |
| "step": 26250 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.058890586169946e-05, | |
| "loss": 0.4629, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.0531907617245392e-05, | |
| "loss": 0.471, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0474909372791317e-05, | |
| "loss": 0.4667, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0417911128337248e-05, | |
| "loss": 0.4684, | |
| "step": 26450 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0360912883883176e-05, | |
| "loss": 0.4879, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0303914639429104e-05, | |
| "loss": 0.4789, | |
| "step": 26550 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0246916394975036e-05, | |
| "loss": 0.4743, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0189918150520964e-05, | |
| "loss": 0.4621, | |
| "step": 26650 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0132919906066895e-05, | |
| "loss": 0.4463, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0075921661612823e-05, | |
| "loss": 0.4662, | |
| "step": 26750 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.001892341715875e-05, | |
| "loss": 0.4968, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.9961925172704683e-05, | |
| "loss": 0.4652, | |
| "step": 26850 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.990492692825061e-05, | |
| "loss": 0.4772, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.984792868379654e-05, | |
| "loss": 0.4855, | |
| "step": 26950 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.979093043934247e-05, | |
| "loss": 0.4771, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.97339321948884e-05, | |
| "loss": 0.4758, | |
| "step": 27050 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.967693395043433e-05, | |
| "loss": 0.4901, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.9619935705980258e-05, | |
| "loss": 0.4627, | |
| "step": 27150 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.9562937461526186e-05, | |
| "loss": 0.4656, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.9505939217072114e-05, | |
| "loss": 0.4754, | |
| "step": 27250 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.9448940972618042e-05, | |
| "loss": 0.4963, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.9391942728163973e-05, | |
| "loss": 0.4665, | |
| "step": 27350 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.93349444837099e-05, | |
| "loss": 0.4733, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9277946239255833e-05, | |
| "loss": 0.4678, | |
| "step": 27450 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.922094799480176e-05, | |
| "loss": 0.4889, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.916394975034769e-05, | |
| "loss": 0.4806, | |
| "step": 27550 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.910695150589362e-05, | |
| "loss": 0.4547, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.904995326143955e-05, | |
| "loss": 0.492, | |
| "step": 27650 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.8992955016985477e-05, | |
| "loss": 0.4799, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.8935956772531408e-05, | |
| "loss": 0.4741, | |
| "step": 27750 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.8878958528077336e-05, | |
| "loss": 0.4816, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.8821960283623268e-05, | |
| "loss": 0.4761, | |
| "step": 27850 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.8764962039169196e-05, | |
| "loss": 0.4596, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.8707963794715124e-05, | |
| "loss": 0.4823, | |
| "step": 27950 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.8650965550261055e-05, | |
| "loss": 0.4787, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.859396730580698e-05, | |
| "loss": 0.464, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.853696906135291e-05, | |
| "loss": 0.4784, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.847997081689884e-05, | |
| "loss": 0.4931, | |
| "step": 28150 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.8422972572444767e-05, | |
| "loss": 0.4663, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.83659743279907e-05, | |
| "loss": 0.468, | |
| "step": 28250 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.8308976083536627e-05, | |
| "loss": 0.4927, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 1.8251977839082558e-05, | |
| "loss": 0.4759, | |
| "step": 28350 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 1.8194979594628486e-05, | |
| "loss": 0.4773, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 1.8137981350174414e-05, | |
| "loss": 0.4849, | |
| "step": 28450 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.8080983105720346e-05, | |
| "loss": 0.4894, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.8023984861266274e-05, | |
| "loss": 0.4584, | |
| "step": 28550 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.7966986616812205e-05, | |
| "loss": 0.4739, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.7909988372358133e-05, | |
| "loss": 0.4971, | |
| "step": 28650 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.785299012790406e-05, | |
| "loss": 0.4737, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.7795991883449993e-05, | |
| "loss": 0.4674, | |
| "step": 28750 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.773899363899592e-05, | |
| "loss": 0.5095, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.768199539454185e-05, | |
| "loss": 0.6674, | |
| "step": 28850 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.7624997150087777e-05, | |
| "loss": 0.683, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.7567998905633705e-05, | |
| "loss": 0.554, | |
| "step": 28950 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.7511000661179637e-05, | |
| "loss": 0.529, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.7454002416725565e-05, | |
| "loss": 0.5206, | |
| "step": 29050 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 1.7397004172271496e-05, | |
| "loss": 0.5223, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 1.7340005927817424e-05, | |
| "loss": 0.5134, | |
| "step": 29150 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 1.7283007683363352e-05, | |
| "loss": 0.5048, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 1.7226009438909284e-05, | |
| "loss": 0.5268, | |
| "step": 29250 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 1.716901119445521e-05, | |
| "loss": 0.5021, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 1.711201295000114e-05, | |
| "loss": 0.507, | |
| "step": 29350 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.705501470554707e-05, | |
| "loss": 0.5111, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.6998016461093e-05, | |
| "loss": 0.5057, | |
| "step": 29450 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.694101821663893e-05, | |
| "loss": 0.5187, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.688401997218486e-05, | |
| "loss": 0.5258, | |
| "step": 29550 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.6827021727730787e-05, | |
| "loss": 0.5441, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.6770023483276718e-05, | |
| "loss": 0.513, | |
| "step": 29650 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.6713025238822643e-05, | |
| "loss": 0.5211, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.6656026994368574e-05, | |
| "loss": 0.5033, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.6599028749914502e-05, | |
| "loss": 0.5205, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.654203050546043e-05, | |
| "loss": 0.5044, | |
| "step": 29850 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.6485032261006362e-05, | |
| "loss": 0.5123, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.642803401655229e-05, | |
| "loss": 0.5146, | |
| "step": 29950 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.637103577209822e-05, | |
| "loss": 0.5034, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.631403752764415e-05, | |
| "loss": 0.5002, | |
| "step": 30050 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.6257039283190077e-05, | |
| "loss": 0.5213, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.620004103873601e-05, | |
| "loss": 0.5083, | |
| "step": 30150 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.6143042794281937e-05, | |
| "loss": 0.5096, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.608604454982787e-05, | |
| "loss": 0.5147, | |
| "step": 30250 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.6029046305373796e-05, | |
| "loss": 0.4939, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.5972048060919724e-05, | |
| "loss": 0.5027, | |
| "step": 30350 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.5915049816465656e-05, | |
| "loss": 0.5152, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.5858051572011584e-05, | |
| "loss": 0.517, | |
| "step": 30450 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.5801053327557512e-05, | |
| "loss": 0.4968, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.574405508310344e-05, | |
| "loss": 0.5201, | |
| "step": 30550 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.5687056838649368e-05, | |
| "loss": 0.5068, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.56300585941953e-05, | |
| "loss": 0.4969, | |
| "step": 30650 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.5573060349741228e-05, | |
| "loss": 0.5006, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.551606210528716e-05, | |
| "loss": 0.5208, | |
| "step": 30750 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.5459063860833087e-05, | |
| "loss": 0.5175, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.5402065616379015e-05, | |
| "loss": 0.4906, | |
| "step": 30850 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.5345067371924947e-05, | |
| "loss": 0.5021, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.5288069127470875e-05, | |
| "loss": 0.5058, | |
| "step": 30950 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.5231070883016804e-05, | |
| "loss": 0.5035, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.5174072638562734e-05, | |
| "loss": 0.4971, | |
| "step": 31050 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.5117074394108662e-05, | |
| "loss": 0.5042, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.5060076149654592e-05, | |
| "loss": 0.5126, | |
| "step": 31150 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.5003077905200522e-05, | |
| "loss": 0.5091, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.4946079660746451e-05, | |
| "loss": 0.5054, | |
| "step": 31250 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.488908141629238e-05, | |
| "loss": 0.5025, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.4832083171838308e-05, | |
| "loss": 0.5099, | |
| "step": 31350 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.4775084927384236e-05, | |
| "loss": 0.4915, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.4718086682930165e-05, | |
| "loss": 0.5073, | |
| "step": 31450 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.4661088438476095e-05, | |
| "loss": 0.5031, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.4604090194022025e-05, | |
| "loss": 0.5181, | |
| "step": 31550 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.4547091949567953e-05, | |
| "loss": 0.5257, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.4490093705113883e-05, | |
| "loss": 0.5142, | |
| "step": 31650 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.4433095460659812e-05, | |
| "loss": 0.5039, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.4376097216205742e-05, | |
| "loss": 0.5099, | |
| "step": 31750 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.431909897175167e-05, | |
| "loss": 0.5162, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.42621007272976e-05, | |
| "loss": 0.505, | |
| "step": 31850 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.420510248284353e-05, | |
| "loss": 0.503, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.414810423838946e-05, | |
| "loss": 0.5164, | |
| "step": 31950 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.409110599393539e-05, | |
| "loss": 0.5147, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.4034107749481317e-05, | |
| "loss": 0.4865, | |
| "step": 32050 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.3977109505027247e-05, | |
| "loss": 0.5074, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.3920111260573173e-05, | |
| "loss": 0.4984, | |
| "step": 32150 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.3863113016119103e-05, | |
| "loss": 0.4858, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.3806114771665033e-05, | |
| "loss": 0.5167, | |
| "step": 32250 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.3749116527210963e-05, | |
| "loss": 0.5001, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.369211828275689e-05, | |
| "loss": 0.5178, | |
| "step": 32350 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.363512003830282e-05, | |
| "loss": 0.5179, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.357812179384875e-05, | |
| "loss": 0.4965, | |
| "step": 32450 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.352112354939468e-05, | |
| "loss": 0.503, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.3464125304940608e-05, | |
| "loss": 0.5151, | |
| "step": 32550 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.3407127060486538e-05, | |
| "loss": 0.4917, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.3350128816032467e-05, | |
| "loss": 0.4839, | |
| "step": 32650 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.3293130571578397e-05, | |
| "loss": 0.5002, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.3236132327124325e-05, | |
| "loss": 0.5082, | |
| "step": 32750 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.3179134082670255e-05, | |
| "loss": 0.5221, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.3122135838216185e-05, | |
| "loss": 0.5093, | |
| "step": 32850 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.3065137593762114e-05, | |
| "loss": 0.5144, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.300813934930804e-05, | |
| "loss": 0.4986, | |
| "step": 32950 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.295114110485397e-05, | |
| "loss": 0.5137, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.2894142860399899e-05, | |
| "loss": 0.5076, | |
| "step": 33050 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.2837144615945828e-05, | |
| "loss": 0.5125, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.2780146371491758e-05, | |
| "loss": 0.4967, | |
| "step": 33150 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.2723148127037688e-05, | |
| "loss": 0.5228, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.2666149882583616e-05, | |
| "loss": 0.4989, | |
| "step": 33250 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.2609151638129546e-05, | |
| "loss": 0.5216, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.2552153393675475e-05, | |
| "loss": 0.5039, | |
| "step": 33350 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.2495155149221405e-05, | |
| "loss": 0.5053, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.2438156904767335e-05, | |
| "loss": 0.5013, | |
| "step": 33450 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.2381158660313263e-05, | |
| "loss": 0.513, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.2324160415859193e-05, | |
| "loss": 0.4902, | |
| "step": 33550 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.226716217140512e-05, | |
| "loss": 0.519, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.221016392695105e-05, | |
| "loss": 0.4956, | |
| "step": 33650 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.215316568249698e-05, | |
| "loss": 0.5088, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.2096167438042908e-05, | |
| "loss": 0.5174, | |
| "step": 33750 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.2039169193588838e-05, | |
| "loss": 0.5183, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.1982170949134768e-05, | |
| "loss": 0.4934, | |
| "step": 33850 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.1925172704680698e-05, | |
| "loss": 0.5109, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.1868174460226626e-05, | |
| "loss": 0.5163, | |
| "step": 33950 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.1811176215772554e-05, | |
| "loss": 0.5088, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.1754177971318483e-05, | |
| "loss": 0.5143, | |
| "step": 34050 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.1697179726864413e-05, | |
| "loss": 0.5087, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.1640181482410343e-05, | |
| "loss": 0.5253, | |
| "step": 34150 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.1583183237956271e-05, | |
| "loss": 0.5104, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.15261849935022e-05, | |
| "loss": 0.513, | |
| "step": 34250 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.146918674904813e-05, | |
| "loss": 0.5103, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.141218850459406e-05, | |
| "loss": 0.4975, | |
| "step": 34350 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.1355190260139988e-05, | |
| "loss": 0.5092, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.1298192015685916e-05, | |
| "loss": 0.5011, | |
| "step": 34450 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.1241193771231846e-05, | |
| "loss": 0.5027, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.1184195526777776e-05, | |
| "loss": 0.5034, | |
| "step": 34550 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.1127197282323706e-05, | |
| "loss": 0.4982, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.1070199037869634e-05, | |
| "loss": 0.5037, | |
| "step": 34650 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.1013200793415563e-05, | |
| "loss": 0.4958, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.0956202548961493e-05, | |
| "loss": 0.5199, | |
| "step": 34750 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.0899204304507421e-05, | |
| "loss": 0.5108, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.0842206060053351e-05, | |
| "loss": 0.505, | |
| "step": 34850 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.0785207815599279e-05, | |
| "loss": 0.5053, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.0728209571145209e-05, | |
| "loss": 0.5072, | |
| "step": 34950 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.0671211326691138e-05, | |
| "loss": 0.523, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.0614213082237068e-05, | |
| "loss": 0.5121, | |
| "step": 35050 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.0557214837782998e-05, | |
| "loss": 0.5072, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.0500216593328926e-05, | |
| "loss": 0.4977, | |
| "step": 35150 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.0443218348874856e-05, | |
| "loss": 0.5165, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.0386220104420784e-05, | |
| "loss": 0.5042, | |
| "step": 35250 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.0329221859966714e-05, | |
| "loss": 0.516, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.0272223615512643e-05, | |
| "loss": 0.4964, | |
| "step": 35350 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.0215225371058571e-05, | |
| "loss": 0.4903, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.0158227126604501e-05, | |
| "loss": 0.5009, | |
| "step": 35450 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.010122888215043e-05, | |
| "loss": 0.5015, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.004423063769636e-05, | |
| "loss": 0.4986, | |
| "step": 35550 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 9.987232393242289e-06, | |
| "loss": 0.5018, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 9.930234148788217e-06, | |
| "loss": 0.5197, | |
| "step": 35650 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 9.873235904334146e-06, | |
| "loss": 0.5129, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 9.816237659880076e-06, | |
| "loss": 0.5091, | |
| "step": 35750 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 9.759239415426006e-06, | |
| "loss": 0.5197, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 9.702241170971934e-06, | |
| "loss": 0.4794, | |
| "step": 35850 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 9.645242926517864e-06, | |
| "loss": 0.5027, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 9.588244682063793e-06, | |
| "loss": 0.5051, | |
| "step": 35950 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 9.531246437609723e-06, | |
| "loss": 0.5007, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 9.474248193155651e-06, | |
| "loss": 0.5028, | |
| "step": 36050 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 9.41724994870158e-06, | |
| "loss": 0.5173, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 9.360251704247509e-06, | |
| "loss": 0.4957, | |
| "step": 36150 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 9.303253459793439e-06, | |
| "loss": 0.4869, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 9.246255215339369e-06, | |
| "loss": 0.5006, | |
| "step": 36250 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 9.189256970885297e-06, | |
| "loss": 0.5063, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 9.132258726431226e-06, | |
| "loss": 0.4872, | |
| "step": 36350 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 9.075260481977156e-06, | |
| "loss": 0.5157, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 9.018262237523084e-06, | |
| "loss": 0.4927, | |
| "step": 36450 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 8.961263993069014e-06, | |
| "loss": 0.5193, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 8.904265748614942e-06, | |
| "loss": 0.4993, | |
| "step": 36550 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 8.847267504160872e-06, | |
| "loss": 0.5108, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 8.790269259706801e-06, | |
| "loss": 0.4985, | |
| "step": 36650 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 8.733271015252731e-06, | |
| "loss": 0.5015, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 8.676272770798661e-06, | |
| "loss": 0.4881, | |
| "step": 36750 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 8.619274526344589e-06, | |
| "loss": 0.4971, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 8.562276281890517e-06, | |
| "loss": 0.4989, | |
| "step": 36850 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 8.505278037436447e-06, | |
| "loss": 0.505, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 8.448279792982377e-06, | |
| "loss": 0.5069, | |
| "step": 36950 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 8.391281548528306e-06, | |
| "loss": 0.5175, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 8.334283304074234e-06, | |
| "loss": 0.4933, | |
| "step": 37050 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 8.277285059620164e-06, | |
| "loss": 0.5073, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 8.220286815166094e-06, | |
| "loss": 0.5014, | |
| "step": 37150 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 8.163288570712024e-06, | |
| "loss": 0.5197, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 8.106290326257952e-06, | |
| "loss": 0.5035, | |
| "step": 37250 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 8.04929208180388e-06, | |
| "loss": 0.4902, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 7.99229383734981e-06, | |
| "loss": 0.5123, | |
| "step": 37350 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 7.93529559289574e-06, | |
| "loss": 0.5182, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 7.878297348441669e-06, | |
| "loss": 0.5093, | |
| "step": 37450 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 7.821299103987597e-06, | |
| "loss": 0.4936, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 7.764300859533527e-06, | |
| "loss": 0.4883, | |
| "step": 37550 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 7.707302615079457e-06, | |
| "loss": 0.4968, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 7.650304370625386e-06, | |
| "loss": 0.4925, | |
| "step": 37650 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 7.5933061261713135e-06, | |
| "loss": 0.5, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 7.536307881717243e-06, | |
| "loss": 0.5009, | |
| "step": 37750 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 7.479309637263172e-06, | |
| "loss": 0.5161, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 7.422311392809102e-06, | |
| "loss": 0.5119, | |
| "step": 37850 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 7.365313148355031e-06, | |
| "loss": 0.4972, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 7.3083149039009605e-06, | |
| "loss": 0.5116, | |
| "step": 37950 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 7.251316659446889e-06, | |
| "loss": 0.5006, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 7.194318414992819e-06, | |
| "loss": 0.5207, | |
| "step": 38050 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 7.137320170538747e-06, | |
| "loss": 0.505, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 7.080321926084676e-06, | |
| "loss": 0.5112, | |
| "step": 38150 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 7.023323681630606e-06, | |
| "loss": 0.5169, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 6.966325437176536e-06, | |
| "loss": 0.5127, | |
| "step": 38250 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 6.9093271927224645e-06, | |
| "loss": 0.5065, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 6.852328948268394e-06, | |
| "loss": 0.4879, | |
| "step": 38350 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 6.795330703814323e-06, | |
| "loss": 0.5281, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 6.738332459360253e-06, | |
| "loss": 0.5051, | |
| "step": 38450 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 6.681334214906181e-06, | |
| "loss": 0.5217, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 6.62433597045211e-06, | |
| "loss": 0.4942, | |
| "step": 38550 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 6.56733772599804e-06, | |
| "loss": 0.5241, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 6.5103394815439685e-06, | |
| "loss": 0.4991, | |
| "step": 38650 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 6.453341237089898e-06, | |
| "loss": 0.5034, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 6.396342992635827e-06, | |
| "loss": 0.5008, | |
| "step": 38750 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 6.339344748181757e-06, | |
| "loss": 0.5183, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 6.282346503727686e-06, | |
| "loss": 0.5006, | |
| "step": 38850 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 6.225348259273615e-06, | |
| "loss": 0.4936, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 6.1683500148195444e-06, | |
| "loss": 0.4758, | |
| "step": 38950 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 6.111351770365473e-06, | |
| "loss": 0.4654, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 6.054353525911402e-06, | |
| "loss": 0.4584, | |
| "step": 39050 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 5.997355281457331e-06, | |
| "loss": 0.4929, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 5.940357037003261e-06, | |
| "loss": 0.4793, | |
| "step": 39150 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 5.88335879254919e-06, | |
| "loss": 0.4625, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 5.826360548095119e-06, | |
| "loss": 0.462, | |
| "step": 39250 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 5.7693623036410484e-06, | |
| "loss": 0.4598, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 5.712364059186977e-06, | |
| "loss": 0.4778, | |
| "step": 39350 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 5.655365814732907e-06, | |
| "loss": 0.4707, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.598367570278835e-06, | |
| "loss": 0.4468, | |
| "step": 39450 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.541369325824765e-06, | |
| "loss": 0.454, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.484371081370694e-06, | |
| "loss": 0.4454, | |
| "step": 39550 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.4273728369166235e-06, | |
| "loss": 0.4493, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.3703745924625524e-06, | |
| "loss": 0.451, | |
| "step": 39650 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.313376348008481e-06, | |
| "loss": 0.4515, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.256378103554411e-06, | |
| "loss": 0.4714, | |
| "step": 39750 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.19937985910034e-06, | |
| "loss": 0.4658, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.142381614646269e-06, | |
| "loss": 0.4694, | |
| "step": 39850 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.085383370192199e-06, | |
| "loss": 0.4487, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.0283851257381275e-06, | |
| "loss": 0.4528, | |
| "step": 39950 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.971386881284057e-06, | |
| "loss": 0.4515, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.914388636829985e-06, | |
| "loss": 0.4878, | |
| "step": 40050 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.857390392375915e-06, | |
| "loss": 0.4668, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.800392147921844e-06, | |
| "loss": 0.4651, | |
| "step": 40150 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.743393903467774e-06, | |
| "loss": 0.4722, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.686395659013703e-06, | |
| "loss": 0.47, | |
| "step": 40250 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.6293974145596315e-06, | |
| "loss": 0.4629, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.572399170105561e-06, | |
| "loss": 0.4534, | |
| "step": 40350 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.51540092565149e-06, | |
| "loss": 0.4816, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.458402681197419e-06, | |
| "loss": 0.4601, | |
| "step": 40450 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.401404436743348e-06, | |
| "loss": 0.4642, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.344406192289278e-06, | |
| "loss": 0.4699, | |
| "step": 40550 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.2874079478352075e-06, | |
| "loss": 0.4459, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.2304097033811355e-06, | |
| "loss": 0.4494, | |
| "step": 40650 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.173411458927065e-06, | |
| "loss": 0.4565, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.116413214472994e-06, | |
| "loss": 0.4567, | |
| "step": 40750 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.059414970018924e-06, | |
| "loss": 0.4602, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.002416725564853e-06, | |
| "loss": 0.4396, | |
| "step": 40850 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.945418481110782e-06, | |
| "loss": 0.459, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.8884202366567115e-06, | |
| "loss": 0.4686, | |
| "step": 40950 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.83142199220264e-06, | |
| "loss": 0.462, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 3.7744237477485697e-06, | |
| "loss": 0.4658, | |
| "step": 41050 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 3.7174255032944986e-06, | |
| "loss": 0.4596, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 3.660427258840428e-06, | |
| "loss": 0.4557, | |
| "step": 41150 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 3.6034290143863572e-06, | |
| "loss": 0.4635, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 3.5464307699322866e-06, | |
| "loss": 0.4548, | |
| "step": 41250 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 3.489432525478215e-06, | |
| "loss": 0.4472, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.4324342810241448e-06, | |
| "loss": 0.4563, | |
| "step": 41350 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.375436036570074e-06, | |
| "loss": 0.4541, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.3184377921160034e-06, | |
| "loss": 0.4619, | |
| "step": 41450 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.261439547661932e-06, | |
| "loss": 0.4267, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.2044413032078612e-06, | |
| "loss": 0.4374, | |
| "step": 41550 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.1474430587537906e-06, | |
| "loss": 0.4332, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.0904448142997195e-06, | |
| "loss": 0.4487, | |
| "step": 41650 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.033446569845649e-06, | |
| "loss": 0.4467, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 2.976448325391578e-06, | |
| "loss": 0.4488, | |
| "step": 41750 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 2.9194500809375074e-06, | |
| "loss": 0.4473, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 2.8624518364834363e-06, | |
| "loss": 0.4497, | |
| "step": 41850 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 2.8054535920293657e-06, | |
| "loss": 0.4458, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 2.7484553475752946e-06, | |
| "loss": 0.4496, | |
| "step": 41950 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 2.691457103121224e-06, | |
| "loss": 0.4375, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 2.634458858667153e-06, | |
| "loss": 0.4403, | |
| "step": 42050 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.5774606142130825e-06, | |
| "loss": 0.4317, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.5204623697590114e-06, | |
| "loss": 0.4473, | |
| "step": 42150 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.4634641253049407e-06, | |
| "loss": 0.4407, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.4064658808508696e-06, | |
| "loss": 0.414, | |
| "step": 42250 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.349467636396799e-06, | |
| "loss": 0.4204, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.2924693919427283e-06, | |
| "loss": 0.4546, | |
| "step": 42350 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.2354711474886576e-06, | |
| "loss": 0.4494, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.1784729030345865e-06, | |
| "loss": 0.4347, | |
| "step": 42450 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.121474658580516e-06, | |
| "loss": 0.461, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.0644764141264447e-06, | |
| "loss": 0.4579, | |
| "step": 42550 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.007478169672374e-06, | |
| "loss": 0.4155, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 1.9504799252183034e-06, | |
| "loss": 0.4498, | |
| "step": 42650 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 1.8934816807642325e-06, | |
| "loss": 0.4504, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 1.8364834363101616e-06, | |
| "loss": 0.4387, | |
| "step": 42750 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 1.779485191856091e-06, | |
| "loss": 0.4304, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.7224869474020203e-06, | |
| "loss": 0.4286, | |
| "step": 42850 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.6654887029479494e-06, | |
| "loss": 0.4177, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.6084904584938787e-06, | |
| "loss": 0.4518, | |
| "step": 42950 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.5514922140398076e-06, | |
| "loss": 0.4462, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.4944939695857367e-06, | |
| "loss": 0.432, | |
| "step": 43050 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.437495725131666e-06, | |
| "loss": 0.4284, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.3804974806775951e-06, | |
| "loss": 0.4183, | |
| "step": 43150 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.3234992362235243e-06, | |
| "loss": 0.4254, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.2665009917694536e-06, | |
| "loss": 0.4459, | |
| "step": 43250 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.2095027473153827e-06, | |
| "loss": 0.433, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.1525045028613118e-06, | |
| "loss": 0.4376, | |
| "step": 43350 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.0955062584072411e-06, | |
| "loss": 0.4389, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.0385080139531702e-06, | |
| "loss": 0.4332, | |
| "step": 43450 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 9.815097694990994e-07, | |
| "loss": 0.4354, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 9.245115250450286e-07, | |
| "loss": 0.4511, | |
| "step": 43550 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 8.675132805909578e-07, | |
| "loss": 0.4149, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 8.105150361368869e-07, | |
| "loss": 0.4322, | |
| "step": 43650 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.535167916828162e-07, | |
| "loss": 0.4128, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 6.965185472287453e-07, | |
| "loss": 0.4194, | |
| "step": 43750 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 6.395203027746746e-07, | |
| "loss": 0.4325, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.825220583206038e-07, | |
| "loss": 0.4287, | |
| "step": 43850 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.255238138665329e-07, | |
| "loss": 0.4353, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.685255694124621e-07, | |
| "loss": 0.4207, | |
| "step": 43950 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.1152732495839127e-07, | |
| "loss": 0.4465, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.545290805043205e-07, | |
| "loss": 0.4238, | |
| "step": 44050 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.9753083605024965e-07, | |
| "loss": 0.4226, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.4053259159617887e-07, | |
| "loss": 0.4465, | |
| "step": 44150 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.8353434714210804e-07, | |
| "loss": 0.4094, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.265361026880372e-07, | |
| "loss": 0.4491, | |
| "step": 44250 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 6.95378582339664e-08, | |
| "loss": 0.4208, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 1.253961377989558e-08, | |
| "loss": 0.4375, | |
| "step": 44350 | |
| } | |
| ], | |
| "max_steps": 44361, | |
| "num_train_epochs": 3, | |
| "total_flos": 1.77824607798456e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |