| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 8.695652173913043, |
| "global_step": 1000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04, |
| "learning_rate": 5.000000000000001e-07, |
| "loss": 0.7601, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.0000000000000002e-06, |
| "loss": 0.7662, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.5e-06, |
| "loss": 0.7385, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9000000000000002e-06, |
| "loss": 0.707, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 2.4000000000000003e-06, |
| "loss": 0.6551, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 2.8000000000000003e-06, |
| "loss": 0.5934, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 3.3000000000000006e-06, |
| "loss": 0.5183, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 3.8000000000000005e-06, |
| "loss": 0.487, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 4.3e-06, |
| "loss": 0.4695, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 4.800000000000001e-06, |
| "loss": 0.4183, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 5.300000000000001e-06, |
| "loss": 0.4144, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 5.8e-06, |
| "loss": 0.3759, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 6.300000000000001e-06, |
| "loss": 0.3933, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.800000000000001e-06, |
| "loss": 0.406, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 7.3e-06, |
| "loss": 0.3771, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 7.800000000000002e-06, |
| "loss": 0.3719, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 8.3e-06, |
| "loss": 0.366, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 8.8e-06, |
| "loss": 0.3465, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 9.3e-06, |
| "loss": 0.3979, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 9.800000000000001e-06, |
| "loss": 0.3745, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 9.966666666666667e-06, |
| "loss": 0.388, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 9.911111111111113e-06, |
| "loss": 0.4033, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 9.855555555555555e-06, |
| "loss": 0.3652, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 9.800000000000001e-06, |
| "loss": 0.3252, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 9.744444444444445e-06, |
| "loss": 0.357, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 9.688888888888889e-06, |
| "loss": 0.3843, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 9.633333333333335e-06, |
| "loss": 0.3602, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 9.577777777777779e-06, |
| "loss": 0.2971, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 9.522222222222223e-06, |
| "loss": 0.3397, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 9.466666666666667e-06, |
| "loss": 0.3062, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 9.411111111111113e-06, |
| "loss": 0.327, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 9.355555555555557e-06, |
| "loss": 0.3006, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 9.3e-06, |
| "loss": 0.304, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 9.244444444444445e-06, |
| "loss": 0.2943, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 9.188888888888889e-06, |
| "loss": 0.252, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.133333333333335e-06, |
| "loss": 0.3139, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.077777777777779e-06, |
| "loss": 0.2665, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 9.022222222222223e-06, |
| "loss": 0.302, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.966666666666667e-06, |
| "loss": 0.2248, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 8.91111111111111e-06, |
| "loss": 0.2442, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 8.855555555555556e-06, |
| "loss": 0.2233, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 8.8e-06, |
| "loss": 0.1981, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 8.744444444444446e-06, |
| "loss": 0.2188, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 8.68888888888889e-06, |
| "loss": 0.2156, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 8.633333333333334e-06, |
| "loss": 0.1932, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 8.577777777777778e-06, |
| "loss": 0.2024, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 8.522222222222222e-06, |
| "loss": 0.2008, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 8.466666666666668e-06, |
| "loss": 0.2096, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 8.411111111111112e-06, |
| "loss": 0.1882, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 8.355555555555556e-06, |
| "loss": 0.2387, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 8.3e-06, |
| "loss": 0.1566, |
| "step": 255 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 8.244444444444444e-06, |
| "loss": 0.2002, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 8.18888888888889e-06, |
| "loss": 0.1623, |
| "step": 265 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 8.133333333333334e-06, |
| "loss": 0.2073, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 8.077777777777778e-06, |
| "loss": 0.2038, |
| "step": 275 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 8.022222222222222e-06, |
| "loss": 0.1519, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 7.966666666666668e-06, |
| "loss": 0.1626, |
| "step": 285 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 7.911111111111112e-06, |
| "loss": 0.1737, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 7.855555555555556e-06, |
| "loss": 0.1992, |
| "step": 295 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 7.800000000000002e-06, |
| "loss": 0.1539, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.744444444444446e-06, |
| "loss": 0.1764, |
| "step": 305 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 7.68888888888889e-06, |
| "loss": 0.1601, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 7.633333333333334e-06, |
| "loss": 0.1793, |
| "step": 315 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 7.5777777777777785e-06, |
| "loss": 0.1314, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 7.5222222222222226e-06, |
| "loss": 0.1543, |
| "step": 325 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 7.4666666666666675e-06, |
| "loss": 0.1393, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 7.411111111111112e-06, |
| "loss": 0.1847, |
| "step": 335 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 7.3555555555555555e-06, |
| "loss": 0.2043, |
| "step": 340 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 7.3e-06, |
| "loss": 0.1308, |
| "step": 345 |
| }, |
| { |
| "epoch": 3.04, |
| "learning_rate": 7.244444444444445e-06, |
| "loss": 0.1352, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.09, |
| "learning_rate": 7.188888888888889e-06, |
| "loss": 0.1378, |
| "step": 355 |
| }, |
| { |
| "epoch": 3.13, |
| "learning_rate": 7.133333333333334e-06, |
| "loss": 0.1581, |
| "step": 360 |
| }, |
| { |
| "epoch": 3.17, |
| "learning_rate": 7.077777777777778e-06, |
| "loss": 0.1399, |
| "step": 365 |
| }, |
| { |
| "epoch": 3.22, |
| "learning_rate": 7.022222222222222e-06, |
| "loss": 0.149, |
| "step": 370 |
| }, |
| { |
| "epoch": 3.26, |
| "learning_rate": 6.966666666666667e-06, |
| "loss": 0.0899, |
| "step": 375 |
| }, |
| { |
| "epoch": 3.3, |
| "learning_rate": 6.911111111111112e-06, |
| "loss": 0.1157, |
| "step": 380 |
| }, |
| { |
| "epoch": 3.35, |
| "learning_rate": 6.855555555555556e-06, |
| "loss": 0.137, |
| "step": 385 |
| }, |
| { |
| "epoch": 3.39, |
| "learning_rate": 6.800000000000001e-06, |
| "loss": 0.109, |
| "step": 390 |
| }, |
| { |
| "epoch": 3.43, |
| "learning_rate": 6.744444444444444e-06, |
| "loss": 0.0956, |
| "step": 395 |
| }, |
| { |
| "epoch": 3.48, |
| "learning_rate": 6.688888888888889e-06, |
| "loss": 0.1121, |
| "step": 400 |
| }, |
| { |
| "epoch": 3.52, |
| "learning_rate": 6.633333333333334e-06, |
| "loss": 0.125, |
| "step": 405 |
| }, |
| { |
| "epoch": 3.57, |
| "learning_rate": 6.577777777777779e-06, |
| "loss": 0.1222, |
| "step": 410 |
| }, |
| { |
| "epoch": 3.61, |
| "learning_rate": 6.522222222222223e-06, |
| "loss": 0.0806, |
| "step": 415 |
| }, |
| { |
| "epoch": 3.65, |
| "learning_rate": 6.466666666666667e-06, |
| "loss": 0.1144, |
| "step": 420 |
| }, |
| { |
| "epoch": 3.7, |
| "learning_rate": 6.411111111111111e-06, |
| "loss": 0.1139, |
| "step": 425 |
| }, |
| { |
| "epoch": 3.74, |
| "learning_rate": 6.355555555555556e-06, |
| "loss": 0.0836, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.78, |
| "learning_rate": 6.300000000000001e-06, |
| "loss": 0.1699, |
| "step": 435 |
| }, |
| { |
| "epoch": 3.83, |
| "learning_rate": 6.2444444444444456e-06, |
| "loss": 0.1187, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.87, |
| "learning_rate": 6.18888888888889e-06, |
| "loss": 0.0888, |
| "step": 445 |
| }, |
| { |
| "epoch": 3.91, |
| "learning_rate": 6.133333333333334e-06, |
| "loss": 0.1015, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.96, |
| "learning_rate": 6.077777777777778e-06, |
| "loss": 0.0988, |
| "step": 455 |
| }, |
| { |
| "epoch": 4.0, |
| "learning_rate": 6.0222222222222225e-06, |
| "loss": 0.1092, |
| "step": 460 |
| }, |
| { |
| "epoch": 4.04, |
| "learning_rate": 5.966666666666667e-06, |
| "loss": 0.0985, |
| "step": 465 |
| }, |
| { |
| "epoch": 4.09, |
| "learning_rate": 5.911111111111112e-06, |
| "loss": 0.0934, |
| "step": 470 |
| }, |
| { |
| "epoch": 4.13, |
| "learning_rate": 5.855555555555556e-06, |
| "loss": 0.0739, |
| "step": 475 |
| }, |
| { |
| "epoch": 4.17, |
| "learning_rate": 5.8e-06, |
| "loss": 0.0905, |
| "step": 480 |
| }, |
| { |
| "epoch": 4.22, |
| "learning_rate": 5.744444444444444e-06, |
| "loss": 0.0695, |
| "step": 485 |
| }, |
| { |
| "epoch": 4.26, |
| "learning_rate": 5.688888888888889e-06, |
| "loss": 0.0875, |
| "step": 490 |
| }, |
| { |
| "epoch": 4.3, |
| "learning_rate": 5.633333333333334e-06, |
| "loss": 0.0896, |
| "step": 495 |
| }, |
| { |
| "epoch": 4.35, |
| "learning_rate": 5.577777777777778e-06, |
| "loss": 0.0751, |
| "step": 500 |
| }, |
| { |
| "epoch": 4.39, |
| "learning_rate": 5.522222222222222e-06, |
| "loss": 0.0923, |
| "step": 505 |
| }, |
| { |
| "epoch": 4.43, |
| "learning_rate": 5.466666666666667e-06, |
| "loss": 0.0841, |
| "step": 510 |
| }, |
| { |
| "epoch": 4.48, |
| "learning_rate": 5.411111111111111e-06, |
| "loss": 0.0978, |
| "step": 515 |
| }, |
| { |
| "epoch": 4.52, |
| "learning_rate": 5.355555555555556e-06, |
| "loss": 0.0831, |
| "step": 520 |
| }, |
| { |
| "epoch": 4.57, |
| "learning_rate": 5.300000000000001e-06, |
| "loss": 0.0817, |
| "step": 525 |
| }, |
| { |
| "epoch": 4.61, |
| "learning_rate": 5.244444444444445e-06, |
| "loss": 0.0612, |
| "step": 530 |
| }, |
| { |
| "epoch": 4.65, |
| "learning_rate": 5.188888888888889e-06, |
| "loss": 0.0601, |
| "step": 535 |
| }, |
| { |
| "epoch": 4.7, |
| "learning_rate": 5.133333333333334e-06, |
| "loss": 0.0664, |
| "step": 540 |
| }, |
| { |
| "epoch": 4.74, |
| "learning_rate": 5.077777777777778e-06, |
| "loss": 0.0893, |
| "step": 545 |
| }, |
| { |
| "epoch": 4.78, |
| "learning_rate": 5.022222222222223e-06, |
| "loss": 0.0725, |
| "step": 550 |
| }, |
| { |
| "epoch": 4.83, |
| "learning_rate": 4.966666666666667e-06, |
| "loss": 0.0761, |
| "step": 555 |
| }, |
| { |
| "epoch": 4.87, |
| "learning_rate": 4.911111111111112e-06, |
| "loss": 0.0509, |
| "step": 560 |
| }, |
| { |
| "epoch": 4.91, |
| "learning_rate": 4.855555555555556e-06, |
| "loss": 0.0841, |
| "step": 565 |
| }, |
| { |
| "epoch": 4.96, |
| "learning_rate": 4.800000000000001e-06, |
| "loss": 0.1143, |
| "step": 570 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 4.744444444444445e-06, |
| "loss": 0.0867, |
| "step": 575 |
| }, |
| { |
| "epoch": 5.04, |
| "learning_rate": 4.6888888888888895e-06, |
| "loss": 0.0829, |
| "step": 580 |
| }, |
| { |
| "epoch": 5.09, |
| "learning_rate": 4.633333333333334e-06, |
| "loss": 0.0609, |
| "step": 585 |
| }, |
| { |
| "epoch": 5.13, |
| "learning_rate": 4.5777777777777785e-06, |
| "loss": 0.0551, |
| "step": 590 |
| }, |
| { |
| "epoch": 5.17, |
| "learning_rate": 4.5222222222222225e-06, |
| "loss": 0.0653, |
| "step": 595 |
| }, |
| { |
| "epoch": 5.22, |
| "learning_rate": 4.4666666666666665e-06, |
| "loss": 0.067, |
| "step": 600 |
| }, |
| { |
| "epoch": 5.26, |
| "learning_rate": 4.411111111111111e-06, |
| "loss": 0.0561, |
| "step": 605 |
| }, |
| { |
| "epoch": 5.3, |
| "learning_rate": 4.3555555555555555e-06, |
| "loss": 0.0465, |
| "step": 610 |
| }, |
| { |
| "epoch": 5.35, |
| "learning_rate": 4.3e-06, |
| "loss": 0.0484, |
| "step": 615 |
| }, |
| { |
| "epoch": 5.39, |
| "learning_rate": 4.244444444444445e-06, |
| "loss": 0.0667, |
| "step": 620 |
| }, |
| { |
| "epoch": 5.43, |
| "learning_rate": 4.188888888888889e-06, |
| "loss": 0.0692, |
| "step": 625 |
| }, |
| { |
| "epoch": 5.48, |
| "learning_rate": 4.133333333333333e-06, |
| "loss": 0.0602, |
| "step": 630 |
| }, |
| { |
| "epoch": 5.52, |
| "learning_rate": 4.077777777777778e-06, |
| "loss": 0.0665, |
| "step": 635 |
| }, |
| { |
| "epoch": 5.57, |
| "learning_rate": 4.022222222222222e-06, |
| "loss": 0.0974, |
| "step": 640 |
| }, |
| { |
| "epoch": 5.61, |
| "learning_rate": 3.966666666666667e-06, |
| "loss": 0.0573, |
| "step": 645 |
| }, |
| { |
| "epoch": 5.65, |
| "learning_rate": 3.911111111111112e-06, |
| "loss": 0.064, |
| "step": 650 |
| }, |
| { |
| "epoch": 5.7, |
| "learning_rate": 3.855555555555556e-06, |
| "loss": 0.0556, |
| "step": 655 |
| }, |
| { |
| "epoch": 5.74, |
| "learning_rate": 3.8000000000000005e-06, |
| "loss": 0.0401, |
| "step": 660 |
| }, |
| { |
| "epoch": 5.78, |
| "learning_rate": 3.744444444444445e-06, |
| "loss": 0.0626, |
| "step": 665 |
| }, |
| { |
| "epoch": 5.83, |
| "learning_rate": 3.688888888888889e-06, |
| "loss": 0.0483, |
| "step": 670 |
| }, |
| { |
| "epoch": 5.87, |
| "learning_rate": 3.633333333333334e-06, |
| "loss": 0.0847, |
| "step": 675 |
| }, |
| { |
| "epoch": 5.91, |
| "learning_rate": 3.577777777777778e-06, |
| "loss": 0.0492, |
| "step": 680 |
| }, |
| { |
| "epoch": 5.96, |
| "learning_rate": 3.5222222222222223e-06, |
| "loss": 0.0515, |
| "step": 685 |
| }, |
| { |
| "epoch": 6.0, |
| "learning_rate": 3.4666666666666672e-06, |
| "loss": 0.0372, |
| "step": 690 |
| }, |
| { |
| "epoch": 6.04, |
| "learning_rate": 3.4111111111111113e-06, |
| "loss": 0.0418, |
| "step": 695 |
| }, |
| { |
| "epoch": 6.09, |
| "learning_rate": 3.3555555555555557e-06, |
| "loss": 0.0368, |
| "step": 700 |
| }, |
| { |
| "epoch": 6.13, |
| "learning_rate": 3.3000000000000006e-06, |
| "loss": 0.0476, |
| "step": 705 |
| }, |
| { |
| "epoch": 6.17, |
| "learning_rate": 3.2444444444444446e-06, |
| "loss": 0.044, |
| "step": 710 |
| }, |
| { |
| "epoch": 6.22, |
| "learning_rate": 3.188888888888889e-06, |
| "loss": 0.0591, |
| "step": 715 |
| }, |
| { |
| "epoch": 6.26, |
| "learning_rate": 3.133333333333334e-06, |
| "loss": 0.0493, |
| "step": 720 |
| }, |
| { |
| "epoch": 6.3, |
| "learning_rate": 3.077777777777778e-06, |
| "loss": 0.0306, |
| "step": 725 |
| }, |
| { |
| "epoch": 6.35, |
| "learning_rate": 3.0222222222222225e-06, |
| "loss": 0.0481, |
| "step": 730 |
| }, |
| { |
| "epoch": 6.39, |
| "learning_rate": 2.9666666666666673e-06, |
| "loss": 0.0469, |
| "step": 735 |
| }, |
| { |
| "epoch": 6.43, |
| "learning_rate": 2.9111111111111114e-06, |
| "loss": 0.0412, |
| "step": 740 |
| }, |
| { |
| "epoch": 6.48, |
| "learning_rate": 2.855555555555556e-06, |
| "loss": 0.0325, |
| "step": 745 |
| }, |
| { |
| "epoch": 6.52, |
| "learning_rate": 2.8000000000000003e-06, |
| "loss": 0.0563, |
| "step": 750 |
| }, |
| { |
| "epoch": 6.57, |
| "learning_rate": 2.7444444444444448e-06, |
| "loss": 0.0402, |
| "step": 755 |
| }, |
| { |
| "epoch": 6.61, |
| "learning_rate": 2.6888888888888892e-06, |
| "loss": 0.0781, |
| "step": 760 |
| }, |
| { |
| "epoch": 6.65, |
| "learning_rate": 2.6333333333333332e-06, |
| "loss": 0.0459, |
| "step": 765 |
| }, |
| { |
| "epoch": 6.7, |
| "learning_rate": 2.577777777777778e-06, |
| "loss": 0.0339, |
| "step": 770 |
| }, |
| { |
| "epoch": 6.74, |
| "learning_rate": 2.5222222222222226e-06, |
| "loss": 0.0388, |
| "step": 775 |
| }, |
| { |
| "epoch": 6.78, |
| "learning_rate": 2.466666666666667e-06, |
| "loss": 0.0359, |
| "step": 780 |
| }, |
| { |
| "epoch": 6.83, |
| "learning_rate": 2.411111111111111e-06, |
| "loss": 0.088, |
| "step": 785 |
| }, |
| { |
| "epoch": 6.87, |
| "learning_rate": 2.3555555555555555e-06, |
| "loss": 0.0376, |
| "step": 790 |
| }, |
| { |
| "epoch": 6.91, |
| "learning_rate": 2.3000000000000004e-06, |
| "loss": 0.0361, |
| "step": 795 |
| }, |
| { |
| "epoch": 6.96, |
| "learning_rate": 2.2444444444444445e-06, |
| "loss": 0.0489, |
| "step": 800 |
| }, |
| { |
| "epoch": 7.0, |
| "learning_rate": 2.188888888888889e-06, |
| "loss": 0.0337, |
| "step": 805 |
| }, |
| { |
| "epoch": 7.04, |
| "learning_rate": 2.133333333333334e-06, |
| "loss": 0.0488, |
| "step": 810 |
| }, |
| { |
| "epoch": 7.09, |
| "learning_rate": 2.077777777777778e-06, |
| "loss": 0.0367, |
| "step": 815 |
| }, |
| { |
| "epoch": 7.13, |
| "learning_rate": 2.0222222222222223e-06, |
| "loss": 0.0514, |
| "step": 820 |
| }, |
| { |
| "epoch": 7.17, |
| "learning_rate": 1.9666666666666668e-06, |
| "loss": 0.0274, |
| "step": 825 |
| }, |
| { |
| "epoch": 7.22, |
| "learning_rate": 1.9111111111111112e-06, |
| "loss": 0.0356, |
| "step": 830 |
| }, |
| { |
| "epoch": 7.26, |
| "learning_rate": 1.8555555555555557e-06, |
| "loss": 0.03, |
| "step": 835 |
| }, |
| { |
| "epoch": 7.3, |
| "learning_rate": 1.8000000000000001e-06, |
| "loss": 0.0339, |
| "step": 840 |
| }, |
| { |
| "epoch": 7.35, |
| "learning_rate": 1.7444444444444448e-06, |
| "loss": 0.0269, |
| "step": 845 |
| }, |
| { |
| "epoch": 7.39, |
| "learning_rate": 1.688888888888889e-06, |
| "loss": 0.0379, |
| "step": 850 |
| }, |
| { |
| "epoch": 7.43, |
| "learning_rate": 1.6333333333333335e-06, |
| "loss": 0.0584, |
| "step": 855 |
| }, |
| { |
| "epoch": 7.48, |
| "learning_rate": 1.5777777777777778e-06, |
| "loss": 0.0282, |
| "step": 860 |
| }, |
| { |
| "epoch": 7.52, |
| "learning_rate": 1.5222222222222224e-06, |
| "loss": 0.0304, |
| "step": 865 |
| }, |
| { |
| "epoch": 7.57, |
| "learning_rate": 1.4666666666666669e-06, |
| "loss": 0.0289, |
| "step": 870 |
| }, |
| { |
| "epoch": 7.61, |
| "learning_rate": 1.4111111111111111e-06, |
| "loss": 0.0393, |
| "step": 875 |
| }, |
| { |
| "epoch": 7.65, |
| "learning_rate": 1.3555555555555558e-06, |
| "loss": 0.0318, |
| "step": 880 |
| }, |
| { |
| "epoch": 7.7, |
| "learning_rate": 1.3e-06, |
| "loss": 0.0704, |
| "step": 885 |
| }, |
| { |
| "epoch": 7.74, |
| "learning_rate": 1.2444444444444445e-06, |
| "loss": 0.0261, |
| "step": 890 |
| }, |
| { |
| "epoch": 7.78, |
| "learning_rate": 1.188888888888889e-06, |
| "loss": 0.0389, |
| "step": 895 |
| }, |
| { |
| "epoch": 7.83, |
| "learning_rate": 1.1333333333333334e-06, |
| "loss": 0.031, |
| "step": 900 |
| }, |
| { |
| "epoch": 7.87, |
| "learning_rate": 1.0777777777777779e-06, |
| "loss": 0.0261, |
| "step": 905 |
| }, |
| { |
| "epoch": 7.91, |
| "learning_rate": 1.0222222222222223e-06, |
| "loss": 0.0349, |
| "step": 910 |
| }, |
| { |
| "epoch": 7.96, |
| "learning_rate": 9.666666666666668e-07, |
| "loss": 0.0326, |
| "step": 915 |
| }, |
| { |
| "epoch": 8.0, |
| "learning_rate": 9.111111111111113e-07, |
| "loss": 0.0742, |
| "step": 920 |
| }, |
| { |
| "epoch": 8.04, |
| "learning_rate": 8.555555555555556e-07, |
| "loss": 0.0343, |
| "step": 925 |
| }, |
| { |
| "epoch": 8.09, |
| "learning_rate": 8.000000000000001e-07, |
| "loss": 0.0348, |
| "step": 930 |
| }, |
| { |
| "epoch": 8.13, |
| "learning_rate": 7.444444444444444e-07, |
| "loss": 0.0293, |
| "step": 935 |
| }, |
| { |
| "epoch": 8.17, |
| "learning_rate": 6.88888888888889e-07, |
| "loss": 0.0357, |
| "step": 940 |
| }, |
| { |
| "epoch": 8.22, |
| "learning_rate": 6.333333333333334e-07, |
| "loss": 0.024, |
| "step": 945 |
| }, |
| { |
| "epoch": 8.26, |
| "learning_rate": 5.777777777777778e-07, |
| "loss": 0.0336, |
| "step": 950 |
| }, |
| { |
| "epoch": 8.3, |
| "learning_rate": 5.222222222222223e-07, |
| "loss": 0.0202, |
| "step": 955 |
| }, |
| { |
| "epoch": 8.35, |
| "learning_rate": 4.666666666666667e-07, |
| "loss": 0.0379, |
| "step": 960 |
| }, |
| { |
| "epoch": 8.39, |
| "learning_rate": 4.111111111111112e-07, |
| "loss": 0.0388, |
| "step": 965 |
| }, |
| { |
| "epoch": 8.43, |
| "learning_rate": 3.555555555555556e-07, |
| "loss": 0.0287, |
| "step": 970 |
| }, |
| { |
| "epoch": 8.48, |
| "learning_rate": 3.0000000000000004e-07, |
| "loss": 0.021, |
| "step": 975 |
| }, |
| { |
| "epoch": 8.52, |
| "learning_rate": 2.444444444444445e-07, |
| "loss": 0.0342, |
| "step": 980 |
| }, |
| { |
| "epoch": 8.57, |
| "learning_rate": 1.888888888888889e-07, |
| "loss": 0.0329, |
| "step": 985 |
| }, |
| { |
| "epoch": 8.61, |
| "learning_rate": 1.3333333333333336e-07, |
| "loss": 0.0517, |
| "step": 990 |
| }, |
| { |
| "epoch": 8.65, |
| "learning_rate": 7.777777777777778e-08, |
| "loss": 0.024, |
| "step": 995 |
| }, |
| { |
| "epoch": 8.7, |
| "learning_rate": 2.2222222222222224e-08, |
| "loss": 0.0259, |
| "step": 1000 |
| }, |
| { |
| "epoch": 8.7, |
| "step": 1000, |
| "total_flos": 1850846955153408.0, |
| "train_loss": 0.15014069636166097, |
| "train_runtime": 142.3624, |
| "train_samples_per_second": 56.195, |
| "train_steps_per_second": 7.024 |
| } |
| ], |
| "max_steps": 1000, |
| "num_train_epochs": 9, |
| "total_flos": 1850846955153408.0, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|