| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 7497, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.999997804999731e-05, | |
| "loss": 2.5926, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9999912200085595e-05, | |
| "loss": 2.726, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.999980245055394e-05, | |
| "loss": 2.7046, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9999648801884152e-05, | |
| "loss": 2.5412, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9999451254750736e-05, | |
| "loss": 2.4895, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.999920981002093e-05, | |
| "loss": 2.5147, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9998924468754672e-05, | |
| "loss": 2.3979, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9998595232204618e-05, | |
| "loss": 2.3704, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9998222101816113e-05, | |
| "loss": 2.3891, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.99978050792272e-05, | |
| "loss": 2.3385, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.999734416626861e-05, | |
| "loss": 2.3395, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9996839364963744e-05, | |
| "loss": 2.2868, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9996290677528688e-05, | |
| "loss": 2.2824, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9995698106372176e-05, | |
| "loss": 2.2458, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.99950616540956e-05, | |
| "loss": 2.2594, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.999438132349298e-05, | |
| "loss": 2.2887, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9993657117550972e-05, | |
| "loss": 2.2976, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9992889039448838e-05, | |
| "loss": 2.2611, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9992077092558443e-05, | |
| "loss": 2.2257, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9991221280444237e-05, | |
| "loss": 2.3287, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9990321606863224e-05, | |
| "loss": 2.3082, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9989378075764988e-05, | |
| "loss": 2.251, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.998839069129162e-05, | |
| "loss": 2.2239, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.998735945777774e-05, | |
| "loss": 2.2154, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9986284379750462e-05, | |
| "loss": 2.2939, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9985165461929385e-05, | |
| "loss": 2.2763, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9984002709226558e-05, | |
| "loss": 2.1672, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.998279612674646e-05, | |
| "loss": 2.1404, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9981545719785994e-05, | |
| "loss": 2.1851, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9980251493834447e-05, | |
| "loss": 2.1481, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.997891345457347e-05, | |
| "loss": 2.1352, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9977531607877055e-05, | |
| "loss": 2.2454, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9976105959811512e-05, | |
| "loss": 2.3146, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9974636516635436e-05, | |
| "loss": 2.208, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9973123284799683e-05, | |
| "loss": 2.1468, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.997156627094734e-05, | |
| "loss": 2.1792, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9969965481913704e-05, | |
| "loss": 2.3303, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9968320924726236e-05, | |
| "loss": 2.2114, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9966632606604543e-05, | |
| "loss": 2.2289, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.996490053496034e-05, | |
| "loss": 2.2365, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9963124717397425e-05, | |
| "loss": 2.1721, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9961305161711643e-05, | |
| "loss": 2.1837, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9959441875890834e-05, | |
| "loss": 2.1694, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9957534868114835e-05, | |
| "loss": 2.1377, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.99555841467554e-05, | |
| "loss": 2.1963, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9953589720376204e-05, | |
| "loss": 2.2826, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.995155159773278e-05, | |
| "loss": 2.2644, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9949469787772484e-05, | |
| "loss": 2.1792, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9947344299634464e-05, | |
| "loss": 2.2485, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9945175142649617e-05, | |
| "loss": 2.2065, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.994296232634054e-05, | |
| "loss": 2.2887, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9940705860421493e-05, | |
| "loss": 2.1542, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9938405754798377e-05, | |
| "loss": 2.1692, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9936062019568646e-05, | |
| "loss": 2.1624, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9933674665021295e-05, | |
| "loss": 2.19, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9931243701636826e-05, | |
| "loss": 2.1613, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9928769140087157e-05, | |
| "loss": 2.2509, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.992625099123562e-05, | |
| "loss": 2.1907, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.992368926613689e-05, | |
| "loss": 2.1528, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9921083976036937e-05, | |
| "loss": 2.1551, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9918435132372992e-05, | |
| "loss": 2.2466, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.991574274677348e-05, | |
| "loss": 2.2388, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9913006831057967e-05, | |
| "loss": 2.1348, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9910227397237136e-05, | |
| "loss": 2.1659, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9907404457512693e-05, | |
| "loss": 2.1995, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.990453802427735e-05, | |
| "loss": 2.2362, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.990162811011475e-05, | |
| "loss": 2.1695, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9898674727799418e-05, | |
| "loss": 2.2246, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9895677890296704e-05, | |
| "loss": 2.1766, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9892637610762723e-05, | |
| "loss": 2.1765, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9889553902544307e-05, | |
| "loss": 2.251, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9886426779178935e-05, | |
| "loss": 2.1699, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.988325625439468e-05, | |
| "loss": 2.2406, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.988004234211015e-05, | |
| "loss": 2.1664, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9876785056434422e-05, | |
| "loss": 2.1711, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.987348441166698e-05, | |
| "loss": 2.1874, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9870140422297655e-05, | |
| "loss": 2.1686, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9866753103006564e-05, | |
| "loss": 2.2158, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9863322468664038e-05, | |
| "loss": 2.147, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.985984853433057e-05, | |
| "loss": 2.1805, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9856331315256728e-05, | |
| "loss": 2.2011, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9852770826883106e-05, | |
| "loss": 2.0972, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.984916708484025e-05, | |
| "loss": 2.243, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.984552010494859e-05, | |
| "loss": 2.2098, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9841829903218377e-05, | |
| "loss": 2.1151, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9838096495849582e-05, | |
| "loss": 2.2735, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9834319899231884e-05, | |
| "loss": 2.1943, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.983050012994453e-05, | |
| "loss": 2.2244, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9826637204756312e-05, | |
| "loss": 2.178, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.982273114062548e-05, | |
| "loss": 2.1652, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9818781954699656e-05, | |
| "loss": 2.2102, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9814789664315762e-05, | |
| "loss": 2.1286, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.981075428699996e-05, | |
| "loss": 2.2028, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.980667584046756e-05, | |
| "loss": 2.122, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.980255434262294e-05, | |
| "loss": 2.0737, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.979838981155948e-05, | |
| "loss": 2.1987, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.979418226555947e-05, | |
| "loss": 2.2296, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9789931723094046e-05, | |
| "loss": 2.1883, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.978563820282309e-05, | |
| "loss": 2.0761, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.9781301723595154e-05, | |
| "loss": 2.1523, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.977692230444739e-05, | |
| "loss": 2.2492, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.977249996460544e-05, | |
| "loss": 2.179, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.976803472348339e-05, | |
| "loss": 2.1714, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.976352660068365e-05, | |
| "loss": 2.1702, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.9758975615996874e-05, | |
| "loss": 2.0948, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.975438178940189e-05, | |
| "loss": 2.1055, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.97497451410656e-05, | |
| "loss": 2.1615, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.97450656913429e-05, | |
| "loss": 2.1442, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9740343460776564e-05, | |
| "loss": 2.1537, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9735578470097195e-05, | |
| "loss": 2.2454, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.9730770740223098e-05, | |
| "loss": 2.2078, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.972592029226022e-05, | |
| "loss": 2.1602, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.972102714750202e-05, | |
| "loss": 2.229, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.971609132742941e-05, | |
| "loss": 2.1704, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9711112853710645e-05, | |
| "loss": 2.2131, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9706091748201227e-05, | |
| "loss": 2.2287, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.9701028032943805e-05, | |
| "loss": 2.2162, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.96959217301681e-05, | |
| "loss": 2.2365, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.969077286229078e-05, | |
| "loss": 2.1714, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9685581451915382e-05, | |
| "loss": 2.1122, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.968034752183219e-05, | |
| "loss": 2.1883, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.9675071095018175e-05, | |
| "loss": 2.1136, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.966975219463684e-05, | |
| "loss": 2.2358, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9664390844038168e-05, | |
| "loss": 2.1575, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.965898706675849e-05, | |
| "loss": 2.176, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.9653540886520387e-05, | |
| "loss": 2.0719, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.96480523272326e-05, | |
| "loss": 2.1178, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9642521412989902e-05, | |
| "loss": 2.1896, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.963694816807301e-05, | |
| "loss": 2.1458, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9631332616948472e-05, | |
| "loss": 2.1789, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9625674784268564e-05, | |
| "loss": 2.1564, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.9619974694871173e-05, | |
| "loss": 2.1995, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9614232373779692e-05, | |
| "loss": 2.1413, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9608447846202918e-05, | |
| "loss": 2.1179, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9602621137534923e-05, | |
| "loss": 2.1635, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.959675227335497e-05, | |
| "loss": 2.187, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.9590841279427368e-05, | |
| "loss": 2.2399, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9584888181701388e-05, | |
| "loss": 2.1589, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9578893006311135e-05, | |
| "loss": 2.1393, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9572855779575427e-05, | |
| "loss": 2.1048, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9566776527997694e-05, | |
| "loss": 2.1851, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.9560655278265852e-05, | |
| "loss": 2.1909, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9554492057252196e-05, | |
| "loss": 2.1493, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9548286892013262e-05, | |
| "loss": 2.1845, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9542039809789736e-05, | |
| "loss": 2.2492, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9535750838006307e-05, | |
| "loss": 2.1798, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.9529420004271568e-05, | |
| "loss": 2.1765, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.952304733637788e-05, | |
| "loss": 2.1119, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9516632862301258e-05, | |
| "loss": 2.0904, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9510176610201254e-05, | |
| "loss": 2.1828, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.950367860842081e-05, | |
| "loss": 2.1519, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.9497138885486157e-05, | |
| "loss": 2.0773, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.949055747010669e-05, | |
| "loss": 2.1506, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9483934391174816e-05, | |
| "loss": 2.2158, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9477269677765863e-05, | |
| "loss": 2.1949, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9470563359137923e-05, | |
| "loss": 2.1424, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9463815464731737e-05, | |
| "loss": 2.2106, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.945702602417057e-05, | |
| "loss": 2.1789, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.945019506726007e-05, | |
| "loss": 2.0718, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9443322623988133e-05, | |
| "loss": 2.1993, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9436408724524794e-05, | |
| "loss": 2.1222, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9429453399222077e-05, | |
| "loss": 2.2176, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.942245667861386e-05, | |
| "loss": 2.1398, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9415418593415755e-05, | |
| "loss": 2.1975, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9408339174524957e-05, | |
| "loss": 2.0749, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9401218453020118e-05, | |
| "loss": 2.1371, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.939405646016121e-05, | |
| "loss": 2.162, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.938685322738939e-05, | |
| "loss": 2.1034, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9379608786326843e-05, | |
| "loss": 2.1339, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.937232316877668e-05, | |
| "loss": 2.1118, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.936499640672276e-05, | |
| "loss": 2.1367, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9357628532329574e-05, | |
| "loss": 2.1633, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9350219577942095e-05, | |
| "loss": 2.0912, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9342769576085633e-05, | |
| "loss": 2.1311, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9335278559465703e-05, | |
| "loss": 2.1007, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9327746560967878e-05, | |
| "loss": 2.1426, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.9320173613657628e-05, | |
| "loss": 2.1446, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9312559750780196e-05, | |
| "loss": 2.1508, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9304905005760447e-05, | |
| "loss": 2.1822, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9297209412202712e-05, | |
| "loss": 2.0216, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9289473003890658e-05, | |
| "loss": 2.2102, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.9281695814787113e-05, | |
| "loss": 2.0846, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9273877879033946e-05, | |
| "loss": 2.1596, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.92660192309519e-05, | |
| "loss": 2.2179, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9258119905040443e-05, | |
| "loss": 2.1029, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9250179935977617e-05, | |
| "loss": 2.1343, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9242199358619897e-05, | |
| "loss": 2.0859, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.9234178208002014e-05, | |
| "loss": 2.1138, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.922611651933683e-05, | |
| "loss": 2.0537, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.921801432801516e-05, | |
| "loss": 2.1042, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.920987166960563e-05, | |
| "loss": 2.1944, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.920168857985451e-05, | |
| "loss": 2.0814, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9193465094685577e-05, | |
| "loss": 2.1275, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9185201250199926e-05, | |
| "loss": 2.0843, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9176897082675845e-05, | |
| "loss": 2.119, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9168552628568632e-05, | |
| "loss": 2.1449, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.916016792451044e-05, | |
| "loss": 2.1665, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9151743007310134e-05, | |
| "loss": 2.0731, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.91432779139531e-05, | |
| "loss": 2.1782, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9134772681601102e-05, | |
| "loss": 2.2128, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9126227347592113e-05, | |
| "loss": 2.1159, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9117641949440156e-05, | |
| "loss": 2.1405, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9109016524835136e-05, | |
| "loss": 2.1809, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9100351111642666e-05, | |
| "loss": 2.092, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.909164574790392e-05, | |
| "loss": 2.1187, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9082900471835448e-05, | |
| "loss": 2.1296, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9074115321829015e-05, | |
| "loss": 2.0347, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9065290336451435e-05, | |
| "loss": 2.1343, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.90564255544444e-05, | |
| "loss": 2.1648, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9047521014724303e-05, | |
| "loss": 2.0968, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9038576756382084e-05, | |
| "loss": 2.1439, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.9029592818683042e-05, | |
| "loss": 2.1372, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9020569241066665e-05, | |
| "loss": 2.0636, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.901150606314646e-05, | |
| "loss": 2.1161, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.9002403324709793e-05, | |
| "loss": 2.1448, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.8993261065717685e-05, | |
| "loss": 2.1542, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.8984079326304653e-05, | |
| "loss": 2.1861, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.8974858146778547e-05, | |
| "loss": 2.1522, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.8965597567620344e-05, | |
| "loss": 2.1143, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.8956297629483995e-05, | |
| "loss": 2.1022, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.894695837319623e-05, | |
| "loss": 2.1759, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.893757983975639e-05, | |
| "loss": 2.1132, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.892816207033625e-05, | |
| "loss": 2.1244, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.8918705106279812e-05, | |
| "loss": 2.1174, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.8909208989103155e-05, | |
| "loss": 2.1678, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.889967376049424e-05, | |
| "loss": 2.1666, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.8890099462312732e-05, | |
| "loss": 2.0998, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.8880486136589794e-05, | |
| "loss": 2.1457, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.887083382552794e-05, | |
| "loss": 2.1633, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.8861142571500813e-05, | |
| "loss": 2.1287, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.8851412417053033e-05, | |
| "loss": 2.0356, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.8841643404899974e-05, | |
| "loss": 2.1444, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.883183557792761e-05, | |
| "loss": 2.1141, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.8821988979192307e-05, | |
| "loss": 2.1039, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.8812103651920633e-05, | |
| "loss": 2.1084, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.8802179639509188e-05, | |
| "loss": 2.1054, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.879221698552439e-05, | |
| "loss": 2.0563, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.8782215733702286e-05, | |
| "loss": 2.2365, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.877217592794839e-05, | |
| "loss": 2.177, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.8762097612337444e-05, | |
| "loss": 2.1924, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.875198083111327e-05, | |
| "loss": 2.1563, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.8741825628688534e-05, | |
| "loss": 2.0484, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.873163204964458e-05, | |
| "loss": 2.0747, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.872140013873123e-05, | |
| "loss": 2.0507, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.8711129940866577e-05, | |
| "loss": 2.1103, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.8700821501136797e-05, | |
| "loss": 2.1322, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.8690474864795942e-05, | |
| "loss": 2.0776, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.8680090077265753e-05, | |
| "loss": 2.1604, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.8669667184135454e-05, | |
| "loss": 2.1689, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.8659206231161548e-05, | |
| "loss": 2.0437, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.8648707264267628e-05, | |
| "loss": 2.1611, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.8638170329544164e-05, | |
| "loss": 2.1317, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8627595473248303e-05, | |
| "loss": 2.0664, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.861698274180367e-05, | |
| "loss": 2.0979, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8606332181800165e-05, | |
| "loss": 2.0975, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8595643839993747e-05, | |
| "loss": 2.0339, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.8584917763306246e-05, | |
| "loss": 2.1365, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8574153998825145e-05, | |
| "loss": 2.1832, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.856335259380337e-05, | |
| "loss": 2.1306, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8552513595659103e-05, | |
| "loss": 2.1845, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8541637051975544e-05, | |
| "loss": 2.1158, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.8530723010500733e-05, | |
| "loss": 2.1405, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8519771519147312e-05, | |
| "loss": 2.0784, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.850878262599234e-05, | |
| "loss": 2.1533, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8497756379277058e-05, | |
| "loss": 2.1382, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8486692827406694e-05, | |
| "loss": 2.129, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.8475592018950256e-05, | |
| "loss": 2.1339, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.846445400264029e-05, | |
| "loss": 2.161, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.8453278827372697e-05, | |
| "loss": 2.0769, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.84420665422065e-05, | |
| "loss": 2.07, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.843081719636364e-05, | |
| "loss": 2.2284, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.841953083922875e-05, | |
| "loss": 2.0727, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8408207520348945e-05, | |
| "loss": 2.1778, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.83968472894336e-05, | |
| "loss": 2.0996, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.838545019635413e-05, | |
| "loss": 2.0714, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8374016291143792e-05, | |
| "loss": 2.1285, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.8362545623997428e-05, | |
| "loss": 2.0806, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.835103824527127e-05, | |
| "loss": 2.079, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8339494205482722e-05, | |
| "loss": 2.1132, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8327913555310125e-05, | |
| "loss": 2.0798, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8316296345592534e-05, | |
| "loss": 2.1899, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.8304642627329513e-05, | |
| "loss": 2.0772, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8292952451680886e-05, | |
| "loss": 2.1043, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8281225869966532e-05, | |
| "loss": 2.1246, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8269462933666157e-05, | |
| "loss": 2.026, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8257663694419045e-05, | |
| "loss": 2.1274, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.8245828204023875e-05, | |
| "loss": 2.0544, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.823395651443845e-05, | |
| "loss": 2.0656, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8222048677779495e-05, | |
| "loss": 2.164, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.821010474632242e-05, | |
| "loss": 2.0952, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.8198124772501085e-05, | |
| "loss": 2.1039, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.818610880890759e-05, | |
| "loss": 2.0748, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.817405690829202e-05, | |
| "loss": 2.0959, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.816196912356222e-05, | |
| "loss": 2.1587, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.8149845507783574e-05, | |
| "loss": 2.0942, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.8137686114178763e-05, | |
| "loss": 2.1103, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.812549099612753e-05, | |
| "loss": 2.2153, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8113260207166453e-05, | |
| "loss": 2.0573, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8100993800988696e-05, | |
| "loss": 2.0931, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8088691831443798e-05, | |
| "loss": 2.0398, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8076354352537402e-05, | |
| "loss": 2.0999, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.806398141843105e-05, | |
| "loss": 2.1188, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8051573083441935e-05, | |
| "loss": 2.1364, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8039129402042648e-05, | |
| "loss": 2.05, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8026650428860958e-05, | |
| "loss": 2.0816, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8014136218679566e-05, | |
| "loss": 2.1499, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.8001586826435863e-05, | |
| "loss": 2.2182, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.7989002307221685e-05, | |
| "loss": 2.0472, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.7976382716283075e-05, | |
| "loss": 2.1268, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.7963728109020047e-05, | |
| "loss": 2.1228, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.795103854098634e-05, | |
| "loss": 2.0701, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.7938314067889154e-05, | |
| "loss": 2.1223, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.7925554745588938e-05, | |
| "loss": 2.0372, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.791276063009912e-05, | |
| "loss": 2.0967, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.789993177758588e-05, | |
| "loss": 2.0381, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.7887068244367887e-05, | |
| "loss": 2.1974, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.787417008691605e-05, | |
| "loss": 2.1828, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7861237361853297e-05, | |
| "loss": 2.1431, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7848270125954295e-05, | |
| "loss": 2.0985, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7835268436145217e-05, | |
| "loss": 2.1519, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7822232349503485e-05, | |
| "loss": 2.1818, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.780916192325753e-05, | |
| "loss": 2.1099, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.779605721478652e-05, | |
| "loss": 2.1819, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.778291828162015e-05, | |
| "loss": 2.1122, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7769745181438333e-05, | |
| "loss": 2.1088, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.775653797207099e-05, | |
| "loss": 2.1283, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.774329671149777e-05, | |
| "loss": 2.1591, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7730021457847822e-05, | |
| "loss": 2.0889, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7716712269399512e-05, | |
| "loss": 2.1279, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7703369204580187e-05, | |
| "loss": 2.1395, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7689992321965907e-05, | |
| "loss": 2.1582, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7676581680281194e-05, | |
| "loss": 2.0378, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.766313733839877e-05, | |
| "loss": 2.0885, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7649659355339313e-05, | |
| "loss": 2.1027, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7636147790271163e-05, | |
| "loss": 2.1033, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7622602702510104e-05, | |
| "loss": 2.1079, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7609024151519082e-05, | |
| "loss": 1.9935, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.7595412196907937e-05, | |
| "loss": 2.1249, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.758176689843316e-05, | |
| "loss": 2.1409, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.756808831599762e-05, | |
| "loss": 2.0591, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.75543765096503e-05, | |
| "loss": 2.0628, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.754063153958603e-05, | |
| "loss": 2.1537, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.7526853466145248e-05, | |
| "loss": 2.1149, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.7513042349813692e-05, | |
| "loss": 2.1868, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.7499198251222184e-05, | |
| "loss": 2.1345, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.7485321231146314e-05, | |
| "loss": 2.1213, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.747141135050621e-05, | |
| "loss": 2.0826, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7457468670366258e-05, | |
| "loss": 2.1221, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7443493251934826e-05, | |
| "loss": 2.1497, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7429485156564014e-05, | |
| "loss": 1.9985, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.7415444445749367e-05, | |
| "loss": 2.1146, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.740137118112961e-05, | |
| "loss": 2.0794, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7387265424486384e-05, | |
| "loss": 2.1051, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7373127237743966e-05, | |
| "loss": 2.0684, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7358956682969012e-05, | |
| "loss": 2.1082, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7344753822370253e-05, | |
| "loss": 2.1632, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.7330518718298263e-05, | |
| "loss": 2.1144, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7316251433245157e-05, | |
| "loss": 2.1013, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.7301952029844318e-05, | |
| "loss": 2.0976, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.728762057087014e-05, | |
| "loss": 2.1355, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.727325711923773e-05, | |
| "loss": 2.1501, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.725886173800265e-05, | |
| "loss": 2.1784, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.724443449036064e-05, | |
| "loss": 2.1633, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7229975439647316e-05, | |
| "loss": 2.0991, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7215484649337918e-05, | |
| "loss": 2.1432, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7200962183047028e-05, | |
| "loss": 2.0887, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.7186408104528278e-05, | |
| "loss": 2.168, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.717182247767408e-05, | |
| "loss": 2.1152, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7157205366515347e-05, | |
| "loss": 2.0208, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.71425568352212e-05, | |
| "loss": 2.0639, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.7127876948098706e-05, | |
| "loss": 2.1513, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.711316576959257e-05, | |
| "loss": 2.0711, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7098423364284878e-05, | |
| "loss": 2.0285, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7083649796894798e-05, | |
| "loss": 2.1167, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7068845132278294e-05, | |
| "loss": 2.1325, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7054009435427855e-05, | |
| "loss": 2.1438, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.7039142771472197e-05, | |
| "loss": 2.1716, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.7024245205675986e-05, | |
| "loss": 2.0852, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.700931680343954e-05, | |
| "loss": 2.0742, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.6994357630298558e-05, | |
| "loss": 2.1573, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.697936775192381e-05, | |
| "loss": 2.1373, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.6964347234120874e-05, | |
| "loss": 2.1254, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.6949296142829833e-05, | |
| "loss": 2.1806, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.6934214544124986e-05, | |
| "loss": 2.0781, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.6919102504214557e-05, | |
| "loss": 2.1924, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.690396008944041e-05, | |
| "loss": 2.0889, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.6888787366277752e-05, | |
| "loss": 2.0604, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.6873584401334856e-05, | |
| "loss": 2.102, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.685835126135273e-05, | |
| "loss": 2.0902, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.684308801320488e-05, | |
| "loss": 2.1354, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.6827794723896968e-05, | |
| "loss": 2.0281, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.681247146056654e-05, | |
| "loss": 2.1993, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.6797118290482737e-05, | |
| "loss": 2.0583, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.678173528104598e-05, | |
| "loss": 2.1452, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.676632249978769e-05, | |
| "loss": 2.0835, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.6750880014369983e-05, | |
| "loss": 2.0589, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.6735407892585378e-05, | |
| "loss": 2.0813, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.6719906202356502e-05, | |
| "loss": 2.0503, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.670437501173578e-05, | |
| "loss": 2.0837, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.6688814388905146e-05, | |
| "loss": 2.081, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.6673224402175743e-05, | |
| "loss": 2.2051, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.6657605119987626e-05, | |
| "loss": 2.137, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.6641956610909453e-05, | |
| "loss": 2.0797, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.6626278943638177e-05, | |
| "loss": 2.1248, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.6610572186998772e-05, | |
| "loss": 2.075, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.659483640994391e-05, | |
| "loss": 2.1157, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.657907168155366e-05, | |
| "loss": 2.0207, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6563278071035182e-05, | |
| "loss": 2.1333, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6547455647722443e-05, | |
| "loss": 2.1026, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6531604481075886e-05, | |
| "loss": 2.0887, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6515724640682146e-05, | |
| "loss": 2.1504, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.649981619625372e-05, | |
| "loss": 2.0906, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6483879217628696e-05, | |
| "loss": 2.0802, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6467913774770416e-05, | |
| "loss": 2.093, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6451919937767184e-05, | |
| "loss": 2.0915, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.6435897776831956e-05, | |
| "loss": 2.1309, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.641984736230202e-05, | |
| "loss": 2.1177, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6403768764638708e-05, | |
| "loss": 2.0697, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6387662054427073e-05, | |
| "loss": 2.115, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.637152730237558e-05, | |
| "loss": 2.1686, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.63553645793158e-05, | |
| "loss": 2.1476, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.6339173956202098e-05, | |
| "loss": 2.0028, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.632295550411132e-05, | |
| "loss": 2.15, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.630670929424247e-05, | |
| "loss": 2.0369, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6290435397916426e-05, | |
| "loss": 2.0505, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6274133886575597e-05, | |
| "loss": 2.1644, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.6257804831783633e-05, | |
| "loss": 2.1103, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6241448305225085e-05, | |
| "loss": 2.073, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6225064378705124e-05, | |
| "loss": 2.0924, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6208653124149183e-05, | |
| "loss": 2.1405, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6192214613602687e-05, | |
| "loss": 2.0792, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.6175748919230708e-05, | |
| "loss": 2.1245, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.6159256113317646e-05, | |
| "loss": 2.0774, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.614273626826693e-05, | |
| "loss": 2.0218, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.6126189456600694e-05, | |
| "loss": 2.189, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.6109615750959445e-05, | |
| "loss": 2.0766, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.609301522410176e-05, | |
| "loss": 2.0824, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6076387948903965e-05, | |
| "loss": 2.0341, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.60597339983598e-05, | |
| "loss": 2.1769, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.604305344558012e-05, | |
| "loss": 2.0102, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6026346363792565e-05, | |
| "loss": 2.1207, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.6009612826341226e-05, | |
| "loss": 2.1376, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.5992852906686346e-05, | |
| "loss": 2.0659, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.597606667840398e-05, | |
| "loss": 2.1084, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.5959254215185683e-05, | |
| "loss": 2.0434, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.594241559083817e-05, | |
| "loss": 2.0074, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.5925550879283017e-05, | |
| "loss": 2.0913, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.5908660154556313e-05, | |
| "loss": 2.0843, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.589174349080835e-05, | |
| "loss": 2.0662, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.587480096230329e-05, | |
| "loss": 2.0527, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.5857832643418844e-05, | |
| "loss": 2.0494, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.5840838608645943e-05, | |
| "loss": 2.121, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.5823818932588406e-05, | |
| "loss": 2.1162, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.580677368996262e-05, | |
| "loss": 2.1505, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.578970295559721e-05, | |
| "loss": 2.1857, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.5772606804432704e-05, | |
| "loss": 2.159, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.5755485311521223e-05, | |
| "loss": 2.0198, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.5738338552026126e-05, | |
| "loss": 2.0953, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.5721166601221697e-05, | |
| "loss": 2.0743, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.5703969534492812e-05, | |
| "loss": 2.0906, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.56867474273346e-05, | |
| "loss": 2.0602, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.5669500355352117e-05, | |
| "loss": 2.1485, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.5652228394260026e-05, | |
| "loss": 2.1229, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.563493161988224e-05, | |
| "loss": 2.1125, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.561761010815161e-05, | |
| "loss": 2.2295, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.5600263935109584e-05, | |
| "loss": 2.0598, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.5582893176905867e-05, | |
| "loss": 2.1363, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.55654979097981e-05, | |
| "loss": 2.0272, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.554807821015151e-05, | |
| "loss": 2.0747, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.5530634154438592e-05, | |
| "loss": 2.0505, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.5513165819238764e-05, | |
| "loss": 2.0423, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.5495673281238016e-05, | |
| "loss": 2.0228, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.547815661722861e-05, | |
| "loss": 2.0991, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5460615904108704e-05, | |
| "loss": 2.0544, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5443051218882042e-05, | |
| "loss": 2.0873, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.5425462638657597e-05, | |
| "loss": 2.0415, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.540785024064925e-05, | |
| "loss": 2.1257, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5390214102175432e-05, | |
| "loss": 2.0762, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5372554300658807e-05, | |
| "loss": 2.0193, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5354870913625907e-05, | |
| "loss": 2.0364, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5337164018706822e-05, | |
| "loss": 2.1337, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.5319433693634814e-05, | |
| "loss": 2.0862, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.530168001624603e-05, | |
| "loss": 2.1117, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5283903064479125e-05, | |
| "loss": 2.1142, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5266102916374918e-05, | |
| "loss": 2.07, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.524827965007608e-05, | |
| "loss": 2.0325, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5230433343826748e-05, | |
| "loss": 2.106, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.521256407597222e-05, | |
| "loss": 2.0486, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5194671924958597e-05, | |
| "loss": 2.0351, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5176756969332428e-05, | |
| "loss": 2.1353, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5158819287740372e-05, | |
| "loss": 2.097, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5140858958928872e-05, | |
| "loss": 2.0776, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5122876061743772e-05, | |
| "loss": 2.1002, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5104870675130006e-05, | |
| "loss": 2.0693, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5086842878131224e-05, | |
| "loss": 1.9931, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5068792749889473e-05, | |
| "loss": 2.0622, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.5050720369644822e-05, | |
| "loss": 2.0894, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.5032625816735027e-05, | |
| "loss": 1.9941, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.5014509170595193e-05, | |
| "loss": 2.0138, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.4996370510757397e-05, | |
| "loss": 2.1484, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.4978209916850368e-05, | |
| "loss": 2.0602, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.4960027468599128e-05, | |
| "loss": 2.1208, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.494182324582463e-05, | |
| "loss": 2.0889, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.4923597328443423e-05, | |
| "loss": 2.0358, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.4905349796467295e-05, | |
| "loss": 2.11, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.4887080730002918e-05, | |
| "loss": 2.0893, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.4868790209251509e-05, | |
| "loss": 2.0928, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.4850478314508459e-05, | |
| "loss": 2.0905, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.4832145126162996e-05, | |
| "loss": 2.0338, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.4813790724697832e-05, | |
| "loss": 2.07, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.479541519068879e-05, | |
| "loss": 2.0586, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.4777018604804484e-05, | |
| "loss": 2.064, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.4758601047805934e-05, | |
| "loss": 2.0792, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.4740162600546217e-05, | |
| "loss": 2.0141, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.4721703343970137e-05, | |
| "loss": 2.1187, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.4703223359113836e-05, | |
| "loss": 2.1028, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.4684722727104455e-05, | |
| "loss": 1.9526, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.4666201529159781e-05, | |
| "loss": 2.1122, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.464765984658788e-05, | |
| "loss": 2.0869, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.4629097760786755e-05, | |
| "loss": 2.0335, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.4610515353243964e-05, | |
| "loss": 2.1524, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.4591912705536294e-05, | |
| "loss": 2.0777, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.4573289899329374e-05, | |
| "loss": 2.1048, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.4554647016377334e-05, | |
| "loss": 2.1319, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.4535984138522442e-05, | |
| "loss": 1.9869, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.4517301347694743e-05, | |
| "loss": 1.9912, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.4498598725911693e-05, | |
| "loss": 2.0666, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4479876355277815e-05, | |
| "loss": 2.1449, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4461134317984326e-05, | |
| "loss": 2.1128, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4442372696308782e-05, | |
| "loss": 2.1342, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4423591572614705e-05, | |
| "loss": 2.0594, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.4404791029351249e-05, | |
| "loss": 2.098, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4385971149052803e-05, | |
| "loss": 2.0177, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4367132014338648e-05, | |
| "loss": 2.0658, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4348273707912601e-05, | |
| "loss": 2.1465, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.4329396312562638e-05, | |
| "loss": 2.0353, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.431049991116053e-05, | |
| "loss": 2.1074, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4291584586661494e-05, | |
| "loss": 2.0332, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.427265042210381e-05, | |
| "loss": 2.1348, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4253697500608475e-05, | |
| "loss": 2.0878, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4234725905378826e-05, | |
| "loss": 2.0234, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.4215735719700167e-05, | |
| "loss": 2.2253, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4196727026939434e-05, | |
| "loss": 2.0872, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4177699910544793e-05, | |
| "loss": 2.0715, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4158654454045292e-05, | |
| "loss": 2.0865, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4139590741050505e-05, | |
| "loss": 2.1196, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.4120508855250134e-05, | |
| "loss": 2.0734, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4101408880413672e-05, | |
| "loss": 2.1182, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4082290900390016e-05, | |
| "loss": 2.0678, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.406315499910711e-05, | |
| "loss": 2.1042, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4044001260571574e-05, | |
| "loss": 2.0094, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.4024829768868324e-05, | |
| "loss": 1.9978, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.4005640608160223e-05, | |
| "loss": 2.1296, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.3986433862687699e-05, | |
| "loss": 2.0618, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.3967209616768369e-05, | |
| "loss": 2.1228, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.394796795479669e-05, | |
| "loss": 2.0857, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.392870896124356e-05, | |
| "loss": 2.0438, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3909432720655976e-05, | |
| "loss": 2.137, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3890139317656645e-05, | |
| "loss": 2.0582, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3870828836943617e-05, | |
| "loss": 2.0735, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3851501363289907e-05, | |
| "loss": 2.1025, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.3832156981543145e-05, | |
| "loss": 2.0914, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.381279577662517e-05, | |
| "loss": 2.1104, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.3793417833531683e-05, | |
| "loss": 2.0286, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.377402323733186e-05, | |
| "loss": 2.0399, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.3754612073168002e-05, | |
| "loss": 2.0269, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.3735184426255117e-05, | |
| "loss": 2.0914, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.3715740381880592e-05, | |
| "loss": 2.0965, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.369628002540379e-05, | |
| "loss": 2.049, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.3676803442255688e-05, | |
| "loss": 2.1164, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.3657310717938495e-05, | |
| "loss": 2.1314, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.3637801938025282e-05, | |
| "loss": 2.0716, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.3618277188159601e-05, | |
| "loss": 2.1287, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.3598736554055116e-05, | |
| "loss": 2.0827, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.3579180121495225e-05, | |
| "loss": 2.1128, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.3559607976332675e-05, | |
| "loss": 2.0894, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.354002020448919e-05, | |
| "loss": 2.0323, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.3520416891955101e-05, | |
| "loss": 2.1541, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.3500798124788965e-05, | |
| "loss": 2.1082, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.3481163989117177e-05, | |
| "loss": 2.1331, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.3461514571133598e-05, | |
| "loss": 2.0552, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.3441849957099193e-05, | |
| "loss": 2.1692, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3422170233341625e-05, | |
| "loss": 2.1162, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3402475486254887e-05, | |
| "loss": 2.109, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3382765802298934e-05, | |
| "loss": 2.0635, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.3363041267999291e-05, | |
| "loss": 2.137, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.334330196994667e-05, | |
| "loss": 2.0237, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3323547994796597e-05, | |
| "loss": 2.0276, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.330377942926904e-05, | |
| "loss": 2.0646, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3283996360148015e-05, | |
| "loss": 2.1255, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3264198874281198e-05, | |
| "loss": 2.0739, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3244387058579566e-05, | |
| "loss": 2.1738, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3224561000017002e-05, | |
| "loss": 2.1162, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.320472078562991e-05, | |
| "loss": 2.0098, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3184866502516846e-05, | |
| "loss": 2.0545, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3164998237838122e-05, | |
| "loss": 2.0972, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.3145116078815427e-05, | |
| "loss": 2.0943, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3125220112731458e-05, | |
| "loss": 2.0336, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.310531042692951e-05, | |
| "loss": 2.0876, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3085387108813116e-05, | |
| "loss": 2.0413, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3065450245845652e-05, | |
| "loss": 2.1667, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.3045499925549962e-05, | |
| "loss": 2.082, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.302553623550796e-05, | |
| "loss": 2.0349, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.300555926336025e-05, | |
| "loss": 2.1066, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.298556909680576e-05, | |
| "loss": 2.1254, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.296556582360133e-05, | |
| "loss": 2.078, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.2945549531561334e-05, | |
| "loss": 2.1126, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2925520308557316e-05, | |
| "loss": 2.0396, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2905478242517563e-05, | |
| "loss": 2.0167, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2885423421426762e-05, | |
| "loss": 2.0488, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.286535593332559e-05, | |
| "loss": 2.1691, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.2845275866310325e-05, | |
| "loss": 2.1113, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2825183308532478e-05, | |
| "loss": 2.1532, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2805078348198387e-05, | |
| "loss": 2.0869, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2784961073568835e-05, | |
| "loss": 2.1402, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.2764831572958674e-05, | |
| "loss": 2.045, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.274468993473642e-05, | |
| "loss": 2.1223, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.2724536247323868e-05, | |
| "loss": 2.085, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.270437059919573e-05, | |
| "loss": 2.1571, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.2684193078879205e-05, | |
| "loss": 2.1077, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.266400377495362e-05, | |
| "loss": 2.0425, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.2643802776050027e-05, | |
| "loss": 2.1263, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2623590170850826e-05, | |
| "loss": 2.1301, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2603366048089363e-05, | |
| "loss": 2.0984, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2583130496549545e-05, | |
| "loss": 2.0693, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2562883605065456e-05, | |
| "loss": 2.1245, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.2542625462520968e-05, | |
| "loss": 2.1125, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2522356157849323e-05, | |
| "loss": 2.1237, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2502075780032792e-05, | |
| "loss": 2.0992, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.248178441810224e-05, | |
| "loss": 2.0085, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.2461482161136756e-05, | |
| "loss": 2.1386, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.244116909826326e-05, | |
| "loss": 2.0545, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2420845318656107e-05, | |
| "loss": 2.0876, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2400510911536705e-05, | |
| "loss": 2.0853, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2380165966173105e-05, | |
| "loss": 2.1409, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2359810571879638e-05, | |
| "loss": 2.071, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.2339444818016488e-05, | |
| "loss": 2.1455, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2319068793989326e-05, | |
| "loss": 2.1625, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.229868258924891e-05, | |
| "loss": 2.0579, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2278286293290693e-05, | |
| "loss": 2.045, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.225787999565442e-05, | |
| "loss": 2.0943, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.2237463785923752e-05, | |
| "loss": 2.0928, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.221703775372586e-05, | |
| "loss": 2.071, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2196601988731035e-05, | |
| "loss": 2.21, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2176156580652293e-05, | |
| "loss": 2.0664, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.2155701619244997e-05, | |
| "loss": 1.9691, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.213523719430643e-05, | |
| "loss": 2.0508, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.211476339567543e-05, | |
| "loss": 1.9978, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2094280313231985e-05, | |
| "loss": 2.1024, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2073788036896842e-05, | |
| "loss": 2.0739, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2053286656631093e-05, | |
| "loss": 2.1285, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.2032776262435824e-05, | |
| "loss": 2.0807, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.2012256944351664e-05, | |
| "loss": 2.0765, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1991728792458432e-05, | |
| "loss": 2.0483, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1971191896874734e-05, | |
| "loss": 2.0142, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1950646347757548e-05, | |
| "loss": 1.9939, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1930092235301847e-05, | |
| "loss": 2.17, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1909529649740188e-05, | |
| "loss": 2.1462, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1888958681342345e-05, | |
| "loss": 2.1365, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.186837942041487e-05, | |
| "loss": 2.0604, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1847791957300735e-05, | |
| "loss": 2.1323, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1827196382378914e-05, | |
| "loss": 2.0574, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1806592786063991e-05, | |
| "loss": 2.1837, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1785981258805764e-05, | |
| "loss": 2.0659, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1765361891088851e-05, | |
| "loss": 2.0591, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1744734773432282e-05, | |
| "loss": 2.1358, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.172409999638912e-05, | |
| "loss": 2.0352, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.170345765054605e-05, | |
| "loss": 2.0648, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1682807826522973e-05, | |
| "loss": 1.956, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1662150614972633e-05, | |
| "loss": 2.0617, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.16414861065802e-05, | |
| "loss": 2.1478, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1620814392062873e-05, | |
| "loss": 2.097, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1600135562169494e-05, | |
| "loss": 2.0129, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1579449707680137e-05, | |
| "loss": 2.0587, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1558756919405712e-05, | |
| "loss": 2.1292, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1538057288187573e-05, | |
| "loss": 2.1235, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.151735090489711e-05, | |
| "loss": 2.0787, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1496637860435355e-05, | |
| "loss": 2.0729, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1475918245732588e-05, | |
| "loss": 2.0815, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1455192151747931e-05, | |
| "loss": 2.1627, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1434459669468944e-05, | |
| "loss": 2.1101, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.1413720889911232e-05, | |
| "loss": 2.0376, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1392975904118058e-05, | |
| "loss": 2.0389, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1372224803159912e-05, | |
| "loss": 2.0645, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.135146767813414e-05, | |
| "loss": 2.0577, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.133070462016454e-05, | |
| "loss": 2.0651, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.1309935720400934e-05, | |
| "loss": 2.1368, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1289161070018813e-05, | |
| "loss": 1.9803, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1268380760218901e-05, | |
| "loss": 2.1219, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.124759488222677e-05, | |
| "loss": 2.0363, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1226803527292435e-05, | |
| "loss": 2.0921, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.1206006786689951e-05, | |
| "loss": 2.0454, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.118520475171703e-05, | |
| "loss": 2.0804, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1164397513694608e-05, | |
| "loss": 2.0569, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.114358516396647e-05, | |
| "loss": 2.128, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1122767793898853e-05, | |
| "loss": 2.0417, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.1101945494880013e-05, | |
| "loss": 2.0421, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1081118358319857e-05, | |
| "loss": 2.108, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1060286475649529e-05, | |
| "loss": 2.0822, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1039449938321e-05, | |
| "loss": 2.0601, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.1018608837806685e-05, | |
| "loss": 2.1049, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.0997763265599023e-05, | |
| "loss": 2.1129, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.0976913313210084e-05, | |
| "loss": 2.1099, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.095605907217118e-05, | |
| "loss": 2.0894, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.093520063403243e-05, | |
| "loss": 2.1987, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.0914338090362395e-05, | |
| "loss": 2.0848, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.0893471532747651e-05, | |
| "loss": 2.1345, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.0872601052792398e-05, | |
| "loss": 2.1285, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.0851726742118051e-05, | |
| "loss": 2.1164, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.0830848692362852e-05, | |
| "loss": 2.1502, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.0809966995181443e-05, | |
| "loss": 2.082, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.078908174224449e-05, | |
| "loss": 2.0757, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0768193025238265e-05, | |
| "loss": 2.1279, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0747300935864245e-05, | |
| "loss": 2.0593, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0726405565838713e-05, | |
| "loss": 2.0968, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0705507006892356e-05, | |
| "loss": 2.1628, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.0684605350769862e-05, | |
| "loss": 2.0794, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0663700689229506e-05, | |
| "loss": 1.9637, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.064279311404277e-05, | |
| "loss": 2.146, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0621882716993916e-05, | |
| "loss": 2.1017, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.06009695898796e-05, | |
| "loss": 2.1323, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0580053824508463e-05, | |
| "loss": 2.0876, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.055913551270072e-05, | |
| "loss": 2.1592, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0538214746287775e-05, | |
| "loss": 2.0061, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0517291617111809e-05, | |
| "loss": 2.0367, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0496366217025362e-05, | |
| "loss": 2.0932, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.0475438637890958e-05, | |
| "loss": 2.061, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0454508971580675e-05, | |
| "loss": 2.0244, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0433577309975763e-05, | |
| "loss": 2.0368, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0412643744966226e-05, | |
| "loss": 2.1809, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0391708368450429e-05, | |
| "loss": 2.1164, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0370771272334679e-05, | |
| "loss": 2.111, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0349832548532848e-05, | |
| "loss": 2.1022, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0328892288965939e-05, | |
| "loss": 2.1016, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0307950585561705e-05, | |
| "loss": 2.1116, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.0287007530254233e-05, | |
| "loss": 2.0237, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.026606321498355e-05, | |
| "loss": 2.0666, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0245117731695213e-05, | |
| "loss": 2.023, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0224171172339899e-05, | |
| "loss": 2.1712, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0203223628873019e-05, | |
| "loss": 2.1528, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0182275193254296e-05, | |
| "loss": 2.056, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.0161325957447377e-05, | |
| "loss": 2.092, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0140376013419419e-05, | |
| "loss": 2.0285, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0119425453140687e-05, | |
| "loss": 2.1339, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0098474368584153e-05, | |
| "loss": 2.0977, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0077522851725086e-05, | |
| "loss": 2.0364, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.0056570994540658e-05, | |
| "loss": 2.2121, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0035618889009535e-05, | |
| "loss": 2.0931, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0014666627111467e-05, | |
| "loss": 2.0551, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.993714300826901e-06, | |
| "loss": 2.0648, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.972762002136554e-06, | |
| "loss": 2.0979, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 9.951809823021037e-06, | |
| "loss": 2.0781, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.930857855460421e-06, | |
| "loss": 2.1147, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.909906191433855e-06, | |
| "loss": 2.0669, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.88895492291916e-06, | |
| "loss": 2.0526, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.868004141892412e-06, | |
| "loss": 2.145, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.847053940327556e-06, | |
| "loss": 2.0464, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.826104410195976e-06, | |
| "loss": 2.1414, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.805155643466133e-06, | |
| "loss": 2.0413, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.784207732103122e-06, | |
| "loss": 2.0787, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.76326076806828e-06, | |
| "loss": 2.0104, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.742314843318796e-06, | |
| "loss": 2.0562, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.721370049807287e-06, | |
| "loss": 2.0541, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.700426479481408e-06, | |
| "loss": 2.1302, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.67948422428345e-06, | |
| "loss": 2.0897, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.658543376149917e-06, | |
| "loss": 2.0729, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.637604027011145e-06, | |
| "loss": 2.0358, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.61666626879089e-06, | |
| "loss": 2.025, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.595730193405924e-06, | |
| "loss": 2.0552, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.574795892765628e-06, | |
| "loss": 2.1235, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.553863458771585e-06, | |
| "loss": 2.097, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.532932983317202e-06, | |
| "loss": 2.1525, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.512004558287277e-06, | |
| "loss": 2.0289, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.491078275557596e-06, | |
| "loss": 2.0842, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.470154226994564e-06, | |
| "loss": 1.9881, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.449232504454763e-06, | |
| "loss": 2.0582, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.428313199784557e-06, | |
| "loss": 2.1515, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.407396404819717e-06, | |
| "loss": 2.0348, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.38648221138498e-06, | |
| "loss": 2.1124, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.365570711293658e-06, | |
| "loss": 2.1524, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.34466199634726e-06, | |
| "loss": 2.0611, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.323756158335054e-06, | |
| "loss": 2.0999, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.302853289033673e-06, | |
| "loss": 2.102, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.281953480206725e-06, | |
| "loss": 2.0639, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.26105682360439e-06, | |
| "loss": 2.0317, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.240163410962999e-06, | |
| "loss": 2.0991, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.219273334004635e-06, | |
| "loss": 2.1076, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.19838668443676e-06, | |
| "loss": 1.9736, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.177503553951775e-06, | |
| "loss": 1.9861, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.156624034226623e-06, | |
| "loss": 2.0982, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.13574821692242e-06, | |
| "loss": 2.0654, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.114876193684011e-06, | |
| "loss": 2.1508, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.094008056139587e-06, | |
| "loss": 2.0808, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.07314389590028e-06, | |
| "loss": 2.1509, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.052283804559774e-06, | |
| "loss": 2.0443, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.031427873693875e-06, | |
| "loss": 2.0152, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.010576194860129e-06, | |
| "loss": 2.0672, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 8.989728859597424e-06, | |
| "loss": 2.1208, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 8.968885959425567e-06, | |
| "loss": 2.1253, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 8.9480475858449e-06, | |
| "loss": 2.0516, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 8.927213830335904e-06, | |
| "loss": 2.0799, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 8.906384784358767e-06, | |
| "loss": 2.0959, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.88556053935301e-06, | |
| "loss": 2.0965, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.864741186737094e-06, | |
| "loss": 2.0327, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.843926817907976e-06, | |
| "loss": 2.1073, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.823117524240751e-06, | |
| "loss": 1.9879, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.802313397088223e-06, | |
| "loss": 2.0548, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.781514527780531e-06, | |
| "loss": 2.1127, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.760721007624718e-06, | |
| "loss": 2.0538, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.739932927904345e-06, | |
| "loss": 2.1022, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.719150379879099e-06, | |
| "loss": 2.1688, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.698373454784374e-06, | |
| "loss": 2.0854, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.677602243830882e-06, | |
| "loss": 2.1103, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.656836838204255e-06, | |
| "loss": 2.0988, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.63607732906463e-06, | |
| "loss": 2.0324, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.615323807546258e-06, | |
| "loss": 2.0913, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.594576364757126e-06, | |
| "loss": 2.0598, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.573835091778505e-06, | |
| "loss": 2.0585, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.553100079664598e-06, | |
| "loss": 2.0281, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.53237141944212e-06, | |
| "loss": 2.1304, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.511649202109902e-06, | |
| "loss": 2.1373, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.490933518638488e-06, | |
| "loss": 2.0538, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.470224459969737e-06, | |
| "loss": 2.1275, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.449522117016432e-06, | |
| "loss": 2.088, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.428826580661867e-06, | |
| "loss": 2.1329, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.408137941759459e-06, | |
| "loss": 2.1218, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.387456291132343e-06, | |
| "loss": 2.0591, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.366781719572978e-06, | |
| "loss": 2.2194, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.34611431784274e-06, | |
| "loss": 2.0977, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.325454176671541e-06, | |
| "loss": 2.1123, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.304801386757407e-06, | |
| "loss": 2.0763, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.284156038766095e-06, | |
| "loss": 2.0852, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.263518223330698e-06, | |
| "loss": 2.0575, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.242888031051234e-06, | |
| "loss": 2.075, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.222265552494262e-06, | |
| "loss": 2.0556, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.201650878192468e-06, | |
| "loss": 2.0116, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.18104409864429e-06, | |
| "loss": 1.9719, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.160445304313497e-06, | |
| "loss": 2.0438, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.139854585628806e-06, | |
| "loss": 2.0374, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.119272032983487e-06, | |
| "loss": 2.1279, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.098697736734957e-06, | |
| "loss": 2.0143, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.078131787204382e-06, | |
| "loss": 2.1391, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.0575742746763e-06, | |
| "loss": 2.0629, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.037025289398196e-06, | |
| "loss": 2.1205, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.016484921580127e-06, | |
| "loss": 2.0495, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.995953261394321e-06, | |
| "loss": 2.0628, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.975430398974775e-06, | |
| "loss": 2.0952, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.954916424416869e-06, | |
| "loss": 2.0941, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.934411427776958e-06, | |
| "loss": 2.1818, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.913915499071994e-06, | |
| "loss": 2.0785, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.89342872827911e-06, | |
| "loss": 2.1087, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.872951205335242e-06, | |
| "loss": 2.0482, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.85248302013673e-06, | |
| "loss": 2.2156, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.832024262538914e-06, | |
| "loss": 2.04, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.811575022355749e-06, | |
| "loss": 2.0792, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.791135389359417e-06, | |
| "loss": 1.9923, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.770705453279916e-06, | |
| "loss": 2.1058, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.750285303804674e-06, | |
| "loss": 2.0794, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.729875030578157e-06, | |
| "loss": 2.1222, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.70947472320148e-06, | |
| "loss": 2.0841, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.689084471232001e-06, | |
| "loss": 2.0811, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.668704364182935e-06, | |
| "loss": 2.0769, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.648334491522966e-06, | |
| "loss": 2.1287, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.627974942675846e-06, | |
| "loss": 2.0796, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.607625807020003e-06, | |
| "loss": 2.013, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.587287173888156e-06, | |
| "loss": 2.082, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.566959132566914e-06, | |
| "loss": 2.0406, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.546641772296388e-06, | |
| "loss": 1.9931, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.526335182269804e-06, | |
| "loss": 2.019, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.506039451633101e-06, | |
| "loss": 2.0648, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.485754669484549e-06, | |
| "loss": 1.9721, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.465480924874348e-06, | |
| "loss": 2.0578, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.445218306804253e-06, | |
| "loss": 2.1054, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.424966904227167e-06, | |
| "loss": 2.0934, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.404726806046753e-06, | |
| "loss": 2.0193, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.384498101117062e-06, | |
| "loss": 2.0823, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.364280878242114e-06, | |
| "loss": 2.1134, | |
| "step": 4385 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.344075226175527e-06, | |
| "loss": 2.1058, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.323881233620129e-06, | |
| "loss": 2.1015, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.303698989227554e-06, | |
| "loss": 2.0341, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.283528581597867e-06, | |
| "loss": 2.1213, | |
| "step": 4405 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.263370099279173e-06, | |
| "loss": 2.0839, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.2432236307672136e-06, | |
| "loss": 2.0384, | |
| "step": 4415 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.223089264505001e-06, | |
| "loss": 2.0842, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.202967088882405e-06, | |
| "loss": 2.087, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.1828571922358e-06, | |
| "loss": 1.9524, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.162759662847638e-06, | |
| "loss": 2.0555, | |
| "step": 4435 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.142674588946079e-06, | |
| "loss": 2.1134, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.122602058704615e-06, | |
| "loss": 2.0735, | |
| "step": 4445 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.102542160241664e-06, | |
| "loss": 2.1126, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.082494981620184e-06, | |
| "loss": 2.0787, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.062460610847308e-06, | |
| "loss": 1.9744, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.042439135873932e-06, | |
| "loss": 2.1265, | |
| "step": 4465 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.022430644594339e-06, | |
| "loss": 1.9675, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.002435224845822e-06, | |
| "loss": 2.0836, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.98245296440828e-06, | |
| "loss": 2.0921, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.962483951003851e-06, | |
| "loss": 2.0507, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.942528272296505e-06, | |
| "loss": 1.9563, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.922586015891695e-06, | |
| "loss": 2.1548, | |
| "step": 4495 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.902657269335933e-06, | |
| "loss": 2.1145, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.882742120116419e-06, | |
| "loss": 2.0528, | |
| "step": 4505 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.862840655660679e-06, | |
| "loss": 2.1135, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.842952963336154e-06, | |
| "loss": 2.0965, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.82307913044981e-06, | |
| "loss": 2.042, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.803219244247798e-06, | |
| "loss": 2.1888, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.783373391915027e-06, | |
| "loss": 2.0114, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.763541660574795e-06, | |
| "loss": 2.0276, | |
| "step": 4535 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.74372413728841e-06, | |
| "loss": 2.0472, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.723920909054822e-06, | |
| "loss": 2.1113, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.704132062810204e-06, | |
| "loss": 2.1455, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.6843576854276005e-06, | |
| "loss": 2.0414, | |
| "step": 4555 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.6645978637165485e-06, | |
| "loss": 2.0696, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.644852684422669e-06, | |
| "loss": 2.0612, | |
| "step": 4565 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.625122234227306e-06, | |
| "loss": 1.9787, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.605406599747156e-06, | |
| "loss": 2.115, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.585705867533859e-06, | |
| "loss": 2.1017, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.566020124073638e-06, | |
| "loss": 2.0864, | |
| "step": 4585 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.546349455786926e-06, | |
| "loss": 2.1133, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.52669394902796e-06, | |
| "loss": 2.0318, | |
| "step": 4595 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.507053690084426e-06, | |
| "loss": 2.0634, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.487428765177073e-06, | |
| "loss": 2.0525, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.467819260459332e-06, | |
| "loss": 2.13, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.448225262016938e-06, | |
| "loss": 2.044, | |
| "step": 4615 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.428646855867553e-06, | |
| "loss": 2.1884, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.409084127960398e-06, | |
| "loss": 2.1139, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.389537164175852e-06, | |
| "loss": 2.1151, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.370006050325095e-06, | |
| "loss": 2.0269, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.350490872149737e-06, | |
| "loss": 2.1715, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.3309917153214115e-06, | |
| "loss": 2.048, | |
| "step": 4645 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.311508665441428e-06, | |
| "loss": 2.1836, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.292041808040393e-06, | |
| "loss": 1.9866, | |
| "step": 4655 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.272591228577816e-06, | |
| "loss": 2.1363, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.253157012441751e-06, | |
| "loss": 2.0852, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.233739244948416e-06, | |
| "loss": 2.0689, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.214338011341825e-06, | |
| "loss": 2.0112, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.194953396793401e-06, | |
| "loss": 2.0693, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.175585486401612e-06, | |
| "loss": 2.0233, | |
| "step": 4685 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.156234365191598e-06, | |
| "loss": 2.0183, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.136900118114789e-06, | |
| "loss": 1.9348, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.117582830048539e-06, | |
| "loss": 2.0473, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.0982825857957565e-06, | |
| "loss": 2.0919, | |
| "step": 4705 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.0789994700845234e-06, | |
| "loss": 2.0806, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.059733567567725e-06, | |
| "loss": 2.0353, | |
| "step": 4715 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.040484962822686e-06, | |
| "loss": 2.0148, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.021253740350793e-06, | |
| "loss": 2.0244, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.002039984577121e-06, | |
| "loss": 2.1046, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 5.982843779850067e-06, | |
| "loss": 2.0956, | |
| "step": 4735 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 5.963665210440983e-06, | |
| "loss": 2.0678, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 5.944504360543797e-06, | |
| "loss": 2.0827, | |
| "step": 4745 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 5.9253613142746505e-06, | |
| "loss": 2.0919, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 5.906236155671529e-06, | |
| "loss": 2.1765, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 5.887128968693887e-06, | |
| "loss": 2.0592, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 5.8680398372222834e-06, | |
| "loss": 2.015, | |
| "step": 4765 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 5.8489688450580215e-06, | |
| "loss": 2.0643, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 5.829916075922766e-06, | |
| "loss": 2.1475, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 5.810881613458176e-06, | |
| "loss": 2.0388, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 5.791865541225561e-06, | |
| "loss": 2.0519, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 5.772867942705485e-06, | |
| "loss": 2.0604, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 5.753888901297416e-06, | |
| "loss": 2.0753, | |
| "step": 4795 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 5.734928500319352e-06, | |
| "loss": 1.9892, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 5.71598682300747e-06, | |
| "loss": 2.124, | |
| "step": 4805 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 5.697063952515745e-06, | |
| "loss": 2.0879, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.678159971915578e-06, | |
| "loss": 2.0136, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.659274964195463e-06, | |
| "loss": 2.1105, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.6404090122605925e-06, | |
| "loss": 2.0453, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.621562198932499e-06, | |
| "loss": 2.0919, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 5.602734606948714e-06, | |
| "loss": 1.9485, | |
| "step": 4835 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.583926318962376e-06, | |
| "loss": 2.124, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.565137417541866e-06, | |
| "loss": 2.139, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.546367985170483e-06, | |
| "loss": 2.0599, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.527618104246044e-06, | |
| "loss": 2.0426, | |
| "step": 4855 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 5.5088878570805336e-06, | |
| "loss": 2.0937, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.490177325899744e-06, | |
| "loss": 2.1013, | |
| "step": 4865 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.471486592842927e-06, | |
| "loss": 2.056, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.452815739962404e-06, | |
| "loss": 2.0075, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.434164849223225e-06, | |
| "loss": 2.061, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 5.415534002502818e-06, | |
| "loss": 2.1069, | |
| "step": 4885 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.396923281590609e-06, | |
| "loss": 2.1454, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.378332768187667e-06, | |
| "loss": 2.0682, | |
| "step": 4895 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.3597625439063685e-06, | |
| "loss": 2.0854, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.341212690269994e-06, | |
| "loss": 2.0926, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 5.322683288712416e-06, | |
| "loss": 2.0883, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.304174420577719e-06, | |
| "loss": 2.082, | |
| "step": 4915 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.285686167119845e-06, | |
| "loss": 2.0835, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.267218609502234e-06, | |
| "loss": 2.1362, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.248771828797474e-06, | |
| "loss": 2.0765, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.230345905986945e-06, | |
| "loss": 2.1331, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.211940921960455e-06, | |
| "loss": 2.131, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.193556957515888e-06, | |
| "loss": 2.002, | |
| "step": 4945 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.1751940933588714e-06, | |
| "loss": 2.1444, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.156852410102382e-06, | |
| "loss": 2.0563, | |
| "step": 4955 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.1385319882664135e-06, | |
| "loss": 2.068, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.120232908277642e-06, | |
| "loss": 2.0141, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.101955250469025e-06, | |
| "loss": 2.0439, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.0836990950794906e-06, | |
| "loss": 2.0652, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.06546452225358e-06, | |
| "loss": 2.067, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 5.047251612041072e-06, | |
| "loss": 2.049, | |
| "step": 4985 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.029060444396653e-06, | |
| "loss": 2.0319, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 5.010891099179558e-06, | |
| "loss": 2.1232, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.992743656153224e-06, | |
| "loss": 2.1031, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.974618194984934e-06, | |
| "loss": 2.1366, | |
| "step": 5005 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.956514795245469e-06, | |
| "loss": 2.0794, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.938433536408771e-06, | |
| "loss": 1.9702, | |
| "step": 5015 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.9203744978515735e-06, | |
| "loss": 2.1197, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.9023377588530626e-06, | |
| "loss": 2.0746, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.884323398594536e-06, | |
| "loss": 2.0547, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.866331496159042e-06, | |
| "loss": 2.1689, | |
| "step": 5035 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.848362130531039e-06, | |
| "loss": 2.0292, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.8304153805960595e-06, | |
| "loss": 1.977, | |
| "step": 5045 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.812491325140343e-06, | |
| "loss": 2.0431, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.7945900428505e-06, | |
| "loss": 2.0355, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.776711612313171e-06, | |
| "loss": 2.1682, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.758856112014675e-06, | |
| "loss": 1.9793, | |
| "step": 5065 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.741023620340668e-06, | |
| "loss": 2.0138, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.723214215575796e-06, | |
| "loss": 1.9691, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.705427975903359e-06, | |
| "loss": 2.0467, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.687664979404959e-06, | |
| "loss": 2.0741, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.669925304060159e-06, | |
| "loss": 2.0443, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.652209027746143e-06, | |
| "loss": 2.0996, | |
| "step": 5095 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.634516228237372e-06, | |
| "loss": 2.0144, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.616846983205246e-06, | |
| "loss": 2.0942, | |
| "step": 5105 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.599201370217763e-06, | |
| "loss": 2.063, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.581579466739171e-06, | |
| "loss": 2.0336, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.563981350129637e-06, | |
| "loss": 2.117, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.546407097644903e-06, | |
| "loss": 2.0373, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.528856786435945e-06, | |
| "loss": 2.0733, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.511330493548638e-06, | |
| "loss": 2.0492, | |
| "step": 5135 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.493828295923416e-06, | |
| "loss": 2.1111, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.476350270394942e-06, | |
| "loss": 2.037, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.458896493691754e-06, | |
| "loss": 2.0392, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.441467042435941e-06, | |
| "loss": 2.0423, | |
| "step": 5155 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.424061993142803e-06, | |
| "loss": 2.0407, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.406681422220518e-06, | |
| "loss": 2.0928, | |
| "step": 5165 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.389325405969799e-06, | |
| "loss": 2.1026, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.371994020583563e-06, | |
| "loss": 2.077, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.3546873421466104e-06, | |
| "loss": 2.1275, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.337405446635264e-06, | |
| "loss": 2.0782, | |
| "step": 5185 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.320148409917055e-06, | |
| "loss": 1.9864, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.302916307750383e-06, | |
| "loss": 2.007, | |
| "step": 5195 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.285709215784187e-06, | |
| "loss": 2.0399, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.268527209557605e-06, | |
| "loss": 2.01, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.2513703644996615e-06, | |
| "loss": 2.0768, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.234238755928913e-06, | |
| "loss": 2.0703, | |
| "step": 5215 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.217132459053129e-06, | |
| "loss": 2.1422, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.200051548968962e-06, | |
| "loss": 2.1622, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.182996100661616e-06, | |
| "loss": 2.1308, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.16596618900452e-06, | |
| "loss": 2.1187, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.148961888758991e-06, | |
| "loss": 2.0699, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.131983274573922e-06, | |
| "loss": 2.0442, | |
| "step": 5245 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.1150304209854375e-06, | |
| "loss": 2.1129, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.098103402416571e-06, | |
| "loss": 2.0754, | |
| "step": 5255 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.081202293176945e-06, | |
| "loss": 2.037, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.064327167462437e-06, | |
| "loss": 2.087, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.047478099354857e-06, | |
| "loss": 2.0902, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.030655162821626e-06, | |
| "loss": 2.0168, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.013858431715445e-06, | |
| "loss": 2.0462, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 3.9970879797739695e-06, | |
| "loss": 2.077, | |
| "step": 5285 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.980343880619494e-06, | |
| "loss": 2.0406, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.9636262077586225e-06, | |
| "loss": 2.1191, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.946935034581948e-06, | |
| "loss": 1.9797, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.9302704343637265e-06, | |
| "loss": 2.0947, | |
| "step": 5305 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.913632480261567e-06, | |
| "loss": 2.1328, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 3.897021245316101e-06, | |
| "loss": 2.1311, | |
| "step": 5315 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 3.880436802450645e-06, | |
| "loss": 2.0713, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 3.863879224470922e-06, | |
| "loss": 2.0167, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 3.847348584064709e-06, | |
| "loss": 2.1206, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 3.830844953801518e-06, | |
| "loss": 2.0814, | |
| "step": 5335 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.814368406132304e-06, | |
| "loss": 2.0911, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.7979190133891163e-06, | |
| "loss": 2.1073, | |
| "step": 5345 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.781496847784798e-06, | |
| "loss": 2.0458, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.7651019814126656e-06, | |
| "loss": 2.0481, | |
| "step": 5355 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.7487344862461907e-06, | |
| "loss": 2.0473, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.732394434138685e-06, | |
| "loss": 2.0432, | |
| "step": 5365 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.716081896822984e-06, | |
| "loss": 2.0772, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.6997969459111414e-06, | |
| "loss": 2.0831, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.6835396528941003e-06, | |
| "loss": 2.1639, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 3.667310089141376e-06, | |
| "loss": 1.9198, | |
| "step": 5385 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.651108325900773e-06, | |
| "loss": 2.1591, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.6349344342980365e-06, | |
| "loss": 2.0509, | |
| "step": 5395 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.618788485336555e-06, | |
| "loss": 2.0814, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.6026705498970617e-06, | |
| "loss": 2.1067, | |
| "step": 5405 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 3.5865806987372996e-06, | |
| "loss": 2.1184, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.570519002491717e-06, | |
| "loss": 1.9818, | |
| "step": 5415 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.5544855316711768e-06, | |
| "loss": 2.0269, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.538480356662621e-06, | |
| "loss": 2.0397, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.5225035477287783e-06, | |
| "loss": 2.0323, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 3.506555175007843e-06, | |
| "loss": 2.1048, | |
| "step": 5435 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.4906353085131917e-06, | |
| "loss": 2.0833, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.474744018133036e-06, | |
| "loss": 2.0813, | |
| "step": 5445 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.4588813736301507e-06, | |
| "loss": 2.137, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.443047444641557e-06, | |
| "loss": 2.0274, | |
| "step": 5455 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.427242300678213e-06, | |
| "loss": 2.1431, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.411466011124701e-06, | |
| "loss": 2.0952, | |
| "step": 5465 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.3957186452389545e-06, | |
| "loss": 2.0546, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.3800002721519064e-06, | |
| "loss": 2.0325, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.364310960867223e-06, | |
| "loss": 2.1614, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 3.348650780260997e-06, | |
| "loss": 2.0274, | |
| "step": 5485 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.333019799081425e-06, | |
| "loss": 2.0534, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.317418085948524e-06, | |
| "loss": 2.0118, | |
| "step": 5495 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.301845709353818e-06, | |
| "loss": 2.0819, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.286302737660061e-06, | |
| "loss": 1.9773, | |
| "step": 5505 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.2707892391008945e-06, | |
| "loss": 2.1252, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.2553052817805887e-06, | |
| "loss": 2.0567, | |
| "step": 5515 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.239850933673728e-06, | |
| "loss": 2.0363, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.224426262624908e-06, | |
| "loss": 2.0662, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.2090313363484383e-06, | |
| "loss": 2.1223, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.1936662224280634e-06, | |
| "loss": 2.1169, | |
| "step": 5535 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.1783309883166327e-06, | |
| "loss": 2.1307, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.16302570133583e-06, | |
| "loss": 1.9699, | |
| "step": 5545 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.147750428675883e-06, | |
| "loss": 2.1088, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.1325052373952424e-06, | |
| "loss": 2.1042, | |
| "step": 5555 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.117290194420306e-06, | |
| "loss": 2.1419, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.1021053665451204e-06, | |
| "loss": 2.045, | |
| "step": 5565 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.0869508204310883e-06, | |
| "loss": 2.1147, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.0718266226066753e-06, | |
| "loss": 2.0688, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.0567328394671146e-06, | |
| "loss": 2.1113, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.041669537274129e-06, | |
| "loss": 2.0537, | |
| "step": 5585 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0266367821556196e-06, | |
| "loss": 2.0977, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0116346401053908e-06, | |
| "loss": 2.0842, | |
| "step": 5595 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 2.9966631769828524e-06, | |
| "loss": 2.0854, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 2.981722458512738e-06, | |
| "loss": 1.9946, | |
| "step": 5605 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 2.9668125502848035e-06, | |
| "loss": 2.0781, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.9519335177535624e-06, | |
| "loss": 2.0189, | |
| "step": 5615 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.9370854262379723e-06, | |
| "loss": 2.0735, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.9222683409211618e-06, | |
| "loss": 2.1514, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.907482326850144e-06, | |
| "loss": 2.0128, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.8927274489355296e-06, | |
| "loss": 2.1505, | |
| "step": 5635 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.878003771951239e-06, | |
| "loss": 2.1523, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.86331136053422e-06, | |
| "loss": 2.0934, | |
| "step": 5645 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.848650279184173e-06, | |
| "loss": 2.0922, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.834020592263249e-06, | |
| "loss": 2.0501, | |
| "step": 5655 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.8194223639957816e-06, | |
| "loss": 2.0705, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.804855658468002e-06, | |
| "loss": 2.1016, | |
| "step": 5665 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.7903205396277546e-06, | |
| "loss": 2.0355, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.7758170712842147e-06, | |
| "loss": 2.0765, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.7613453171076233e-06, | |
| "loss": 2.0409, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.7469053406289857e-06, | |
| "loss": 2.0348, | |
| "step": 5685 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.732497205239807e-06, | |
| "loss": 1.9971, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.7181209741918093e-06, | |
| "loss": 2.0887, | |
| "step": 5695 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.703776710596654e-06, | |
| "loss": 2.091, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.6894644774256663e-06, | |
| "loss": 2.097, | |
| "step": 5705 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.6751843375095543e-06, | |
| "loss": 1.9934, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.660936353538146e-06, | |
| "loss": 2.0891, | |
| "step": 5715 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.6467205880600954e-06, | |
| "loss": 2.102, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.632537103482622e-06, | |
| "loss": 2.1345, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.6183859620712294e-06, | |
| "loss": 2.1523, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.6042672259494352e-06, | |
| "loss": 2.1234, | |
| "step": 5735 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.590180957098498e-06, | |
| "loss": 2.0256, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.5761272173571507e-06, | |
| "loss": 2.0513, | |
| "step": 5745 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.5621060684213138e-06, | |
| "loss": 2.0805, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.54811757184384e-06, | |
| "loss": 2.001, | |
| "step": 5755 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 2.5341617890342375e-06, | |
| "loss": 2.0726, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.5202387812583984e-06, | |
| "loss": 2.1048, | |
| "step": 5765 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.506348609638336e-06, | |
| "loss": 2.0817, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.492491335151909e-06, | |
| "loss": 2.0392, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.478667018632562e-06, | |
| "loss": 1.9889, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 2.4648757207690533e-06, | |
| "loss": 2.0698, | |
| "step": 5785 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.451117502105185e-06, | |
| "loss": 2.0412, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.437392423039546e-06, | |
| "loss": 2.0682, | |
| "step": 5795 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.42370054382524e-06, | |
| "loss": 2.0883, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.410041924569625e-06, | |
| "loss": 2.0525, | |
| "step": 5805 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.3964166252340438e-06, | |
| "loss": 2.0261, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.3828247056335718e-06, | |
| "loss": 2.0894, | |
| "step": 5815 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.3692662254367426e-06, | |
| "loss": 2.0458, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.3557412441652903e-06, | |
| "loss": 2.0402, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.3422498211938918e-06, | |
| "loss": 2.1306, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.328792015749899e-06, | |
| "loss": 2.082, | |
| "step": 5835 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.3153678869130834e-06, | |
| "loss": 2.1265, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.3019774936153827e-06, | |
| "loss": 2.128, | |
| "step": 5845 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.28862089464063e-06, | |
| "loss": 2.0362, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.2752981486242985e-06, | |
| "loss": 2.1417, | |
| "step": 5855 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.2620093140532528e-06, | |
| "loss": 2.0554, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.2487544492654832e-06, | |
| "loss": 2.0155, | |
| "step": 5865 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.235533612449854e-06, | |
| "loss": 2.1008, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.222346861645841e-06, | |
| "loss": 2.105, | |
| "step": 5875 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.209194254743295e-06, | |
| "loss": 2.0809, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.196075849482164e-06, | |
| "loss": 2.1432, | |
| "step": 5885 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.1829917034522552e-06, | |
| "loss": 2.1098, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.1699418740929744e-06, | |
| "loss": 2.0564, | |
| "step": 5895 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.1569264186930817e-06, | |
| "loss": 2.1031, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.143945394390429e-06, | |
| "loss": 2.1048, | |
| "step": 5905 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.1309988581717257e-06, | |
| "loss": 2.1031, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.118086866872271e-06, | |
| "loss": 2.1187, | |
| "step": 5915 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.1052094771757125e-06, | |
| "loss": 2.0922, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.0923667456137987e-06, | |
| "loss": 2.056, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.0795587285661278e-06, | |
| "loss": 2.1164, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.0667854822599022e-06, | |
| "loss": 2.0951, | |
| "step": 5935 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.054047062769675e-06, | |
| "loss": 2.0682, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.041343526017123e-06, | |
| "loss": 1.9337, | |
| "step": 5945 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.0286749277707783e-06, | |
| "loss": 2.0883, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.0160413236457855e-06, | |
| "loss": 2.1396, | |
| "step": 5955 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.0034427691036817e-06, | |
| "loss": 2.0591, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.990879319452125e-06, | |
| "loss": 2.0639, | |
| "step": 5965 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.978351029844665e-06, | |
| "loss": 2.1451, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.9658579552805036e-06, | |
| "loss": 2.0165, | |
| "step": 5975 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.9534001506042455e-06, | |
| "loss": 2.0768, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.9409776705056514e-06, | |
| "loss": 2.0771, | |
| "step": 5985 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.928590569519425e-06, | |
| "loss": 2.0962, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.9162389020249416e-06, | |
| "loss": 2.007, | |
| "step": 5995 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.90392272224603e-06, | |
| "loss": 2.0936, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.8916420842507232e-06, | |
| "loss": 2.0272, | |
| "step": 6005 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.8793970419510333e-06, | |
| "loss": 2.1202, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.867187649102702e-06, | |
| "loss": 2.1284, | |
| "step": 6015 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.8550139593049631e-06, | |
| "loss": 1.9971, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.8428760260003297e-06, | |
| "loss": 2.0249, | |
| "step": 6025 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.830773902474332e-06, | |
| "loss": 2.0139, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.8187076418552974e-06, | |
| "loss": 2.1007, | |
| "step": 6035 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.806677297114121e-06, | |
| "loss": 2.0515, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.7946829210640238e-06, | |
| "loss": 2.0003, | |
| "step": 6045 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.7827245663603133e-06, | |
| "loss": 2.115, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.7708022855001816e-06, | |
| "loss": 2.0915, | |
| "step": 6055 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.7589161308224466e-06, | |
| "loss": 1.9925, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.7470661545073319e-06, | |
| "loss": 2.059, | |
| "step": 6065 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.735252408576239e-06, | |
| "loss": 2.0153, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.7234749448915255e-06, | |
| "loss": 2.1008, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.7117338151562568e-06, | |
| "loss": 2.1203, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.7000290709139977e-06, | |
| "loss": 2.0918, | |
| "step": 6085 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.6883607635485877e-06, | |
| "loss": 2.1489, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.6767289442838985e-06, | |
| "loss": 2.0792, | |
| "step": 6095 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.6651336641836214e-06, | |
| "loss": 2.0856, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.65357497415105e-06, | |
| "loss": 2.0075, | |
| "step": 6105 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.642052924928832e-06, | |
| "loss": 1.9995, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.6305675670987686e-06, | |
| "loss": 2.0732, | |
| "step": 6115 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.6191189510815942e-06, | |
| "loss": 2.1005, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.607707127136734e-06, | |
| "loss": 2.1333, | |
| "step": 6125 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.596332145362104e-06, | |
| "loss": 2.1098, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.5849940556938782e-06, | |
| "loss": 2.0098, | |
| "step": 6135 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.5736929079062768e-06, | |
| "loss": 2.1078, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.5624287516113457e-06, | |
| "loss": 1.9704, | |
| "step": 6145 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.5512016362587346e-06, | |
| "loss": 2.023, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.540011611135489e-06, | |
| "loss": 2.1033, | |
| "step": 6155 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.5288587253658248e-06, | |
| "loss": 2.0905, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.5177430279109128e-06, | |
| "loss": 2.1233, | |
| "step": 6165 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.5066645675686787e-06, | |
| "loss": 2.1036, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.495623392973562e-06, | |
| "loss": 2.0099, | |
| "step": 6175 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.4846195525963247e-06, | |
| "loss": 1.9788, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.4736530947438377e-06, | |
| "loss": 2.0596, | |
| "step": 6185 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.4627240675588539e-06, | |
| "loss": 2.0375, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.451832519019808e-06, | |
| "loss": 2.0944, | |
| "step": 6195 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.4409784969406049e-06, | |
| "loss": 2.0623, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.4301620489704072e-06, | |
| "loss": 2.0963, | |
| "step": 6205 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.4193832225934267e-06, | |
| "loss": 2.1514, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.4086420651287157e-06, | |
| "loss": 2.0243, | |
| "step": 6215 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.3979386237299641e-06, | |
| "loss": 2.0516, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.387272945385285e-06, | |
| "loss": 2.067, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.3766450769170115e-06, | |
| "loss": 2.038, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.3660550649814918e-06, | |
| "loss": 2.1, | |
| "step": 6235 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.3555029560688838e-06, | |
| "loss": 1.9969, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.3449887965029485e-06, | |
| "loss": 2.0904, | |
| "step": 6245 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.3345126324408575e-06, | |
| "loss": 2.0902, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.3240745098729745e-06, | |
| "loss": 2.0967, | |
| "step": 6255 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3136744746226626e-06, | |
| "loss": 1.9768, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.3033125723460826e-06, | |
| "loss": 2.0477, | |
| "step": 6265 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.2929888485319908e-06, | |
| "loss": 2.1532, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.2827033485015406e-06, | |
| "loss": 1.9443, | |
| "step": 6275 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.2724561174080796e-06, | |
| "loss": 2.1782, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.2622472002369635e-06, | |
| "loss": 2.1286, | |
| "step": 6285 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.2520766418053408e-06, | |
| "loss": 2.0444, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.241944486761969e-06, | |
| "loss": 2.0266, | |
| "step": 6295 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.2318507795870138e-06, | |
| "loss": 2.1015, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.2217955645918568e-06, | |
| "loss": 2.0997, | |
| "step": 6305 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2117788859188928e-06, | |
| "loss": 2.0902, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.2018007875413518e-06, | |
| "loss": 2.163, | |
| "step": 6315 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.1918613132630896e-06, | |
| "loss": 1.9922, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.1819605067184025e-06, | |
| "loss": 2.0706, | |
| "step": 6325 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.1720984113718382e-06, | |
| "loss": 2.0752, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1622750705179986e-06, | |
| "loss": 2.0683, | |
| "step": 6335 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1524905272813558e-06, | |
| "loss": 2.1766, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1427448246160588e-06, | |
| "loss": 2.064, | |
| "step": 6345 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.133038005305751e-06, | |
| "loss": 2.1203, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.1233701119633723e-06, | |
| "loss": 2.0746, | |
| "step": 6355 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.1137411870309788e-06, | |
| "loss": 1.995, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.104151272779559e-06, | |
| "loss": 1.9195, | |
| "step": 6365 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.0946004113088381e-06, | |
| "loss": 2.0753, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.0850886445471055e-06, | |
| "loss": 2.087, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.0756160142510197e-06, | |
| "loss": 2.1446, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0661825620054366e-06, | |
| "loss": 2.1177, | |
| "step": 6385 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.056788329223215e-06, | |
| "loss": 2.13, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0474333571450412e-06, | |
| "loss": 2.0877, | |
| "step": 6395 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0381176868392483e-06, | |
| "loss": 2.1238, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.0288413592016345e-06, | |
| "loss": 2.1037, | |
| "step": 6405 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0196044149552798e-06, | |
| "loss": 2.0599, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0104068946503787e-06, | |
| "loss": 2.1005, | |
| "step": 6415 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.0012488386640484e-06, | |
| "loss": 2.0853, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 9.921302872001603e-07, | |
| "loss": 2.1091, | |
| "step": 6425 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 9.830512802891601e-07, | |
| "loss": 2.0541, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.74011857787892e-07, | |
| "loss": 2.0986, | |
| "step": 6435 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.650120593794266e-07, | |
| "loss": 2.1166, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.560519245728816e-07, | |
| "loss": 2.0038, | |
| "step": 6445 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.471314927032571e-07, | |
| "loss": 2.0662, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 9.382508029312521e-07, | |
| "loss": 2.0667, | |
| "step": 6455 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.294098942430996e-07, | |
| "loss": 2.132, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.206088054503937e-07, | |
| "loss": 2.1023, | |
| "step": 6465 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.118475751899192e-07, | |
| "loss": 2.0209, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 9.031262419234787e-07, | |
| "loss": 2.04, | |
| "step": 6475 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.94444843937734e-07, | |
| "loss": 1.9607, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.858034193440257e-07, | |
| "loss": 2.0577, | |
| "step": 6485 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.772020060782127e-07, | |
| "loss": 2.1076, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.686406419005022e-07, | |
| "loss": 2.101, | |
| "step": 6495 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.601193643952888e-07, | |
| "loss": 2.0948, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 8.516382109709854e-07, | |
| "loss": 2.0716, | |
| "step": 6505 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.431972188598581e-07, | |
| "loss": 1.972, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.347964251178697e-07, | |
| "loss": 1.9502, | |
| "step": 6515 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.264358666245098e-07, | |
| "loss": 1.9629, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.181155800826301e-07, | |
| "loss": 2.0311, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.098356020182973e-07, | |
| "loss": 1.9672, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 8.015959687806174e-07, | |
| "loss": 2.1489, | |
| "step": 6535 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.933967165415846e-07, | |
| "loss": 2.14, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.852378812959227e-07, | |
| "loss": 2.0682, | |
| "step": 6545 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.771194988609221e-07, | |
| "loss": 2.124, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.690416048762861e-07, | |
| "loss": 2.1139, | |
| "step": 6555 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.610042348039715e-07, | |
| "loss": 2.0541, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.530074239280405e-07, | |
| "loss": 2.0691, | |
| "step": 6565 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.450512073544946e-07, | |
| "loss": 2.0822, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.371356200111279e-07, | |
| "loss": 2.0511, | |
| "step": 6575 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.29260696647377e-07, | |
| "loss": 2.1127, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.214264718341591e-07, | |
| "loss": 2.0741, | |
| "step": 6585 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.13632979963721e-07, | |
| "loss": 2.0819, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.058802552495004e-07, | |
| "loss": 2.0707, | |
| "step": 6595 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 6.981683317259613e-07, | |
| "loss": 2.0275, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 6.904972432484513e-07, | |
| "loss": 2.0088, | |
| "step": 6605 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.828670234930556e-07, | |
| "loss": 2.1961, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.752777059564431e-07, | |
| "loss": 2.0664, | |
| "step": 6615 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.677293239557181e-07, | |
| "loss": 2.0626, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.602219106282836e-07, | |
| "loss": 2.0819, | |
| "step": 6625 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.527554989316898e-07, | |
| "loss": 2.0876, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.453301216434871e-07, | |
| "loss": 2.0798, | |
| "step": 6635 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.379458113610825e-07, | |
| "loss": 2.0956, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.306026005016086e-07, | |
| "loss": 2.1568, | |
| "step": 6645 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.233005213017607e-07, | |
| "loss": 2.0234, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.160396058176687e-07, | |
| "loss": 1.9598, | |
| "step": 6655 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.088198859247596e-07, | |
| "loss": 2.0365, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.016413933176057e-07, | |
| "loss": 2.0681, | |
| "step": 6665 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.945041595097922e-07, | |
| "loss": 2.1563, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.874082158337835e-07, | |
| "loss": 2.0175, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 5.803535934407734e-07, | |
| "loss": 2.0223, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.733403233005563e-07, | |
| "loss": 2.0616, | |
| "step": 6685 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.663684362013944e-07, | |
| "loss": 2.0521, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.594379627498747e-07, | |
| "loss": 2.0198, | |
| "step": 6695 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.525489333707812e-07, | |
| "loss": 2.077, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 5.45701378306952e-07, | |
| "loss": 2.1654, | |
| "step": 6705 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.388953276191644e-07, | |
| "loss": 2.0618, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.321308111859791e-07, | |
| "loss": 2.0267, | |
| "step": 6715 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.254078587036282e-07, | |
| "loss": 2.1178, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.187264996858776e-07, | |
| "loss": 2.1309, | |
| "step": 6725 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 5.120867634638971e-07, | |
| "loss": 2.0864, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.054886791861291e-07, | |
| "loss": 2.0438, | |
| "step": 6735 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.989322758181736e-07, | |
| "loss": 2.0632, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.924175821426392e-07, | |
| "loss": 2.0679, | |
| "step": 6745 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.859446267590351e-07, | |
| "loss": 2.0548, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.795134380836419e-07, | |
| "loss": 1.9926, | |
| "step": 6755 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.7312404434938007e-07, | |
| "loss": 2.09, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.667764736056912e-07, | |
| "loss": 2.0877, | |
| "step": 6765 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.604707537184139e-07, | |
| "loss": 2.1266, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.542069123696624e-07, | |
| "loss": 2.1788, | |
| "step": 6775 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.479849770577038e-07, | |
| "loss": 2.0806, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.418049750968356e-07, | |
| "loss": 2.0333, | |
| "step": 6785 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.3566693361727254e-07, | |
| "loss": 2.0392, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.2957087956501885e-07, | |
| "loss": 2.0723, | |
| "step": 6795 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.235168397017542e-07, | |
| "loss": 2.1018, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.1750484060471885e-07, | |
| "loss": 2.0483, | |
| "step": 6805 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.1153490866658983e-07, | |
| "loss": 2.0559, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.0560707009537184e-07, | |
| "loss": 2.0806, | |
| "step": 6815 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 3.9972135091428165e-07, | |
| "loss": 2.0804, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 3.9387777696162757e-07, | |
| "loss": 2.0594, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 3.880763738907034e-07, | |
| "loss": 2.0933, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.823171671696713e-07, | |
| "loss": 1.9843, | |
| "step": 6835 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.766001820814513e-07, | |
| "loss": 2.1302, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.7092544372361314e-07, | |
| "loss": 2.0555, | |
| "step": 6845 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.6529297700825694e-07, | |
| "loss": 2.0146, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 3.59702806661919e-07, | |
| "loss": 2.138, | |
| "step": 6855 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.541549572254488e-07, | |
| "loss": 2.1426, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.4864945305390863e-07, | |
| "loss": 2.0187, | |
| "step": 6865 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.4318631831646297e-07, | |
| "loss": 2.0497, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.3776557699627844e-07, | |
| "loss": 1.9925, | |
| "step": 6875 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.323872528904104e-07, | |
| "loss": 2.0441, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.2705136960970554e-07, | |
| "loss": 2.0205, | |
| "step": 6885 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.2175795057869494e-07, | |
| "loss": 2.0661, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.165070190354913e-07, | |
| "loss": 2.0878, | |
| "step": 6895 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.1129859803168516e-07, | |
| "loss": 2.0933, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.061327104322487e-07, | |
| "loss": 2.0879, | |
| "step": 6905 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.0100937891543116e-07, | |
| "loss": 2.0346, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.9592862597266013e-07, | |
| "loss": 2.0495, | |
| "step": 6915 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.9089047390844395e-07, | |
| "loss": 1.9751, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.858949448402737e-07, | |
| "loss": 2.0457, | |
| "step": 6925 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 2.809420606985236e-07, | |
| "loss": 2.0559, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.760318432263587e-07, | |
| "loss": 2.0475, | |
| "step": 6935 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.7116431397963604e-07, | |
| "loss": 2.0471, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.663394943268094e-07, | |
| "loss": 2.0074, | |
| "step": 6945 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.615574054488434e-07, | |
| "loss": 1.9663, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.568180683391097e-07, | |
| "loss": 2.1215, | |
| "step": 6955 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.521215038032987e-07, | |
| "loss": 2.0805, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.474677324593333e-07, | |
| "loss": 2.0014, | |
| "step": 6965 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.4285677473727123e-07, | |
| "loss": 2.0507, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.3828865087922038e-07, | |
| "loss": 2.1024, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.3376338093924477e-07, | |
| "loss": 2.1282, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.2928098478328442e-07, | |
| "loss": 2.0834, | |
| "step": 6985 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.2484148208905987e-07, | |
| "loss": 2.0063, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.2044489234599008e-07, | |
| "loss": 1.9253, | |
| "step": 6995 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.1609123485510697e-07, | |
| "loss": 2.0383, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.117805287289687e-07, | |
| "loss": 2.0034, | |
| "step": 7005 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.0751279289157656e-07, | |
| "loss": 2.0936, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.0328804607829488e-07, | |
| "loss": 2.0141, | |
| "step": 7015 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.9910630683576458e-07, | |
| "loss": 2.1539, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.9496759352182204e-07, | |
| "loss": 2.0881, | |
| "step": 7025 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.9087192430542135e-07, | |
| "loss": 2.1645, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.868193171665522e-07, | |
| "loss": 2.0837, | |
| "step": 7035 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.8280978989616338e-07, | |
| "loss": 2.0468, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.7884336009607928e-07, | |
| "loss": 2.1323, | |
| "step": 7045 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.7492004517893125e-07, | |
| "loss": 2.0991, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 1.7103986236807312e-07, | |
| "loss": 2.072, | |
| "step": 7055 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.6720282869751025e-07, | |
| "loss": 2.0069, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.6340896101182058e-07, | |
| "loss": 2.0095, | |
| "step": 7065 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.5965827596608695e-07, | |
| "loss": 2.0803, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.5595079002581836e-07, | |
| "loss": 2.0156, | |
| "step": 7075 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.522865194668799e-07, | |
| "loss": 2.131, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.486654803754206e-07, | |
| "loss": 2.1106, | |
| "step": 7085 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.4508768864780586e-07, | |
| "loss": 2.0862, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.4155315999054176e-07, | |
| "loss": 2.0843, | |
| "step": 7095 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.3806190992021185e-07, | |
| "loss": 2.0172, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.3461395376340502e-07, | |
| "loss": 2.1321, | |
| "step": 7105 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.3120930665665e-07, | |
| "loss": 2.1629, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.2784798354635196e-07, | |
| "loss": 2.1107, | |
| "step": 7115 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.2452999918871943e-07, | |
| "loss": 2.0813, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.212553681497064e-07, | |
| "loss": 2.0589, | |
| "step": 7125 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.1802410480494353e-07, | |
| "loss": 2.035, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.1483622333967936e-07, | |
| "loss": 2.0216, | |
| "step": 7135 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.1169173774871478e-07, | |
| "loss": 1.978, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.0859066183634414e-07, | |
| "loss": 2.0482, | |
| "step": 7145 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.0553300921629206e-07, | |
| "loss": 2.1282, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.0251879331165559e-07, | |
| "loss": 2.0081, | |
| "step": 7155 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.954802735484214e-08, | |
| "loss": 2.1178, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.662072438751835e-08, | |
| "loss": 2.0898, | |
| "step": 7165 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.373689726054458e-08, | |
| "loss": 2.032, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 9.089655863392389e-08, | |
| "loss": 1.996, | |
| "step": 7175 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 8.809972097674424e-08, | |
| "loss": 2.1137, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.53463965671253e-08, | |
| "loss": 1.9833, | |
| "step": 7185 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 8.263659749215946e-08, | |
| "loss": 2.1127, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.997033564787093e-08, | |
| "loss": 2.0218, | |
| "step": 7195 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.734762273914675e-08, | |
| "loss": 2.1112, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 7.476847027970024e-08, | |
| "loss": 2.0171, | |
| "step": 7205 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 7.223288959200991e-08, | |
| "loss": 2.0654, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.97408918072795e-08, | |
| "loss": 2.012, | |
| "step": 7215 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.729248786537912e-08, | |
| "loss": 2.1098, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.488768851480087e-08, | |
| "loss": 2.0205, | |
| "step": 7225 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 6.252650431261886e-08, | |
| "loss": 2.0984, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 6.020894562443146e-08, | |
| "loss": 2.0464, | |
| "step": 7235 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.793502262432249e-08, | |
| "loss": 2.103, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.5704745294815624e-08, | |
| "loss": 2.0956, | |
| "step": 7245 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.3518123426830046e-08, | |
| "loss": 2.0641, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 5.137516661963493e-08, | |
| "loss": 2.1449, | |
| "step": 7255 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.927588428081387e-08, | |
| "loss": 2.0376, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.7220285626216054e-08, | |
| "loss": 2.0383, | |
| "step": 7265 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.5208379679921865e-08, | |
| "loss": 2.1728, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.3240175274197325e-08, | |
| "loss": 2.093, | |
| "step": 7275 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.131568104946415e-08, | |
| "loss": 2.0098, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.943490545425088e-08, | |
| "loss": 1.998, | |
| "step": 7285 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.759785674516292e-08, | |
| "loss": 2.099, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.580454298684699e-08, | |
| "loss": 2.041, | |
| "step": 7295 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.405497205195007e-08, | |
| "loss": 2.1463, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.234915162108942e-08, | |
| "loss": 2.1113, | |
| "step": 7305 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 3.0687089182819264e-08, | |
| "loss": 2.0994, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.906879203359303e-08, | |
| "loss": 1.9752, | |
| "step": 7315 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.7494267277735632e-08, | |
| "loss": 1.9797, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.5963521827412352e-08, | |
| "loss": 2.0584, | |
| "step": 7325 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.447656240259777e-08, | |
| "loss": 2.0813, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 2.3033395531042448e-08, | |
| "loss": 2.0428, | |
| "step": 7335 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 2.1634027548250723e-08, | |
| "loss": 2.0691, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 2.0278464597449644e-08, | |
| "loss": 2.0987, | |
| "step": 7345 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.896671262955896e-08, | |
| "loss": 2.1581, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.7698777403173383e-08, | |
| "loss": 2.0984, | |
| "step": 7355 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.6474664484527058e-08, | |
| "loss": 2.0336, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.5294379247476898e-08, | |
| "loss": 2.1252, | |
| "step": 7365 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.4157926873475947e-08, | |
| "loss": 2.0692, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.3065312351552283e-08, | |
| "loss": 2.1003, | |
| "step": 7375 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.201654047828127e-08, | |
| "loss": 2.1239, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.1011615857773328e-08, | |
| "loss": 2.1577, | |
| "step": 7385 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.0050542901648419e-08, | |
| "loss": 2.1533, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 9.13332582901716e-09, | |
| "loss": 2.0182, | |
| "step": 7395 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.259968666463058e-09, | |
| "loss": 2.1697, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.43047524802476e-09, | |
| "loss": 2.0327, | |
| "step": 7405 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.644849215178273e-09, | |
| "loss": 2.0703, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.903094016822542e-09, | |
| "loss": 2.0087, | |
| "step": 7415 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.205212909262791e-09, | |
| "loss": 2.0258, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.551208956198317e-09, | |
| "loss": 1.9995, | |
| "step": 7425 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 3.941085028705827e-09, | |
| "loss": 2.054, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.3748438052305656e-09, | |
| "loss": 2.1216, | |
| "step": 7435 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.852487771571877e-09, | |
| "loss": 2.0869, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.3740192208721034e-09, | |
| "loss": 2.0625, | |
| "step": 7445 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 1.9394402536099255e-09, | |
| "loss": 2.0759, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 1.5487527775848165e-09, | |
| "loss": 2.0811, | |
| "step": 7455 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.201958507918155e-09, | |
| "loss": 2.1014, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 8.990589670343497e-10, | |
| "loss": 2.0307, | |
| "step": 7465 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 6.400554846641705e-10, | |
| "loss": 2.0364, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.2494919783142574e-10, | |
| "loss": 2.0335, | |
| "step": 7475 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.5374105085518297e-10, | |
| "loss": 2.0692, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 1.264317953364458e-10, | |
| "loss": 2.0747, | |
| "step": 7485 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.3021990164815455e-11, | |
| "loss": 2.0857, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.5120015096090644e-12, | |
| "loss": 2.1121, | |
| "step": 7495 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 7497, | |
| "total_flos": 4.954742374840074e+17, | |
| "train_loss": 2.0977668136346335, | |
| "train_runtime": 13801.1295, | |
| "train_samples_per_second": 8.691, | |
| "train_steps_per_second": 0.543 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 7497, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "total_flos": 4.954742374840074e+17, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |