Spaces:
Build error
Build error
| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9399674144629653, | |
| "global_step": 7500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 0.7141, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 0.7055, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3e-06, | |
| "loss": 0.6885, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.6619, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6147, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6e-06, | |
| "loss": 0.5332, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 7.000000000000001e-06, | |
| "loss": 0.4607, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.3797, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 9e-06, | |
| "loss": 0.3388, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1e-05, | |
| "loss": 0.27, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 0.2613, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.2139, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 0.2052, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 0.1641, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.1887, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.1547, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.7000000000000003e-05, | |
| "loss": 0.1562, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.1266, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9e-05, | |
| "loss": 0.1668, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2e-05, | |
| "loss": 0.15, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.1e-05, | |
| "loss": 0.1001, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 0.1054, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.3000000000000003e-05, | |
| "loss": 0.0753, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.4e-05, | |
| "loss": 0.0855, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.0987, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 0.091, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.7000000000000002e-05, | |
| "loss": 0.0597, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 0.082, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 2.9e-05, | |
| "loss": 0.0953, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3e-05, | |
| "loss": 0.0678, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.1e-05, | |
| "loss": 0.0692, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 0.0521, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.3e-05, | |
| "loss": 0.0749, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 0.0855, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.0697, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.6e-05, | |
| "loss": 0.0881, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.7e-05, | |
| "loss": 0.0754, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.8e-05, | |
| "loss": 0.0619, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.9000000000000006e-05, | |
| "loss": 0.0509, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4e-05, | |
| "loss": 0.0641, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.1e-05, | |
| "loss": 0.0473, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.2e-05, | |
| "loss": 0.0624, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.3e-05, | |
| "loss": 0.0774, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 0.0608, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.0658, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 0.0696, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.7e-05, | |
| "loss": 0.0577, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.8e-05, | |
| "loss": 0.0439, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9e-05, | |
| "loss": 0.0345, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 5e-05, | |
| "loss": 0.075, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.993314614253243e-05, | |
| "loss": 0.0508, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.986629228506485e-05, | |
| "loss": 0.0911, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9799438427597275e-05, | |
| "loss": 0.0621, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.97325845701297e-05, | |
| "loss": 0.0669, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9665730712662124e-05, | |
| "loss": 0.045, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.959887685519455e-05, | |
| "loss": 0.0557, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.953202299772697e-05, | |
| "loss": 0.057, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.946516914025939e-05, | |
| "loss": 0.0718, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.939831528279182e-05, | |
| "loss": 0.0379, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9331461425324245e-05, | |
| "loss": 0.0516, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.926460756785667e-05, | |
| "loss": 0.0429, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9197753710389094e-05, | |
| "loss": 0.0652, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.913089985292151e-05, | |
| "loss": 0.0538, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.9064045995453936e-05, | |
| "loss": 0.0708, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.899719213798637e-05, | |
| "loss": 0.0316, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.893033828051879e-05, | |
| "loss": 0.0678, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8863484423051216e-05, | |
| "loss": 0.0645, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.879663056558363e-05, | |
| "loss": 0.0879, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.872977670811606e-05, | |
| "loss": 0.0391, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.866292285064848e-05, | |
| "loss": 0.0945, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.859606899318091e-05, | |
| "loss": 0.0726, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.852921513571334e-05, | |
| "loss": 0.0898, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8462361278245755e-05, | |
| "loss": 0.0664, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.839550742077818e-05, | |
| "loss": 0.0519, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.83286535633106e-05, | |
| "loss": 0.0645, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.826179970584303e-05, | |
| "loss": 0.0328, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.819494584837546e-05, | |
| "loss": 0.042, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8128091990907876e-05, | |
| "loss": 0.0897, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.80612381334403e-05, | |
| "loss": 0.0702, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.7994384275972725e-05, | |
| "loss": 0.0582, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.792753041850515e-05, | |
| "loss": 0.0381, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.786067656103757e-05, | |
| "loss": 0.0623, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.779382270357e-05, | |
| "loss": 0.0628, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.772696884610242e-05, | |
| "loss": 0.0543, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.7660114988634846e-05, | |
| "loss": 0.0455, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.759326113116727e-05, | |
| "loss": 0.0511, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.7526407273699695e-05, | |
| "loss": 0.0421, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.745955341623212e-05, | |
| "loss": 0.0781, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.7392699558764544e-05, | |
| "loss": 0.0389, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.732584570129697e-05, | |
| "loss": 0.0512, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.725899184382939e-05, | |
| "loss": 0.0558, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7192137986361816e-05, | |
| "loss": 0.0535, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7125284128894234e-05, | |
| "loss": 0.0811, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7058430271426665e-05, | |
| "loss": 0.0587, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.699157641395909e-05, | |
| "loss": 0.0865, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.6924722556491514e-05, | |
| "loss": 0.0483, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.685786869902394e-05, | |
| "loss": 0.0479, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.6791014841556355e-05, | |
| "loss": 0.0682, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.672416098408878e-05, | |
| "loss": 0.0807, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.665730712662121e-05, | |
| "loss": 0.0487, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.6590453269153635e-05, | |
| "loss": 0.0375, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.652359941168606e-05, | |
| "loss": 0.0318, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.645674555421848e-05, | |
| "loss": 0.0575, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.63898916967509e-05, | |
| "loss": 0.0312, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.6323037839283326e-05, | |
| "loss": 0.0328, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.625618398181576e-05, | |
| "loss": 0.0448, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.618933012434818e-05, | |
| "loss": 0.0525, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.61224762668806e-05, | |
| "loss": 0.0384, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.605562240941302e-05, | |
| "loss": 0.0598, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.598876855194545e-05, | |
| "loss": 0.0279, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.592191469447787e-05, | |
| "loss": 0.0765, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.58550608370103e-05, | |
| "loss": 0.0526, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.578820697954272e-05, | |
| "loss": 0.0595, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.5721353122075144e-05, | |
| "loss": 0.0373, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.565449926460757e-05, | |
| "loss": 0.0443, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.558764540713999e-05, | |
| "loss": 0.041, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.552079154967242e-05, | |
| "loss": 0.0664, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.545393769220484e-05, | |
| "loss": 0.0544, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.5387083834737266e-05, | |
| "loss": 0.0832, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.532022997726969e-05, | |
| "loss": 0.0731, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.5253376119802115e-05, | |
| "loss": 0.0496, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.518652226233454e-05, | |
| "loss": 0.0478, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.511966840486696e-05, | |
| "loss": 0.0502, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.505281454739939e-05, | |
| "loss": 0.0392, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.498596068993181e-05, | |
| "loss": 0.0393, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.4919106832464236e-05, | |
| "loss": 0.051, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.485225297499666e-05, | |
| "loss": 0.0487, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.478539911752908e-05, | |
| "loss": 0.0411, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.471854526006151e-05, | |
| "loss": 0.0466, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.465169140259393e-05, | |
| "loss": 0.0604, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.458483754512636e-05, | |
| "loss": 0.0457, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.451798368765878e-05, | |
| "loss": 0.037, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.44511298301912e-05, | |
| "loss": 0.0573, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.4384275972723624e-05, | |
| "loss": 0.0548, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.4317422115256055e-05, | |
| "loss": 0.0381, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.425056825778848e-05, | |
| "loss": 0.1086, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.4183714400320903e-05, | |
| "loss": 0.039, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.411686054285332e-05, | |
| "loss": 0.0499, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.4050006685385745e-05, | |
| "loss": 0.0326, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.398315282791817e-05, | |
| "loss": 0.0159, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.39162989704506e-05, | |
| "loss": 0.0553, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.3849445112983025e-05, | |
| "loss": 0.0654, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.378259125551544e-05, | |
| "loss": 0.0732, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.371573739804787e-05, | |
| "loss": 0.0598, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.364888354058029e-05, | |
| "loss": 0.0495, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.3582029683112715e-05, | |
| "loss": 0.0773, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.3515175825645146e-05, | |
| "loss": 0.0368, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.3448321968177564e-05, | |
| "loss": 0.0349, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.338146811070999e-05, | |
| "loss": 0.0587, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.331461425324241e-05, | |
| "loss": 0.0426, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.324776039577484e-05, | |
| "loss": 0.0483, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.318090653830726e-05, | |
| "loss": 0.0628, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.3114052680839686e-05, | |
| "loss": 0.0569, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.304719882337211e-05, | |
| "loss": 0.0485, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.2980344965904534e-05, | |
| "loss": 0.0457, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.291349110843696e-05, | |
| "loss": 0.0529, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.284663725096938e-05, | |
| "loss": 0.0471, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.277978339350181e-05, | |
| "loss": 0.0494, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.271292953603423e-05, | |
| "loss": 0.0402, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.2646075678566656e-05, | |
| "loss": 0.0277, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.257922182109908e-05, | |
| "loss": 0.0748, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.2512367963631504e-05, | |
| "loss": 0.0439, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.244551410616392e-05, | |
| "loss": 0.0468, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.237866024869635e-05, | |
| "loss": 0.0555, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.231180639122878e-05, | |
| "loss": 0.0473, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.22449525337612e-05, | |
| "loss": 0.0458, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.2178098676293626e-05, | |
| "loss": 0.0583, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.211124481882604e-05, | |
| "loss": 0.0405, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.204439096135847e-05, | |
| "loss": 0.034, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.19775371038909e-05, | |
| "loss": 0.0857, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.191068324642332e-05, | |
| "loss": 0.0415, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.184382938895575e-05, | |
| "loss": 0.0621, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.1776975531488165e-05, | |
| "loss": 0.055, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.171012167402059e-05, | |
| "loss": 0.0267, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.1643267816553013e-05, | |
| "loss": 0.0433, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.1576413959085445e-05, | |
| "loss": 0.0384, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.150956010161787e-05, | |
| "loss": 0.0485, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.1442706244150286e-05, | |
| "loss": 0.0278, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.137585238668271e-05, | |
| "loss": 0.0669, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.1308998529215135e-05, | |
| "loss": 0.0554, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.124214467174756e-05, | |
| "loss": 0.0392, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.117529081427999e-05, | |
| "loss": 0.0632, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.1108436956812415e-05, | |
| "loss": 0.0494, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.104158309934483e-05, | |
| "loss": 0.0334, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.0974729241877256e-05, | |
| "loss": 0.0658, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.090787538440968e-05, | |
| "loss": 0.0456, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.0841021526942105e-05, | |
| "loss": 0.0359, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.0774167669474536e-05, | |
| "loss": 0.0513, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.0707313812006954e-05, | |
| "loss": 0.0391, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.064045995453938e-05, | |
| "loss": 0.0507, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.05736060970718e-05, | |
| "loss": 0.0458, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.050675223960423e-05, | |
| "loss": 0.0572, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.043989838213665e-05, | |
| "loss": 0.0326, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.0373044524669075e-05, | |
| "loss": 0.0446, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.03061906672015e-05, | |
| "loss": 0.0359, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.0239336809733924e-05, | |
| "loss": 0.0689, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.017248295226635e-05, | |
| "loss": 0.0165, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.010562909479877e-05, | |
| "loss": 0.0442, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.00387752373312e-05, | |
| "loss": 0.0343, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.997192137986362e-05, | |
| "loss": 0.0684, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.9905067522396045e-05, | |
| "loss": 0.0565, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.983821366492847e-05, | |
| "loss": 0.0534, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.9771359807460894e-05, | |
| "loss": 0.0362, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.970450594999331e-05, | |
| "loss": 0.0464, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.963765209252574e-05, | |
| "loss": 0.021, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.957079823505817e-05, | |
| "loss": 0.0531, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.950394437759059e-05, | |
| "loss": 0.0669, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.9437090520123016e-05, | |
| "loss": 0.0509, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.937023666265543e-05, | |
| "loss": 0.0409, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.930338280518786e-05, | |
| "loss": 0.0518, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.923652894772029e-05, | |
| "loss": 0.0316, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.916967509025271e-05, | |
| "loss": 0.0479, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.910282123278514e-05, | |
| "loss": 0.0329, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.9035967375317555e-05, | |
| "loss": 0.032, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.896911351784998e-05, | |
| "loss": 0.053, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.89022596603824e-05, | |
| "loss": 0.0545, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.8835405802914834e-05, | |
| "loss": 0.0269, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.876855194544726e-05, | |
| "loss": 0.0408, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.8701698087979676e-05, | |
| "loss": 0.0334, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.86348442305121e-05, | |
| "loss": 0.041, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.8567990373044525e-05, | |
| "loss": 0.0353, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.850113651557695e-05, | |
| "loss": 0.025, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.843428265810938e-05, | |
| "loss": 0.0365, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.83674288006418e-05, | |
| "loss": 0.0366, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.830057494317422e-05, | |
| "loss": 0.0451, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.8233721085706646e-05, | |
| "loss": 0.0495, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.816686722823907e-05, | |
| "loss": 0.0451, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.8100013370771495e-05, | |
| "loss": 0.0441, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.803315951330392e-05, | |
| "loss": 0.0532, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.7966305655836343e-05, | |
| "loss": 0.0666, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.789945179836877e-05, | |
| "loss": 0.0434, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.783259794090119e-05, | |
| "loss": 0.0522, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.7765744083433616e-05, | |
| "loss": 0.0368, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.769889022596604e-05, | |
| "loss": 0.0386, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.7632036368498465e-05, | |
| "loss": 0.0639, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.756518251103089e-05, | |
| "loss": 0.0348, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.7498328653563314e-05, | |
| "loss": 0.0332, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.743147479609574e-05, | |
| "loss": 0.0439, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.7364620938628155e-05, | |
| "loss": 0.0433, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.7297767081160587e-05, | |
| "loss": 0.0506, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.723091322369301e-05, | |
| "loss": 0.0618, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.7164059366225435e-05, | |
| "loss": 0.0326, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.709720550875786e-05, | |
| "loss": 0.0589, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.703035165129028e-05, | |
| "loss": 0.0586, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.69634977938227e-05, | |
| "loss": 0.0616, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.689664393635513e-05, | |
| "loss": 0.0355, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.682979007888756e-05, | |
| "loss": 0.0399, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.676293622141998e-05, | |
| "loss": 0.0609, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.66960823639524e-05, | |
| "loss": 0.0679, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.662922850648482e-05, | |
| "loss": 0.0427, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.656237464901725e-05, | |
| "loss": 0.0333, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.649552079154968e-05, | |
| "loss": 0.0329, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.64286669340821e-05, | |
| "loss": 0.048, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.636181307661452e-05, | |
| "loss": 0.0617, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.6294959219146944e-05, | |
| "loss": 0.061, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.622810536167937e-05, | |
| "loss": 0.0551, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.616125150421179e-05, | |
| "loss": 0.05, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.6094397646744224e-05, | |
| "loss": 0.0198, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.602754378927664e-05, | |
| "loss": 0.0502, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.5960689931809066e-05, | |
| "loss": 0.0549, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.589383607434149e-05, | |
| "loss": 0.0385, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.5826982216873914e-05, | |
| "loss": 0.0594, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.576012835940634e-05, | |
| "loss": 0.0599, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.569327450193876e-05, | |
| "loss": 0.0456, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.562642064447119e-05, | |
| "loss": 0.0379, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.555956678700361e-05, | |
| "loss": 0.0476, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.5492712929536036e-05, | |
| "loss": 0.055, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.542585907206846e-05, | |
| "loss": 0.0425, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.5359005214600885e-05, | |
| "loss": 0.0428, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.529215135713331e-05, | |
| "loss": 0.0264, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.522529749966573e-05, | |
| "loss": 0.0526, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.515844364219816e-05, | |
| "loss": 0.0324, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.509158978473058e-05, | |
| "loss": 0.0385, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.5024735927263e-05, | |
| "loss": 0.0675, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.495788206979543e-05, | |
| "loss": 0.0556, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.4891028212327855e-05, | |
| "loss": 0.0486, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.482417435486028e-05, | |
| "loss": 0.0347, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.47573204973927e-05, | |
| "loss": 0.0359, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.469046663992512e-05, | |
| "loss": 0.0684, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.4623612782457545e-05, | |
| "loss": 0.0524, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.4556758924989976e-05, | |
| "loss": 0.0604, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.44899050675224e-05, | |
| "loss": 0.0377, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.4423051210054825e-05, | |
| "loss": 0.0592, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.435619735258724e-05, | |
| "loss": 0.0483, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.428934349511967e-05, | |
| "loss": 0.0544, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.422248963765209e-05, | |
| "loss": 0.0554, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.415563578018452e-05, | |
| "loss": 0.0427, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.4088781922716946e-05, | |
| "loss": 0.0514, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.4021928065249364e-05, | |
| "loss": 0.0445, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.395507420778179e-05, | |
| "loss": 0.0453, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.388822035031421e-05, | |
| "loss": 0.0435, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.382136649284664e-05, | |
| "loss": 0.027, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.375451263537907e-05, | |
| "loss": 0.0369, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.3687658777911485e-05, | |
| "loss": 0.0629, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.362080492044391e-05, | |
| "loss": 0.0589, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.3553951062976334e-05, | |
| "loss": 0.0449, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.348709720550876e-05, | |
| "loss": 0.0464, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.342024334804118e-05, | |
| "loss": 0.0477, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.335338949057361e-05, | |
| "loss": 0.0263, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.328653563310603e-05, | |
| "loss": 0.0477, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.3219681775638456e-05, | |
| "loss": 0.038, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.315282791817088e-05, | |
| "loss": 0.0581, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.3085974060703304e-05, | |
| "loss": 0.0237, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.301912020323573e-05, | |
| "loss": 0.0469, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.295226634576815e-05, | |
| "loss": 0.0262, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.288541248830058e-05, | |
| "loss": 0.0294, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.2818558630833e-05, | |
| "loss": 0.0421, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.2751704773365426e-05, | |
| "loss": 0.0295, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.268485091589785e-05, | |
| "loss": 0.0309, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.2617997058430274e-05, | |
| "loss": 0.0434, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.25511432009627e-05, | |
| "loss": 0.0609, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.248428934349512e-05, | |
| "loss": 0.0401, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.241743548602755e-05, | |
| "loss": 0.0235, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.2350581628559965e-05, | |
| "loss": 0.0369, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.228372777109239e-05, | |
| "loss": 0.0446, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.221687391362482e-05, | |
| "loss": 0.0289, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.2150020056157244e-05, | |
| "loss": 0.0369, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.208316619868967e-05, | |
| "loss": 0.0243, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.2016312341222086e-05, | |
| "loss": 0.0308, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.194945848375451e-05, | |
| "loss": 0.0564, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.1882604626286935e-05, | |
| "loss": 0.0855, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.1815750768819366e-05, | |
| "loss": 0.045, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.174889691135179e-05, | |
| "loss": 0.0563, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.168204305388421e-05, | |
| "loss": 0.052, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.161518919641663e-05, | |
| "loss": 0.0467, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.1548335338949056e-05, | |
| "loss": 0.0346, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.148148148148148e-05, | |
| "loss": 0.041, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.141462762401391e-05, | |
| "loss": 0.0345, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.134777376654633e-05, | |
| "loss": 0.043, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.1280919909078754e-05, | |
| "loss": 0.0356, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 3.121406605161118e-05, | |
| "loss": 0.052, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.11472121941436e-05, | |
| "loss": 0.0477, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.1080358336676027e-05, | |
| "loss": 0.0307, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.101350447920845e-05, | |
| "loss": 0.0896, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.0946650621740875e-05, | |
| "loss": 0.0426, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.08797967642733e-05, | |
| "loss": 0.0581, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.0812942906805724e-05, | |
| "loss": 0.0531, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.074608904933815e-05, | |
| "loss": 0.0313, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 3.067923519187057e-05, | |
| "loss": 0.0296, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.0612381334403e-05, | |
| "loss": 0.0459, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.054552747693542e-05, | |
| "loss": 0.0298, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.0478673619467845e-05, | |
| "loss": 0.062, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.041181976200027e-05, | |
| "loss": 0.0329, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.034496590453269e-05, | |
| "loss": 0.0534, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.0278112047065115e-05, | |
| "loss": 0.045, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.0211258189597543e-05, | |
| "loss": 0.0533, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 3.0144404332129967e-05, | |
| "loss": 0.0584, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 3.007755047466239e-05, | |
| "loss": 0.0691, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 3.0010696617194812e-05, | |
| "loss": 0.0473, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.9943842759727236e-05, | |
| "loss": 0.0391, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.987698890225966e-05, | |
| "loss": 0.0325, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.981013504479209e-05, | |
| "loss": 0.0619, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.9743281187324513e-05, | |
| "loss": 0.0249, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.967642732985693e-05, | |
| "loss": 0.0431, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.9609573472389358e-05, | |
| "loss": 0.0391, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9542719614921782e-05, | |
| "loss": 0.0374, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9475865757454206e-05, | |
| "loss": 0.0354, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9409011899986634e-05, | |
| "loss": 0.0467, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9342158042519052e-05, | |
| "loss": 0.0186, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9275304185051476e-05, | |
| "loss": 0.0408, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9208450327583904e-05, | |
| "loss": 0.0468, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9141596470116328e-05, | |
| "loss": 0.053, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.9074742612648752e-05, | |
| "loss": 0.0489, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.9007888755181173e-05, | |
| "loss": 0.0854, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8941034897713598e-05, | |
| "loss": 0.0521, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8874181040246022e-05, | |
| "loss": 0.0383, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.880732718277845e-05, | |
| "loss": 0.0524, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8740473325310874e-05, | |
| "loss": 0.0164, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8673619467843295e-05, | |
| "loss": 0.0402, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.860676561037572e-05, | |
| "loss": 0.0559, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8539911752908143e-05, | |
| "loss": 0.0442, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8473057895440568e-05, | |
| "loss": 0.0563, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8406204037972995e-05, | |
| "loss": 0.0336, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8339350180505413e-05, | |
| "loss": 0.0225, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.827249632303784e-05, | |
| "loss": 0.0447, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8205642465570265e-05, | |
| "loss": 0.0523, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.813878860810269e-05, | |
| "loss": 0.0383, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8071934750635114e-05, | |
| "loss": 0.0623, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.8005080893167534e-05, | |
| "loss": 0.0593, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.793822703569996e-05, | |
| "loss": 0.0557, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.7871373178232386e-05, | |
| "loss": 0.0378, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.780451932076481e-05, | |
| "loss": 0.0453, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.7737665463297235e-05, | |
| "loss": 0.0486, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.7670811605829656e-05, | |
| "loss": 0.0371, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.760395774836208e-05, | |
| "loss": 0.0785, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.7537103890894505e-05, | |
| "loss": 0.0258, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7470250033426932e-05, | |
| "loss": 0.0417, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7403396175959357e-05, | |
| "loss": 0.0419, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7336542318491774e-05, | |
| "loss": 0.025, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7269688461024202e-05, | |
| "loss": 0.0528, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7202834603556626e-05, | |
| "loss": 0.0461, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.713598074608905e-05, | |
| "loss": 0.0362, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7069126888621478e-05, | |
| "loss": 0.031, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.7002273031153896e-05, | |
| "loss": 0.0961, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.693541917368632e-05, | |
| "loss": 0.0249, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6868565316218748e-05, | |
| "loss": 0.0622, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6801711458751172e-05, | |
| "loss": 0.0567, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6734857601283596e-05, | |
| "loss": 0.0346, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6668003743816017e-05, | |
| "loss": 0.0555, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.660114988634844e-05, | |
| "loss": 0.0486, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6534296028880866e-05, | |
| "loss": 0.0431, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.6467442171413293e-05, | |
| "loss": 0.0649, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.6400588313945718e-05, | |
| "loss": 0.0697, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.633373445647814e-05, | |
| "loss": 0.0458, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.6266880599010563e-05, | |
| "loss": 0.0246, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.6200026741542987e-05, | |
| "loss": 0.0644, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.613317288407541e-05, | |
| "loss": 0.0487, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.606631902660784e-05, | |
| "loss": 0.0341, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.5999465169140257e-05, | |
| "loss": 0.0605, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.5932611311672685e-05, | |
| "loss": 0.0392, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.586575745420511e-05, | |
| "loss": 0.0438, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.5798903596737533e-05, | |
| "loss": 0.0518, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.5732049739269957e-05, | |
| "loss": 0.0386, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.566519588180238e-05, | |
| "loss": 0.0507, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.5598342024334803e-05, | |
| "loss": 0.0363, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.553148816686723e-05, | |
| "loss": 0.0262, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.5464634309399655e-05, | |
| "loss": 0.0399, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.539778045193208e-05, | |
| "loss": 0.0493, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.53309265944645e-05, | |
| "loss": 0.042, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.5264072736996924e-05, | |
| "loss": 0.038, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.519721887952935e-05, | |
| "loss": 0.0331, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.5130365022061776e-05, | |
| "loss": 0.0246, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.50635111645942e-05, | |
| "loss": 0.0578, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.499665730712662e-05, | |
| "loss": 0.0292, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.492980344965905e-05, | |
| "loss": 0.0358, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.486294959219147e-05, | |
| "loss": 0.0508, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4796095734723894e-05, | |
| "loss": 0.0461, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.472924187725632e-05, | |
| "loss": 0.073, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4662388019788743e-05, | |
| "loss": 0.0451, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4595534162321167e-05, | |
| "loss": 0.0582, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.452868030485359e-05, | |
| "loss": 0.0356, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4461826447386016e-05, | |
| "loss": 0.0308, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4394972589918437e-05, | |
| "loss": 0.0531, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.4328118732450864e-05, | |
| "loss": 0.036, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.426126487498329e-05, | |
| "loss": 0.0445, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.419441101751571e-05, | |
| "loss": 0.0286, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.4127557160048137e-05, | |
| "loss": 0.0394, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.4060703302580558e-05, | |
| "loss": 0.0366, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.3993849445112983e-05, | |
| "loss": 0.0482, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.392699558764541e-05, | |
| "loss": 0.0431, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.386014173017783e-05, | |
| "loss": 0.0311, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.3793287872710256e-05, | |
| "loss": 0.053, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.372643401524268e-05, | |
| "loss": 0.0655, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.3659580157775104e-05, | |
| "loss": 0.0537, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.359272630030753e-05, | |
| "loss": 0.0314, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.3525872442839953e-05, | |
| "loss": 0.0506, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.3459018585372377e-05, | |
| "loss": 0.0239, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.33921647279048e-05, | |
| "loss": 0.0306, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.3325310870437226e-05, | |
| "loss": 0.0385, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.325845701296965e-05, | |
| "loss": 0.0336, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.3191603155502074e-05, | |
| "loss": 0.0592, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.31247492980345e-05, | |
| "loss": 0.04, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.305789544056692e-05, | |
| "loss": 0.0441, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.2991041583099347e-05, | |
| "loss": 0.0392, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.292418772563177e-05, | |
| "loss": 0.0373, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.2857333868164192e-05, | |
| "loss": 0.0291, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.279048001069662e-05, | |
| "loss": 0.0489, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.272362615322904e-05, | |
| "loss": 0.0714, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2656772295761465e-05, | |
| "loss": 0.0389, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2589918438293893e-05, | |
| "loss": 0.0391, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2523064580826314e-05, | |
| "loss": 0.0314, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2456210723358738e-05, | |
| "loss": 0.0548, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2389356865891163e-05, | |
| "loss": 0.043, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2322503008423587e-05, | |
| "loss": 0.039, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.225564915095601e-05, | |
| "loss": 0.0338, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.2188795293488435e-05, | |
| "loss": 0.0368, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.212194143602086e-05, | |
| "loss": 0.0324, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.205508757855328e-05, | |
| "loss": 0.0423, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.198823372108571e-05, | |
| "loss": 0.0333, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.1921379863618133e-05, | |
| "loss": 0.045, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.1854526006150554e-05, | |
| "loss": 0.0508, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.178767214868298e-05, | |
| "loss": 0.0314, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.1720818291215402e-05, | |
| "loss": 0.0305, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.1653964433747826e-05, | |
| "loss": 0.0338, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1587110576280254e-05, | |
| "loss": 0.0562, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1520256718812675e-05, | |
| "loss": 0.04, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.14534028613451e-05, | |
| "loss": 0.0362, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1386549003877524e-05, | |
| "loss": 0.0468, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1319695146409948e-05, | |
| "loss": 0.039, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1252841288942372e-05, | |
| "loss": 0.0533, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.1185987431474797e-05, | |
| "loss": 0.0505, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.111913357400722e-05, | |
| "loss": 0.0399, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.1052279716539645e-05, | |
| "loss": 0.0371, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.098542585907207e-05, | |
| "loss": 0.0468, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.0918572001604494e-05, | |
| "loss": 0.0379, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.0851718144136918e-05, | |
| "loss": 0.0713, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.0784864286669342e-05, | |
| "loss": 0.0629, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.0718010429201763e-05, | |
| "loss": 0.0429, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.065115657173419e-05, | |
| "loss": 0.0408, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 2.0584302714266615e-05, | |
| "loss": 0.0369, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0517448856799036e-05, | |
| "loss": 0.0362, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0450594999331464e-05, | |
| "loss": 0.0354, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0383741141863885e-05, | |
| "loss": 0.0308, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.031688728439631e-05, | |
| "loss": 0.0465, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0250033426928737e-05, | |
| "loss": 0.0451, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0183179569461158e-05, | |
| "loss": 0.0187, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0116325711993582e-05, | |
| "loss": 0.052, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.0049471854526006e-05, | |
| "loss": 0.0427, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.998261799705843e-05, | |
| "loss": 0.0458, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.9915764139590855e-05, | |
| "loss": 0.0323, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.984891028212328e-05, | |
| "loss": 0.0265, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.9782056424655704e-05, | |
| "loss": 0.046, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.9715202567188128e-05, | |
| "loss": 0.0222, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.9648348709720552e-05, | |
| "loss": 0.0612, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.9581494852252977e-05, | |
| "loss": 0.0557, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.95146409947854e-05, | |
| "loss": 0.0448, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.9447787137317825e-05, | |
| "loss": 0.0312, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.938093327985025e-05, | |
| "loss": 0.037, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.9314079422382674e-05, | |
| "loss": 0.0418, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.9247225564915098e-05, | |
| "loss": 0.0362, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.918037170744752e-05, | |
| "loss": 0.0473, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.9113517849979943e-05, | |
| "loss": 0.0587, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.904666399251237e-05, | |
| "loss": 0.039, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8979810135044792e-05, | |
| "loss": 0.0611, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8912956277577216e-05, | |
| "loss": 0.0516, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.884610242010964e-05, | |
| "loss": 0.0362, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8779248562642065e-05, | |
| "loss": 0.0436, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.871239470517449e-05, | |
| "loss": 0.0346, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8645540847706913e-05, | |
| "loss": 0.0366, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8578686990239338e-05, | |
| "loss": 0.0281, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8511833132771762e-05, | |
| "loss": 0.0262, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.8444979275304186e-05, | |
| "loss": 0.0686, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.837812541783661e-05, | |
| "loss": 0.0257, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8311271560369035e-05, | |
| "loss": 0.0298, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.824441770290146e-05, | |
| "loss": 0.0454, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.817756384543388e-05, | |
| "loss": 0.0373, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8110709987966308e-05, | |
| "loss": 0.042, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.8043856130498732e-05, | |
| "loss": 0.0522, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7977002273031153e-05, | |
| "loss": 0.0364, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.791014841556358e-05, | |
| "loss": 0.0548, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7843294558096002e-05, | |
| "loss": 0.0292, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7776440700628426e-05, | |
| "loss": 0.0331, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7709586843160854e-05, | |
| "loss": 0.037, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7642732985693275e-05, | |
| "loss": 0.0227, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.75758791282257e-05, | |
| "loss": 0.0362, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7509025270758123e-05, | |
| "loss": 0.044, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7442171413290548e-05, | |
| "loss": 0.0458, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.7375317555822972e-05, | |
| "loss": 0.0301, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.7308463698355396e-05, | |
| "loss": 0.0474, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.724160984088782e-05, | |
| "loss": 0.0314, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.7174755983420245e-05, | |
| "loss": 0.0613, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.710790212595267e-05, | |
| "loss": 0.0323, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.7041048268485093e-05, | |
| "loss": 0.05, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6974194411017518e-05, | |
| "loss": 0.0334, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6907340553549942e-05, | |
| "loss": 0.0488, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6840486696082363e-05, | |
| "loss": 0.0313, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.677363283861479e-05, | |
| "loss": 0.045, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6706778981147215e-05, | |
| "loss": 0.068, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6639925123679636e-05, | |
| "loss": 0.0409, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6573071266212064e-05, | |
| "loss": 0.0402, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6506217408744484e-05, | |
| "loss": 0.0411, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.643936355127691e-05, | |
| "loss": 0.0608, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6372509693809336e-05, | |
| "loss": 0.0374, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.6305655836341757e-05, | |
| "loss": 0.0521, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.623880197887418e-05, | |
| "loss": 0.0253, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.6171948121406606e-05, | |
| "loss": 0.0399, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.610509426393903e-05, | |
| "loss": 0.0196, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.6038240406471455e-05, | |
| "loss": 0.041, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.597138654900388e-05, | |
| "loss": 0.0209, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5904532691536303e-05, | |
| "loss": 0.0408, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5837678834068724e-05, | |
| "loss": 0.0419, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5770824976601152e-05, | |
| "loss": 0.024, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5703971119133576e-05, | |
| "loss": 0.0261, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5637117261665997e-05, | |
| "loss": 0.0511, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5570263404198425e-05, | |
| "loss": 0.0469, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5503409546730846e-05, | |
| "loss": 0.029, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.543655568926327e-05, | |
| "loss": 0.0364, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5369701831795698e-05, | |
| "loss": 0.05, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.530284797432812e-05, | |
| "loss": 0.0304, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5235994116860545e-05, | |
| "loss": 0.0303, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.5169140259392967e-05, | |
| "loss": 0.0626, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.5102286401925391e-05, | |
| "loss": 0.0532, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.5035432544457817e-05, | |
| "loss": 0.0292, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.496857868699024e-05, | |
| "loss": 0.0319, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4901724829522664e-05, | |
| "loss": 0.0505, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4834870972055087e-05, | |
| "loss": 0.0361, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4768017114587513e-05, | |
| "loss": 0.0574, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4701163257119937e-05, | |
| "loss": 0.0465, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.463430939965236e-05, | |
| "loss": 0.0368, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4567455542184786e-05, | |
| "loss": 0.0273, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4500601684717209e-05, | |
| "loss": 0.0281, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4433747827249633e-05, | |
| "loss": 0.0449, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4366893969782059e-05, | |
| "loss": 0.0325, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4300040112314481e-05, | |
| "loss": 0.0382, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4233186254846906e-05, | |
| "loss": 0.0376, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.4166332397379328e-05, | |
| "loss": 0.0325, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.4099478539911753e-05, | |
| "loss": 0.0423, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.4032624682444179e-05, | |
| "loss": 0.0333, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3965770824976601e-05, | |
| "loss": 0.0266, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3898916967509026e-05, | |
| "loss": 0.0388, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3832063110041448e-05, | |
| "loss": 0.0321, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3765209252573874e-05, | |
| "loss": 0.0298, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3698355395106298e-05, | |
| "loss": 0.0443, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3631501537638721e-05, | |
| "loss": 0.026, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.3564647680171147e-05, | |
| "loss": 0.0296, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.349779382270357e-05, | |
| "loss": 0.0472, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.3430939965235994e-05, | |
| "loss": 0.031, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.336408610776842e-05, | |
| "loss": 0.0501, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.3297232250300843e-05, | |
| "loss": 0.0452, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.3230378392833267e-05, | |
| "loss": 0.0374, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.316352453536569e-05, | |
| "loss": 0.0368, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.3096670677898116e-05, | |
| "loss": 0.0329, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.302981682043054e-05, | |
| "loss": 0.024, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2962962962962962e-05, | |
| "loss": 0.0291, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2896109105495388e-05, | |
| "loss": 0.0576, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2829255248027811e-05, | |
| "loss": 0.0353, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2762401390560235e-05, | |
| "loss": 0.0228, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2695547533092661e-05, | |
| "loss": 0.0425, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2628693675625084e-05, | |
| "loss": 0.0274, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2561839818157508e-05, | |
| "loss": 0.0352, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2494985960689933e-05, | |
| "loss": 0.0261, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2428132103222357e-05, | |
| "loss": 0.0299, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.236127824575478e-05, | |
| "loss": 0.0283, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2294424388287206e-05, | |
| "loss": 0.0364, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.222757053081963e-05, | |
| "loss": 0.0488, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2160716673352052e-05, | |
| "loss": 0.0447, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2093862815884477e-05, | |
| "loss": 0.0486, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.2027008958416901e-05, | |
| "loss": 0.0478, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1960155100949325e-05, | |
| "loss": 0.0663, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.189330124348175e-05, | |
| "loss": 0.0213, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1826447386014174e-05, | |
| "loss": 0.0611, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1759593528546598e-05, | |
| "loss": 0.0648, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1692739671079021e-05, | |
| "loss": 0.0485, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1625885813611447e-05, | |
| "loss": 0.029, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1559031956143871e-05, | |
| "loss": 0.0242, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1492178098676294e-05, | |
| "loss": 0.0401, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1425324241208718e-05, | |
| "loss": 0.0677, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1358470383741142e-05, | |
| "loss": 0.0189, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1291616526273567e-05, | |
| "loss": 0.0302, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1224762668805991e-05, | |
| "loss": 0.0405, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1157908811338415e-05, | |
| "loss": 0.059, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1091054953870838e-05, | |
| "loss": 0.0259, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.1024201096403262e-05, | |
| "loss": 0.0365, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0957347238935688e-05, | |
| "loss": 0.0304, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.089049338146811e-05, | |
| "loss": 0.0412, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0823639524000535e-05, | |
| "loss": 0.0424, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.075678566653296e-05, | |
| "loss": 0.0343, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0689931809065384e-05, | |
| "loss": 0.0444, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0623077951597808e-05, | |
| "loss": 0.0412, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0556224094130232e-05, | |
| "loss": 0.0346, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0489370236662657e-05, | |
| "loss": 0.0568, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.042251637919508e-05, | |
| "loss": 0.0407, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.0355662521727504e-05, | |
| "loss": 0.0439, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.028880866425993e-05, | |
| "loss": 0.0605, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.0221954806792352e-05, | |
| "loss": 0.0544, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.0155100949324776e-05, | |
| "loss": 0.0437, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.00882470918572e-05, | |
| "loss": 0.0369, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.0021393234389623e-05, | |
| "loss": 0.0342, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.95453937692205e-06, | |
| "loss": 0.0448, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.887685519454474e-06, | |
| "loss": 0.0538, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.820831661986896e-06, | |
| "loss": 0.0244, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.75397780451932e-06, | |
| "loss": 0.0824, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.687123947051745e-06, | |
| "loss": 0.0502, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.62027008958417e-06, | |
| "loss": 0.0338, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.553416232116594e-06, | |
| "loss": 0.0419, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.486562374649018e-06, | |
| "loss": 0.0576, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.419708517181442e-06, | |
| "loss": 0.039, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.352854659713865e-06, | |
| "loss": 0.0662, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.28600080224629e-06, | |
| "loss": 0.0398, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.219146944778715e-06, | |
| "loss": 0.0349, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.152293087311138e-06, | |
| "loss": 0.0314, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.085439229843562e-06, | |
| "loss": 0.0388, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.018585372375988e-06, | |
| "loss": 0.0285, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.95173151490841e-06, | |
| "loss": 0.0473, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.884877657440835e-06, | |
| "loss": 0.0462, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.81802379997326e-06, | |
| "loss": 0.043, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.751169942505682e-06, | |
| "loss": 0.0387, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.684316085038108e-06, | |
| "loss": 0.0545, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.617462227570532e-06, | |
| "loss": 0.0785, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.550608370102955e-06, | |
| "loss": 0.0413, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.483754512635379e-06, | |
| "loss": 0.0433, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.416900655167803e-06, | |
| "loss": 0.0467, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.350046797700228e-06, | |
| "loss": 0.0201, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.283192940232652e-06, | |
| "loss": 0.0233, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.216339082765076e-06, | |
| "loss": 0.0421, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.1494852252975e-06, | |
| "loss": 0.0279, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.082631367829923e-06, | |
| "loss": 0.0358, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.01577751036235e-06, | |
| "loss": 0.0391, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.948923652894773e-06, | |
| "loss": 0.0345, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.882069795427196e-06, | |
| "loss": 0.0546, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.81521593795962e-06, | |
| "loss": 0.0252, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.748362080492045e-06, | |
| "loss": 0.0382, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.681508223024469e-06, | |
| "loss": 0.0392, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.614654365556893e-06, | |
| "loss": 0.0318, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.547800508089317e-06, | |
| "loss": 0.0471, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.480946650621741e-06, | |
| "loss": 0.0433, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.4140927931541645e-06, | |
| "loss": 0.0531, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.34723893568659e-06, | |
| "loss": 0.0278, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.280385078219014e-06, | |
| "loss": 0.0209, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.2135312207514375e-06, | |
| "loss": 0.0532, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 7.146677363283862e-06, | |
| "loss": 0.0439, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 7.079823505816285e-06, | |
| "loss": 0.04, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 7.01296964834871e-06, | |
| "loss": 0.0464, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.946115790881135e-06, | |
| "loss": 0.032, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.879261933413558e-06, | |
| "loss": 0.0284, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.8124080759459824e-06, | |
| "loss": 0.0267, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.745554218478406e-06, | |
| "loss": 0.0363, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.678700361010831e-06, | |
| "loss": 0.0317, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.611846503543255e-06, | |
| "loss": 0.0336, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.544992646075679e-06, | |
| "loss": 0.028, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.478138788608102e-06, | |
| "loss": 0.0347, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.411284931140527e-06, | |
| "loss": 0.0359, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.344431073672952e-06, | |
| "loss": 0.0381, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.277577216205375e-06, | |
| "loss": 0.0175, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.2107233587377995e-06, | |
| "loss": 0.0453, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.143869501270223e-06, | |
| "loss": 0.038, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.077015643802648e-06, | |
| "loss": 0.029, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.0101617863350716e-06, | |
| "loss": 0.0458, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.943307928867496e-06, | |
| "loss": 0.039, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.87645407139992e-06, | |
| "loss": 0.0384, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.8096002139323445e-06, | |
| "loss": 0.0621, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.742746356464768e-06, | |
| "loss": 0.0291, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.675892498997192e-06, | |
| "loss": 0.054, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.6090386415296165e-06, | |
| "loss": 0.0325, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.542184784062041e-06, | |
| "loss": 0.0417, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.475330926594465e-06, | |
| "loss": 0.0361, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.408477069126889e-06, | |
| "loss": 0.0531, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.341623211659313e-06, | |
| "loss": 0.0265, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.274769354191737e-06, | |
| "loss": 0.0435, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.2079154967241615e-06, | |
| "loss": 0.0276, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.141061639256586e-06, | |
| "loss": 0.0264, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.074207781789009e-06, | |
| "loss": 0.0234, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 5.007353924321434e-06, | |
| "loss": 0.0362, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.940500066853858e-06, | |
| "loss": 0.0304, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.873646209386281e-06, | |
| "loss": 0.0339, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.8067923519187065e-06, | |
| "loss": 0.0381, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.73993849445113e-06, | |
| "loss": 0.0472, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.673084636983554e-06, | |
| "loss": 0.029, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.606230779515979e-06, | |
| "loss": 0.0503, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.539376922048402e-06, | |
| "loss": 0.0446, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.472523064580827e-06, | |
| "loss": 0.0506, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.405669207113251e-06, | |
| "loss": 0.0322, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.338815349645674e-06, | |
| "loss": 0.0477, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.271961492178099e-06, | |
| "loss": 0.0276, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.205107634710523e-06, | |
| "loss": 0.0483, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.138253777242947e-06, | |
| "loss": 0.049, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.071399919775371e-06, | |
| "loss": 0.0388, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.004546062307795e-06, | |
| "loss": 0.034, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.93769220484022e-06, | |
| "loss": 0.0408, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.870838347372643e-06, | |
| "loss": 0.0448, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.803984489905068e-06, | |
| "loss": 0.0377, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.7371306324374916e-06, | |
| "loss": 0.0175, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.6702767749699155e-06, | |
| "loss": 0.0432, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.60342291750234e-06, | |
| "loss": 0.0381, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.536569060034764e-06, | |
| "loss": 0.0425, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.4697152025671884e-06, | |
| "loss": 0.0432, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.4028613450996123e-06, | |
| "loss": 0.0355, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.336007487632037e-06, | |
| "loss": 0.0802, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.269153630164461e-06, | |
| "loss": 0.0484, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.2022997726968843e-06, | |
| "loss": 0.0535, | |
| "step": 7500 | |
| } | |
| ], | |
| "max_steps": 7979, | |
| "num_train_epochs": 1, | |
| "total_flos": 3974305443840000.0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |