| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.49847046802871836, | |
| "global_step": 803500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9997932086836637e-05, | |
| "loss": 1.8211, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9995864173673275e-05, | |
| "loss": 1.8484, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9993796260509913e-05, | |
| "loss": 1.832, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9991728347346548e-05, | |
| "loss": 1.8583, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9989660434183183e-05, | |
| "loss": 1.8529, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.998759252101982e-05, | |
| "loss": 1.8312, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.998552460785646e-05, | |
| "loss": 1.8302, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9983456694693095e-05, | |
| "loss": 1.8357, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.998138878152973e-05, | |
| "loss": 1.8555, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9979320868366368e-05, | |
| "loss": 1.8106, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9977252955203007e-05, | |
| "loss": 1.7992, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.997518504203964e-05, | |
| "loss": 1.8011, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.997311712887628e-05, | |
| "loss": 1.8131, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9971049215712918e-05, | |
| "loss": 1.8333, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9968981302549553e-05, | |
| "loss": 1.8279, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.9966913389386188e-05, | |
| "loss": 1.8247, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9964845476222826e-05, | |
| "loss": 1.7957, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9962777563059465e-05, | |
| "loss": 1.8243, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.99607096498961e-05, | |
| "loss": 1.7959, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9958641736732735e-05, | |
| "loss": 1.8339, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9956573823569373e-05, | |
| "loss": 1.8002, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.995450591040601e-05, | |
| "loss": 1.8096, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9952437997242646e-05, | |
| "loss": 1.7897, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.995037008407928e-05, | |
| "loss": 1.7928, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.994830217091592e-05, | |
| "loss": 1.788, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9946234257752558e-05, | |
| "loss": 1.776, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9944166344589193e-05, | |
| "loss": 1.7895, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9942098431425828e-05, | |
| "loss": 1.7952, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9940030518262466e-05, | |
| "loss": 1.8282, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9937962605099105e-05, | |
| "loss": 1.7913, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.993589469193574e-05, | |
| "loss": 1.7591, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9933826778772375e-05, | |
| "loss": 1.7948, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9931758865609013e-05, | |
| "loss": 1.8091, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.992969095244565e-05, | |
| "loss": 1.7862, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9927623039282286e-05, | |
| "loss": 1.8192, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.992555512611892e-05, | |
| "loss": 1.8088, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.992348721295556e-05, | |
| "loss": 1.7959, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9921419299792198e-05, | |
| "loss": 1.803, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9919351386628833e-05, | |
| "loss": 1.7873, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.991728347346547e-05, | |
| "loss": 1.7821, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.991521556030211e-05, | |
| "loss": 1.7863, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9913147647138745e-05, | |
| "loss": 1.8086, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.991107973397538e-05, | |
| "loss": 1.7648, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9909011820812018e-05, | |
| "loss": 1.774, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9906943907648656e-05, | |
| "loss": 1.8018, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.990487599448529e-05, | |
| "loss": 1.7961, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9902808081321926e-05, | |
| "loss": 1.7624, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.9900740168158564e-05, | |
| "loss": 1.7494, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9898672254995203e-05, | |
| "loss": 1.7513, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9896604341831838e-05, | |
| "loss": 1.7641, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9894536428668473e-05, | |
| "loss": 1.7823, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.989246851550511e-05, | |
| "loss": 1.7488, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.989040060234175e-05, | |
| "loss": 1.7275, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9888332689178384e-05, | |
| "loss": 1.78, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.988626477601502e-05, | |
| "loss": 1.7703, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9884196862851658e-05, | |
| "loss": 1.759, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9882128949688296e-05, | |
| "loss": 1.7876, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.988006103652493e-05, | |
| "loss": 1.7472, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9877993123361566e-05, | |
| "loss": 1.786, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9875925210198204e-05, | |
| "loss": 1.7707, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9873857297034843e-05, | |
| "loss": 1.749, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9871789383871478e-05, | |
| "loss": 1.7522, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9869721470708113e-05, | |
| "loss": 1.7915, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9867653557544754e-05, | |
| "loss": 1.7413, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.986558564438139e-05, | |
| "loss": 1.735, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9863517731218024e-05, | |
| "loss": 1.7725, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9861449818054663e-05, | |
| "loss": 1.79, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.98593819048913e-05, | |
| "loss": 1.7809, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9857313991727936e-05, | |
| "loss": 1.7557, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.985524607856457e-05, | |
| "loss": 1.7474, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.985317816540121e-05, | |
| "loss": 1.7717, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9851110252237848e-05, | |
| "loss": 1.761, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9849042339074483e-05, | |
| "loss": 1.763, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9846974425911118e-05, | |
| "loss": 1.7306, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9844906512747756e-05, | |
| "loss": 1.7409, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9842838599584394e-05, | |
| "loss": 1.7731, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.984077068642103e-05, | |
| "loss": 1.782, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9838702773257664e-05, | |
| "loss": 1.7416, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9836634860094302e-05, | |
| "loss": 1.7537, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.983456694693094e-05, | |
| "loss": 1.7449, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9832499033767576e-05, | |
| "loss": 1.7629, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.983043112060421e-05, | |
| "loss": 1.7539, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.982836320744085e-05, | |
| "loss": 1.7311, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9826295294277487e-05, | |
| "loss": 1.7415, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9824227381114122e-05, | |
| "loss": 1.7529, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9822159467950757e-05, | |
| "loss": 1.7372, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9820091554787396e-05, | |
| "loss": 1.7288, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9818023641624034e-05, | |
| "loss": 1.7581, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.981595572846067e-05, | |
| "loss": 1.7325, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9813887815297307e-05, | |
| "loss": 1.7552, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9811819902133946e-05, | |
| "loss": 1.7689, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.980975198897058e-05, | |
| "loss": 1.7547, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9807684075807216e-05, | |
| "loss": 1.7676, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9805616162643854e-05, | |
| "loss": 1.7406, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9803548249480492e-05, | |
| "loss": 1.7515, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9801480336317127e-05, | |
| "loss": 1.7379, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9799412423153762e-05, | |
| "loss": 1.7673, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.97973445099904e-05, | |
| "loss": 1.7142, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.979527659682704e-05, | |
| "loss": 1.7152, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9793208683663674e-05, | |
| "loss": 1.7406, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.979114077050031e-05, | |
| "loss": 1.752, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9789072857336947e-05, | |
| "loss": 1.7744, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9787004944173586e-05, | |
| "loss": 1.754, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.978493703101022e-05, | |
| "loss": 1.7402, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9782869117846856e-05, | |
| "loss": 1.7219, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9780801204683494e-05, | |
| "loss": 1.7076, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9778733291520132e-05, | |
| "loss": 1.7554, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9776665378356767e-05, | |
| "loss": 1.7478, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9774597465193402e-05, | |
| "loss": 1.7663, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.977252955203004e-05, | |
| "loss": 1.753, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.977046163886668e-05, | |
| "loss": 1.7328, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9768393725703314e-05, | |
| "loss": 1.7462, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.976632581253995e-05, | |
| "loss": 1.7636, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9764257899376587e-05, | |
| "loss": 1.7446, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9762189986213225e-05, | |
| "loss": 1.7484, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.976012207304986e-05, | |
| "loss": 1.7466, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.97580541598865e-05, | |
| "loss": 1.7511, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9755986246723137e-05, | |
| "loss": 1.7163, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9753918333559772e-05, | |
| "loss": 1.7215, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9751850420396407e-05, | |
| "loss": 1.7513, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9749782507233045e-05, | |
| "loss": 1.7199, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9747714594069684e-05, | |
| "loss": 1.7385, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.974564668090632e-05, | |
| "loss": 1.7485, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9743578767742954e-05, | |
| "loss": 1.7382, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9741510854579592e-05, | |
| "loss": 1.7108, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.973944294141623e-05, | |
| "loss": 1.7437, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9737375028252865e-05, | |
| "loss": 1.7325, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.97353071150895e-05, | |
| "loss": 1.775, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.973323920192614e-05, | |
| "loss": 1.7149, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9731171288762777e-05, | |
| "loss": 1.7412, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9729103375599412e-05, | |
| "loss": 1.7411, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9727035462436047e-05, | |
| "loss": 1.716, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9724967549272685e-05, | |
| "loss": 1.722, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9722899636109324e-05, | |
| "loss": 1.7307, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.972083172294596e-05, | |
| "loss": 1.7038, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9718763809782594e-05, | |
| "loss": 1.7468, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9716695896619232e-05, | |
| "loss": 1.7389, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.971462798345587e-05, | |
| "loss": 1.7329, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9712560070292505e-05, | |
| "loss": 1.7209, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9710492157129144e-05, | |
| "loss": 1.715, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9708424243965782e-05, | |
| "loss": 1.7383, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9706356330802417e-05, | |
| "loss": 1.7256, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9704288417639052e-05, | |
| "loss": 1.7646, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.970222050447569e-05, | |
| "loss": 1.7673, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.970015259131233e-05, | |
| "loss": 1.7027, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9698084678148963e-05, | |
| "loss": 1.7549, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.96960167649856e-05, | |
| "loss": 1.7388, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9693948851822237e-05, | |
| "loss": 1.7039, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9691880938658875e-05, | |
| "loss": 1.7451, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.968981302549551e-05, | |
| "loss": 1.7338, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9687745112332145e-05, | |
| "loss": 1.7265, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9685677199168783e-05, | |
| "loss": 1.7263, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9683609286005422e-05, | |
| "loss": 1.7253, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9681541372842057e-05, | |
| "loss": 1.728, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.967947345967869e-05, | |
| "loss": 1.7316, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.967740554651533e-05, | |
| "loss": 1.7585, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.967533763335197e-05, | |
| "loss": 1.7129, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9673269720188603e-05, | |
| "loss": 1.7137, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9671201807025238e-05, | |
| "loss": 1.7334, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9669133893861877e-05, | |
| "loss": 1.7323, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9667065980698515e-05, | |
| "loss": 1.7381, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.966499806753515e-05, | |
| "loss": 1.722, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9662930154371785e-05, | |
| "loss": 1.7139, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9660862241208423e-05, | |
| "loss": 1.7404, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.965879432804506e-05, | |
| "loss": 1.7501, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9656726414881697e-05, | |
| "loss": 1.7188, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9654658501718335e-05, | |
| "loss": 1.7154, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9652590588554973e-05, | |
| "loss": 1.7433, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9650522675391608e-05, | |
| "loss": 1.6987, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9648454762228243e-05, | |
| "loss": 1.7371, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.964638684906488e-05, | |
| "loss": 1.7143, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.964431893590152e-05, | |
| "loss": 1.7327, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9642251022738155e-05, | |
| "loss": 1.7334, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.964018310957479e-05, | |
| "loss": 1.7079, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9638115196411428e-05, | |
| "loss": 1.7426, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9636047283248067e-05, | |
| "loss": 1.7066, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.96339793700847e-05, | |
| "loss": 1.7613, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9631911456921336e-05, | |
| "loss": 1.7257, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9629843543757975e-05, | |
| "loss": 1.7413, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9627775630594613e-05, | |
| "loss": 1.7053, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9625707717431248e-05, | |
| "loss": 1.7156, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9623639804267883e-05, | |
| "loss": 1.7176, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.962157189110452e-05, | |
| "loss": 1.7132, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.961950397794116e-05, | |
| "loss": 1.7286, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9617436064777795e-05, | |
| "loss": 1.7361, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.961536815161443e-05, | |
| "loss": 1.7267, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9613300238451068e-05, | |
| "loss": 1.7048, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9611232325287706e-05, | |
| "loss": 1.7163, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.960916441212434e-05, | |
| "loss": 1.7627, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9607096498960976e-05, | |
| "loss": 1.7167, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9605028585797618e-05, | |
| "loss": 1.7274, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9602960672634253e-05, | |
| "loss": 1.7079, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9600892759470888e-05, | |
| "loss": 1.7104, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9598824846307526e-05, | |
| "loss": 1.7126, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9596756933144165e-05, | |
| "loss": 1.7055, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.95946890199808e-05, | |
| "loss": 1.6893, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9592621106817435e-05, | |
| "loss": 1.7458, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9590553193654073e-05, | |
| "loss": 1.7388, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.958848528049071e-05, | |
| "loss": 1.7242, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9586417367327346e-05, | |
| "loss": 1.7237, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.958434945416398e-05, | |
| "loss": 1.7167, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.958228154100062e-05, | |
| "loss": 1.7259, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9580213627837258e-05, | |
| "loss": 1.719, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9578145714673893e-05, | |
| "loss": 1.7311, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9576077801510528e-05, | |
| "loss": 1.6825, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9574009888347166e-05, | |
| "loss": 1.7203, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9571941975183805e-05, | |
| "loss": 1.7543, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.956987406202044e-05, | |
| "loss": 1.7323, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9567806148857074e-05, | |
| "loss": 1.7179, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9565738235693713e-05, | |
| "loss": 1.6992, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.956367032253035e-05, | |
| "loss": 1.7269, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9561602409366986e-05, | |
| "loss": 1.7197, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.955953449620362e-05, | |
| "loss": 1.7175, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.955746658304026e-05, | |
| "loss": 1.731, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9555398669876898e-05, | |
| "loss": 1.6981, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9553330756713533e-05, | |
| "loss": 1.6977, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.955126284355017e-05, | |
| "loss": 1.7397, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.954919493038681e-05, | |
| "loss": 1.692, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9547127017223444e-05, | |
| "loss": 1.7, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.954505910406008e-05, | |
| "loss": 1.733, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9542991190896718e-05, | |
| "loss": 1.703, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9540923277733356e-05, | |
| "loss": 1.714, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.953885536456999e-05, | |
| "loss": 1.7271, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9536787451406626e-05, | |
| "loss": 1.7385, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9534719538243264e-05, | |
| "loss": 1.7427, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9532651625079903e-05, | |
| "loss": 1.7234, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9530583711916538e-05, | |
| "loss": 1.7279, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9528515798753173e-05, | |
| "loss": 1.6994, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.952644788558981e-05, | |
| "loss": 1.7076, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.952437997242645e-05, | |
| "loss": 1.7031, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9522312059263084e-05, | |
| "loss": 1.7195, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.952024414609972e-05, | |
| "loss": 1.6933, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9518176232936358e-05, | |
| "loss": 1.7005, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9516108319772996e-05, | |
| "loss": 1.7131, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.951404040660963e-05, | |
| "loss": 1.7252, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9511972493446266e-05, | |
| "loss": 1.7333, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9509904580282904e-05, | |
| "loss": 1.6931, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9507836667119543e-05, | |
| "loss": 1.7487, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9505768753956178e-05, | |
| "loss": 1.7313, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9503700840792812e-05, | |
| "loss": 1.6854, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.950163292762945e-05, | |
| "loss": 1.7253, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.949956501446609e-05, | |
| "loss": 1.7121, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9497497101302724e-05, | |
| "loss": 1.7068, | |
| "step": 121500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9495429188139362e-05, | |
| "loss": 1.7373, | |
| "step": 122000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9493361274976e-05, | |
| "loss": 1.731, | |
| "step": 122500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9491293361812636e-05, | |
| "loss": 1.6952, | |
| "step": 123000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.948922544864927e-05, | |
| "loss": 1.7152, | |
| "step": 123500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.948715753548591e-05, | |
| "loss": 1.718, | |
| "step": 124000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9485089622322547e-05, | |
| "loss": 1.6984, | |
| "step": 124500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9483021709159182e-05, | |
| "loss": 1.7289, | |
| "step": 125000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9480953795995817e-05, | |
| "loss": 1.7152, | |
| "step": 125500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9478885882832456e-05, | |
| "loss": 1.7138, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9476817969669094e-05, | |
| "loss": 1.7083, | |
| "step": 126500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.947475005650573e-05, | |
| "loss": 1.7135, | |
| "step": 127000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9472682143342364e-05, | |
| "loss": 1.6916, | |
| "step": 127500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9470614230179002e-05, | |
| "loss": 1.7172, | |
| "step": 128000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.946854631701564e-05, | |
| "loss": 1.7369, | |
| "step": 128500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9466478403852276e-05, | |
| "loss": 1.7164, | |
| "step": 129000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.946441049068891e-05, | |
| "loss": 1.7208, | |
| "step": 129500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.946234257752555e-05, | |
| "loss": 1.7315, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9460274664362187e-05, | |
| "loss": 1.7234, | |
| "step": 130500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9458206751198822e-05, | |
| "loss": 1.73, | |
| "step": 131000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9456138838035457e-05, | |
| "loss": 1.7105, | |
| "step": 131500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9454070924872096e-05, | |
| "loss": 1.7049, | |
| "step": 132000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9452003011708734e-05, | |
| "loss": 1.7187, | |
| "step": 132500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.944993509854537e-05, | |
| "loss": 1.7466, | |
| "step": 133000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9447867185382007e-05, | |
| "loss": 1.6927, | |
| "step": 133500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9445799272218646e-05, | |
| "loss": 1.7113, | |
| "step": 134000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.944373135905528e-05, | |
| "loss": 1.7202, | |
| "step": 134500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9441663445891916e-05, | |
| "loss": 1.6841, | |
| "step": 135000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9439595532728554e-05, | |
| "loss": 1.7161, | |
| "step": 135500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9437527619565192e-05, | |
| "loss": 1.7156, | |
| "step": 136000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9435459706401827e-05, | |
| "loss": 1.7228, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9433391793238462e-05, | |
| "loss": 1.7018, | |
| "step": 137000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.94313238800751e-05, | |
| "loss": 1.7207, | |
| "step": 137500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.942925596691174e-05, | |
| "loss": 1.7134, | |
| "step": 138000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9427188053748374e-05, | |
| "loss": 1.6893, | |
| "step": 138500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.942512014058501e-05, | |
| "loss": 1.7211, | |
| "step": 139000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9423052227421647e-05, | |
| "loss": 1.7061, | |
| "step": 139500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9420984314258285e-05, | |
| "loss": 1.7193, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.941891640109492e-05, | |
| "loss": 1.706, | |
| "step": 140500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9416848487931555e-05, | |
| "loss": 1.7256, | |
| "step": 141000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9414780574768194e-05, | |
| "loss": 1.7544, | |
| "step": 141500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9412712661604832e-05, | |
| "loss": 1.7104, | |
| "step": 142000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9410644748441467e-05, | |
| "loss": 1.7459, | |
| "step": 142500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9408576835278102e-05, | |
| "loss": 1.7049, | |
| "step": 143000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.940650892211474e-05, | |
| "loss": 1.7099, | |
| "step": 143500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.940444100895138e-05, | |
| "loss": 1.697, | |
| "step": 144000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9402373095788014e-05, | |
| "loss": 1.6939, | |
| "step": 144500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.940030518262465e-05, | |
| "loss": 1.7076, | |
| "step": 145000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9398237269461287e-05, | |
| "loss": 1.7028, | |
| "step": 145500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9396169356297925e-05, | |
| "loss": 1.6949, | |
| "step": 146000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.939410144313456e-05, | |
| "loss": 1.7152, | |
| "step": 146500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.93920335299712e-05, | |
| "loss": 1.7053, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9389965616807837e-05, | |
| "loss": 1.7175, | |
| "step": 147500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9387897703644472e-05, | |
| "loss": 1.6896, | |
| "step": 148000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9385829790481107e-05, | |
| "loss": 1.6895, | |
| "step": 148500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9383761877317745e-05, | |
| "loss": 1.7108, | |
| "step": 149000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9381693964154384e-05, | |
| "loss": 1.7032, | |
| "step": 149500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.937962605099102e-05, | |
| "loss": 1.7051, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9377558137827654e-05, | |
| "loss": 1.7253, | |
| "step": 150500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9375490224664292e-05, | |
| "loss": 1.6897, | |
| "step": 151000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.937342231150093e-05, | |
| "loss": 1.6875, | |
| "step": 151500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9371354398337565e-05, | |
| "loss": 1.6933, | |
| "step": 152000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.93692864851742e-05, | |
| "loss": 1.6915, | |
| "step": 152500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.936721857201084e-05, | |
| "loss": 1.7162, | |
| "step": 153000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9365150658847477e-05, | |
| "loss": 1.6907, | |
| "step": 153500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9363082745684112e-05, | |
| "loss": 1.7201, | |
| "step": 154000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9361014832520747e-05, | |
| "loss": 1.6888, | |
| "step": 154500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9358946919357385e-05, | |
| "loss": 1.6645, | |
| "step": 155000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9356879006194023e-05, | |
| "loss": 1.7112, | |
| "step": 155500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.935481109303066e-05, | |
| "loss": 1.7021, | |
| "step": 156000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9352743179867293e-05, | |
| "loss": 1.6912, | |
| "step": 156500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9350675266703932e-05, | |
| "loss": 1.7171, | |
| "step": 157000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.934860735354057e-05, | |
| "loss": 1.6884, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9346539440377205e-05, | |
| "loss": 1.6607, | |
| "step": 158000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.934447152721384e-05, | |
| "loss": 1.6743, | |
| "step": 158500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.934240361405048e-05, | |
| "loss": 1.6976, | |
| "step": 159000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9340335700887117e-05, | |
| "loss": 1.6957, | |
| "step": 159500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.933826778772375e-05, | |
| "loss": 1.7103, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.933619987456039e-05, | |
| "loss": 1.7008, | |
| "step": 160500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.933413196139703e-05, | |
| "loss": 1.7077, | |
| "step": 161000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9332064048233663e-05, | |
| "loss": 1.7027, | |
| "step": 161500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9329996135070298e-05, | |
| "loss": 1.7152, | |
| "step": 162000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9327928221906937e-05, | |
| "loss": 1.6772, | |
| "step": 162500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9325860308743575e-05, | |
| "loss": 1.6789, | |
| "step": 163000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.932379239558021e-05, | |
| "loss": 1.679, | |
| "step": 163500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9321724482416845e-05, | |
| "loss": 1.7125, | |
| "step": 164000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9319656569253483e-05, | |
| "loss": 1.7123, | |
| "step": 164500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.931758865609012e-05, | |
| "loss": 1.6982, | |
| "step": 165000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9315520742926757e-05, | |
| "loss": 1.7348, | |
| "step": 165500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.931345282976339e-05, | |
| "loss": 1.6996, | |
| "step": 166000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.931138491660003e-05, | |
| "loss": 1.712, | |
| "step": 166500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9309317003436668e-05, | |
| "loss": 1.7365, | |
| "step": 167000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9307249090273303e-05, | |
| "loss": 1.7034, | |
| "step": 167500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9305181177109938e-05, | |
| "loss": 1.7086, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9303113263946577e-05, | |
| "loss": 1.6911, | |
| "step": 168500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9301045350783215e-05, | |
| "loss": 1.6881, | |
| "step": 169000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.929897743761985e-05, | |
| "loss": 1.7006, | |
| "step": 169500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9296909524456485e-05, | |
| "loss": 1.7111, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9294841611293123e-05, | |
| "loss": 1.6961, | |
| "step": 170500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.929277369812976e-05, | |
| "loss": 1.7163, | |
| "step": 171000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9290705784966396e-05, | |
| "loss": 1.6837, | |
| "step": 171500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9288637871803035e-05, | |
| "loss": 1.6935, | |
| "step": 172000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.928656995863967e-05, | |
| "loss": 1.7296, | |
| "step": 172500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9284502045476308e-05, | |
| "loss": 1.7189, | |
| "step": 173000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9282434132312943e-05, | |
| "loss": 1.7131, | |
| "step": 173500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.928036621914958e-05, | |
| "loss": 1.6656, | |
| "step": 174000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.927829830598622e-05, | |
| "loss": 1.7097, | |
| "step": 174500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9276230392822855e-05, | |
| "loss": 1.7183, | |
| "step": 175000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.927416247965949e-05, | |
| "loss": 1.6925, | |
| "step": 175500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9272094566496128e-05, | |
| "loss": 1.7156, | |
| "step": 176000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9270026653332766e-05, | |
| "loss": 1.7002, | |
| "step": 176500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.92679587401694e-05, | |
| "loss": 1.7181, | |
| "step": 177000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9265890827006036e-05, | |
| "loss": 1.6844, | |
| "step": 177500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9263822913842675e-05, | |
| "loss": 1.714, | |
| "step": 178000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9261755000679313e-05, | |
| "loss": 1.6939, | |
| "step": 178500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9259687087515948e-05, | |
| "loss": 1.7179, | |
| "step": 179000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9257619174352583e-05, | |
| "loss": 1.6719, | |
| "step": 179500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.925555126118922e-05, | |
| "loss": 1.7066, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.925348334802586e-05, | |
| "loss": 1.7037, | |
| "step": 180500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9251415434862495e-05, | |
| "loss": 1.7064, | |
| "step": 181000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.924934752169913e-05, | |
| "loss": 1.6772, | |
| "step": 181500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9247279608535768e-05, | |
| "loss": 1.6937, | |
| "step": 182000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9245211695372406e-05, | |
| "loss": 1.6681, | |
| "step": 182500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.924314378220904e-05, | |
| "loss": 1.7035, | |
| "step": 183000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9241075869045676e-05, | |
| "loss": 1.7057, | |
| "step": 183500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9239007955882315e-05, | |
| "loss": 1.6688, | |
| "step": 184000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9236940042718953e-05, | |
| "loss": 1.6797, | |
| "step": 184500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9234872129555588e-05, | |
| "loss": 1.6937, | |
| "step": 185000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9232804216392226e-05, | |
| "loss": 1.7253, | |
| "step": 185500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.923073630322886e-05, | |
| "loss": 1.7019, | |
| "step": 186000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.92286683900655e-05, | |
| "loss": 1.685, | |
| "step": 186500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9226600476902134e-05, | |
| "loss": 1.7247, | |
| "step": 187000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9224532563738773e-05, | |
| "loss": 1.706, | |
| "step": 187500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.922246465057541e-05, | |
| "loss": 1.7113, | |
| "step": 188000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9220396737412046e-05, | |
| "loss": 1.7225, | |
| "step": 188500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.921832882424868e-05, | |
| "loss": 1.6968, | |
| "step": 189000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.921626091108532e-05, | |
| "loss": 1.6904, | |
| "step": 189500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9214192997921958e-05, | |
| "loss": 1.6812, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9212125084758593e-05, | |
| "loss": 1.6845, | |
| "step": 190500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9210057171595228e-05, | |
| "loss": 1.7069, | |
| "step": 191000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9207989258431866e-05, | |
| "loss": 1.6864, | |
| "step": 191500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9205921345268504e-05, | |
| "loss": 1.6812, | |
| "step": 192000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.920385343210514e-05, | |
| "loss": 1.6656, | |
| "step": 192500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9201785518941774e-05, | |
| "loss": 1.7053, | |
| "step": 193000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9199717605778413e-05, | |
| "loss": 1.6912, | |
| "step": 193500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.919764969261505e-05, | |
| "loss": 1.6788, | |
| "step": 194000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9195581779451686e-05, | |
| "loss": 1.7159, | |
| "step": 194500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.919351386628832e-05, | |
| "loss": 1.6926, | |
| "step": 195000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.919144595312496e-05, | |
| "loss": 1.6993, | |
| "step": 195500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9189378039961598e-05, | |
| "loss": 1.6873, | |
| "step": 196000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9187310126798233e-05, | |
| "loss": 1.6958, | |
| "step": 196500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.918524221363487e-05, | |
| "loss": 1.707, | |
| "step": 197000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9183174300471506e-05, | |
| "loss": 1.7106, | |
| "step": 197500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9181106387308144e-05, | |
| "loss": 1.6955, | |
| "step": 198000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.917903847414478e-05, | |
| "loss": 1.6852, | |
| "step": 198500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9176970560981418e-05, | |
| "loss": 1.6947, | |
| "step": 199000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9174902647818056e-05, | |
| "loss": 1.6855, | |
| "step": 199500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.917283473465469e-05, | |
| "loss": 1.6682, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9170766821491326e-05, | |
| "loss": 1.6652, | |
| "step": 200500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9168698908327964e-05, | |
| "loss": 1.6933, | |
| "step": 201000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9166630995164603e-05, | |
| "loss": 1.6796, | |
| "step": 201500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9164563082001237e-05, | |
| "loss": 1.6792, | |
| "step": 202000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9162495168837872e-05, | |
| "loss": 1.6971, | |
| "step": 202500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.916042725567451e-05, | |
| "loss": 1.68, | |
| "step": 203000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.915835934251115e-05, | |
| "loss": 1.7343, | |
| "step": 203500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9156291429347784e-05, | |
| "loss": 1.6952, | |
| "step": 204000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.915422351618442e-05, | |
| "loss": 1.6982, | |
| "step": 204500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9152155603021057e-05, | |
| "loss": 1.6913, | |
| "step": 205000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9150087689857696e-05, | |
| "loss": 1.7325, | |
| "step": 205500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.914801977669433e-05, | |
| "loss": 1.702, | |
| "step": 206000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9145951863530966e-05, | |
| "loss": 1.7036, | |
| "step": 206500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9143883950367604e-05, | |
| "loss": 1.6976, | |
| "step": 207000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9141816037204242e-05, | |
| "loss": 1.6641, | |
| "step": 207500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9139748124040877e-05, | |
| "loss": 1.6834, | |
| "step": 208000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9137680210877512e-05, | |
| "loss": 1.6544, | |
| "step": 208500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.913561229771415e-05, | |
| "loss": 1.6664, | |
| "step": 209000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.913354438455079e-05, | |
| "loss": 1.7106, | |
| "step": 209500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9131476471387424e-05, | |
| "loss": 1.7085, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9129408558224062e-05, | |
| "loss": 1.7068, | |
| "step": 210500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9127340645060697e-05, | |
| "loss": 1.6862, | |
| "step": 211000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9125272731897336e-05, | |
| "loss": 1.7104, | |
| "step": 211500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.912320481873397e-05, | |
| "loss": 1.7098, | |
| "step": 212000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.912113690557061e-05, | |
| "loss": 1.724, | |
| "step": 212500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9119068992407247e-05, | |
| "loss": 1.6682, | |
| "step": 213000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9117001079243882e-05, | |
| "loss": 1.6904, | |
| "step": 213500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9114933166080517e-05, | |
| "loss": 1.6882, | |
| "step": 214000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9112865252917156e-05, | |
| "loss": 1.6934, | |
| "step": 214500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9110797339753794e-05, | |
| "loss": 1.661, | |
| "step": 215000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.910872942659043e-05, | |
| "loss": 1.7105, | |
| "step": 215500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9106661513427064e-05, | |
| "loss": 1.6951, | |
| "step": 216000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9104593600263702e-05, | |
| "loss": 1.6618, | |
| "step": 216500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.910252568710034e-05, | |
| "loss": 1.6655, | |
| "step": 217000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9100457773936976e-05, | |
| "loss": 1.6697, | |
| "step": 217500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.909838986077361e-05, | |
| "loss": 1.697, | |
| "step": 218000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.909632194761025e-05, | |
| "loss": 1.7151, | |
| "step": 218500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9094254034446887e-05, | |
| "loss": 1.6973, | |
| "step": 219000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9092186121283522e-05, | |
| "loss": 1.6952, | |
| "step": 219500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9090118208120157e-05, | |
| "loss": 1.6862, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9088050294956795e-05, | |
| "loss": 1.6556, | |
| "step": 220500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9085982381793434e-05, | |
| "loss": 1.6698, | |
| "step": 221000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.908391446863007e-05, | |
| "loss": 1.7042, | |
| "step": 221500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9081846555466704e-05, | |
| "loss": 1.675, | |
| "step": 222000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9079778642303342e-05, | |
| "loss": 1.6962, | |
| "step": 222500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.907771072913998e-05, | |
| "loss": 1.7022, | |
| "step": 223000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9075642815976615e-05, | |
| "loss": 1.705, | |
| "step": 223500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9073574902813254e-05, | |
| "loss": 1.6785, | |
| "step": 224000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.907150698964989e-05, | |
| "loss": 1.6881, | |
| "step": 224500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9069439076486527e-05, | |
| "loss": 1.6617, | |
| "step": 225000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9067371163323162e-05, | |
| "loss": 1.6976, | |
| "step": 225500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.90653032501598e-05, | |
| "loss": 1.67, | |
| "step": 226000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.906323533699644e-05, | |
| "loss": 1.6941, | |
| "step": 226500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9061167423833074e-05, | |
| "loss": 1.6841, | |
| "step": 227000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.905909951066971e-05, | |
| "loss": 1.6705, | |
| "step": 227500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9057031597506347e-05, | |
| "loss": 1.6706, | |
| "step": 228000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9054963684342985e-05, | |
| "loss": 1.6915, | |
| "step": 228500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.905289577117962e-05, | |
| "loss": 1.6837, | |
| "step": 229000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9050827858016255e-05, | |
| "loss": 1.6841, | |
| "step": 229500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9048759944852894e-05, | |
| "loss": 1.6806, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9046692031689532e-05, | |
| "loss": 1.6783, | |
| "step": 230500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9044624118526167e-05, | |
| "loss": 1.6878, | |
| "step": 231000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9042556205362802e-05, | |
| "loss": 1.6912, | |
| "step": 231500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.904048829219944e-05, | |
| "loss": 1.6867, | |
| "step": 232000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.903842037903608e-05, | |
| "loss": 1.6788, | |
| "step": 232500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9036352465872714e-05, | |
| "loss": 1.6876, | |
| "step": 233000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.903428455270935e-05, | |
| "loss": 1.6774, | |
| "step": 233500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9032216639545987e-05, | |
| "loss": 1.6914, | |
| "step": 234000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9030148726382625e-05, | |
| "loss": 1.6734, | |
| "step": 234500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.902808081321926e-05, | |
| "loss": 1.6902, | |
| "step": 235000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.90260129000559e-05, | |
| "loss": 1.6824, | |
| "step": 235500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9023944986892533e-05, | |
| "loss": 1.7159, | |
| "step": 236000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9021877073729172e-05, | |
| "loss": 1.6848, | |
| "step": 236500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9019809160565807e-05, | |
| "loss": 1.6727, | |
| "step": 237000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9017741247402445e-05, | |
| "loss": 1.6647, | |
| "step": 237500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.901567333423908e-05, | |
| "loss": 1.666, | |
| "step": 238000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.901360542107572e-05, | |
| "loss": 1.7109, | |
| "step": 238500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9011537507912353e-05, | |
| "loss": 1.6651, | |
| "step": 239000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9009469594748992e-05, | |
| "loss": 1.6668, | |
| "step": 239500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.900740168158563e-05, | |
| "loss": 1.682, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9005333768422265e-05, | |
| "loss": 1.7031, | |
| "step": 240500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.90032658552589e-05, | |
| "loss": 1.6993, | |
| "step": 241000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.900119794209554e-05, | |
| "loss": 1.6602, | |
| "step": 241500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8999130028932177e-05, | |
| "loss": 1.6772, | |
| "step": 242000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.899706211576881e-05, | |
| "loss": 1.7098, | |
| "step": 242500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8994994202605447e-05, | |
| "loss": 1.6754, | |
| "step": 243000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8992926289442085e-05, | |
| "loss": 1.6923, | |
| "step": 243500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8990858376278723e-05, | |
| "loss": 1.6801, | |
| "step": 244000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8988790463115358e-05, | |
| "loss": 1.6797, | |
| "step": 244500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8986722549951993e-05, | |
| "loss": 1.6852, | |
| "step": 245000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.898465463678863e-05, | |
| "loss": 1.6869, | |
| "step": 245500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.898258672362527e-05, | |
| "loss": 1.6933, | |
| "step": 246000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8980518810461905e-05, | |
| "loss": 1.6877, | |
| "step": 246500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.897845089729854e-05, | |
| "loss": 1.6923, | |
| "step": 247000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8976382984135178e-05, | |
| "loss": 1.7072, | |
| "step": 247500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8974315070971817e-05, | |
| "loss": 1.7022, | |
| "step": 248000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.897224715780845e-05, | |
| "loss": 1.6557, | |
| "step": 248500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.897017924464509e-05, | |
| "loss": 1.679, | |
| "step": 249000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.8968111331481725e-05, | |
| "loss": 1.6773, | |
| "step": 249500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8966043418318363e-05, | |
| "loss": 1.6589, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8963975505154998e-05, | |
| "loss": 1.6512, | |
| "step": 250500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8961907591991636e-05, | |
| "loss": 1.677, | |
| "step": 251000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.895983967882827e-05, | |
| "loss": 1.6426, | |
| "step": 251500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.895777176566491e-05, | |
| "loss": 1.6895, | |
| "step": 252000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8955703852501545e-05, | |
| "loss": 1.7015, | |
| "step": 252500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8953635939338183e-05, | |
| "loss": 1.6814, | |
| "step": 253000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.895156802617482e-05, | |
| "loss": 1.6872, | |
| "step": 253500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8949500113011456e-05, | |
| "loss": 1.6808, | |
| "step": 254000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.894743219984809e-05, | |
| "loss": 1.6724, | |
| "step": 254500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.894536428668473e-05, | |
| "loss": 1.6583, | |
| "step": 255000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8943296373521368e-05, | |
| "loss": 1.6886, | |
| "step": 255500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8941228460358003e-05, | |
| "loss": 1.6697, | |
| "step": 256000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8939160547194638e-05, | |
| "loss": 1.6663, | |
| "step": 256500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8937092634031276e-05, | |
| "loss": 1.6772, | |
| "step": 257000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8935024720867915e-05, | |
| "loss": 1.686, | |
| "step": 257500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.893295680770455e-05, | |
| "loss": 1.6776, | |
| "step": 258000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8930888894541185e-05, | |
| "loss": 1.6644, | |
| "step": 258500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8928820981377823e-05, | |
| "loss": 1.709, | |
| "step": 259000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.892675306821446e-05, | |
| "loss": 1.6555, | |
| "step": 259500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8924685155051096e-05, | |
| "loss": 1.7072, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8922617241887735e-05, | |
| "loss": 1.6783, | |
| "step": 260500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.892054932872437e-05, | |
| "loss": 1.6766, | |
| "step": 261000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8918481415561008e-05, | |
| "loss": 1.6712, | |
| "step": 261500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8916413502397643e-05, | |
| "loss": 1.6923, | |
| "step": 262000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.891434558923428e-05, | |
| "loss": 1.699, | |
| "step": 262500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8912277676070916e-05, | |
| "loss": 1.6557, | |
| "step": 263000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8910209762907555e-05, | |
| "loss": 1.6932, | |
| "step": 263500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.890814184974419e-05, | |
| "loss": 1.6695, | |
| "step": 264000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8906073936580828e-05, | |
| "loss": 1.6708, | |
| "step": 264500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.8904006023417463e-05, | |
| "loss": 1.6846, | |
| "step": 265000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.89019381102541e-05, | |
| "loss": 1.6674, | |
| "step": 265500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8899870197090736e-05, | |
| "loss": 1.6888, | |
| "step": 266000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8897802283927375e-05, | |
| "loss": 1.6255, | |
| "step": 266500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8895734370764013e-05, | |
| "loss": 1.6834, | |
| "step": 267000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8893666457600648e-05, | |
| "loss": 1.6467, | |
| "step": 267500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8891598544437283e-05, | |
| "loss": 1.6775, | |
| "step": 268000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.888953063127392e-05, | |
| "loss": 1.6667, | |
| "step": 268500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.888746271811056e-05, | |
| "loss": 1.6797, | |
| "step": 269000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8885394804947194e-05, | |
| "loss": 1.6653, | |
| "step": 269500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.888332689178383e-05, | |
| "loss": 1.6725, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8881258978620468e-05, | |
| "loss": 1.6932, | |
| "step": 270500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8879191065457106e-05, | |
| "loss": 1.7109, | |
| "step": 271000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.887712315229374e-05, | |
| "loss": 1.6952, | |
| "step": 271500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8875055239130376e-05, | |
| "loss": 1.6407, | |
| "step": 272000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8872987325967014e-05, | |
| "loss": 1.6619, | |
| "step": 272500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8870919412803653e-05, | |
| "loss": 1.7035, | |
| "step": 273000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8868851499640288e-05, | |
| "loss": 1.6827, | |
| "step": 273500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8866783586476926e-05, | |
| "loss": 1.6668, | |
| "step": 274000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.886471567331356e-05, | |
| "loss": 1.6379, | |
| "step": 274500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.88626477601502e-05, | |
| "loss": 1.6813, | |
| "step": 275000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8860579846986834e-05, | |
| "loss": 1.7091, | |
| "step": 275500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8858511933823473e-05, | |
| "loss": 1.6762, | |
| "step": 276000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8856444020660108e-05, | |
| "loss": 1.688, | |
| "step": 276500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8854376107496746e-05, | |
| "loss": 1.6393, | |
| "step": 277000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.885230819433338e-05, | |
| "loss": 1.6629, | |
| "step": 277500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.885024028117002e-05, | |
| "loss": 1.6881, | |
| "step": 278000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8848172368006654e-05, | |
| "loss": 1.7069, | |
| "step": 278500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8846104454843293e-05, | |
| "loss": 1.6752, | |
| "step": 279000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8844036541679928e-05, | |
| "loss": 1.678, | |
| "step": 279500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8841968628516566e-05, | |
| "loss": 1.6643, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8839900715353204e-05, | |
| "loss": 1.6525, | |
| "step": 280500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.883783280218984e-05, | |
| "loss": 1.7002, | |
| "step": 281000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8835764889026474e-05, | |
| "loss": 1.6542, | |
| "step": 281500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8833696975863113e-05, | |
| "loss": 1.7043, | |
| "step": 282000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.883162906269975e-05, | |
| "loss": 1.6496, | |
| "step": 282500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8829561149536386e-05, | |
| "loss": 1.6587, | |
| "step": 283000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.882749323637302e-05, | |
| "loss": 1.6738, | |
| "step": 283500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.882542532320966e-05, | |
| "loss": 1.6681, | |
| "step": 284000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8823357410046297e-05, | |
| "loss": 1.633, | |
| "step": 284500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8821289496882932e-05, | |
| "loss": 1.6639, | |
| "step": 285000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8819221583719567e-05, | |
| "loss": 1.6569, | |
| "step": 285500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8817153670556206e-05, | |
| "loss": 1.7017, | |
| "step": 286000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8815085757392844e-05, | |
| "loss": 1.6748, | |
| "step": 286500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.881301784422948e-05, | |
| "loss": 1.6641, | |
| "step": 287000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8810949931066117e-05, | |
| "loss": 1.6913, | |
| "step": 287500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8808882017902752e-05, | |
| "loss": 1.6746, | |
| "step": 288000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.880681410473939e-05, | |
| "loss": 1.7007, | |
| "step": 288500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8804746191576026e-05, | |
| "loss": 1.6874, | |
| "step": 289000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8802678278412664e-05, | |
| "loss": 1.6801, | |
| "step": 289500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.88006103652493e-05, | |
| "loss": 1.6809, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8798542452085937e-05, | |
| "loss": 1.6858, | |
| "step": 290500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8796474538922572e-05, | |
| "loss": 1.6333, | |
| "step": 291000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.879440662575921e-05, | |
| "loss": 1.6655, | |
| "step": 291500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8792338712595846e-05, | |
| "loss": 1.6815, | |
| "step": 292000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8790270799432484e-05, | |
| "loss": 1.6881, | |
| "step": 292500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.878820288626912e-05, | |
| "loss": 1.6599, | |
| "step": 293000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8786134973105757e-05, | |
| "loss": 1.6517, | |
| "step": 293500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8784067059942396e-05, | |
| "loss": 1.6852, | |
| "step": 294000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.878199914677903e-05, | |
| "loss": 1.6945, | |
| "step": 294500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8779931233615666e-05, | |
| "loss": 1.6562, | |
| "step": 295000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8777863320452304e-05, | |
| "loss": 1.6828, | |
| "step": 295500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8775795407288942e-05, | |
| "loss": 1.6758, | |
| "step": 296000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8773727494125577e-05, | |
| "loss": 1.6738, | |
| "step": 296500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8771659580962212e-05, | |
| "loss": 1.6369, | |
| "step": 297000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.876959166779885e-05, | |
| "loss": 1.6627, | |
| "step": 297500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.876752375463549e-05, | |
| "loss": 1.6529, | |
| "step": 298000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8765455841472124e-05, | |
| "loss": 1.6635, | |
| "step": 298500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8763387928308762e-05, | |
| "loss": 1.6566, | |
| "step": 299000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8761320015145397e-05, | |
| "loss": 1.6997, | |
| "step": 299500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8759252101982035e-05, | |
| "loss": 1.6699, | |
| "step": 300000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.875718418881867e-05, | |
| "loss": 1.6844, | |
| "step": 300500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.875511627565531e-05, | |
| "loss": 1.6682, | |
| "step": 301000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8753048362491944e-05, | |
| "loss": 1.6541, | |
| "step": 301500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8750980449328582e-05, | |
| "loss": 1.6626, | |
| "step": 302000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8748912536165217e-05, | |
| "loss": 1.6739, | |
| "step": 302500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8746844623001855e-05, | |
| "loss": 1.6582, | |
| "step": 303000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.874477670983849e-05, | |
| "loss": 1.6743, | |
| "step": 303500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.874270879667513e-05, | |
| "loss": 1.6741, | |
| "step": 304000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8740640883511764e-05, | |
| "loss": 1.6588, | |
| "step": 304500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8738572970348402e-05, | |
| "loss": 1.6654, | |
| "step": 305000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8736505057185037e-05, | |
| "loss": 1.6849, | |
| "step": 305500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8734437144021675e-05, | |
| "loss": 1.6606, | |
| "step": 306000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.873236923085831e-05, | |
| "loss": 1.7128, | |
| "step": 306500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.873030131769495e-05, | |
| "loss": 1.6696, | |
| "step": 307000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8728233404531587e-05, | |
| "loss": 1.6637, | |
| "step": 307500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8726165491368222e-05, | |
| "loss": 1.6955, | |
| "step": 308000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8724097578204857e-05, | |
| "loss": 1.6834, | |
| "step": 308500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8722029665041495e-05, | |
| "loss": 1.6335, | |
| "step": 309000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8719961751878134e-05, | |
| "loss": 1.6748, | |
| "step": 309500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.871789383871477e-05, | |
| "loss": 1.6522, | |
| "step": 310000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8715825925551404e-05, | |
| "loss": 1.6767, | |
| "step": 310500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8713758012388042e-05, | |
| "loss": 1.6893, | |
| "step": 311000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.871169009922468e-05, | |
| "loss": 1.6616, | |
| "step": 311500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8709622186061315e-05, | |
| "loss": 1.6518, | |
| "step": 312000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8707554272897954e-05, | |
| "loss": 1.6703, | |
| "step": 312500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.870548635973459e-05, | |
| "loss": 1.6511, | |
| "step": 313000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8703418446571227e-05, | |
| "loss": 1.712, | |
| "step": 313500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8701350533407862e-05, | |
| "loss": 1.684, | |
| "step": 314000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.86992826202445e-05, | |
| "loss": 1.6989, | |
| "step": 314500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8697214707081135e-05, | |
| "loss": 1.6937, | |
| "step": 315000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8695146793917774e-05, | |
| "loss": 1.679, | |
| "step": 315500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.869307888075441e-05, | |
| "loss": 1.6523, | |
| "step": 316000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8691010967591047e-05, | |
| "loss": 1.6638, | |
| "step": 316500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8688943054427682e-05, | |
| "loss": 1.6578, | |
| "step": 317000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.868687514126432e-05, | |
| "loss": 1.668, | |
| "step": 317500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8684807228100955e-05, | |
| "loss": 1.6638, | |
| "step": 318000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8682739314937593e-05, | |
| "loss": 1.6585, | |
| "step": 318500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8680671401774232e-05, | |
| "loss": 1.6564, | |
| "step": 319000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8678603488610867e-05, | |
| "loss": 1.679, | |
| "step": 319500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8676535575447502e-05, | |
| "loss": 1.6669, | |
| "step": 320000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.867446766228414e-05, | |
| "loss": 1.6687, | |
| "step": 320500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.867239974912078e-05, | |
| "loss": 1.6963, | |
| "step": 321000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8670331835957413e-05, | |
| "loss": 1.6617, | |
| "step": 321500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.866826392279405e-05, | |
| "loss": 1.6586, | |
| "step": 322000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8666196009630687e-05, | |
| "loss": 1.6685, | |
| "step": 322500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8664128096467325e-05, | |
| "loss": 1.6909, | |
| "step": 323000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.866206018330396e-05, | |
| "loss": 1.6458, | |
| "step": 323500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.86599922701406e-05, | |
| "loss": 1.6812, | |
| "step": 324000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8657924356977233e-05, | |
| "loss": 1.6596, | |
| "step": 324500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.865585644381387e-05, | |
| "loss": 1.6848, | |
| "step": 325000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8653788530650507e-05, | |
| "loss": 1.6796, | |
| "step": 325500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8651720617487145e-05, | |
| "loss": 1.6883, | |
| "step": 326000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.864965270432378e-05, | |
| "loss": 1.6402, | |
| "step": 326500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8647584791160418e-05, | |
| "loss": 1.6754, | |
| "step": 327000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8645516877997053e-05, | |
| "loss": 1.6419, | |
| "step": 327500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.864344896483369e-05, | |
| "loss": 1.6601, | |
| "step": 328000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8641381051670327e-05, | |
| "loss": 1.6689, | |
| "step": 328500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8639313138506965e-05, | |
| "loss": 1.6372, | |
| "step": 329000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.86372452253436e-05, | |
| "loss": 1.6361, | |
| "step": 329500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8635177312180238e-05, | |
| "loss": 1.6986, | |
| "step": 330000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8633109399016873e-05, | |
| "loss": 1.6626, | |
| "step": 330500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.863104148585351e-05, | |
| "loss": 1.6471, | |
| "step": 331000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8628973572690146e-05, | |
| "loss": 1.6749, | |
| "step": 331500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8626905659526785e-05, | |
| "loss": 1.6646, | |
| "step": 332000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8624837746363423e-05, | |
| "loss": 1.6745, | |
| "step": 332500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8622769833200058e-05, | |
| "loss": 1.6759, | |
| "step": 333000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8620701920036693e-05, | |
| "loss": 1.6502, | |
| "step": 333500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.861863400687333e-05, | |
| "loss": 1.6745, | |
| "step": 334000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.861656609370997e-05, | |
| "loss": 1.6366, | |
| "step": 334500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8614498180546605e-05, | |
| "loss": 1.6612, | |
| "step": 335000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.861243026738324e-05, | |
| "loss": 1.6645, | |
| "step": 335500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8610362354219878e-05, | |
| "loss": 1.6613, | |
| "step": 336000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8608294441056516e-05, | |
| "loss": 1.6703, | |
| "step": 336500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.860622652789315e-05, | |
| "loss": 1.6633, | |
| "step": 337000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.860415861472979e-05, | |
| "loss": 1.6559, | |
| "step": 337500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8602090701566425e-05, | |
| "loss": 1.6805, | |
| "step": 338000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8600022788403063e-05, | |
| "loss": 1.6817, | |
| "step": 338500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8597954875239698e-05, | |
| "loss": 1.7016, | |
| "step": 339000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8595886962076336e-05, | |
| "loss": 1.6582, | |
| "step": 339500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.859381904891297e-05, | |
| "loss": 1.6472, | |
| "step": 340000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.859175113574961e-05, | |
| "loss": 1.6682, | |
| "step": 340500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8589683222586245e-05, | |
| "loss": 1.7143, | |
| "step": 341000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8587615309422883e-05, | |
| "loss": 1.6578, | |
| "step": 341500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8585547396259518e-05, | |
| "loss": 1.662, | |
| "step": 342000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8583479483096156e-05, | |
| "loss": 1.6454, | |
| "step": 342500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.858141156993279e-05, | |
| "loss": 1.6586, | |
| "step": 343000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.857934365676943e-05, | |
| "loss": 1.6859, | |
| "step": 343500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8577275743606065e-05, | |
| "loss": 1.6667, | |
| "step": 344000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8575207830442703e-05, | |
| "loss": 1.6875, | |
| "step": 344500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8573139917279338e-05, | |
| "loss": 1.6784, | |
| "step": 345000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8571072004115976e-05, | |
| "loss": 1.6744, | |
| "step": 345500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8569004090952615e-05, | |
| "loss": 1.6499, | |
| "step": 346000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.856693617778925e-05, | |
| "loss": 1.6434, | |
| "step": 346500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8564868264625884e-05, | |
| "loss": 1.6592, | |
| "step": 347000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8562800351462523e-05, | |
| "loss": 1.6665, | |
| "step": 347500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.856073243829916e-05, | |
| "loss": 1.6438, | |
| "step": 348000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8558664525135796e-05, | |
| "loss": 1.6838, | |
| "step": 348500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.855659661197243e-05, | |
| "loss": 1.6925, | |
| "step": 349000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.855452869880907e-05, | |
| "loss": 1.6616, | |
| "step": 349500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8552460785645708e-05, | |
| "loss": 1.6478, | |
| "step": 350000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8550392872482343e-05, | |
| "loss": 1.6539, | |
| "step": 350500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.854832495931898e-05, | |
| "loss": 1.6667, | |
| "step": 351000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8546257046155616e-05, | |
| "loss": 1.698, | |
| "step": 351500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8544189132992254e-05, | |
| "loss": 1.6305, | |
| "step": 352000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.854212121982889e-05, | |
| "loss": 1.6525, | |
| "step": 352500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8540053306665528e-05, | |
| "loss": 1.6506, | |
| "step": 353000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8537985393502163e-05, | |
| "loss": 1.6907, | |
| "step": 353500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.85359174803388e-05, | |
| "loss": 1.6641, | |
| "step": 354000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8533849567175436e-05, | |
| "loss": 1.6587, | |
| "step": 354500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8531781654012074e-05, | |
| "loss": 1.6549, | |
| "step": 355000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.852971374084871e-05, | |
| "loss": 1.6669, | |
| "step": 355500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8527645827685348e-05, | |
| "loss": 1.6615, | |
| "step": 356000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8525577914521983e-05, | |
| "loss": 1.7056, | |
| "step": 356500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.852351000135862e-05, | |
| "loss": 1.6965, | |
| "step": 357000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8521442088195256e-05, | |
| "loss": 1.65, | |
| "step": 357500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8519374175031894e-05, | |
| "loss": 1.662, | |
| "step": 358000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.851730626186853e-05, | |
| "loss": 1.6659, | |
| "step": 358500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8515238348705168e-05, | |
| "loss": 1.668, | |
| "step": 359000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8513170435541806e-05, | |
| "loss": 1.6712, | |
| "step": 359500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.851110252237844e-05, | |
| "loss": 1.655, | |
| "step": 360000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8509034609215076e-05, | |
| "loss": 1.6694, | |
| "step": 360500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8506966696051714e-05, | |
| "loss": 1.6588, | |
| "step": 361000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8504898782888353e-05, | |
| "loss": 1.6963, | |
| "step": 361500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8502830869724988e-05, | |
| "loss": 1.67, | |
| "step": 362000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8500762956561626e-05, | |
| "loss": 1.6612, | |
| "step": 362500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.849869504339826e-05, | |
| "loss": 1.6683, | |
| "step": 363000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.84966271302349e-05, | |
| "loss": 1.6851, | |
| "step": 363500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8494559217071534e-05, | |
| "loss": 1.6498, | |
| "step": 364000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8492491303908173e-05, | |
| "loss": 1.644, | |
| "step": 364500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8490423390744807e-05, | |
| "loss": 1.6754, | |
| "step": 365000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8488355477581446e-05, | |
| "loss": 1.6836, | |
| "step": 365500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.848628756441808e-05, | |
| "loss": 1.6537, | |
| "step": 366000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.848421965125472e-05, | |
| "loss": 1.651, | |
| "step": 366500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8482151738091354e-05, | |
| "loss": 1.6182, | |
| "step": 367000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8480083824927992e-05, | |
| "loss": 1.6661, | |
| "step": 367500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8478015911764627e-05, | |
| "loss": 1.6385, | |
| "step": 368000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8475947998601266e-05, | |
| "loss": 1.6284, | |
| "step": 368500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.84738800854379e-05, | |
| "loss": 1.6813, | |
| "step": 369000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.847181217227454e-05, | |
| "loss": 1.6621, | |
| "step": 369500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8469744259111174e-05, | |
| "loss": 1.663, | |
| "step": 370000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8467676345947812e-05, | |
| "loss": 1.6451, | |
| "step": 370500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8465608432784447e-05, | |
| "loss": 1.6663, | |
| "step": 371000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8463540519621086e-05, | |
| "loss": 1.63, | |
| "step": 371500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.846147260645772e-05, | |
| "loss": 1.6637, | |
| "step": 372000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.845940469329436e-05, | |
| "loss": 1.6485, | |
| "step": 372500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8457336780130997e-05, | |
| "loss": 1.6679, | |
| "step": 373000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8455268866967632e-05, | |
| "loss": 1.6681, | |
| "step": 373500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8453200953804267e-05, | |
| "loss": 1.6559, | |
| "step": 374000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8451133040640906e-05, | |
| "loss": 1.6885, | |
| "step": 374500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8449065127477544e-05, | |
| "loss": 1.6863, | |
| "step": 375000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.844699721431418e-05, | |
| "loss": 1.6402, | |
| "step": 375500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8444929301150817e-05, | |
| "loss": 1.6368, | |
| "step": 376000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8442861387987452e-05, | |
| "loss": 1.6648, | |
| "step": 376500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.844079347482409e-05, | |
| "loss": 1.6663, | |
| "step": 377000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8438725561660726e-05, | |
| "loss": 1.6628, | |
| "step": 377500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8436657648497364e-05, | |
| "loss": 1.6763, | |
| "step": 378000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8434589735334e-05, | |
| "loss": 1.6907, | |
| "step": 378500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8432521822170637e-05, | |
| "loss": 1.6456, | |
| "step": 379000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8430453909007272e-05, | |
| "loss": 1.6476, | |
| "step": 379500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.842838599584391e-05, | |
| "loss": 1.6527, | |
| "step": 380000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8426318082680545e-05, | |
| "loss": 1.6693, | |
| "step": 380500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8424250169517184e-05, | |
| "loss": 1.6449, | |
| "step": 381000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.842218225635382e-05, | |
| "loss": 1.6448, | |
| "step": 381500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8420114343190457e-05, | |
| "loss": 1.6584, | |
| "step": 382000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8418046430027092e-05, | |
| "loss": 1.6626, | |
| "step": 382500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.841597851686373e-05, | |
| "loss": 1.6332, | |
| "step": 383000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8413910603700365e-05, | |
| "loss": 1.6822, | |
| "step": 383500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8411842690537004e-05, | |
| "loss": 1.6918, | |
| "step": 384000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.840977477737364e-05, | |
| "loss": 1.6372, | |
| "step": 384500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8407706864210277e-05, | |
| "loss": 1.6666, | |
| "step": 385000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8405638951046912e-05, | |
| "loss": 1.6548, | |
| "step": 385500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.840357103788355e-05, | |
| "loss": 1.6538, | |
| "step": 386000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.840150312472019e-05, | |
| "loss": 1.6661, | |
| "step": 386500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8399435211556824e-05, | |
| "loss": 1.6869, | |
| "step": 387000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8397367298393462e-05, | |
| "loss": 1.6475, | |
| "step": 387500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8395299385230097e-05, | |
| "loss": 1.6752, | |
| "step": 388000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8393231472066735e-05, | |
| "loss": 1.6536, | |
| "step": 388500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.839116355890337e-05, | |
| "loss": 1.6247, | |
| "step": 389000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.838909564574001e-05, | |
| "loss": 1.6753, | |
| "step": 389500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8387027732576644e-05, | |
| "loss": 1.6638, | |
| "step": 390000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8384959819413282e-05, | |
| "loss": 1.6797, | |
| "step": 390500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8382891906249917e-05, | |
| "loss": 1.6426, | |
| "step": 391000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8380823993086555e-05, | |
| "loss": 1.6834, | |
| "step": 391500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.837875607992319e-05, | |
| "loss": 1.6424, | |
| "step": 392000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.837668816675983e-05, | |
| "loss": 1.6996, | |
| "step": 392500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8374620253596464e-05, | |
| "loss": 1.6475, | |
| "step": 393000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8372552340433102e-05, | |
| "loss": 1.668, | |
| "step": 393500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8370484427269737e-05, | |
| "loss": 1.6578, | |
| "step": 394000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.8368416514106375e-05, | |
| "loss": 1.6566, | |
| "step": 394500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.836634860094301e-05, | |
| "loss": 1.6448, | |
| "step": 395000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.836428068777965e-05, | |
| "loss": 1.6473, | |
| "step": 395500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8362212774616283e-05, | |
| "loss": 1.6196, | |
| "step": 396000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8360144861452922e-05, | |
| "loss": 1.6599, | |
| "step": 396500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8358076948289557e-05, | |
| "loss": 1.6793, | |
| "step": 397000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8356009035126195e-05, | |
| "loss": 1.6576, | |
| "step": 397500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.835394112196283e-05, | |
| "loss": 1.6568, | |
| "step": 398000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.835187320879947e-05, | |
| "loss": 1.6515, | |
| "step": 398500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8349805295636103e-05, | |
| "loss": 1.6276, | |
| "step": 399000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8347737382472742e-05, | |
| "loss": 1.6365, | |
| "step": 399500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.834566946930938e-05, | |
| "loss": 1.658, | |
| "step": 400000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8343601556146015e-05, | |
| "loss": 1.6499, | |
| "step": 400500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8341533642982653e-05, | |
| "loss": 1.653, | |
| "step": 401000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.833946572981929e-05, | |
| "loss": 1.635, | |
| "step": 401500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8337397816655927e-05, | |
| "loss": 1.675, | |
| "step": 402000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8335329903492562e-05, | |
| "loss": 1.6724, | |
| "step": 402500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.83332619903292e-05, | |
| "loss": 1.6298, | |
| "step": 403000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8331194077165835e-05, | |
| "loss": 1.6759, | |
| "step": 403500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8329126164002473e-05, | |
| "loss": 1.6709, | |
| "step": 404000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.832705825083911e-05, | |
| "loss": 1.6522, | |
| "step": 404500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8324990337675747e-05, | |
| "loss": 1.6463, | |
| "step": 405000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.832292242451238e-05, | |
| "loss": 1.6643, | |
| "step": 405500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.832085451134902e-05, | |
| "loss": 1.6055, | |
| "step": 406000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8318786598185655e-05, | |
| "loss": 1.6392, | |
| "step": 406500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8316718685022293e-05, | |
| "loss": 1.6522, | |
| "step": 407000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8314650771858928e-05, | |
| "loss": 1.6631, | |
| "step": 407500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8312582858695567e-05, | |
| "loss": 1.6838, | |
| "step": 408000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.83105149455322e-05, | |
| "loss": 1.6631, | |
| "step": 408500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.830844703236884e-05, | |
| "loss": 1.6685, | |
| "step": 409000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8306379119205475e-05, | |
| "loss": 1.653, | |
| "step": 409500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8304311206042113e-05, | |
| "loss": 1.6389, | |
| "step": 410000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8302243292878748e-05, | |
| "loss": 1.675, | |
| "step": 410500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.8300175379715387e-05, | |
| "loss": 1.6373, | |
| "step": 411000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.829810746655202e-05, | |
| "loss": 1.6883, | |
| "step": 411500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.829603955338866e-05, | |
| "loss": 1.6283, | |
| "step": 412000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8293971640225295e-05, | |
| "loss": 1.6554, | |
| "step": 412500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8291903727061933e-05, | |
| "loss": 1.6459, | |
| "step": 413000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.828983581389857e-05, | |
| "loss": 1.6659, | |
| "step": 413500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8287767900735206e-05, | |
| "loss": 1.6516, | |
| "step": 414000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8285699987571845e-05, | |
| "loss": 1.6383, | |
| "step": 414500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.828363207440848e-05, | |
| "loss": 1.6741, | |
| "step": 415000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8281564161245118e-05, | |
| "loss": 1.6367, | |
| "step": 415500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8279496248081753e-05, | |
| "loss": 1.643, | |
| "step": 416000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.827742833491839e-05, | |
| "loss": 1.6696, | |
| "step": 416500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8275360421755026e-05, | |
| "loss": 1.6456, | |
| "step": 417000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8273292508591665e-05, | |
| "loss": 1.6417, | |
| "step": 417500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.82712245954283e-05, | |
| "loss": 1.6375, | |
| "step": 418000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8269156682264938e-05, | |
| "loss": 1.6403, | |
| "step": 418500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8267088769101573e-05, | |
| "loss": 1.66, | |
| "step": 419000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.826502085593821e-05, | |
| "loss": 1.6382, | |
| "step": 419500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8262952942774846e-05, | |
| "loss": 1.6647, | |
| "step": 420000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8260885029611485e-05, | |
| "loss": 1.6479, | |
| "step": 420500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.825881711644812e-05, | |
| "loss": 1.6657, | |
| "step": 421000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8256749203284758e-05, | |
| "loss": 1.614, | |
| "step": 421500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8254681290121393e-05, | |
| "loss": 1.6482, | |
| "step": 422000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.825261337695803e-05, | |
| "loss": 1.6583, | |
| "step": 422500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8250545463794666e-05, | |
| "loss": 1.6448, | |
| "step": 423000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8248477550631305e-05, | |
| "loss": 1.6754, | |
| "step": 423500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.824640963746794e-05, | |
| "loss": 1.6287, | |
| "step": 424000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8244341724304578e-05, | |
| "loss": 1.6743, | |
| "step": 424500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8242273811141213e-05, | |
| "loss": 1.665, | |
| "step": 425000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.824020589797785e-05, | |
| "loss": 1.6557, | |
| "step": 425500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.823813798481449e-05, | |
| "loss": 1.6336, | |
| "step": 426000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8236070071651125e-05, | |
| "loss": 1.6197, | |
| "step": 426500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.8234002158487763e-05, | |
| "loss": 1.6367, | |
| "step": 427000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8231934245324398e-05, | |
| "loss": 1.624, | |
| "step": 427500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8229866332161036e-05, | |
| "loss": 1.6275, | |
| "step": 428000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.822779841899767e-05, | |
| "loss": 1.6556, | |
| "step": 428500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.822573050583431e-05, | |
| "loss": 1.6741, | |
| "step": 429000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8223662592670944e-05, | |
| "loss": 1.6421, | |
| "step": 429500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8221594679507583e-05, | |
| "loss": 1.6785, | |
| "step": 430000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8219526766344218e-05, | |
| "loss": 1.6487, | |
| "step": 430500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8217458853180856e-05, | |
| "loss": 1.6681, | |
| "step": 431000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.821539094001749e-05, | |
| "loss": 1.6533, | |
| "step": 431500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.821332302685413e-05, | |
| "loss": 1.6671, | |
| "step": 432000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8211255113690764e-05, | |
| "loss": 1.6432, | |
| "step": 432500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8209187200527403e-05, | |
| "loss": 1.6305, | |
| "step": 433000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8207119287364038e-05, | |
| "loss": 1.6523, | |
| "step": 433500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8205051374200676e-05, | |
| "loss": 1.6561, | |
| "step": 434000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.820298346103731e-05, | |
| "loss": 1.6446, | |
| "step": 434500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.820091554787395e-05, | |
| "loss": 1.6156, | |
| "step": 435000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8198847634710584e-05, | |
| "loss": 1.6394, | |
| "step": 435500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8196779721547223e-05, | |
| "loss": 1.6407, | |
| "step": 436000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8194711808383858e-05, | |
| "loss": 1.6342, | |
| "step": 436500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8192643895220496e-05, | |
| "loss": 1.6749, | |
| "step": 437000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.819057598205713e-05, | |
| "loss": 1.6508, | |
| "step": 437500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.818850806889377e-05, | |
| "loss": 1.6519, | |
| "step": 438000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8186440155730404e-05, | |
| "loss": 1.6573, | |
| "step": 438500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8184372242567043e-05, | |
| "loss": 1.667, | |
| "step": 439000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.818230432940368e-05, | |
| "loss": 1.6384, | |
| "step": 439500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8180236416240316e-05, | |
| "loss": 1.6238, | |
| "step": 440000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8178168503076954e-05, | |
| "loss": 1.6725, | |
| "step": 440500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.817610058991359e-05, | |
| "loss": 1.6449, | |
| "step": 441000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8174032676750228e-05, | |
| "loss": 1.6505, | |
| "step": 441500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8171964763586863e-05, | |
| "loss": 1.6707, | |
| "step": 442000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.81698968504235e-05, | |
| "loss": 1.6427, | |
| "step": 442500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8167828937260136e-05, | |
| "loss": 1.6475, | |
| "step": 443000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8165761024096774e-05, | |
| "loss": 1.6559, | |
| "step": 443500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.816369311093341e-05, | |
| "loss": 1.6613, | |
| "step": 444000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8161625197770048e-05, | |
| "loss": 1.6699, | |
| "step": 444500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8159557284606682e-05, | |
| "loss": 1.6719, | |
| "step": 445000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.815748937144332e-05, | |
| "loss": 1.6476, | |
| "step": 445500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8155421458279956e-05, | |
| "loss": 1.6824, | |
| "step": 446000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8153353545116594e-05, | |
| "loss": 1.6431, | |
| "step": 446500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.815128563195323e-05, | |
| "loss": 1.6254, | |
| "step": 447000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8149217718789867e-05, | |
| "loss": 1.6346, | |
| "step": 447500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8147149805626502e-05, | |
| "loss": 1.6573, | |
| "step": 448000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.814508189246314e-05, | |
| "loss": 1.6255, | |
| "step": 448500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8143013979299776e-05, | |
| "loss": 1.6647, | |
| "step": 449000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8140946066136414e-05, | |
| "loss": 1.6665, | |
| "step": 449500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.813887815297305e-05, | |
| "loss": 1.6266, | |
| "step": 450000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8136810239809687e-05, | |
| "loss": 1.6343, | |
| "step": 450500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8134742326646326e-05, | |
| "loss": 1.6593, | |
| "step": 451000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.813267441348296e-05, | |
| "loss": 1.6333, | |
| "step": 451500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.81306065003196e-05, | |
| "loss": 1.6328, | |
| "step": 452000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8128538587156234e-05, | |
| "loss": 1.6466, | |
| "step": 452500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8126470673992872e-05, | |
| "loss": 1.6815, | |
| "step": 453000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8124402760829507e-05, | |
| "loss": 1.638, | |
| "step": 453500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8122334847666146e-05, | |
| "loss": 1.6642, | |
| "step": 454000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.812026693450278e-05, | |
| "loss": 1.667, | |
| "step": 454500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.811819902133942e-05, | |
| "loss": 1.6592, | |
| "step": 455000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8116131108176054e-05, | |
| "loss": 1.6665, | |
| "step": 455500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8114063195012692e-05, | |
| "loss": 1.638, | |
| "step": 456000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8111995281849327e-05, | |
| "loss": 1.678, | |
| "step": 456500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8109927368685966e-05, | |
| "loss": 1.6459, | |
| "step": 457000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.81078594555226e-05, | |
| "loss": 1.626, | |
| "step": 457500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.810579154235924e-05, | |
| "loss": 1.6603, | |
| "step": 458000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8103723629195874e-05, | |
| "loss": 1.6515, | |
| "step": 458500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8101655716032512e-05, | |
| "loss": 1.6363, | |
| "step": 459000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8099587802869147e-05, | |
| "loss": 1.6601, | |
| "step": 459500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8097519889705786e-05, | |
| "loss": 1.6861, | |
| "step": 460000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.809545197654242e-05, | |
| "loss": 1.6588, | |
| "step": 460500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.809338406337906e-05, | |
| "loss": 1.639, | |
| "step": 461000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8091316150215694e-05, | |
| "loss": 1.6452, | |
| "step": 461500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8089248237052332e-05, | |
| "loss": 1.6557, | |
| "step": 462000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8087180323888967e-05, | |
| "loss": 1.6331, | |
| "step": 462500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8085112410725605e-05, | |
| "loss": 1.667, | |
| "step": 463000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.808304449756224e-05, | |
| "loss": 1.6583, | |
| "step": 463500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.808097658439888e-05, | |
| "loss": 1.6553, | |
| "step": 464000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8078908671235517e-05, | |
| "loss": 1.6531, | |
| "step": 464500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8076840758072152e-05, | |
| "loss": 1.6146, | |
| "step": 465000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.807477284490879e-05, | |
| "loss": 1.648, | |
| "step": 465500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8072704931745425e-05, | |
| "loss": 1.6284, | |
| "step": 466000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8070637018582064e-05, | |
| "loss": 1.6384, | |
| "step": 466500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.80685691054187e-05, | |
| "loss": 1.6111, | |
| "step": 467000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8066501192255337e-05, | |
| "loss": 1.6365, | |
| "step": 467500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8064433279091972e-05, | |
| "loss": 1.6376, | |
| "step": 468000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.806236536592861e-05, | |
| "loss": 1.6842, | |
| "step": 468500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8060297452765245e-05, | |
| "loss": 1.6258, | |
| "step": 469000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8058229539601884e-05, | |
| "loss": 1.6524, | |
| "step": 469500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.805616162643852e-05, | |
| "loss": 1.672, | |
| "step": 470000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8054093713275157e-05, | |
| "loss": 1.6601, | |
| "step": 470500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8052025800111792e-05, | |
| "loss": 1.6481, | |
| "step": 471000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.804995788694843e-05, | |
| "loss": 1.6419, | |
| "step": 471500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8047889973785065e-05, | |
| "loss": 1.6351, | |
| "step": 472000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8045822060621704e-05, | |
| "loss": 1.7028, | |
| "step": 472500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.804375414745834e-05, | |
| "loss": 1.6305, | |
| "step": 473000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8041686234294977e-05, | |
| "loss": 1.6622, | |
| "step": 473500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8039618321131612e-05, | |
| "loss": 1.6547, | |
| "step": 474000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.803755040796825e-05, | |
| "loss": 1.6331, | |
| "step": 474500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8035482494804885e-05, | |
| "loss": 1.6587, | |
| "step": 475000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8033414581641524e-05, | |
| "loss": 1.6488, | |
| "step": 475500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.803134666847816e-05, | |
| "loss": 1.6752, | |
| "step": 476000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8029278755314797e-05, | |
| "loss": 1.6321, | |
| "step": 476500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8027210842151432e-05, | |
| "loss": 1.6399, | |
| "step": 477000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.802514292898807e-05, | |
| "loss": 1.6356, | |
| "step": 477500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.802307501582471e-05, | |
| "loss": 1.6602, | |
| "step": 478000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8021007102661343e-05, | |
| "loss": 1.645, | |
| "step": 478500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8018939189497982e-05, | |
| "loss": 1.6481, | |
| "step": 479000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8016871276334617e-05, | |
| "loss": 1.6616, | |
| "step": 479500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8014803363171255e-05, | |
| "loss": 1.6492, | |
| "step": 480000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.801273545000789e-05, | |
| "loss": 1.6401, | |
| "step": 480500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.801066753684453e-05, | |
| "loss": 1.6495, | |
| "step": 481000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8008599623681163e-05, | |
| "loss": 1.638, | |
| "step": 481500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8006531710517802e-05, | |
| "loss": 1.6777, | |
| "step": 482000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8004463797354437e-05, | |
| "loss": 1.623, | |
| "step": 482500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.8002395884191075e-05, | |
| "loss": 1.6565, | |
| "step": 483000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.800032797102771e-05, | |
| "loss": 1.652, | |
| "step": 483500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.799826005786435e-05, | |
| "loss": 1.6234, | |
| "step": 484000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7996192144700983e-05, | |
| "loss": 1.6683, | |
| "step": 484500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.799412423153762e-05, | |
| "loss": 1.6404, | |
| "step": 485000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7992056318374257e-05, | |
| "loss": 1.6737, | |
| "step": 485500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7989988405210895e-05, | |
| "loss": 1.6639, | |
| "step": 486000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.798792049204753e-05, | |
| "loss": 1.6638, | |
| "step": 486500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.798585257888417e-05, | |
| "loss": 1.6462, | |
| "step": 487000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7983784665720803e-05, | |
| "loss": 1.636, | |
| "step": 487500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.798171675255744e-05, | |
| "loss": 1.6169, | |
| "step": 488000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7979648839394077e-05, | |
| "loss": 1.6738, | |
| "step": 488500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7977580926230715e-05, | |
| "loss": 1.6582, | |
| "step": 489000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7975513013067353e-05, | |
| "loss": 1.6413, | |
| "step": 489500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7973445099903988e-05, | |
| "loss": 1.6314, | |
| "step": 490000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.7971377186740623e-05, | |
| "loss": 1.6558, | |
| "step": 490500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.796930927357726e-05, | |
| "loss": 1.6365, | |
| "step": 491000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.79672413604139e-05, | |
| "loss": 1.6507, | |
| "step": 491500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7965173447250535e-05, | |
| "loss": 1.6312, | |
| "step": 492000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7963105534087173e-05, | |
| "loss": 1.6446, | |
| "step": 492500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7961037620923808e-05, | |
| "loss": 1.617, | |
| "step": 493000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7958969707760447e-05, | |
| "loss": 1.6339, | |
| "step": 493500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.795690179459708e-05, | |
| "loss": 1.6548, | |
| "step": 494000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.795483388143372e-05, | |
| "loss": 1.6461, | |
| "step": 494500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7952765968270355e-05, | |
| "loss": 1.6437, | |
| "step": 495000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7950698055106993e-05, | |
| "loss": 1.6661, | |
| "step": 495500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7948630141943628e-05, | |
| "loss": 1.6442, | |
| "step": 496000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7946562228780266e-05, | |
| "loss": 1.6483, | |
| "step": 496500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.79444943156169e-05, | |
| "loss": 1.633, | |
| "step": 497000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.794242640245354e-05, | |
| "loss": 1.6162, | |
| "step": 497500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7940358489290175e-05, | |
| "loss": 1.662, | |
| "step": 498000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7938290576126813e-05, | |
| "loss": 1.6637, | |
| "step": 498500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7936222662963448e-05, | |
| "loss": 1.6344, | |
| "step": 499000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7934154749800086e-05, | |
| "loss": 1.6536, | |
| "step": 499500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.793208683663672e-05, | |
| "loss": 1.6527, | |
| "step": 500000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.793001892347336e-05, | |
| "loss": 1.6556, | |
| "step": 500500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7927951010309995e-05, | |
| "loss": 1.65, | |
| "step": 501000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7925883097146633e-05, | |
| "loss": 1.6368, | |
| "step": 501500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7923815183983268e-05, | |
| "loss": 1.6548, | |
| "step": 502000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7921747270819906e-05, | |
| "loss": 1.6426, | |
| "step": 502500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7919679357656545e-05, | |
| "loss": 1.653, | |
| "step": 503000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.791761144449318e-05, | |
| "loss": 1.6329, | |
| "step": 503500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7915543531329815e-05, | |
| "loss": 1.6587, | |
| "step": 504000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7913475618166453e-05, | |
| "loss": 1.6381, | |
| "step": 504500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.791140770500309e-05, | |
| "loss": 1.6562, | |
| "step": 505000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7909339791839726e-05, | |
| "loss": 1.6149, | |
| "step": 505500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7907271878676365e-05, | |
| "loss": 1.6132, | |
| "step": 506000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7905203965513e-05, | |
| "loss": 1.6335, | |
| "step": 506500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7903136052349638e-05, | |
| "loss": 1.6401, | |
| "step": 507000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.7901068139186273e-05, | |
| "loss": 1.6226, | |
| "step": 507500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.789900022602291e-05, | |
| "loss": 1.6317, | |
| "step": 508000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7896932312859546e-05, | |
| "loss": 1.6391, | |
| "step": 508500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7894864399696185e-05, | |
| "loss": 1.6428, | |
| "step": 509000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.789279648653282e-05, | |
| "loss": 1.6388, | |
| "step": 509500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7890728573369458e-05, | |
| "loss": 1.6535, | |
| "step": 510000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7888660660206093e-05, | |
| "loss": 1.6337, | |
| "step": 510500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.788659274704273e-05, | |
| "loss": 1.6459, | |
| "step": 511000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7884524833879366e-05, | |
| "loss": 1.6372, | |
| "step": 511500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7882456920716004e-05, | |
| "loss": 1.6263, | |
| "step": 512000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.788038900755264e-05, | |
| "loss": 1.6255, | |
| "step": 512500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7878321094389278e-05, | |
| "loss": 1.6396, | |
| "step": 513000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7876253181225913e-05, | |
| "loss": 1.6647, | |
| "step": 513500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.787418526806255e-05, | |
| "loss": 1.6595, | |
| "step": 514000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.787211735489919e-05, | |
| "loss": 1.6335, | |
| "step": 514500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7870049441735824e-05, | |
| "loss": 1.6682, | |
| "step": 515000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.786798152857246e-05, | |
| "loss": 1.6296, | |
| "step": 515500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7865913615409098e-05, | |
| "loss": 1.5955, | |
| "step": 516000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7863845702245736e-05, | |
| "loss": 1.6446, | |
| "step": 516500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.786177778908237e-05, | |
| "loss": 1.6368, | |
| "step": 517000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7859709875919006e-05, | |
| "loss": 1.6246, | |
| "step": 517500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7857641962755644e-05, | |
| "loss": 1.6548, | |
| "step": 518000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7855574049592283e-05, | |
| "loss": 1.6405, | |
| "step": 518500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7853506136428918e-05, | |
| "loss": 1.6844, | |
| "step": 519000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7851438223265556e-05, | |
| "loss": 1.6358, | |
| "step": 519500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.784937031010219e-05, | |
| "loss": 1.6695, | |
| "step": 520000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.784730239693883e-05, | |
| "loss": 1.5935, | |
| "step": 520500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7845234483775464e-05, | |
| "loss": 1.6619, | |
| "step": 521000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7843166570612103e-05, | |
| "loss": 1.6272, | |
| "step": 521500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7841098657448738e-05, | |
| "loss": 1.6283, | |
| "step": 522000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.7839030744285376e-05, | |
| "loss": 1.6553, | |
| "step": 522500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.783696283112201e-05, | |
| "loss": 1.6153, | |
| "step": 523000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.783489491795865e-05, | |
| "loss": 1.6412, | |
| "step": 523500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7832827004795284e-05, | |
| "loss": 1.6449, | |
| "step": 524000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7830759091631923e-05, | |
| "loss": 1.5893, | |
| "step": 524500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7828691178468558e-05, | |
| "loss": 1.6059, | |
| "step": 525000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7826623265305196e-05, | |
| "loss": 1.6054, | |
| "step": 525500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.782455535214183e-05, | |
| "loss": 1.6312, | |
| "step": 526000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.782248743897847e-05, | |
| "loss": 1.6631, | |
| "step": 526500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7820419525815104e-05, | |
| "loss": 1.6257, | |
| "step": 527000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7818351612651742e-05, | |
| "loss": 1.6556, | |
| "step": 527500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.781628369948838e-05, | |
| "loss": 1.6308, | |
| "step": 528000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7814215786325016e-05, | |
| "loss": 1.6606, | |
| "step": 528500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.781214787316165e-05, | |
| "loss": 1.6317, | |
| "step": 529000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.781007995999829e-05, | |
| "loss": 1.6716, | |
| "step": 529500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7808012046834927e-05, | |
| "loss": 1.6568, | |
| "step": 530000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7805944133671562e-05, | |
| "loss": 1.5854, | |
| "step": 530500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7803876220508197e-05, | |
| "loss": 1.6572, | |
| "step": 531000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7801808307344836e-05, | |
| "loss": 1.6564, | |
| "step": 531500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7799740394181474e-05, | |
| "loss": 1.6512, | |
| "step": 532000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.779767248101811e-05, | |
| "loss": 1.6238, | |
| "step": 532500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7795604567854747e-05, | |
| "loss": 1.6626, | |
| "step": 533000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7793536654691382e-05, | |
| "loss": 1.6555, | |
| "step": 533500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.779146874152802e-05, | |
| "loss": 1.633, | |
| "step": 534000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7789400828364656e-05, | |
| "loss": 1.6286, | |
| "step": 534500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7787332915201294e-05, | |
| "loss": 1.6621, | |
| "step": 535000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.778526500203793e-05, | |
| "loss": 1.664, | |
| "step": 535500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7783197088874567e-05, | |
| "loss": 1.6794, | |
| "step": 536000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7781129175711202e-05, | |
| "loss": 1.6317, | |
| "step": 536500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.777906126254784e-05, | |
| "loss": 1.6345, | |
| "step": 537000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7776993349384476e-05, | |
| "loss": 1.6339, | |
| "step": 537500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7774925436221114e-05, | |
| "loss": 1.648, | |
| "step": 538000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.777285752305775e-05, | |
| "loss": 1.6169, | |
| "step": 538500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7770789609894387e-05, | |
| "loss": 1.6781, | |
| "step": 539000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7768721696731022e-05, | |
| "loss": 1.6373, | |
| "step": 539500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.776665378356766e-05, | |
| "loss": 1.6333, | |
| "step": 540000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7764585870404296e-05, | |
| "loss": 1.6021, | |
| "step": 540500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7762517957240934e-05, | |
| "loss": 1.6452, | |
| "step": 541000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7760450044077572e-05, | |
| "loss": 1.6011, | |
| "step": 541500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7758382130914207e-05, | |
| "loss": 1.6314, | |
| "step": 542000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7756314217750842e-05, | |
| "loss": 1.6631, | |
| "step": 542500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.775424630458748e-05, | |
| "loss": 1.6245, | |
| "step": 543000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.775217839142412e-05, | |
| "loss": 1.6129, | |
| "step": 543500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7750110478260754e-05, | |
| "loss": 1.641, | |
| "step": 544000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.774804256509739e-05, | |
| "loss": 1.642, | |
| "step": 544500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7745974651934027e-05, | |
| "loss": 1.6146, | |
| "step": 545000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7743906738770665e-05, | |
| "loss": 1.6357, | |
| "step": 545500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.77418388256073e-05, | |
| "loss": 1.6486, | |
| "step": 546000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.773977091244394e-05, | |
| "loss": 1.6318, | |
| "step": 546500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7737702999280574e-05, | |
| "loss": 1.6513, | |
| "step": 547000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7735635086117212e-05, | |
| "loss": 1.6505, | |
| "step": 547500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7733567172953847e-05, | |
| "loss": 1.6147, | |
| "step": 548000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7731499259790485e-05, | |
| "loss": 1.6534, | |
| "step": 548500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.772943134662712e-05, | |
| "loss": 1.6261, | |
| "step": 549000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.772736343346376e-05, | |
| "loss": 1.6144, | |
| "step": 549500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7725295520300394e-05, | |
| "loss": 1.6204, | |
| "step": 550000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7723227607137032e-05, | |
| "loss": 1.6198, | |
| "step": 550500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7721159693973667e-05, | |
| "loss": 1.6431, | |
| "step": 551000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7719091780810305e-05, | |
| "loss": 1.6504, | |
| "step": 551500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.771702386764694e-05, | |
| "loss": 1.6136, | |
| "step": 552000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.771495595448358e-05, | |
| "loss": 1.6396, | |
| "step": 552500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7712888041320217e-05, | |
| "loss": 1.6491, | |
| "step": 553000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7710820128156852e-05, | |
| "loss": 1.6219, | |
| "step": 553500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7708752214993487e-05, | |
| "loss": 1.6575, | |
| "step": 554000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7706684301830125e-05, | |
| "loss": 1.6177, | |
| "step": 554500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7704616388666764e-05, | |
| "loss": 1.6549, | |
| "step": 555000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.77025484755034e-05, | |
| "loss": 1.6602, | |
| "step": 555500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7700480562340034e-05, | |
| "loss": 1.6244, | |
| "step": 556000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7698412649176672e-05, | |
| "loss": 1.6532, | |
| "step": 556500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.769634473601331e-05, | |
| "loss": 1.6365, | |
| "step": 557000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7694276822849945e-05, | |
| "loss": 1.6516, | |
| "step": 557500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.769220890968658e-05, | |
| "loss": 1.6576, | |
| "step": 558000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.769014099652322e-05, | |
| "loss": 1.6514, | |
| "step": 558500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7688073083359857e-05, | |
| "loss": 1.6329, | |
| "step": 559000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7686005170196492e-05, | |
| "loss": 1.6732, | |
| "step": 559500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.768393725703313e-05, | |
| "loss": 1.6065, | |
| "step": 560000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7681869343869765e-05, | |
| "loss": 1.6091, | |
| "step": 560500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7679801430706403e-05, | |
| "loss": 1.66, | |
| "step": 561000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.767773351754304e-05, | |
| "loss": 1.6477, | |
| "step": 561500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7675665604379677e-05, | |
| "loss": 1.641, | |
| "step": 562000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7673597691216312e-05, | |
| "loss": 1.6301, | |
| "step": 562500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.767152977805295e-05, | |
| "loss": 1.6249, | |
| "step": 563000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7669461864889585e-05, | |
| "loss": 1.6138, | |
| "step": 563500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7667393951726223e-05, | |
| "loss": 1.5931, | |
| "step": 564000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.766532603856286e-05, | |
| "loss": 1.622, | |
| "step": 564500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7663258125399497e-05, | |
| "loss": 1.6384, | |
| "step": 565000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.766119021223613e-05, | |
| "loss": 1.6417, | |
| "step": 565500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.765912229907277e-05, | |
| "loss": 1.6032, | |
| "step": 566000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.765705438590941e-05, | |
| "loss": 1.6349, | |
| "step": 566500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7654986472746043e-05, | |
| "loss": 1.6122, | |
| "step": 567000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7652918559582678e-05, | |
| "loss": 1.6405, | |
| "step": 567500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7650850646419317e-05, | |
| "loss": 1.6618, | |
| "step": 568000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7648782733255955e-05, | |
| "loss": 1.6489, | |
| "step": 568500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.764671482009259e-05, | |
| "loss": 1.6121, | |
| "step": 569000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7644646906929225e-05, | |
| "loss": 1.6462, | |
| "step": 569500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7642578993765863e-05, | |
| "loss": 1.624, | |
| "step": 570000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.76405110806025e-05, | |
| "loss": 1.6397, | |
| "step": 570500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.7638443167439137e-05, | |
| "loss": 1.6166, | |
| "step": 571000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.763637525427577e-05, | |
| "loss": 1.6125, | |
| "step": 571500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.763430734111241e-05, | |
| "loss": 1.6181, | |
| "step": 572000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7632239427949048e-05, | |
| "loss": 1.6209, | |
| "step": 572500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7630171514785683e-05, | |
| "loss": 1.6457, | |
| "step": 573000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.762810360162232e-05, | |
| "loss": 1.6035, | |
| "step": 573500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7626035688458957e-05, | |
| "loss": 1.628, | |
| "step": 574000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7623967775295595e-05, | |
| "loss": 1.6419, | |
| "step": 574500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.762189986213223e-05, | |
| "loss": 1.6502, | |
| "step": 575000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7619831948968868e-05, | |
| "loss": 1.6417, | |
| "step": 575500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7617764035805503e-05, | |
| "loss": 1.6496, | |
| "step": 576000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.761569612264214e-05, | |
| "loss": 1.6203, | |
| "step": 576500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7613628209478776e-05, | |
| "loss": 1.6238, | |
| "step": 577000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7611560296315415e-05, | |
| "loss": 1.6238, | |
| "step": 577500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7609492383152053e-05, | |
| "loss": 1.6371, | |
| "step": 578000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7607424469988688e-05, | |
| "loss": 1.6438, | |
| "step": 578500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7605356556825323e-05, | |
| "loss": 1.6448, | |
| "step": 579000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.760328864366196e-05, | |
| "loss": 1.6503, | |
| "step": 579500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.76012207304986e-05, | |
| "loss": 1.6263, | |
| "step": 580000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7599152817335235e-05, | |
| "loss": 1.6184, | |
| "step": 580500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.759708490417187e-05, | |
| "loss": 1.6244, | |
| "step": 581000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7595016991008508e-05, | |
| "loss": 1.6449, | |
| "step": 581500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7592949077845146e-05, | |
| "loss": 1.62, | |
| "step": 582000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.759088116468178e-05, | |
| "loss": 1.6307, | |
| "step": 582500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7588813251518416e-05, | |
| "loss": 1.6465, | |
| "step": 583000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7586745338355055e-05, | |
| "loss": 1.6325, | |
| "step": 583500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7584677425191693e-05, | |
| "loss": 1.6253, | |
| "step": 584000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7582609512028328e-05, | |
| "loss": 1.6174, | |
| "step": 584500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7580541598864966e-05, | |
| "loss": 1.6509, | |
| "step": 585000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.75784736857016e-05, | |
| "loss": 1.651, | |
| "step": 585500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.757640577253824e-05, | |
| "loss": 1.6609, | |
| "step": 586000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7574337859374875e-05, | |
| "loss": 1.659, | |
| "step": 586500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7572269946211513e-05, | |
| "loss": 1.648, | |
| "step": 587000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7570202033048148e-05, | |
| "loss": 1.6242, | |
| "step": 587500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.7568134119884786e-05, | |
| "loss": 1.6211, | |
| "step": 588000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.756606620672142e-05, | |
| "loss": 1.6353, | |
| "step": 588500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.756399829355806e-05, | |
| "loss": 1.623, | |
| "step": 589000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7561930380394695e-05, | |
| "loss": 1.6158, | |
| "step": 589500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7559862467231333e-05, | |
| "loss": 1.6285, | |
| "step": 590000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7557794554067968e-05, | |
| "loss": 1.6366, | |
| "step": 590500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7555726640904606e-05, | |
| "loss": 1.6492, | |
| "step": 591000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7553658727741245e-05, | |
| "loss": 1.6514, | |
| "step": 591500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.755159081457788e-05, | |
| "loss": 1.6418, | |
| "step": 592000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7549522901414514e-05, | |
| "loss": 1.6543, | |
| "step": 592500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7547454988251153e-05, | |
| "loss": 1.666, | |
| "step": 593000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.754538707508779e-05, | |
| "loss": 1.6268, | |
| "step": 593500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7543319161924426e-05, | |
| "loss": 1.6264, | |
| "step": 594000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.754125124876106e-05, | |
| "loss": 1.6544, | |
| "step": 594500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.75391833355977e-05, | |
| "loss": 1.6506, | |
| "step": 595000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7537115422434338e-05, | |
| "loss": 1.6444, | |
| "step": 595500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7535047509270973e-05, | |
| "loss": 1.6227, | |
| "step": 596000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7532979596107608e-05, | |
| "loss": 1.6318, | |
| "step": 596500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7530911682944246e-05, | |
| "loss": 1.67, | |
| "step": 597000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7528843769780884e-05, | |
| "loss": 1.632, | |
| "step": 597500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.752677585661752e-05, | |
| "loss": 1.6357, | |
| "step": 598000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7524707943454158e-05, | |
| "loss": 1.6154, | |
| "step": 598500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7522640030290793e-05, | |
| "loss": 1.6264, | |
| "step": 599000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.752057211712743e-05, | |
| "loss": 1.6287, | |
| "step": 599500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7518504203964066e-05, | |
| "loss": 1.6764, | |
| "step": 600000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7516436290800704e-05, | |
| "loss": 1.6294, | |
| "step": 600500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.751436837763734e-05, | |
| "loss": 1.6131, | |
| "step": 601000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7512300464473978e-05, | |
| "loss": 1.6096, | |
| "step": 601500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7510232551310613e-05, | |
| "loss": 1.6342, | |
| "step": 602000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.750816463814725e-05, | |
| "loss": 1.6407, | |
| "step": 602500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7506096724983886e-05, | |
| "loss": 1.6358, | |
| "step": 603000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7504028811820524e-05, | |
| "loss": 1.6466, | |
| "step": 603500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.750196089865716e-05, | |
| "loss": 1.6399, | |
| "step": 604000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7499892985493798e-05, | |
| "loss": 1.6246, | |
| "step": 604500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7497825072330436e-05, | |
| "loss": 1.6343, | |
| "step": 605000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.749575715916707e-05, | |
| "loss": 1.6361, | |
| "step": 605500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7493689246003706e-05, | |
| "loss": 1.629, | |
| "step": 606000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7491621332840344e-05, | |
| "loss": 1.6218, | |
| "step": 606500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7489553419676983e-05, | |
| "loss": 1.6061, | |
| "step": 607000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7487485506513617e-05, | |
| "loss": 1.6301, | |
| "step": 607500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7485417593350252e-05, | |
| "loss": 1.6347, | |
| "step": 608000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.748334968018689e-05, | |
| "loss": 1.6262, | |
| "step": 608500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.748128176702353e-05, | |
| "loss": 1.6135, | |
| "step": 609000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7479213853860164e-05, | |
| "loss": 1.6469, | |
| "step": 609500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.74771459406968e-05, | |
| "loss": 1.6475, | |
| "step": 610000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7475078027533437e-05, | |
| "loss": 1.6186, | |
| "step": 610500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7473010114370076e-05, | |
| "loss": 1.6156, | |
| "step": 611000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.747094220120671e-05, | |
| "loss": 1.6327, | |
| "step": 611500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.746887428804335e-05, | |
| "loss": 1.605, | |
| "step": 612000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7466806374879984e-05, | |
| "loss": 1.643, | |
| "step": 612500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7464738461716622e-05, | |
| "loss": 1.6171, | |
| "step": 613000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7462670548553257e-05, | |
| "loss": 1.6221, | |
| "step": 613500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7460602635389896e-05, | |
| "loss": 1.6524, | |
| "step": 614000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.745853472222653e-05, | |
| "loss": 1.6575, | |
| "step": 614500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.745646680906317e-05, | |
| "loss": 1.6086, | |
| "step": 615000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7454398895899804e-05, | |
| "loss": 1.6231, | |
| "step": 615500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7452330982736442e-05, | |
| "loss": 1.623, | |
| "step": 616000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.745026306957308e-05, | |
| "loss": 1.629, | |
| "step": 616500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7448195156409716e-05, | |
| "loss": 1.6311, | |
| "step": 617000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.744612724324635e-05, | |
| "loss": 1.6315, | |
| "step": 617500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.744405933008299e-05, | |
| "loss": 1.6265, | |
| "step": 618000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7441991416919627e-05, | |
| "loss": 1.6621, | |
| "step": 618500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7439923503756262e-05, | |
| "loss": 1.6387, | |
| "step": 619000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7437855590592897e-05, | |
| "loss": 1.6451, | |
| "step": 619500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7435787677429536e-05, | |
| "loss": 1.6337, | |
| "step": 620000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7433719764266174e-05, | |
| "loss": 1.6355, | |
| "step": 620500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.743165185110281e-05, | |
| "loss": 1.601, | |
| "step": 621000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7429583937939444e-05, | |
| "loss": 1.6154, | |
| "step": 621500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7427516024776082e-05, | |
| "loss": 1.6416, | |
| "step": 622000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.742544811161272e-05, | |
| "loss": 1.5854, | |
| "step": 622500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7423380198449356e-05, | |
| "loss": 1.6032, | |
| "step": 623000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.742131228528599e-05, | |
| "loss": 1.6338, | |
| "step": 623500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.741924437212263e-05, | |
| "loss": 1.637, | |
| "step": 624000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7417176458959267e-05, | |
| "loss": 1.6116, | |
| "step": 624500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7415108545795902e-05, | |
| "loss": 1.6159, | |
| "step": 625000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.741304063263254e-05, | |
| "loss": 1.662, | |
| "step": 625500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7410972719469175e-05, | |
| "loss": 1.6573, | |
| "step": 626000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7408904806305814e-05, | |
| "loss": 1.619, | |
| "step": 626500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.740683689314245e-05, | |
| "loss": 1.6382, | |
| "step": 627000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7404768979979087e-05, | |
| "loss": 1.6311, | |
| "step": 627500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7402701066815722e-05, | |
| "loss": 1.6368, | |
| "step": 628000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.740063315365236e-05, | |
| "loss": 1.6384, | |
| "step": 628500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7398565240488995e-05, | |
| "loss": 1.6278, | |
| "step": 629000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7396497327325634e-05, | |
| "loss": 1.6204, | |
| "step": 629500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7394429414162272e-05, | |
| "loss": 1.6479, | |
| "step": 630000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7392361500998907e-05, | |
| "loss": 1.6141, | |
| "step": 630500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7390293587835542e-05, | |
| "loss": 1.6368, | |
| "step": 631000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.738822567467218e-05, | |
| "loss": 1.6722, | |
| "step": 631500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.738615776150882e-05, | |
| "loss": 1.6306, | |
| "step": 632000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7384089848345454e-05, | |
| "loss": 1.603, | |
| "step": 632500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.738202193518209e-05, | |
| "loss": 1.6092, | |
| "step": 633000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7379954022018727e-05, | |
| "loss": 1.6071, | |
| "step": 633500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7377886108855365e-05, | |
| "loss": 1.6586, | |
| "step": 634000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7375818195692e-05, | |
| "loss": 1.6435, | |
| "step": 634500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7373750282528635e-05, | |
| "loss": 1.6015, | |
| "step": 635000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7371682369365274e-05, | |
| "loss": 1.6129, | |
| "step": 635500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7369614456201912e-05, | |
| "loss": 1.6231, | |
| "step": 636000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7367546543038547e-05, | |
| "loss": 1.6303, | |
| "step": 636500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7365478629875182e-05, | |
| "loss": 1.6416, | |
| "step": 637000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.736341071671182e-05, | |
| "loss": 1.6392, | |
| "step": 637500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.736134280354846e-05, | |
| "loss": 1.6424, | |
| "step": 638000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7359274890385094e-05, | |
| "loss": 1.6172, | |
| "step": 638500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7357206977221732e-05, | |
| "loss": 1.6347, | |
| "step": 639000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7355139064058367e-05, | |
| "loss": 1.6149, | |
| "step": 639500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7353071150895005e-05, | |
| "loss": 1.6068, | |
| "step": 640000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.735100323773164e-05, | |
| "loss": 1.6347, | |
| "step": 640500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.734893532456828e-05, | |
| "loss": 1.6446, | |
| "step": 641000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7346867411404917e-05, | |
| "loss": 1.6167, | |
| "step": 641500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7344799498241552e-05, | |
| "loss": 1.6412, | |
| "step": 642000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7342731585078187e-05, | |
| "loss": 1.6243, | |
| "step": 642500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7340663671914825e-05, | |
| "loss": 1.6058, | |
| "step": 643000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7338595758751463e-05, | |
| "loss": 1.6145, | |
| "step": 643500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.73365278455881e-05, | |
| "loss": 1.6231, | |
| "step": 644000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7334459932424733e-05, | |
| "loss": 1.6071, | |
| "step": 644500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7332392019261372e-05, | |
| "loss": 1.6086, | |
| "step": 645000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.733032410609801e-05, | |
| "loss": 1.6516, | |
| "step": 645500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7328256192934645e-05, | |
| "loss": 1.6462, | |
| "step": 646000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.732618827977128e-05, | |
| "loss": 1.6195, | |
| "step": 646500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.732412036660792e-05, | |
| "loss": 1.6253, | |
| "step": 647000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7322052453444557e-05, | |
| "loss": 1.6312, | |
| "step": 647500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.731998454028119e-05, | |
| "loss": 1.609, | |
| "step": 648000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7317916627117827e-05, | |
| "loss": 1.6005, | |
| "step": 648500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7315848713954465e-05, | |
| "loss": 1.6201, | |
| "step": 649000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7313780800791103e-05, | |
| "loss": 1.6343, | |
| "step": 649500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7311712887627738e-05, | |
| "loss": 1.6419, | |
| "step": 650000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7309644974464373e-05, | |
| "loss": 1.6027, | |
| "step": 650500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.730757706130101e-05, | |
| "loss": 1.6262, | |
| "step": 651000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.730550914813765e-05, | |
| "loss": 1.6312, | |
| "step": 651500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7303441234974285e-05, | |
| "loss": 1.6311, | |
| "step": 652000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7301373321810923e-05, | |
| "loss": 1.6296, | |
| "step": 652500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7299305408647558e-05, | |
| "loss": 1.6482, | |
| "step": 653000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7297237495484197e-05, | |
| "loss": 1.66, | |
| "step": 653500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.729516958232083e-05, | |
| "loss": 1.601, | |
| "step": 654000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.729310166915747e-05, | |
| "loss": 1.6438, | |
| "step": 654500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7291033755994108e-05, | |
| "loss": 1.6462, | |
| "step": 655000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7288965842830743e-05, | |
| "loss": 1.6466, | |
| "step": 655500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7286897929667378e-05, | |
| "loss": 1.6312, | |
| "step": 656000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7284830016504016e-05, | |
| "loss": 1.6351, | |
| "step": 656500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7282762103340655e-05, | |
| "loss": 1.6702, | |
| "step": 657000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.728069419017729e-05, | |
| "loss": 1.6228, | |
| "step": 657500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7278626277013925e-05, | |
| "loss": 1.6196, | |
| "step": 658000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7276558363850563e-05, | |
| "loss": 1.6107, | |
| "step": 658500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.72744904506872e-05, | |
| "loss": 1.6189, | |
| "step": 659000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7272422537523836e-05, | |
| "loss": 1.6207, | |
| "step": 659500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.727035462436047e-05, | |
| "loss": 1.6378, | |
| "step": 660000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.726828671119711e-05, | |
| "loss": 1.6469, | |
| "step": 660500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7266218798033748e-05, | |
| "loss": 1.6155, | |
| "step": 661000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7264150884870383e-05, | |
| "loss": 1.6233, | |
| "step": 661500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7262082971707018e-05, | |
| "loss": 1.6268, | |
| "step": 662000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7260015058543656e-05, | |
| "loss": 1.62, | |
| "step": 662500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7257947145380295e-05, | |
| "loss": 1.5998, | |
| "step": 663000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.725587923221693e-05, | |
| "loss": 1.6149, | |
| "step": 663500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7253811319053565e-05, | |
| "loss": 1.6559, | |
| "step": 664000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7251743405890203e-05, | |
| "loss": 1.6312, | |
| "step": 664500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.724967549272684e-05, | |
| "loss": 1.6357, | |
| "step": 665000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7247607579563476e-05, | |
| "loss": 1.6029, | |
| "step": 665500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7245539666400115e-05, | |
| "loss": 1.6246, | |
| "step": 666000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.724347175323675e-05, | |
| "loss": 1.6299, | |
| "step": 666500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7241403840073388e-05, | |
| "loss": 1.6287, | |
| "step": 667000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7239335926910023e-05, | |
| "loss": 1.6016, | |
| "step": 667500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.723726801374666e-05, | |
| "loss": 1.647, | |
| "step": 668000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.72352001005833e-05, | |
| "loss": 1.6383, | |
| "step": 668500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7233132187419935e-05, | |
| "loss": 1.6502, | |
| "step": 669000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.723106427425657e-05, | |
| "loss": 1.6618, | |
| "step": 669500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7228996361093208e-05, | |
| "loss": 1.6288, | |
| "step": 670000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7226928447929846e-05, | |
| "loss": 1.6322, | |
| "step": 670500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.722486053476648e-05, | |
| "loss": 1.6003, | |
| "step": 671000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7222792621603116e-05, | |
| "loss": 1.626, | |
| "step": 671500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7220724708439755e-05, | |
| "loss": 1.6195, | |
| "step": 672000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7218656795276393e-05, | |
| "loss": 1.6523, | |
| "step": 672500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7216588882113028e-05, | |
| "loss": 1.6001, | |
| "step": 673000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7214520968949663e-05, | |
| "loss": 1.6395, | |
| "step": 673500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.72124530557863e-05, | |
| "loss": 1.6355, | |
| "step": 674000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.721038514262294e-05, | |
| "loss": 1.6219, | |
| "step": 674500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7208317229459574e-05, | |
| "loss": 1.6117, | |
| "step": 675000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.720624931629621e-05, | |
| "loss": 1.6335, | |
| "step": 675500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7204181403132848e-05, | |
| "loss": 1.6276, | |
| "step": 676000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7202113489969486e-05, | |
| "loss": 1.6281, | |
| "step": 676500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.720004557680612e-05, | |
| "loss": 1.6304, | |
| "step": 677000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7197977663642756e-05, | |
| "loss": 1.6321, | |
| "step": 677500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7195909750479394e-05, | |
| "loss": 1.6247, | |
| "step": 678000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7193841837316033e-05, | |
| "loss": 1.6113, | |
| "step": 678500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7191773924152668e-05, | |
| "loss": 1.629, | |
| "step": 679000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7189706010989306e-05, | |
| "loss": 1.603, | |
| "step": 679500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7187638097825944e-05, | |
| "loss": 1.5872, | |
| "step": 680000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.718557018466258e-05, | |
| "loss": 1.6194, | |
| "step": 680500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7183502271499214e-05, | |
| "loss": 1.6865, | |
| "step": 681000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7181434358335853e-05, | |
| "loss": 1.6145, | |
| "step": 681500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.717936644517249e-05, | |
| "loss": 1.6291, | |
| "step": 682000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7177298532009126e-05, | |
| "loss": 1.6212, | |
| "step": 682500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.717523061884576e-05, | |
| "loss": 1.6258, | |
| "step": 683000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.71731627056824e-05, | |
| "loss": 1.621, | |
| "step": 683500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7171094792519038e-05, | |
| "loss": 1.6305, | |
| "step": 684000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7169026879355673e-05, | |
| "loss": 1.6323, | |
| "step": 684500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7166958966192308e-05, | |
| "loss": 1.6341, | |
| "step": 685000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7164891053028946e-05, | |
| "loss": 1.6249, | |
| "step": 685500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7162823139865584e-05, | |
| "loss": 1.6121, | |
| "step": 686000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.716075522670222e-05, | |
| "loss": 1.584, | |
| "step": 686500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7158687313538854e-05, | |
| "loss": 1.6156, | |
| "step": 687000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7156619400375493e-05, | |
| "loss": 1.6302, | |
| "step": 687500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.715455148721213e-05, | |
| "loss": 1.6081, | |
| "step": 688000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7152483574048766e-05, | |
| "loss": 1.6222, | |
| "step": 688500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.71504156608854e-05, | |
| "loss": 1.6334, | |
| "step": 689000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.714834774772204e-05, | |
| "loss": 1.6235, | |
| "step": 689500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7146279834558677e-05, | |
| "loss": 1.5851, | |
| "step": 690000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7144211921395312e-05, | |
| "loss": 1.6439, | |
| "step": 690500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7142144008231947e-05, | |
| "loss": 1.6358, | |
| "step": 691000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7140076095068586e-05, | |
| "loss": 1.6366, | |
| "step": 691500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7138008181905224e-05, | |
| "loss": 1.6105, | |
| "step": 692000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.713594026874186e-05, | |
| "loss": 1.6611, | |
| "step": 692500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7133872355578497e-05, | |
| "loss": 1.6261, | |
| "step": 693000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7131804442415136e-05, | |
| "loss": 1.6112, | |
| "step": 693500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.712973652925177e-05, | |
| "loss": 1.6321, | |
| "step": 694000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7127668616088406e-05, | |
| "loss": 1.6309, | |
| "step": 694500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7125600702925044e-05, | |
| "loss": 1.6227, | |
| "step": 695000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7123532789761682e-05, | |
| "loss": 1.6312, | |
| "step": 695500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7121464876598317e-05, | |
| "loss": 1.5981, | |
| "step": 696000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7119396963434952e-05, | |
| "loss": 1.6602, | |
| "step": 696500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.711732905027159e-05, | |
| "loss": 1.6249, | |
| "step": 697000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.711526113710823e-05, | |
| "loss": 1.6476, | |
| "step": 697500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7113193223944864e-05, | |
| "loss": 1.6283, | |
| "step": 698000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.71111253107815e-05, | |
| "loss": 1.6297, | |
| "step": 698500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7109057397618137e-05, | |
| "loss": 1.6111, | |
| "step": 699000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7106989484454776e-05, | |
| "loss": 1.6241, | |
| "step": 699500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.710492157129141e-05, | |
| "loss": 1.6062, | |
| "step": 700000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7102853658128046e-05, | |
| "loss": 1.5961, | |
| "step": 700500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7100785744964684e-05, | |
| "loss": 1.6076, | |
| "step": 701000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7098717831801322e-05, | |
| "loss": 1.6305, | |
| "step": 701500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7096649918637957e-05, | |
| "loss": 1.6294, | |
| "step": 702000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7094582005474592e-05, | |
| "loss": 1.616, | |
| "step": 702500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.709251409231123e-05, | |
| "loss": 1.6168, | |
| "step": 703000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.709044617914787e-05, | |
| "loss": 1.6258, | |
| "step": 703500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7088378265984504e-05, | |
| "loss": 1.6252, | |
| "step": 704000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7086310352821142e-05, | |
| "loss": 1.6517, | |
| "step": 704500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.708424243965778e-05, | |
| "loss": 1.5765, | |
| "step": 705000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7082174526494415e-05, | |
| "loss": 1.6487, | |
| "step": 705500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.708010661333105e-05, | |
| "loss": 1.6345, | |
| "step": 706000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.707803870016769e-05, | |
| "loss": 1.6031, | |
| "step": 706500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7075970787004327e-05, | |
| "loss": 1.6342, | |
| "step": 707000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7073902873840962e-05, | |
| "loss": 1.6206, | |
| "step": 707500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7071834960677597e-05, | |
| "loss": 1.6423, | |
| "step": 708000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7069767047514235e-05, | |
| "loss": 1.6234, | |
| "step": 708500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7067699134350874e-05, | |
| "loss": 1.6136, | |
| "step": 709000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.706563122118751e-05, | |
| "loss": 1.6081, | |
| "step": 709500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7063563308024144e-05, | |
| "loss": 1.6156, | |
| "step": 710000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7061495394860782e-05, | |
| "loss": 1.6147, | |
| "step": 710500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.705942748169742e-05, | |
| "loss": 1.6214, | |
| "step": 711000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7057359568534055e-05, | |
| "loss": 1.6178, | |
| "step": 711500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.705529165537069e-05, | |
| "loss": 1.6208, | |
| "step": 712000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.705322374220733e-05, | |
| "loss": 1.6071, | |
| "step": 712500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7051155829043967e-05, | |
| "loss": 1.5874, | |
| "step": 713000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7049087915880602e-05, | |
| "loss": 1.6096, | |
| "step": 713500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7047020002717237e-05, | |
| "loss": 1.629, | |
| "step": 714000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7044952089553875e-05, | |
| "loss": 1.5995, | |
| "step": 714500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7042884176390514e-05, | |
| "loss": 1.6549, | |
| "step": 715000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.704081626322715e-05, | |
| "loss": 1.6188, | |
| "step": 715500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7038748350063784e-05, | |
| "loss": 1.6344, | |
| "step": 716000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7036680436900422e-05, | |
| "loss": 1.6217, | |
| "step": 716500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.703461252373706e-05, | |
| "loss": 1.6103, | |
| "step": 717000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7032544610573695e-05, | |
| "loss": 1.636, | |
| "step": 717500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7030476697410334e-05, | |
| "loss": 1.635, | |
| "step": 718000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7028408784246972e-05, | |
| "loss": 1.6325, | |
| "step": 718500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7026340871083607e-05, | |
| "loss": 1.6302, | |
| "step": 719000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7024272957920242e-05, | |
| "loss": 1.5954, | |
| "step": 719500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.702220504475688e-05, | |
| "loss": 1.6313, | |
| "step": 720000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.702013713159352e-05, | |
| "loss": 1.6005, | |
| "step": 720500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7018069218430154e-05, | |
| "loss": 1.6064, | |
| "step": 721000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.701600130526679e-05, | |
| "loss": 1.6423, | |
| "step": 721500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7013933392103427e-05, | |
| "loss": 1.6485, | |
| "step": 722000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7011865478940065e-05, | |
| "loss": 1.6033, | |
| "step": 722500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.70097975657767e-05, | |
| "loss": 1.627, | |
| "step": 723000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7007729652613335e-05, | |
| "loss": 1.6045, | |
| "step": 723500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7005661739449973e-05, | |
| "loss": 1.6407, | |
| "step": 724000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7003593826286612e-05, | |
| "loss": 1.6423, | |
| "step": 724500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7001525913123247e-05, | |
| "loss": 1.6242, | |
| "step": 725000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6999457999959882e-05, | |
| "loss": 1.6556, | |
| "step": 725500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.699739008679652e-05, | |
| "loss": 1.6116, | |
| "step": 726000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.699532217363316e-05, | |
| "loss": 1.6284, | |
| "step": 726500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6993254260469793e-05, | |
| "loss": 1.618, | |
| "step": 727000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.699118634730643e-05, | |
| "loss": 1.5947, | |
| "step": 727500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6989118434143067e-05, | |
| "loss": 1.6173, | |
| "step": 728000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6987050520979705e-05, | |
| "loss": 1.6308, | |
| "step": 728500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.698498260781634e-05, | |
| "loss": 1.6409, | |
| "step": 729000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6982914694652975e-05, | |
| "loss": 1.6495, | |
| "step": 729500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6980846781489613e-05, | |
| "loss": 1.6259, | |
| "step": 730000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.697877886832625e-05, | |
| "loss": 1.6356, | |
| "step": 730500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6976710955162887e-05, | |
| "loss": 1.6001, | |
| "step": 731000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6974643041999525e-05, | |
| "loss": 1.6428, | |
| "step": 731500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6972575128836163e-05, | |
| "loss": 1.5979, | |
| "step": 732000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6970507215672798e-05, | |
| "loss": 1.6392, | |
| "step": 732500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.6968439302509433e-05, | |
| "loss": 1.627, | |
| "step": 733000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.696637138934607e-05, | |
| "loss": 1.6309, | |
| "step": 733500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.696430347618271e-05, | |
| "loss": 1.6387, | |
| "step": 734000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6962235563019345e-05, | |
| "loss": 1.6423, | |
| "step": 734500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.696016764985598e-05, | |
| "loss": 1.6428, | |
| "step": 735000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6958099736692618e-05, | |
| "loss": 1.5919, | |
| "step": 735500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6956031823529257e-05, | |
| "loss": 1.6586, | |
| "step": 736000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.695396391036589e-05, | |
| "loss": 1.6269, | |
| "step": 736500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6951895997202526e-05, | |
| "loss": 1.6182, | |
| "step": 737000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6949828084039165e-05, | |
| "loss": 1.6255, | |
| "step": 737500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6947760170875803e-05, | |
| "loss": 1.6356, | |
| "step": 738000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6945692257712438e-05, | |
| "loss": 1.5926, | |
| "step": 738500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6943624344549073e-05, | |
| "loss": 1.6191, | |
| "step": 739000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.694155643138571e-05, | |
| "loss": 1.6305, | |
| "step": 739500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.693948851822235e-05, | |
| "loss": 1.6215, | |
| "step": 740000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6937420605058985e-05, | |
| "loss": 1.6102, | |
| "step": 740500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.693535269189562e-05, | |
| "loss": 1.6127, | |
| "step": 741000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6933284778732258e-05, | |
| "loss": 1.6285, | |
| "step": 741500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6931216865568896e-05, | |
| "loss": 1.621, | |
| "step": 742000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.692914895240553e-05, | |
| "loss": 1.6199, | |
| "step": 742500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.692708103924217e-05, | |
| "loss": 1.6117, | |
| "step": 743000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6925013126078808e-05, | |
| "loss": 1.6241, | |
| "step": 743500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6922945212915443e-05, | |
| "loss": 1.5992, | |
| "step": 744000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6920877299752078e-05, | |
| "loss": 1.6158, | |
| "step": 744500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6918809386588716e-05, | |
| "loss": 1.6026, | |
| "step": 745000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6916741473425355e-05, | |
| "loss": 1.6373, | |
| "step": 745500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.691467356026199e-05, | |
| "loss": 1.5959, | |
| "step": 746000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6912605647098625e-05, | |
| "loss": 1.6257, | |
| "step": 746500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6910537733935263e-05, | |
| "loss": 1.6163, | |
| "step": 747000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.69084698207719e-05, | |
| "loss": 1.5829, | |
| "step": 747500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6906401907608536e-05, | |
| "loss": 1.6134, | |
| "step": 748000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.690433399444517e-05, | |
| "loss": 1.6381, | |
| "step": 748500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.690226608128181e-05, | |
| "loss": 1.6317, | |
| "step": 749000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6900198168118448e-05, | |
| "loss": 1.63, | |
| "step": 749500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6898130254955083e-05, | |
| "loss": 1.6404, | |
| "step": 750000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6896062341791718e-05, | |
| "loss": 1.5971, | |
| "step": 750500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6893994428628356e-05, | |
| "loss": 1.6156, | |
| "step": 751000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6891926515464995e-05, | |
| "loss": 1.612, | |
| "step": 751500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.688985860230163e-05, | |
| "loss": 1.6083, | |
| "step": 752000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6887790689138264e-05, | |
| "loss": 1.6326, | |
| "step": 752500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6885722775974903e-05, | |
| "loss": 1.6283, | |
| "step": 753000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.688365486281154e-05, | |
| "loss": 1.6103, | |
| "step": 753500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6881586949648176e-05, | |
| "loss": 1.5798, | |
| "step": 754000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.687951903648481e-05, | |
| "loss": 1.6315, | |
| "step": 754500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.687745112332145e-05, | |
| "loss": 1.6222, | |
| "step": 755000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6875383210158088e-05, | |
| "loss": 1.6095, | |
| "step": 755500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6873315296994723e-05, | |
| "loss": 1.6415, | |
| "step": 756000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.687124738383136e-05, | |
| "loss": 1.612, | |
| "step": 756500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6869179470668e-05, | |
| "loss": 1.6159, | |
| "step": 757000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6867111557504634e-05, | |
| "loss": 1.6069, | |
| "step": 757500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.686504364434127e-05, | |
| "loss": 1.6274, | |
| "step": 758000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6862975731177908e-05, | |
| "loss": 1.6184, | |
| "step": 758500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6860907818014546e-05, | |
| "loss": 1.6288, | |
| "step": 759000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.685883990485118e-05, | |
| "loss": 1.62, | |
| "step": 759500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6856771991687816e-05, | |
| "loss": 1.6481, | |
| "step": 760000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6854704078524454e-05, | |
| "loss": 1.6165, | |
| "step": 760500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6852636165361093e-05, | |
| "loss": 1.6162, | |
| "step": 761000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6850568252197728e-05, | |
| "loss": 1.6005, | |
| "step": 761500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6848500339034363e-05, | |
| "loss": 1.6317, | |
| "step": 762000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6846432425871e-05, | |
| "loss": 1.6315, | |
| "step": 762500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.684436451270764e-05, | |
| "loss": 1.6227, | |
| "step": 763000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6842296599544274e-05, | |
| "loss": 1.6105, | |
| "step": 763500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.684022868638091e-05, | |
| "loss": 1.6221, | |
| "step": 764000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6838160773217548e-05, | |
| "loss": 1.6395, | |
| "step": 764500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.6836092860054186e-05, | |
| "loss": 1.6034, | |
| "step": 765000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.683402494689082e-05, | |
| "loss": 1.5798, | |
| "step": 765500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6831957033727456e-05, | |
| "loss": 1.6371, | |
| "step": 766000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6829889120564094e-05, | |
| "loss": 1.6359, | |
| "step": 766500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6827821207400733e-05, | |
| "loss": 1.5881, | |
| "step": 767000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6825753294237368e-05, | |
| "loss": 1.6114, | |
| "step": 767500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6823685381074002e-05, | |
| "loss": 1.608, | |
| "step": 768000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6821617467910644e-05, | |
| "loss": 1.595, | |
| "step": 768500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.681954955474728e-05, | |
| "loss": 1.6101, | |
| "step": 769000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6817481641583914e-05, | |
| "loss": 1.6065, | |
| "step": 769500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6815413728420553e-05, | |
| "loss": 1.6117, | |
| "step": 770000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.681334581525719e-05, | |
| "loss": 1.6382, | |
| "step": 770500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6811277902093826e-05, | |
| "loss": 1.6375, | |
| "step": 771000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.680920998893046e-05, | |
| "loss": 1.607, | |
| "step": 771500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.68071420757671e-05, | |
| "loss": 1.6272, | |
| "step": 772000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6805074162603737e-05, | |
| "loss": 1.6165, | |
| "step": 772500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6803006249440372e-05, | |
| "loss": 1.6434, | |
| "step": 773000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6800938336277007e-05, | |
| "loss": 1.6218, | |
| "step": 773500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6798870423113646e-05, | |
| "loss": 1.6327, | |
| "step": 774000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6796802509950284e-05, | |
| "loss": 1.6334, | |
| "step": 774500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.679473459678692e-05, | |
| "loss": 1.6142, | |
| "step": 775000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6792666683623554e-05, | |
| "loss": 1.6505, | |
| "step": 775500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6790598770460192e-05, | |
| "loss": 1.6233, | |
| "step": 776000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.678853085729683e-05, | |
| "loss": 1.6019, | |
| "step": 776500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6786462944133466e-05, | |
| "loss": 1.632, | |
| "step": 777000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.67843950309701e-05, | |
| "loss": 1.6164, | |
| "step": 777500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.678232711780674e-05, | |
| "loss": 1.5926, | |
| "step": 778000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6780259204643377e-05, | |
| "loss": 1.6043, | |
| "step": 778500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6778191291480012e-05, | |
| "loss": 1.6415, | |
| "step": 779000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6776123378316647e-05, | |
| "loss": 1.617, | |
| "step": 779500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6774055465153286e-05, | |
| "loss": 1.6147, | |
| "step": 780000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6771987551989924e-05, | |
| "loss": 1.6235, | |
| "step": 780500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.676991963882656e-05, | |
| "loss": 1.5808, | |
| "step": 781000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.6767851725663197e-05, | |
| "loss": 1.6052, | |
| "step": 781500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6765783812499836e-05, | |
| "loss": 1.6179, | |
| "step": 782000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.676371589933647e-05, | |
| "loss": 1.6264, | |
| "step": 782500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6761647986173106e-05, | |
| "loss": 1.6164, | |
| "step": 783000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6759580073009744e-05, | |
| "loss": 1.6288, | |
| "step": 783500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6757512159846382e-05, | |
| "loss": 1.6343, | |
| "step": 784000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6755444246683017e-05, | |
| "loss": 1.5849, | |
| "step": 784500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6753376333519652e-05, | |
| "loss": 1.604, | |
| "step": 785000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.675130842035629e-05, | |
| "loss": 1.606, | |
| "step": 785500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.674924050719293e-05, | |
| "loss": 1.6209, | |
| "step": 786000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6747172594029564e-05, | |
| "loss": 1.5959, | |
| "step": 786500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.67451046808662e-05, | |
| "loss": 1.6141, | |
| "step": 787000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6743036767702837e-05, | |
| "loss": 1.6136, | |
| "step": 787500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6740968854539475e-05, | |
| "loss": 1.6197, | |
| "step": 788000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.673890094137611e-05, | |
| "loss": 1.5811, | |
| "step": 788500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6736833028212745e-05, | |
| "loss": 1.5927, | |
| "step": 789000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6734765115049384e-05, | |
| "loss": 1.6014, | |
| "step": 789500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6732697201886022e-05, | |
| "loss": 1.6378, | |
| "step": 790000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6730629288722657e-05, | |
| "loss": 1.6117, | |
| "step": 790500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6728561375559292e-05, | |
| "loss": 1.6138, | |
| "step": 791000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.672649346239593e-05, | |
| "loss": 1.6263, | |
| "step": 791500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.672442554923257e-05, | |
| "loss": 1.6092, | |
| "step": 792000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6722357636069204e-05, | |
| "loss": 1.597, | |
| "step": 792500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.672028972290584e-05, | |
| "loss": 1.6182, | |
| "step": 793000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6718221809742477e-05, | |
| "loss": 1.5879, | |
| "step": 793500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6716153896579115e-05, | |
| "loss": 1.6249, | |
| "step": 794000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.671408598341575e-05, | |
| "loss": 1.6142, | |
| "step": 794500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.671201807025239e-05, | |
| "loss": 1.6236, | |
| "step": 795000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6709950157089027e-05, | |
| "loss": 1.6165, | |
| "step": 795500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6707882243925662e-05, | |
| "loss": 1.5858, | |
| "step": 796000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6705814330762297e-05, | |
| "loss": 1.6236, | |
| "step": 796500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6703746417598935e-05, | |
| "loss": 1.6028, | |
| "step": 797000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6701678504435574e-05, | |
| "loss": 1.6031, | |
| "step": 797500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.669961059127221e-05, | |
| "loss": 1.6243, | |
| "step": 798000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6697542678108844e-05, | |
| "loss": 1.5955, | |
| "step": 798500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6695474764945482e-05, | |
| "loss": 1.5934, | |
| "step": 799000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.669340685178212e-05, | |
| "loss": 1.6422, | |
| "step": 799500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6691338938618755e-05, | |
| "loss": 1.5994, | |
| "step": 800000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.668927102545539e-05, | |
| "loss": 1.5911, | |
| "step": 800500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.668720311229203e-05, | |
| "loss": 1.6031, | |
| "step": 801000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6685135199128667e-05, | |
| "loss": 1.6206, | |
| "step": 801500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6683067285965302e-05, | |
| "loss": 1.6125, | |
| "step": 802000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6680999372801937e-05, | |
| "loss": 1.6397, | |
| "step": 802500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6678931459638575e-05, | |
| "loss": 1.6119, | |
| "step": 803000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6676863546475213e-05, | |
| "loss": 1.5956, | |
| "step": 803500 | |
| } | |
| ], | |
| "max_steps": 4835793, | |
| "num_train_epochs": 3, | |
| "total_flos": 5.587487855033885e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |