| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9995057389203141, | |
| "global_step": 728000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9965676313910704e-05, | |
| "loss": 2.299, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9931352627821406e-05, | |
| "loss": 2.1716, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.989702894173211e-05, | |
| "loss": 2.083, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.986270525564282e-05, | |
| "loss": 2.04, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.982838156955352e-05, | |
| "loss": 2.0074, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.979405788346422e-05, | |
| "loss": 1.9771, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.975973419737492e-05, | |
| "loss": 1.9466, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.972541051128563e-05, | |
| "loss": 1.9384, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9691086825196334e-05, | |
| "loss": 1.9446, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9656763139107036e-05, | |
| "loss": 1.9054, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.962243945301774e-05, | |
| "loss": 1.8931, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.958811576692845e-05, | |
| "loss": 1.8744, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.955379208083915e-05, | |
| "loss": 1.8849, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.951946839474985e-05, | |
| "loss": 1.8659, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.948514470866055e-05, | |
| "loss": 1.8463, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.945082102257126e-05, | |
| "loss": 1.8243, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9416497336481957e-05, | |
| "loss": 1.8247, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9382173650392665e-05, | |
| "loss": 1.8389, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.934784996430337e-05, | |
| "loss": 1.8103, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9313526278214076e-05, | |
| "loss": 1.7774, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.927920259212477e-05, | |
| "loss": 1.7758, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.924487890603548e-05, | |
| "loss": 1.7937, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.921055521994618e-05, | |
| "loss": 1.7766, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9176231533856884e-05, | |
| "loss": 1.7564, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9141907847767586e-05, | |
| "loss": 1.7585, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9107584161678295e-05, | |
| "loss": 1.7528, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9073260475589e-05, | |
| "loss": 1.7505, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.90389367894997e-05, | |
| "loss": 1.7548, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.90046131034104e-05, | |
| "loss": 1.7404, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.897028941732111e-05, | |
| "loss": 1.7278, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.893596573123181e-05, | |
| "loss": 1.7325, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.8901642045142514e-05, | |
| "loss": 1.7148, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.8867318359053216e-05, | |
| "loss": 1.7168, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.8832994672963924e-05, | |
| "loss": 1.7354, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.879867098687462e-05, | |
| "loss": 1.712, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.876434730078533e-05, | |
| "loss": 1.7054, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.873002361469603e-05, | |
| "loss": 1.7183, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.869569992860674e-05, | |
| "loss": 1.6779, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.8661376242517434e-05, | |
| "loss": 1.7045, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.862705255642814e-05, | |
| "loss": 1.6761, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.8592728870338845e-05, | |
| "loss": 1.6818, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.855840518424955e-05, | |
| "loss": 1.6836, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.852408149816025e-05, | |
| "loss": 1.6878, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.848975781207096e-05, | |
| "loss": 1.6763, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.845543412598166e-05, | |
| "loss": 1.6657, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.842111043989236e-05, | |
| "loss": 1.6568, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.8386786753803064e-05, | |
| "loss": 1.6665, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.835246306771377e-05, | |
| "loss": 1.6809, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.8318139381624475e-05, | |
| "loss": 1.6569, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.8283815695535177e-05, | |
| "loss": 1.6375, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.824949200944588e-05, | |
| "loss": 1.6595, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.821516832335659e-05, | |
| "loss": 1.6409, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.818084463726729e-05, | |
| "loss": 1.6374, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.814652095117799e-05, | |
| "loss": 1.6294, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.811219726508869e-05, | |
| "loss": 1.633, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.80778735789994e-05, | |
| "loss": 1.6388, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.80435498929101e-05, | |
| "loss": 1.6359, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.8009226206820806e-05, | |
| "loss": 1.6297, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.797490252073151e-05, | |
| "loss": 1.635, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.794057883464222e-05, | |
| "loss": 1.629, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.790625514855291e-05, | |
| "loss": 1.609, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.787193146246362e-05, | |
| "loss": 1.6163, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.783760777637432e-05, | |
| "loss": 1.6199, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.7803284090285025e-05, | |
| "loss": 1.5975, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.776896040419573e-05, | |
| "loss": 1.626, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7734636718106436e-05, | |
| "loss": 1.6036, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.770031303201714e-05, | |
| "loss": 1.6077, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.766598934592784e-05, | |
| "loss": 1.6072, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.763166565983854e-05, | |
| "loss": 1.5979, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.759734197374925e-05, | |
| "loss": 1.6057, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.756301828765995e-05, | |
| "loss": 1.5918, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7528694601570654e-05, | |
| "loss": 1.5925, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7494370915481356e-05, | |
| "loss": 1.6056, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7460047229392065e-05, | |
| "loss": 1.5938, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.742572354330276e-05, | |
| "loss": 1.5799, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.739139985721347e-05, | |
| "loss": 1.5874, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.735707617112417e-05, | |
| "loss": 1.5895, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.732275248503488e-05, | |
| "loss": 1.5943, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7288428798945575e-05, | |
| "loss": 1.5695, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.7254105112856284e-05, | |
| "loss": 1.5784, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.7219781426766986e-05, | |
| "loss": 1.5821, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.718545774067769e-05, | |
| "loss": 1.5812, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.715113405458839e-05, | |
| "loss": 1.5745, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.71168103684991e-05, | |
| "loss": 1.5725, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.70824866824098e-05, | |
| "loss": 1.5708, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.70481629963205e-05, | |
| "loss": 1.5654, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.7013839310231204e-05, | |
| "loss": 1.5728, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.697951562414191e-05, | |
| "loss": 1.5469, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.6945191938052615e-05, | |
| "loss": 1.5794, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.691086825196332e-05, | |
| "loss": 1.5555, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.687654456587402e-05, | |
| "loss": 1.5621, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.684222087978473e-05, | |
| "loss": 1.575, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.680789719369543e-05, | |
| "loss": 1.547, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.677357350760613e-05, | |
| "loss": 1.538, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.6739249821516834e-05, | |
| "loss": 1.5478, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.670492613542754e-05, | |
| "loss": 1.5501, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.667060244933824e-05, | |
| "loss": 1.545, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.663627876324895e-05, | |
| "loss": 1.5574, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.660195507715965e-05, | |
| "loss": 1.5398, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.656763139107036e-05, | |
| "loss": 1.5406, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.653330770498105e-05, | |
| "loss": 1.5354, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.649898401889176e-05, | |
| "loss": 1.5448, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.6464660332802463e-05, | |
| "loss": 1.5261, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.6430336646713165e-05, | |
| "loss": 1.5385, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.639601296062387e-05, | |
| "loss": 1.5363, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.6361689274534576e-05, | |
| "loss": 1.5315, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.632736558844528e-05, | |
| "loss": 1.5351, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.629304190235598e-05, | |
| "loss": 1.5171, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.625871821626668e-05, | |
| "loss": 1.5275, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.622439453017739e-05, | |
| "loss": 1.5402, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.619007084408809e-05, | |
| "loss": 1.5287, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.6155747157998795e-05, | |
| "loss": 1.5232, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.61214234719095e-05, | |
| "loss": 1.538, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.6087099785820206e-05, | |
| "loss": 1.522, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.605277609973091e-05, | |
| "loss": 1.5132, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.601845241364161e-05, | |
| "loss": 1.5273, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.598412872755231e-05, | |
| "loss": 1.509, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.594980504146302e-05, | |
| "loss": 1.5257, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.5915481355373716e-05, | |
| "loss": 1.5165, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.5881157669284424e-05, | |
| "loss": 1.5089, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.5846833983195126e-05, | |
| "loss": 1.5302, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.581251029710583e-05, | |
| "loss": 1.5085, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.577818661101653e-05, | |
| "loss": 1.5038, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.574386292492723e-05, | |
| "loss": 1.5175, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.570953923883794e-05, | |
| "loss": 1.5259, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.567521555274864e-05, | |
| "loss": 1.4952, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.5640891866659345e-05, | |
| "loss": 1.4924, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.560656818057005e-05, | |
| "loss": 1.4995, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.5572244494480756e-05, | |
| "loss": 1.4919, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.553792080839145e-05, | |
| "loss": 1.5147, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.550359712230216e-05, | |
| "loss": 1.4928, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.546927343621286e-05, | |
| "loss": 1.4959, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.543494975012357e-05, | |
| "loss": 1.4969, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.5400626064034266e-05, | |
| "loss": 1.5043, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.5366302377944975e-05, | |
| "loss": 1.4929, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.533197869185568e-05, | |
| "loss": 1.4929, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.529765500576638e-05, | |
| "loss": 1.4884, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.526333131967708e-05, | |
| "loss": 1.486, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.522900763358779e-05, | |
| "loss": 1.4958, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.519468394749849e-05, | |
| "loss": 1.4976, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.516036026140919e-05, | |
| "loss": 1.4888, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.5126036575319895e-05, | |
| "loss": 1.4877, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.5091712889230604e-05, | |
| "loss": 1.4691, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.5057389203141306e-05, | |
| "loss": 1.4788, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.502306551705201e-05, | |
| "loss": 1.4976, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.498874183096271e-05, | |
| "loss": 1.499, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.495441814487342e-05, | |
| "loss": 1.4825, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.492009445878412e-05, | |
| "loss": 1.4849, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.488577077269482e-05, | |
| "loss": 1.4852, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.4851447086605525e-05, | |
| "loss": 1.488, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.4817123400516234e-05, | |
| "loss": 1.4779, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.478279971442693e-05, | |
| "loss": 1.4731, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.474847602833764e-05, | |
| "loss": 1.4766, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.471415234224834e-05, | |
| "loss": 1.498, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.467982865615905e-05, | |
| "loss": 1.4696, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.4645504970069744e-05, | |
| "loss": 1.4849, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.461118128398045e-05, | |
| "loss": 1.4802, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.4576857597891154e-05, | |
| "loss": 1.4857, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.4542533911801856e-05, | |
| "loss": 1.4691, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.450821022571256e-05, | |
| "loss": 1.4791, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.447388653962327e-05, | |
| "loss": 1.4448, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.443956285353397e-05, | |
| "loss": 1.4589, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.440523916744467e-05, | |
| "loss": 1.4681, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.437091548135537e-05, | |
| "loss": 1.474, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.433659179526608e-05, | |
| "loss": 1.4692, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.4302268109176784e-05, | |
| "loss": 1.4631, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.4267944423087486e-05, | |
| "loss": 1.4745, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.423362073699819e-05, | |
| "loss": 1.4657, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.4199297050908897e-05, | |
| "loss": 1.4642, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.416497336481959e-05, | |
| "loss": 1.4688, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.41306496787303e-05, | |
| "loss": 1.4422, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.4096325992641e-05, | |
| "loss": 1.4623, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.406200230655171e-05, | |
| "loss": 1.4744, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.4027678620462407e-05, | |
| "loss": 1.4449, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.3993354934373115e-05, | |
| "loss": 1.4662, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.395903124828382e-05, | |
| "loss": 1.4489, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.392470756219452e-05, | |
| "loss": 1.4465, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.389038387610522e-05, | |
| "loss": 1.4508, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.385606019001593e-05, | |
| "loss": 1.4566, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.382173650392663e-05, | |
| "loss": 1.4495, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.3787412817837334e-05, | |
| "loss": 1.4547, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.3753089131748036e-05, | |
| "loss": 1.4525, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.3718765445658745e-05, | |
| "loss": 1.4435, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.368444175956945e-05, | |
| "loss": 1.4497, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.365011807348015e-05, | |
| "loss": 1.4472, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.361579438739085e-05, | |
| "loss": 1.4556, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.358147070130156e-05, | |
| "loss": 1.4483, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.354714701521226e-05, | |
| "loss": 1.4345, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.3512823329122964e-05, | |
| "loss": 1.4402, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.3478499643033666e-05, | |
| "loss": 1.4376, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.3444175956944374e-05, | |
| "loss": 1.4474, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.340985227085507e-05, | |
| "loss": 1.439, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.337552858476578e-05, | |
| "loss": 1.4424, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.334120489867648e-05, | |
| "loss": 1.4493, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.330688121258719e-05, | |
| "loss": 1.4319, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.3272557526497884e-05, | |
| "loss": 1.4312, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.323823384040859e-05, | |
| "loss": 1.4547, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.3203910154319295e-05, | |
| "loss": 1.4464, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.316958646823e-05, | |
| "loss": 1.4292, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.31352627821407e-05, | |
| "loss": 1.4199, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.310093909605141e-05, | |
| "loss": 1.4278, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.306661540996211e-05, | |
| "loss": 1.4218, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.303229172387281e-05, | |
| "loss": 1.4267, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.2997968037783514e-05, | |
| "loss": 1.4229, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.296364435169422e-05, | |
| "loss": 1.4245, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.2929320665604925e-05, | |
| "loss": 1.4369, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.2894996979515627e-05, | |
| "loss": 1.4236, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.286067329342633e-05, | |
| "loss": 1.4441, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.282634960733704e-05, | |
| "loss": 1.43, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.279202592124774e-05, | |
| "loss": 1.427, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.275770223515844e-05, | |
| "loss": 1.4124, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.272337854906914e-05, | |
| "loss": 1.4155, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.268905486297985e-05, | |
| "loss": 1.4292, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.265473117689055e-05, | |
| "loss": 1.4286, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.2620407490801256e-05, | |
| "loss": 1.4254, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.258608380471196e-05, | |
| "loss": 1.4151, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.255176011862266e-05, | |
| "loss": 1.4189, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.251743643253336e-05, | |
| "loss": 1.427, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.248311274644407e-05, | |
| "loss": 1.4256, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.244878906035477e-05, | |
| "loss": 1.4059, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.2414465374265475e-05, | |
| "loss": 1.4165, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.238014168817618e-05, | |
| "loss": 1.4177, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.2345818002086885e-05, | |
| "loss": 1.4236, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.231149431599759e-05, | |
| "loss": 1.4325, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.227717062990829e-05, | |
| "loss": 1.4078, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.224284694381899e-05, | |
| "loss": 1.4134, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.22085232577297e-05, | |
| "loss": 1.4013, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.21741995716404e-05, | |
| "loss": 1.4128, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.2139875885551104e-05, | |
| "loss": 1.4254, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.2105552199461806e-05, | |
| "loss": 1.4164, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.2071228513372515e-05, | |
| "loss": 1.422, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.203690482728321e-05, | |
| "loss": 1.4296, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.200258114119392e-05, | |
| "loss": 1.4151, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.196825745510462e-05, | |
| "loss": 1.3993, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.193393376901533e-05, | |
| "loss": 1.4046, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.1899610082926025e-05, | |
| "loss": 1.4134, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.1865286396836734e-05, | |
| "loss": 1.3971, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.1830962710747436e-05, | |
| "loss": 1.4095, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.179663902465814e-05, | |
| "loss": 1.3892, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.176231533856884e-05, | |
| "loss": 1.3881, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.172799165247955e-05, | |
| "loss": 1.4076, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.169366796639025e-05, | |
| "loss": 1.3896, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.165934428030095e-05, | |
| "loss": 1.4049, | |
| "step": 121500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1625020594211654e-05, | |
| "loss": 1.3841, | |
| "step": 122000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1590696908122356e-05, | |
| "loss": 1.4008, | |
| "step": 122500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1556373222033065e-05, | |
| "loss": 1.3976, | |
| "step": 123000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.152204953594376e-05, | |
| "loss": 1.3941, | |
| "step": 123500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.148772584985447e-05, | |
| "loss": 1.3983, | |
| "step": 124000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.145340216376517e-05, | |
| "loss": 1.3901, | |
| "step": 124500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.141907847767588e-05, | |
| "loss": 1.4039, | |
| "step": 125000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1384754791586575e-05, | |
| "loss": 1.3999, | |
| "step": 125500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1350431105497284e-05, | |
| "loss": 1.3893, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.1316107419407986e-05, | |
| "loss": 1.3875, | |
| "step": 126500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.128178373331869e-05, | |
| "loss": 1.4075, | |
| "step": 127000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.124746004722939e-05, | |
| "loss": 1.4106, | |
| "step": 127500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.12131363611401e-05, | |
| "loss": 1.3838, | |
| "step": 128000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.11788126750508e-05, | |
| "loss": 1.3982, | |
| "step": 128500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.11444889889615e-05, | |
| "loss": 1.4009, | |
| "step": 129000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.1110165302872205e-05, | |
| "loss": 1.3753, | |
| "step": 129500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.1075841616782913e-05, | |
| "loss": 1.379, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.1041517930693615e-05, | |
| "loss": 1.3839, | |
| "step": 130500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.100719424460432e-05, | |
| "loss": 1.3887, | |
| "step": 131000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.097287055851502e-05, | |
| "loss": 1.377, | |
| "step": 131500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.093854687242573e-05, | |
| "loss": 1.3908, | |
| "step": 132000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.090422318633642e-05, | |
| "loss": 1.394, | |
| "step": 132500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.086989950024713e-05, | |
| "loss": 1.4083, | |
| "step": 133000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.0835575814157834e-05, | |
| "loss": 1.3718, | |
| "step": 133500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.080125212806854e-05, | |
| "loss": 1.3849, | |
| "step": 134000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.076692844197924e-05, | |
| "loss": 1.3941, | |
| "step": 134500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.073260475588995e-05, | |
| "loss": 1.3769, | |
| "step": 135000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.069828106980065e-05, | |
| "loss": 1.3808, | |
| "step": 135500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.066395738371135e-05, | |
| "loss": 1.3851, | |
| "step": 136000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.062963369762205e-05, | |
| "loss": 1.3785, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.059531001153276e-05, | |
| "loss": 1.3727, | |
| "step": 137000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.0560986325443464e-05, | |
| "loss": 1.3997, | |
| "step": 137500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.0526662639354166e-05, | |
| "loss": 1.3763, | |
| "step": 138000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.049233895326487e-05, | |
| "loss": 1.3935, | |
| "step": 138500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.0458015267175576e-05, | |
| "loss": 1.3733, | |
| "step": 139000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.042369158108628e-05, | |
| "loss": 1.3802, | |
| "step": 139500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.038936789499698e-05, | |
| "loss": 1.369, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.035504420890768e-05, | |
| "loss": 1.3751, | |
| "step": 140500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.032072052281839e-05, | |
| "loss": 1.3664, | |
| "step": 141000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.028639683672909e-05, | |
| "loss": 1.3892, | |
| "step": 141500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.0252073150639795e-05, | |
| "loss": 1.3661, | |
| "step": 142000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.02177494645505e-05, | |
| "loss": 1.3708, | |
| "step": 142500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.0183425778461206e-05, | |
| "loss": 1.3713, | |
| "step": 143000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.01491020923719e-05, | |
| "loss": 1.3754, | |
| "step": 143500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.011477840628261e-05, | |
| "loss": 1.3628, | |
| "step": 144000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.008045472019331e-05, | |
| "loss": 1.372, | |
| "step": 144500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.004613103410402e-05, | |
| "loss": 1.3484, | |
| "step": 145000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.0011807348014716e-05, | |
| "loss": 1.3768, | |
| "step": 145500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.9977483661925425e-05, | |
| "loss": 1.3821, | |
| "step": 146000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.9943159975836127e-05, | |
| "loss": 1.3691, | |
| "step": 146500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.990883628974683e-05, | |
| "loss": 1.3641, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.987451260365753e-05, | |
| "loss": 1.368, | |
| "step": 147500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.984018891756824e-05, | |
| "loss": 1.3663, | |
| "step": 148000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.980586523147894e-05, | |
| "loss": 1.3698, | |
| "step": 148500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 3.977154154538964e-05, | |
| "loss": 1.3784, | |
| "step": 149000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9737217859300345e-05, | |
| "loss": 1.3594, | |
| "step": 149500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9702894173211054e-05, | |
| "loss": 1.3599, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9668570487121756e-05, | |
| "loss": 1.38, | |
| "step": 150500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.963424680103246e-05, | |
| "loss": 1.3421, | |
| "step": 151000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.959992311494316e-05, | |
| "loss": 1.3661, | |
| "step": 151500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.956559942885387e-05, | |
| "loss": 1.3608, | |
| "step": 152000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.953127574276457e-05, | |
| "loss": 1.368, | |
| "step": 152500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.949695205667527e-05, | |
| "loss": 1.3556, | |
| "step": 153000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9462628370585975e-05, | |
| "loss": 1.3744, | |
| "step": 153500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9428304684496684e-05, | |
| "loss": 1.359, | |
| "step": 154000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.939398099840738e-05, | |
| "loss": 1.3587, | |
| "step": 154500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.935965731231809e-05, | |
| "loss": 1.3647, | |
| "step": 155000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.932533362622879e-05, | |
| "loss": 1.3677, | |
| "step": 155500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.929100994013949e-05, | |
| "loss": 1.3602, | |
| "step": 156000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 3.9256686254050194e-05, | |
| "loss": 1.3743, | |
| "step": 156500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.92223625679609e-05, | |
| "loss": 1.3637, | |
| "step": 157000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.9188038881871604e-05, | |
| "loss": 1.3753, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.9153715195782306e-05, | |
| "loss": 1.3624, | |
| "step": 158000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.911939150969301e-05, | |
| "loss": 1.3525, | |
| "step": 158500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.908506782360372e-05, | |
| "loss": 1.3484, | |
| "step": 159000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.905074413751442e-05, | |
| "loss": 1.3584, | |
| "step": 159500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.901642045142512e-05, | |
| "loss": 1.3617, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.898209676533582e-05, | |
| "loss": 1.3553, | |
| "step": 160500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.894777307924653e-05, | |
| "loss": 1.3524, | |
| "step": 161000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.8913449393157234e-05, | |
| "loss": 1.3708, | |
| "step": 161500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.8879125707067936e-05, | |
| "loss": 1.3372, | |
| "step": 162000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.884480202097864e-05, | |
| "loss": 1.3572, | |
| "step": 162500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.8810478334889347e-05, | |
| "loss": 1.3373, | |
| "step": 163000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 3.877615464880004e-05, | |
| "loss": 1.3364, | |
| "step": 163500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.874183096271075e-05, | |
| "loss": 1.3539, | |
| "step": 164000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.870750727662145e-05, | |
| "loss": 1.3447, | |
| "step": 164500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.867318359053216e-05, | |
| "loss": 1.3547, | |
| "step": 165000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.8638859904442857e-05, | |
| "loss": 1.3542, | |
| "step": 165500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.8604536218353565e-05, | |
| "loss": 1.3433, | |
| "step": 166000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.857021253226427e-05, | |
| "loss": 1.3443, | |
| "step": 166500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.853588884617497e-05, | |
| "loss": 1.3488, | |
| "step": 167000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.850156516008567e-05, | |
| "loss": 1.3495, | |
| "step": 167500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.846724147399638e-05, | |
| "loss": 1.3535, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.843291778790708e-05, | |
| "loss": 1.3285, | |
| "step": 168500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.8398594101817784e-05, | |
| "loss": 1.3522, | |
| "step": 169000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.8364270415728486e-05, | |
| "loss": 1.3613, | |
| "step": 169500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.8329946729639195e-05, | |
| "loss": 1.3612, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.82956230435499e-05, | |
| "loss": 1.339, | |
| "step": 170500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 3.82612993574606e-05, | |
| "loss": 1.3508, | |
| "step": 171000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.82269756713713e-05, | |
| "loss": 1.3573, | |
| "step": 171500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.819265198528201e-05, | |
| "loss": 1.3444, | |
| "step": 172000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.815832829919271e-05, | |
| "loss": 1.3544, | |
| "step": 172500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.8124004613103413e-05, | |
| "loss": 1.3395, | |
| "step": 173000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.8089680927014115e-05, | |
| "loss": 1.363, | |
| "step": 173500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.8055357240924824e-05, | |
| "loss": 1.3448, | |
| "step": 174000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.802103355483552e-05, | |
| "loss": 1.339, | |
| "step": 174500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.798670986874623e-05, | |
| "loss": 1.3348, | |
| "step": 175000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.795238618265693e-05, | |
| "loss": 1.3315, | |
| "step": 175500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.791806249656764e-05, | |
| "loss": 1.3393, | |
| "step": 176000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.7883738810478334e-05, | |
| "loss": 1.3353, | |
| "step": 176500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.784941512438904e-05, | |
| "loss": 1.3398, | |
| "step": 177000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.7815091438299745e-05, | |
| "loss": 1.3396, | |
| "step": 177500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 3.778076775221045e-05, | |
| "loss": 1.3453, | |
| "step": 178000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.774644406612115e-05, | |
| "loss": 1.3311, | |
| "step": 178500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.771212038003186e-05, | |
| "loss": 1.3492, | |
| "step": 179000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.767779669394256e-05, | |
| "loss": 1.32, | |
| "step": 179500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.764347300785326e-05, | |
| "loss": 1.3049, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7609149321763964e-05, | |
| "loss": 1.3506, | |
| "step": 180500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.757482563567467e-05, | |
| "loss": 1.3235, | |
| "step": 181000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7540501949585374e-05, | |
| "loss": 1.3252, | |
| "step": 181500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7506178263496076e-05, | |
| "loss": 1.3228, | |
| "step": 182000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.747185457740678e-05, | |
| "loss": 1.3312, | |
| "step": 182500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.743753089131749e-05, | |
| "loss": 1.343, | |
| "step": 183000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.740320720522818e-05, | |
| "loss": 1.3254, | |
| "step": 183500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7368883519138884e-05, | |
| "loss": 1.3474, | |
| "step": 184000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.733455983304959e-05, | |
| "loss": 1.3312, | |
| "step": 184500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7300236146960295e-05, | |
| "loss": 1.3366, | |
| "step": 185000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 3.7265912460871e-05, | |
| "loss": 1.339, | |
| "step": 185500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.72315887747817e-05, | |
| "loss": 1.3227, | |
| "step": 186000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.719726508869241e-05, | |
| "loss": 1.3249, | |
| "step": 186500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.716294140260311e-05, | |
| "loss": 1.3215, | |
| "step": 187000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.712861771651381e-05, | |
| "loss": 1.3249, | |
| "step": 187500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.7094294030424514e-05, | |
| "loss": 1.3297, | |
| "step": 188000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.705997034433522e-05, | |
| "loss": 1.3154, | |
| "step": 188500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.7025646658245925e-05, | |
| "loss": 1.3291, | |
| "step": 189000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.699132297215663e-05, | |
| "loss": 1.3347, | |
| "step": 189500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.695699928606733e-05, | |
| "loss": 1.3247, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.692267559997804e-05, | |
| "loss": 1.3304, | |
| "step": 190500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.688835191388873e-05, | |
| "loss": 1.3286, | |
| "step": 191000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.685402822779944e-05, | |
| "loss": 1.3131, | |
| "step": 191500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.6819704541710143e-05, | |
| "loss": 1.3191, | |
| "step": 192000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.678538085562085e-05, | |
| "loss": 1.3129, | |
| "step": 192500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 3.675105716953155e-05, | |
| "loss": 1.3022, | |
| "step": 193000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.6716733483442256e-05, | |
| "loss": 1.332, | |
| "step": 193500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.668240979735296e-05, | |
| "loss": 1.3169, | |
| "step": 194000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.664808611126366e-05, | |
| "loss": 1.3113, | |
| "step": 194500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.661376242517436e-05, | |
| "loss": 1.325, | |
| "step": 195000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.657943873908507e-05, | |
| "loss": 1.3363, | |
| "step": 195500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.654511505299577e-05, | |
| "loss": 1.3304, | |
| "step": 196000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.6510791366906475e-05, | |
| "loss": 1.3147, | |
| "step": 196500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.647646768081718e-05, | |
| "loss": 1.3202, | |
| "step": 197000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.6442143994727886e-05, | |
| "loss": 1.3182, | |
| "step": 197500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.640782030863859e-05, | |
| "loss": 1.3251, | |
| "step": 198000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.637349662254929e-05, | |
| "loss": 1.3317, | |
| "step": 198500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.633917293645999e-05, | |
| "loss": 1.3264, | |
| "step": 199000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.63048492503707e-05, | |
| "loss": 1.3281, | |
| "step": 199500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.62705255642814e-05, | |
| "loss": 1.309, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.6236201878192104e-05, | |
| "loss": 1.323, | |
| "step": 200500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.6201878192102806e-05, | |
| "loss": 1.315, | |
| "step": 201000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.6167554506013515e-05, | |
| "loss": 1.3242, | |
| "step": 201500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.613323081992421e-05, | |
| "loss": 1.3092, | |
| "step": 202000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.609890713383492e-05, | |
| "loss": 1.3044, | |
| "step": 202500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.606458344774562e-05, | |
| "loss": 1.3132, | |
| "step": 203000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.603025976165632e-05, | |
| "loss": 1.3191, | |
| "step": 203500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.5995936075567025e-05, | |
| "loss": 1.3096, | |
| "step": 204000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.5961612389477734e-05, | |
| "loss": 1.2949, | |
| "step": 204500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.5927288703388436e-05, | |
| "loss": 1.3083, | |
| "step": 205000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.589296501729914e-05, | |
| "loss": 1.2998, | |
| "step": 205500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.585864133120984e-05, | |
| "loss": 1.3158, | |
| "step": 206000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.582431764512055e-05, | |
| "loss": 1.2995, | |
| "step": 206500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.578999395903125e-05, | |
| "loss": 1.3204, | |
| "step": 207000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.575567027294195e-05, | |
| "loss": 1.3131, | |
| "step": 207500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.5721346586852655e-05, | |
| "loss": 1.2995, | |
| "step": 208000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.568702290076336e-05, | |
| "loss": 1.3058, | |
| "step": 208500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.5652699214674065e-05, | |
| "loss": 1.3073, | |
| "step": 209000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.561837552858477e-05, | |
| "loss": 1.3096, | |
| "step": 209500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.558405184249547e-05, | |
| "loss": 1.3014, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.554972815640618e-05, | |
| "loss": 1.3017, | |
| "step": 210500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.551540447031687e-05, | |
| "loss": 1.2992, | |
| "step": 211000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.548108078422758e-05, | |
| "loss": 1.3256, | |
| "step": 211500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.5446757098138284e-05, | |
| "loss": 1.3012, | |
| "step": 212000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.541243341204899e-05, | |
| "loss": 1.2812, | |
| "step": 212500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.537810972595969e-05, | |
| "loss": 1.3136, | |
| "step": 213000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.53437860398704e-05, | |
| "loss": 1.3052, | |
| "step": 213500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.53094623537811e-05, | |
| "loss": 1.2998, | |
| "step": 214000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.52751386676918e-05, | |
| "loss": 1.2986, | |
| "step": 214500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.52408149816025e-05, | |
| "loss": 1.3083, | |
| "step": 215000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.520649129551321e-05, | |
| "loss": 1.3027, | |
| "step": 215500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.5172167609423914e-05, | |
| "loss": 1.2866, | |
| "step": 216000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.5137843923334616e-05, | |
| "loss": 1.2938, | |
| "step": 216500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.510352023724532e-05, | |
| "loss": 1.3157, | |
| "step": 217000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.5069196551156026e-05, | |
| "loss": 1.286, | |
| "step": 217500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.503487286506673e-05, | |
| "loss": 1.3145, | |
| "step": 218000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.500054917897743e-05, | |
| "loss": 1.2871, | |
| "step": 218500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.496622549288813e-05, | |
| "loss": 1.2767, | |
| "step": 219000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.493190180679884e-05, | |
| "loss": 1.3071, | |
| "step": 219500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.489757812070954e-05, | |
| "loss": 1.2969, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.4863254434620245e-05, | |
| "loss": 1.3044, | |
| "step": 220500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.482893074853095e-05, | |
| "loss": 1.3107, | |
| "step": 221000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.4794607062441656e-05, | |
| "loss": 1.294, | |
| "step": 221500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.476028337635235e-05, | |
| "loss": 1.2912, | |
| "step": 222000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.472595969026306e-05, | |
| "loss": 1.3051, | |
| "step": 222500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.469163600417376e-05, | |
| "loss": 1.3039, | |
| "step": 223000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.465731231808447e-05, | |
| "loss": 1.2867, | |
| "step": 223500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4622988631995166e-05, | |
| "loss": 1.286, | |
| "step": 224000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4588664945905875e-05, | |
| "loss": 1.3003, | |
| "step": 224500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4554341259816577e-05, | |
| "loss": 1.3107, | |
| "step": 225000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.452001757372728e-05, | |
| "loss": 1.2988, | |
| "step": 225500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.448569388763798e-05, | |
| "loss": 1.2908, | |
| "step": 226000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.445137020154869e-05, | |
| "loss": 1.302, | |
| "step": 226500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.441704651545939e-05, | |
| "loss": 1.2844, | |
| "step": 227000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.438272282937009e-05, | |
| "loss": 1.2787, | |
| "step": 227500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4348399143280795e-05, | |
| "loss": 1.3059, | |
| "step": 228000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4314075457191504e-05, | |
| "loss": 1.3013, | |
| "step": 228500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4279751771102206e-05, | |
| "loss": 1.3059, | |
| "step": 229000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.424542808501291e-05, | |
| "loss": 1.2924, | |
| "step": 229500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.421110439892361e-05, | |
| "loss": 1.2929, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.417678071283432e-05, | |
| "loss": 1.2897, | |
| "step": 230500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.4142457026745014e-05, | |
| "loss": 1.3013, | |
| "step": 231000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.410813334065572e-05, | |
| "loss": 1.2833, | |
| "step": 231500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.4073809654566425e-05, | |
| "loss": 1.272, | |
| "step": 232000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.4039485968477134e-05, | |
| "loss": 1.2829, | |
| "step": 232500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.400516228238783e-05, | |
| "loss": 1.2853, | |
| "step": 233000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.397083859629854e-05, | |
| "loss": 1.2925, | |
| "step": 233500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.393651491020924e-05, | |
| "loss": 1.3131, | |
| "step": 234000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.390219122411994e-05, | |
| "loss": 1.2929, | |
| "step": 234500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.3867867538030643e-05, | |
| "loss": 1.2761, | |
| "step": 235000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.383354385194135e-05, | |
| "loss": 1.2832, | |
| "step": 235500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.3799220165852054e-05, | |
| "loss": 1.2793, | |
| "step": 236000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.3764896479762756e-05, | |
| "loss": 1.282, | |
| "step": 236500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.373057279367346e-05, | |
| "loss": 1.2848, | |
| "step": 237000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.369624910758417e-05, | |
| "loss": 1.3026, | |
| "step": 237500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.366192542149487e-05, | |
| "loss": 1.2793, | |
| "step": 238000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.362760173540557e-05, | |
| "loss": 1.2878, | |
| "step": 238500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.359327804931627e-05, | |
| "loss": 1.2957, | |
| "step": 239000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.355895436322698e-05, | |
| "loss": 1.2903, | |
| "step": 239500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.3524630677137684e-05, | |
| "loss": 1.2715, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.3490306991048386e-05, | |
| "loss": 1.2904, | |
| "step": 240500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.345598330495909e-05, | |
| "loss": 1.2847, | |
| "step": 241000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.3421659618869796e-05, | |
| "loss": 1.2684, | |
| "step": 241500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.338733593278049e-05, | |
| "loss": 1.2786, | |
| "step": 242000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.33530122466912e-05, | |
| "loss": 1.2937, | |
| "step": 242500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.33186885606019e-05, | |
| "loss": 1.2848, | |
| "step": 243000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.328436487451261e-05, | |
| "loss": 1.2762, | |
| "step": 243500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.3250041188423306e-05, | |
| "loss": 1.288, | |
| "step": 244000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.321571750233401e-05, | |
| "loss": 1.2726, | |
| "step": 244500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.318139381624472e-05, | |
| "loss": 1.2991, | |
| "step": 245000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.314707013015542e-05, | |
| "loss": 1.2633, | |
| "step": 245500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.311274644406612e-05, | |
| "loss": 1.2817, | |
| "step": 246000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.307842275797682e-05, | |
| "loss": 1.2862, | |
| "step": 246500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.304409907188753e-05, | |
| "loss": 1.267, | |
| "step": 247000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.3009775385798234e-05, | |
| "loss": 1.2825, | |
| "step": 247500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.2975451699708936e-05, | |
| "loss": 1.2701, | |
| "step": 248000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.294112801361964e-05, | |
| "loss": 1.2977, | |
| "step": 248500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.290680432753035e-05, | |
| "loss": 1.2721, | |
| "step": 249000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.287248064144104e-05, | |
| "loss": 1.274, | |
| "step": 249500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.283815695535175e-05, | |
| "loss": 1.2937, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.280383326926245e-05, | |
| "loss": 1.2758, | |
| "step": 250500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.2769509583173155e-05, | |
| "loss": 1.2683, | |
| "step": 251000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.273518589708386e-05, | |
| "loss": 1.2862, | |
| "step": 251500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.2700862210994565e-05, | |
| "loss": 1.2713, | |
| "step": 252000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.266653852490527e-05, | |
| "loss": 1.2652, | |
| "step": 252500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.263221483881597e-05, | |
| "loss": 1.2665, | |
| "step": 253000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.259789115272667e-05, | |
| "loss": 1.2582, | |
| "step": 253500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.256356746663738e-05, | |
| "loss": 1.2558, | |
| "step": 254000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.252924378054808e-05, | |
| "loss": 1.2535, | |
| "step": 254500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.2494920094458784e-05, | |
| "loss": 1.29, | |
| "step": 255000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.2460596408369486e-05, | |
| "loss": 1.2836, | |
| "step": 255500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.2426272722280195e-05, | |
| "loss": 1.2664, | |
| "step": 256000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.23919490361909e-05, | |
| "loss": 1.2936, | |
| "step": 256500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.23576253501016e-05, | |
| "loss": 1.2826, | |
| "step": 257000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.23233016640123e-05, | |
| "loss": 1.2667, | |
| "step": 257500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.228897797792301e-05, | |
| "loss": 1.2803, | |
| "step": 258000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.2254654291833705e-05, | |
| "loss": 1.2748, | |
| "step": 258500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.2220330605744414e-05, | |
| "loss": 1.269, | |
| "step": 259000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.2186006919655116e-05, | |
| "loss": 1.2711, | |
| "step": 259500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.2151683233565824e-05, | |
| "loss": 1.2766, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.211735954747652e-05, | |
| "loss": 1.2703, | |
| "step": 260500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.208303586138723e-05, | |
| "loss": 1.267, | |
| "step": 261000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.204871217529793e-05, | |
| "loss": 1.2844, | |
| "step": 261500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.201438848920863e-05, | |
| "loss": 1.266, | |
| "step": 262000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.1980064803119334e-05, | |
| "loss": 1.2532, | |
| "step": 262500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.194574111703004e-05, | |
| "loss": 1.2806, | |
| "step": 263000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.1911417430940745e-05, | |
| "loss": 1.2638, | |
| "step": 263500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.187709374485145e-05, | |
| "loss": 1.2662, | |
| "step": 264000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.184277005876215e-05, | |
| "loss": 1.2638, | |
| "step": 264500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.180844637267286e-05, | |
| "loss": 1.2673, | |
| "step": 265000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.177412268658356e-05, | |
| "loss": 1.2537, | |
| "step": 265500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.173979900049426e-05, | |
| "loss": 1.258, | |
| "step": 266000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.1705475314404964e-05, | |
| "loss": 1.2689, | |
| "step": 266500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.167115162831567e-05, | |
| "loss": 1.2554, | |
| "step": 267000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.1636827942226375e-05, | |
| "loss": 1.2667, | |
| "step": 267500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.160250425613708e-05, | |
| "loss": 1.276, | |
| "step": 268000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.156818057004778e-05, | |
| "loss": 1.2716, | |
| "step": 268500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.153385688395849e-05, | |
| "loss": 1.265, | |
| "step": 269000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.149953319786918e-05, | |
| "loss": 1.2742, | |
| "step": 269500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.146520951177989e-05, | |
| "loss": 1.2617, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.143088582569059e-05, | |
| "loss": 1.2666, | |
| "step": 270500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.13965621396013e-05, | |
| "loss": 1.2465, | |
| "step": 271000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.1362238453512e-05, | |
| "loss": 1.2708, | |
| "step": 271500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.1327914767422706e-05, | |
| "loss": 1.2571, | |
| "step": 272000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.129359108133341e-05, | |
| "loss": 1.2547, | |
| "step": 272500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 3.125926739524411e-05, | |
| "loss": 1.2629, | |
| "step": 273000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.122494370915481e-05, | |
| "loss": 1.2648, | |
| "step": 273500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.119062002306552e-05, | |
| "loss": 1.2543, | |
| "step": 274000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.115629633697622e-05, | |
| "loss": 1.2653, | |
| "step": 274500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.1121972650886925e-05, | |
| "loss": 1.2709, | |
| "step": 275000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.108764896479763e-05, | |
| "loss": 1.2755, | |
| "step": 275500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.1053325278708336e-05, | |
| "loss": 1.2653, | |
| "step": 276000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.101900159261904e-05, | |
| "loss": 1.249, | |
| "step": 276500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.098467790652974e-05, | |
| "loss": 1.2593, | |
| "step": 277000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.095035422044044e-05, | |
| "loss": 1.2533, | |
| "step": 277500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.091603053435115e-05, | |
| "loss": 1.2724, | |
| "step": 278000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.0881706848261846e-05, | |
| "loss": 1.2583, | |
| "step": 278500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.0847383162172554e-05, | |
| "loss": 1.2428, | |
| "step": 279000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.0813059476083256e-05, | |
| "loss": 1.2461, | |
| "step": 279500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 3.0778735789993965e-05, | |
| "loss": 1.2668, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.074441210390466e-05, | |
| "loss": 1.2578, | |
| "step": 280500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.071008841781537e-05, | |
| "loss": 1.2573, | |
| "step": 281000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.067576473172607e-05, | |
| "loss": 1.2494, | |
| "step": 281500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.064144104563677e-05, | |
| "loss": 1.2602, | |
| "step": 282000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0607117359547475e-05, | |
| "loss": 1.2584, | |
| "step": 282500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0572793673458184e-05, | |
| "loss": 1.261, | |
| "step": 283000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0538469987368886e-05, | |
| "loss": 1.2466, | |
| "step": 283500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.050414630127959e-05, | |
| "loss": 1.2482, | |
| "step": 284000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.046982261519029e-05, | |
| "loss": 1.2479, | |
| "step": 284500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0435498929101e-05, | |
| "loss": 1.2509, | |
| "step": 285000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0401175243011697e-05, | |
| "loss": 1.2528, | |
| "step": 285500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0366851556922406e-05, | |
| "loss": 1.2507, | |
| "step": 286000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0332527870833105e-05, | |
| "loss": 1.25, | |
| "step": 286500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.029820418474381e-05, | |
| "loss": 1.233, | |
| "step": 287000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 3.0263880498654512e-05, | |
| "loss": 1.2248, | |
| "step": 287500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.0229556812565217e-05, | |
| "loss": 1.2268, | |
| "step": 288000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.019523312647592e-05, | |
| "loss": 1.253, | |
| "step": 288500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.0160909440386625e-05, | |
| "loss": 1.235, | |
| "step": 289000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.0126585754297327e-05, | |
| "loss": 1.2512, | |
| "step": 289500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.0092262068208032e-05, | |
| "loss": 1.2603, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.0057938382118734e-05, | |
| "loss": 1.2679, | |
| "step": 290500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 3.002361469602944e-05, | |
| "loss": 1.2452, | |
| "step": 291000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.998929100994014e-05, | |
| "loss": 1.2507, | |
| "step": 291500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.9954967323850847e-05, | |
| "loss": 1.2647, | |
| "step": 292000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.992064363776155e-05, | |
| "loss": 1.2478, | |
| "step": 292500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.9886319951672254e-05, | |
| "loss": 1.2405, | |
| "step": 293000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.9851996265582953e-05, | |
| "loss": 1.2553, | |
| "step": 293500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.981767257949366e-05, | |
| "loss": 1.2543, | |
| "step": 294000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 2.978334889340436e-05, | |
| "loss": 1.2524, | |
| "step": 294500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.974902520731507e-05, | |
| "loss": 1.2478, | |
| "step": 295000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9714701521225768e-05, | |
| "loss": 1.2429, | |
| "step": 295500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9680377835136476e-05, | |
| "loss": 1.2469, | |
| "step": 296000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9646054149047175e-05, | |
| "loss": 1.2442, | |
| "step": 296500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.961173046295788e-05, | |
| "loss": 1.2496, | |
| "step": 297000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9577406776868582e-05, | |
| "loss": 1.2361, | |
| "step": 297500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9543083090779288e-05, | |
| "loss": 1.2271, | |
| "step": 298000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.950875940468999e-05, | |
| "loss": 1.2513, | |
| "step": 298500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9474435718600695e-05, | |
| "loss": 1.2354, | |
| "step": 299000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9440112032511397e-05, | |
| "loss": 1.2414, | |
| "step": 299500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9405788346422102e-05, | |
| "loss": 1.2556, | |
| "step": 300000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9371464660332804e-05, | |
| "loss": 1.2525, | |
| "step": 300500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.933714097424351e-05, | |
| "loss": 1.2477, | |
| "step": 301000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9302817288154212e-05, | |
| "loss": 1.2258, | |
| "step": 301500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 2.9268493602064917e-05, | |
| "loss": 1.2545, | |
| "step": 302000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.923416991597562e-05, | |
| "loss": 1.2424, | |
| "step": 302500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.9199846229886324e-05, | |
| "loss": 1.2384, | |
| "step": 303000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.9165522543797023e-05, | |
| "loss": 1.2437, | |
| "step": 303500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.9131198857707732e-05, | |
| "loss": 1.2501, | |
| "step": 304000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.909687517161843e-05, | |
| "loss": 1.2394, | |
| "step": 304500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.906255148552914e-05, | |
| "loss": 1.2626, | |
| "step": 305000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.9028227799439838e-05, | |
| "loss": 1.2452, | |
| "step": 305500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.899390411335054e-05, | |
| "loss": 1.255, | |
| "step": 306000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8959580427261245e-05, | |
| "loss": 1.2437, | |
| "step": 306500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8925256741171947e-05, | |
| "loss": 1.2505, | |
| "step": 307000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8890933055082653e-05, | |
| "loss": 1.2403, | |
| "step": 307500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8856609368993355e-05, | |
| "loss": 1.254, | |
| "step": 308000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.882228568290406e-05, | |
| "loss": 1.2384, | |
| "step": 308500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8787961996814762e-05, | |
| "loss": 1.2387, | |
| "step": 309000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 2.8753638310725467e-05, | |
| "loss": 1.2377, | |
| "step": 309500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.871931462463617e-05, | |
| "loss": 1.2268, | |
| "step": 310000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8684990938546875e-05, | |
| "loss": 1.2371, | |
| "step": 310500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8650667252457573e-05, | |
| "loss": 1.2299, | |
| "step": 311000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8616343566368282e-05, | |
| "loss": 1.2268, | |
| "step": 311500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.858201988027898e-05, | |
| "loss": 1.2438, | |
| "step": 312000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.854769619418969e-05, | |
| "loss": 1.2413, | |
| "step": 312500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8513372508100388e-05, | |
| "loss": 1.2239, | |
| "step": 313000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8479048822011097e-05, | |
| "loss": 1.2381, | |
| "step": 313500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8444725135921795e-05, | |
| "loss": 1.2397, | |
| "step": 314000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.84104014498325e-05, | |
| "loss": 1.2396, | |
| "step": 314500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8376077763743203e-05, | |
| "loss": 1.2583, | |
| "step": 315000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8341754077653908e-05, | |
| "loss": 1.2382, | |
| "step": 315500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.830743039156461e-05, | |
| "loss": 1.2313, | |
| "step": 316000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 2.8273106705475316e-05, | |
| "loss": 1.2273, | |
| "step": 316500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.8238783019386018e-05, | |
| "loss": 1.2181, | |
| "step": 317000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.8204459333296723e-05, | |
| "loss": 1.2368, | |
| "step": 317500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.8170135647207425e-05, | |
| "loss": 1.241, | |
| "step": 318000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.813581196111813e-05, | |
| "loss": 1.2326, | |
| "step": 318500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.8101488275028832e-05, | |
| "loss": 1.2294, | |
| "step": 319000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.8067164588939538e-05, | |
| "loss": 1.2289, | |
| "step": 319500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.803284090285024e-05, | |
| "loss": 1.2148, | |
| "step": 320000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.7998517216760945e-05, | |
| "loss": 1.2291, | |
| "step": 320500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.7964193530671644e-05, | |
| "loss": 1.2252, | |
| "step": 321000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.7929869844582352e-05, | |
| "loss": 1.2341, | |
| "step": 321500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.789554615849305e-05, | |
| "loss": 1.2222, | |
| "step": 322000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.786122247240376e-05, | |
| "loss": 1.2417, | |
| "step": 322500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.782689878631446e-05, | |
| "loss": 1.2423, | |
| "step": 323000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.7792575100225167e-05, | |
| "loss": 1.2309, | |
| "step": 323500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 2.7758251414135866e-05, | |
| "loss": 1.241, | |
| "step": 324000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.772392772804657e-05, | |
| "loss": 1.2541, | |
| "step": 324500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7689604041957273e-05, | |
| "loss": 1.2289, | |
| "step": 325000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.765528035586798e-05, | |
| "loss": 1.2203, | |
| "step": 325500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.762095666977868e-05, | |
| "loss": 1.2313, | |
| "step": 326000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7586632983689386e-05, | |
| "loss": 1.2185, | |
| "step": 326500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7552309297600088e-05, | |
| "loss": 1.234, | |
| "step": 327000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7517985611510793e-05, | |
| "loss": 1.2222, | |
| "step": 327500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7483661925421495e-05, | |
| "loss": 1.2157, | |
| "step": 328000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.74493382393322e-05, | |
| "loss": 1.2227, | |
| "step": 328500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7415014553242903e-05, | |
| "loss": 1.2353, | |
| "step": 329000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7380690867153608e-05, | |
| "loss": 1.2302, | |
| "step": 329500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.734636718106431e-05, | |
| "loss": 1.2297, | |
| "step": 330000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7312043494975015e-05, | |
| "loss": 1.2354, | |
| "step": 330500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 2.7277719808885714e-05, | |
| "loss": 1.2441, | |
| "step": 331000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.7243396122796423e-05, | |
| "loss": 1.2172, | |
| "step": 331500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.720907243670712e-05, | |
| "loss": 1.2282, | |
| "step": 332000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.717474875061783e-05, | |
| "loss": 1.2104, | |
| "step": 332500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.714042506452853e-05, | |
| "loss": 1.2259, | |
| "step": 333000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.7106101378439238e-05, | |
| "loss": 1.2183, | |
| "step": 333500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.7071777692349936e-05, | |
| "loss": 1.2079, | |
| "step": 334000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.703745400626064e-05, | |
| "loss": 1.2363, | |
| "step": 334500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.7003130320171343e-05, | |
| "loss": 1.2242, | |
| "step": 335000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.696880663408205e-05, | |
| "loss": 1.2268, | |
| "step": 335500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.693448294799275e-05, | |
| "loss": 1.2199, | |
| "step": 336000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.6900159261903456e-05, | |
| "loss": 1.2092, | |
| "step": 336500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.6865835575814158e-05, | |
| "loss": 1.2225, | |
| "step": 337000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.6831511889724864e-05, | |
| "loss": 1.2102, | |
| "step": 337500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.6797188203635566e-05, | |
| "loss": 1.2093, | |
| "step": 338000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.676286451754627e-05, | |
| "loss": 1.2155, | |
| "step": 338500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6728540831456973e-05, | |
| "loss": 1.2225, | |
| "step": 339000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.669421714536768e-05, | |
| "loss": 1.2049, | |
| "step": 339500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.665989345927838e-05, | |
| "loss": 1.2069, | |
| "step": 340000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6625569773189086e-05, | |
| "loss": 1.222, | |
| "step": 340500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6591246087099784e-05, | |
| "loss": 1.204, | |
| "step": 341000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6556922401010493e-05, | |
| "loss": 1.2178, | |
| "step": 341500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6522598714921192e-05, | |
| "loss": 1.2127, | |
| "step": 342000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.64882750288319e-05, | |
| "loss": 1.2064, | |
| "step": 342500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.64539513427426e-05, | |
| "loss": 1.2185, | |
| "step": 343000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6419627656653308e-05, | |
| "loss": 1.2013, | |
| "step": 343500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6385303970564006e-05, | |
| "loss": 1.2292, | |
| "step": 344000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6350980284474712e-05, | |
| "loss": 1.2112, | |
| "step": 344500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.6316656598385414e-05, | |
| "loss": 1.2157, | |
| "step": 345000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 2.628233291229612e-05, | |
| "loss": 1.2207, | |
| "step": 345500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.624800922620682e-05, | |
| "loss": 1.2022, | |
| "step": 346000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.6213685540117527e-05, | |
| "loss": 1.2079, | |
| "step": 346500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.617936185402823e-05, | |
| "loss": 1.2268, | |
| "step": 347000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.6145038167938934e-05, | |
| "loss": 1.2112, | |
| "step": 347500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.6110714481849636e-05, | |
| "loss": 1.2163, | |
| "step": 348000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.607639079576034e-05, | |
| "loss": 1.2091, | |
| "step": 348500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.6042067109671043e-05, | |
| "loss": 1.2165, | |
| "step": 349000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.600774342358175e-05, | |
| "loss": 1.2221, | |
| "step": 349500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.597341973749245e-05, | |
| "loss": 1.2155, | |
| "step": 350000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.5939096051403156e-05, | |
| "loss": 1.2141, | |
| "step": 350500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.5904772365313855e-05, | |
| "loss": 1.2207, | |
| "step": 351000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.5870448679224563e-05, | |
| "loss": 1.2194, | |
| "step": 351500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.5836124993135262e-05, | |
| "loss": 1.2293, | |
| "step": 352000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.580180130704597e-05, | |
| "loss": 1.2218, | |
| "step": 352500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 2.576747762095667e-05, | |
| "loss": 1.2016, | |
| "step": 353000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5733153934867378e-05, | |
| "loss": 1.2126, | |
| "step": 353500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5698830248778077e-05, | |
| "loss": 1.1931, | |
| "step": 354000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5664506562688782e-05, | |
| "loss": 1.2083, | |
| "step": 354500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5630182876599484e-05, | |
| "loss": 1.2098, | |
| "step": 355000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.559585919051019e-05, | |
| "loss": 1.2051, | |
| "step": 355500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.556153550442089e-05, | |
| "loss": 1.2199, | |
| "step": 356000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5527211818331597e-05, | |
| "loss": 1.2235, | |
| "step": 356500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.54928881322423e-05, | |
| "loss": 1.2155, | |
| "step": 357000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5458564446153004e-05, | |
| "loss": 1.2032, | |
| "step": 357500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5424240760063706e-05, | |
| "loss": 1.204, | |
| "step": 358000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.538991707397441e-05, | |
| "loss": 1.2113, | |
| "step": 358500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5355593387885114e-05, | |
| "loss": 1.213, | |
| "step": 359000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.532126970179582e-05, | |
| "loss": 1.21, | |
| "step": 359500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.528694601570652e-05, | |
| "loss": 1.2087, | |
| "step": 360000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 2.5252622329617226e-05, | |
| "loss": 1.2143, | |
| "step": 360500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.521829864352793e-05, | |
| "loss": 1.1881, | |
| "step": 361000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.5183974957438634e-05, | |
| "loss": 1.2134, | |
| "step": 361500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.5149651271349332e-05, | |
| "loss": 1.2089, | |
| "step": 362000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.511532758526004e-05, | |
| "loss": 1.201, | |
| "step": 362500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.508100389917074e-05, | |
| "loss": 1.2026, | |
| "step": 363000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.504668021308145e-05, | |
| "loss": 1.1987, | |
| "step": 363500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.5012356526992147e-05, | |
| "loss": 1.1874, | |
| "step": 364000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.4978032840902853e-05, | |
| "loss": 1.2088, | |
| "step": 364500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.4943709154813554e-05, | |
| "loss": 1.1979, | |
| "step": 365000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.490938546872426e-05, | |
| "loss": 1.2023, | |
| "step": 365500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.4875061782634962e-05, | |
| "loss": 1.203, | |
| "step": 366000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.4840738096545664e-05, | |
| "loss": 1.2179, | |
| "step": 366500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.480641441045637e-05, | |
| "loss": 1.1923, | |
| "step": 367000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 2.477209072436707e-05, | |
| "loss": 1.2023, | |
| "step": 367500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4737767038277777e-05, | |
| "loss": 1.205, | |
| "step": 368000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.470344335218848e-05, | |
| "loss": 1.2018, | |
| "step": 368500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4669119666099184e-05, | |
| "loss": 1.1879, | |
| "step": 369000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4634795980009886e-05, | |
| "loss": 1.1996, | |
| "step": 369500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.460047229392059e-05, | |
| "loss": 1.189, | |
| "step": 370000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4566148607831293e-05, | |
| "loss": 1.2067, | |
| "step": 370500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4531824921742e-05, | |
| "loss": 1.2127, | |
| "step": 371000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.44975012356527e-05, | |
| "loss": 1.1895, | |
| "step": 371500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4463177549563403e-05, | |
| "loss": 1.2007, | |
| "step": 372000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4428853863474108e-05, | |
| "loss": 1.2043, | |
| "step": 372500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.439453017738481e-05, | |
| "loss": 1.1933, | |
| "step": 373000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4360206491295515e-05, | |
| "loss": 1.206, | |
| "step": 373500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4325882805206217e-05, | |
| "loss": 1.2027, | |
| "step": 374000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4291559119116923e-05, | |
| "loss": 1.1793, | |
| "step": 374500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 2.4257235433027625e-05, | |
| "loss": 1.2073, | |
| "step": 375000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.422291174693833e-05, | |
| "loss": 1.1957, | |
| "step": 375500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.4188588060849032e-05, | |
| "loss": 1.1937, | |
| "step": 376000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.4154264374759734e-05, | |
| "loss": 1.1991, | |
| "step": 376500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.411994068867044e-05, | |
| "loss": 1.2111, | |
| "step": 377000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.408561700258114e-05, | |
| "loss": 1.1993, | |
| "step": 377500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.4051293316491847e-05, | |
| "loss": 1.1972, | |
| "step": 378000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.401696963040255e-05, | |
| "loss": 1.1997, | |
| "step": 378500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.3982645944313254e-05, | |
| "loss": 1.1913, | |
| "step": 379000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.3948322258223956e-05, | |
| "loss": 1.1851, | |
| "step": 379500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.3913998572134662e-05, | |
| "loss": 1.189, | |
| "step": 380000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.3879674886045364e-05, | |
| "loss": 1.1992, | |
| "step": 380500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.384535119995607e-05, | |
| "loss": 1.1901, | |
| "step": 381000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.381102751386677e-05, | |
| "loss": 1.2061, | |
| "step": 381500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.3776703827777473e-05, | |
| "loss": 1.1977, | |
| "step": 382000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.374238014168818e-05, | |
| "loss": 1.1913, | |
| "step": 382500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.370805645559888e-05, | |
| "loss": 1.1969, | |
| "step": 383000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3673732769509586e-05, | |
| "loss": 1.1941, | |
| "step": 383500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3639409083420288e-05, | |
| "loss": 1.2025, | |
| "step": 384000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3605085397330993e-05, | |
| "loss": 1.1943, | |
| "step": 384500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3570761711241695e-05, | |
| "loss": 1.1935, | |
| "step": 385000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.35364380251524e-05, | |
| "loss": 1.1848, | |
| "step": 385500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3502114339063103e-05, | |
| "loss": 1.1874, | |
| "step": 386000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3467790652973805e-05, | |
| "loss": 1.1883, | |
| "step": 386500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.343346696688451e-05, | |
| "loss": 1.2087, | |
| "step": 387000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3399143280795212e-05, | |
| "loss": 1.1908, | |
| "step": 387500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3364819594705917e-05, | |
| "loss": 1.189, | |
| "step": 388000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.333049590861662e-05, | |
| "loss": 1.1851, | |
| "step": 388500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3296172222527325e-05, | |
| "loss": 1.1713, | |
| "step": 389000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.3261848536438027e-05, | |
| "loss": 1.1979, | |
| "step": 389500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.3227524850348732e-05, | |
| "loss": 1.1826, | |
| "step": 390000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.3193201164259434e-05, | |
| "loss": 1.1974, | |
| "step": 390500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.315887747817014e-05, | |
| "loss": 1.2038, | |
| "step": 391000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.312455379208084e-05, | |
| "loss": 1.1846, | |
| "step": 391500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.3090230105991543e-05, | |
| "loss": 1.1926, | |
| "step": 392000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.305590641990225e-05, | |
| "loss": 1.1905, | |
| "step": 392500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.302158273381295e-05, | |
| "loss": 1.1834, | |
| "step": 393000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2987259047723656e-05, | |
| "loss": 1.1923, | |
| "step": 393500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2952935361634358e-05, | |
| "loss": 1.1805, | |
| "step": 394000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2918611675545064e-05, | |
| "loss": 1.1868, | |
| "step": 394500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2884287989455766e-05, | |
| "loss": 1.1865, | |
| "step": 395000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.284996430336647e-05, | |
| "loss": 1.1906, | |
| "step": 395500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2815640617277173e-05, | |
| "loss": 1.2035, | |
| "step": 396000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 2.2781316931187878e-05, | |
| "loss": 1.1846, | |
| "step": 396500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.274699324509858e-05, | |
| "loss": 1.1883, | |
| "step": 397000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2712669559009282e-05, | |
| "loss": 1.1768, | |
| "step": 397500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2678345872919984e-05, | |
| "loss": 1.1921, | |
| "step": 398000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2644022186830686e-05, | |
| "loss": 1.191, | |
| "step": 398500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.260969850074139e-05, | |
| "loss": 1.1961, | |
| "step": 399000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2575374814652094e-05, | |
| "loss": 1.1774, | |
| "step": 399500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.25410511285628e-05, | |
| "loss": 1.1892, | |
| "step": 400000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.25067274424735e-05, | |
| "loss": 1.1891, | |
| "step": 400500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2472403756384206e-05, | |
| "loss": 1.1744, | |
| "step": 401000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.243808007029491e-05, | |
| "loss": 1.1823, | |
| "step": 401500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2403756384205614e-05, | |
| "loss": 1.169, | |
| "step": 402000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2369432698116316e-05, | |
| "loss": 1.1818, | |
| "step": 402500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.233510901202702e-05, | |
| "loss": 1.1884, | |
| "step": 403000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2300785325937723e-05, | |
| "loss": 1.178, | |
| "step": 403500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.2266461639848425e-05, | |
| "loss": 1.1778, | |
| "step": 404000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.223213795375913e-05, | |
| "loss": 1.1832, | |
| "step": 404500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.2197814267669832e-05, | |
| "loss": 1.1826, | |
| "step": 405000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.2163490581580538e-05, | |
| "loss": 1.1814, | |
| "step": 405500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.212916689549124e-05, | |
| "loss": 1.1768, | |
| "step": 406000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.2094843209401945e-05, | |
| "loss": 1.1773, | |
| "step": 406500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.2060519523312647e-05, | |
| "loss": 1.1911, | |
| "step": 407000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.2026195837223353e-05, | |
| "loss": 1.1977, | |
| "step": 407500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.1991872151134055e-05, | |
| "loss": 1.1783, | |
| "step": 408000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.1957548465044757e-05, | |
| "loss": 1.1862, | |
| "step": 408500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.1923224778955462e-05, | |
| "loss": 1.1748, | |
| "step": 409000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.1888901092866164e-05, | |
| "loss": 1.1942, | |
| "step": 409500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.185457740677687e-05, | |
| "loss": 1.1826, | |
| "step": 410000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.182025372068757e-05, | |
| "loss": 1.1596, | |
| "step": 410500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.1785930034598277e-05, | |
| "loss": 1.1854, | |
| "step": 411000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 2.175160634850898e-05, | |
| "loss": 1.1754, | |
| "step": 411500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1717282662419684e-05, | |
| "loss": 1.1721, | |
| "step": 412000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1682958976330386e-05, | |
| "loss": 1.1821, | |
| "step": 412500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.164863529024109e-05, | |
| "loss": 1.1721, | |
| "step": 413000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1614311604151793e-05, | |
| "loss": 1.1879, | |
| "step": 413500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1579987918062495e-05, | |
| "loss": 1.1748, | |
| "step": 414000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.15456642319732e-05, | |
| "loss": 1.1784, | |
| "step": 414500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1511340545883903e-05, | |
| "loss": 1.1718, | |
| "step": 415000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1477016859794608e-05, | |
| "loss": 1.1937, | |
| "step": 415500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.144269317370531e-05, | |
| "loss": 1.17, | |
| "step": 416000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1408369487616016e-05, | |
| "loss": 1.1795, | |
| "step": 416500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1374045801526718e-05, | |
| "loss": 1.1762, | |
| "step": 417000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1339722115437423e-05, | |
| "loss": 1.1813, | |
| "step": 417500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1305398429348125e-05, | |
| "loss": 1.1713, | |
| "step": 418000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.127107474325883e-05, | |
| "loss": 1.1697, | |
| "step": 418500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.1236751057169532e-05, | |
| "loss": 1.1785, | |
| "step": 419000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.1202427371080234e-05, | |
| "loss": 1.1781, | |
| "step": 419500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.116810368499094e-05, | |
| "loss": 1.1829, | |
| "step": 420000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.113377999890164e-05, | |
| "loss": 1.175, | |
| "step": 420500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.1099456312812347e-05, | |
| "loss": 1.1739, | |
| "step": 421000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.106513262672305e-05, | |
| "loss": 1.1672, | |
| "step": 421500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.1030808940633754e-05, | |
| "loss": 1.174, | |
| "step": 422000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.0996485254544456e-05, | |
| "loss": 1.1741, | |
| "step": 422500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.0962161568455162e-05, | |
| "loss": 1.1891, | |
| "step": 423000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.0927837882365864e-05, | |
| "loss": 1.1718, | |
| "step": 423500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.0893514196276566e-05, | |
| "loss": 1.176, | |
| "step": 424000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.085919051018727e-05, | |
| "loss": 1.169, | |
| "step": 424500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.0824866824097973e-05, | |
| "loss": 1.1837, | |
| "step": 425000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.079054313800868e-05, | |
| "loss": 1.1749, | |
| "step": 425500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 2.075621945191938e-05, | |
| "loss": 1.161, | |
| "step": 426000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0721895765830086e-05, | |
| "loss": 1.1659, | |
| "step": 426500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0687572079740788e-05, | |
| "loss": 1.1625, | |
| "step": 427000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0653248393651493e-05, | |
| "loss": 1.192, | |
| "step": 427500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0618924707562195e-05, | |
| "loss": 1.1736, | |
| "step": 428000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.05846010214729e-05, | |
| "loss": 1.1756, | |
| "step": 428500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0550277335383603e-05, | |
| "loss": 1.1623, | |
| "step": 429000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0515953649294305e-05, | |
| "loss": 1.1548, | |
| "step": 429500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.048162996320501e-05, | |
| "loss": 1.1696, | |
| "step": 430000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0447306277115712e-05, | |
| "loss": 1.1738, | |
| "step": 430500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0412982591026417e-05, | |
| "loss": 1.1676, | |
| "step": 431000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.037865890493712e-05, | |
| "loss": 1.1667, | |
| "step": 431500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0344335218847825e-05, | |
| "loss": 1.1706, | |
| "step": 432000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0310011532758527e-05, | |
| "loss": 1.1759, | |
| "step": 432500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.0275687846669232e-05, | |
| "loss": 1.1752, | |
| "step": 433000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.0241364160579934e-05, | |
| "loss": 1.1697, | |
| "step": 433500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.0207040474490636e-05, | |
| "loss": 1.1663, | |
| "step": 434000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.017271678840134e-05, | |
| "loss": 1.1607, | |
| "step": 434500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.0138393102312043e-05, | |
| "loss": 1.1808, | |
| "step": 435000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.010406941622275e-05, | |
| "loss": 1.1732, | |
| "step": 435500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.006974573013345e-05, | |
| "loss": 1.1654, | |
| "step": 436000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.0035422044044156e-05, | |
| "loss": 1.1624, | |
| "step": 436500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.0001098357954858e-05, | |
| "loss": 1.1681, | |
| "step": 437000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.9966774671865564e-05, | |
| "loss": 1.1435, | |
| "step": 437500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.9932450985776266e-05, | |
| "loss": 1.1603, | |
| "step": 438000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.989812729968697e-05, | |
| "loss": 1.1743, | |
| "step": 438500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.9863803613597673e-05, | |
| "loss": 1.1566, | |
| "step": 439000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.9829479927508375e-05, | |
| "loss": 1.1655, | |
| "step": 439500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.979515624141908e-05, | |
| "loss": 1.1692, | |
| "step": 440000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.9760832555329782e-05, | |
| "loss": 1.1744, | |
| "step": 440500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9726508869240488e-05, | |
| "loss": 1.1791, | |
| "step": 441000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.969218518315119e-05, | |
| "loss": 1.1512, | |
| "step": 441500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9657861497061895e-05, | |
| "loss": 1.1546, | |
| "step": 442000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9623537810972597e-05, | |
| "loss": 1.1653, | |
| "step": 442500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9589214124883302e-05, | |
| "loss": 1.1763, | |
| "step": 443000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9554890438794004e-05, | |
| "loss": 1.1754, | |
| "step": 443500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.952056675270471e-05, | |
| "loss": 1.1755, | |
| "step": 444000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9486243066615412e-05, | |
| "loss": 1.1707, | |
| "step": 444500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9451919380526114e-05, | |
| "loss": 1.1642, | |
| "step": 445000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.941759569443682e-05, | |
| "loss": 1.149, | |
| "step": 445500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.938327200834752e-05, | |
| "loss": 1.1591, | |
| "step": 446000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9348948322258227e-05, | |
| "loss": 1.1637, | |
| "step": 446500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.931462463616893e-05, | |
| "loss": 1.1537, | |
| "step": 447000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.9280300950079634e-05, | |
| "loss": 1.164, | |
| "step": 447500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.9245977263990336e-05, | |
| "loss": 1.1678, | |
| "step": 448000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.921165357790104e-05, | |
| "loss": 1.1687, | |
| "step": 448500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.9177329891811743e-05, | |
| "loss": 1.1532, | |
| "step": 449000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.9143006205722445e-05, | |
| "loss": 1.1593, | |
| "step": 449500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.910868251963315e-05, | |
| "loss": 1.169, | |
| "step": 450000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.9074358833543853e-05, | |
| "loss": 1.1605, | |
| "step": 450500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.9040035147454558e-05, | |
| "loss": 1.1532, | |
| "step": 451000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.900571146136526e-05, | |
| "loss": 1.1509, | |
| "step": 451500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8971387775275965e-05, | |
| "loss": 1.1675, | |
| "step": 452000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8937064089186667e-05, | |
| "loss": 1.1575, | |
| "step": 452500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8902740403097373e-05, | |
| "loss": 1.1534, | |
| "step": 453000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8868416717008075e-05, | |
| "loss": 1.1656, | |
| "step": 453500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.883409303091878e-05, | |
| "loss": 1.16, | |
| "step": 454000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8799769344829482e-05, | |
| "loss": 1.1664, | |
| "step": 454500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.8765445658740184e-05, | |
| "loss": 1.1549, | |
| "step": 455000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.873112197265089e-05, | |
| "loss": 1.1578, | |
| "step": 455500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.869679828656159e-05, | |
| "loss": 1.1543, | |
| "step": 456000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8662474600472297e-05, | |
| "loss": 1.1582, | |
| "step": 456500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8628150914383e-05, | |
| "loss": 1.1512, | |
| "step": 457000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8593827228293704e-05, | |
| "loss": 1.1682, | |
| "step": 457500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8559503542204406e-05, | |
| "loss": 1.1641, | |
| "step": 458000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8525179856115108e-05, | |
| "loss": 1.1379, | |
| "step": 458500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.849085617002581e-05, | |
| "loss": 1.1676, | |
| "step": 459000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8456532483936516e-05, | |
| "loss": 1.1628, | |
| "step": 459500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8422208797847218e-05, | |
| "loss": 1.1624, | |
| "step": 460000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8387885111757923e-05, | |
| "loss": 1.1337, | |
| "step": 460500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8353561425668625e-05, | |
| "loss": 1.1584, | |
| "step": 461000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8319237739579327e-05, | |
| "loss": 1.1721, | |
| "step": 461500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8284914053490032e-05, | |
| "loss": 1.1605, | |
| "step": 462000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.8250590367400734e-05, | |
| "loss": 1.1683, | |
| "step": 462500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.821626668131144e-05, | |
| "loss": 1.1539, | |
| "step": 463000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8181942995222142e-05, | |
| "loss": 1.1537, | |
| "step": 463500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8147619309132847e-05, | |
| "loss": 1.1487, | |
| "step": 464000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.811329562304355e-05, | |
| "loss": 1.1566, | |
| "step": 464500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8078971936954254e-05, | |
| "loss": 1.1598, | |
| "step": 465000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8044648250864956e-05, | |
| "loss": 1.1632, | |
| "step": 465500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.8010324564775662e-05, | |
| "loss": 1.1617, | |
| "step": 466000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7976000878686364e-05, | |
| "loss": 1.1549, | |
| "step": 466500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7941677192597066e-05, | |
| "loss": 1.1585, | |
| "step": 467000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.790735350650777e-05, | |
| "loss": 1.1581, | |
| "step": 467500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7873029820418473e-05, | |
| "loss": 1.1453, | |
| "step": 468000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.783870613432918e-05, | |
| "loss": 1.1329, | |
| "step": 468500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.780438244823988e-05, | |
| "loss": 1.1644, | |
| "step": 469000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.7770058762150586e-05, | |
| "loss": 1.16, | |
| "step": 469500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7735735076061288e-05, | |
| "loss": 1.1575, | |
| "step": 470000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7701411389971993e-05, | |
| "loss": 1.1594, | |
| "step": 470500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7667087703882695e-05, | |
| "loss": 1.1563, | |
| "step": 471000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7632764017793397e-05, | |
| "loss": 1.1566, | |
| "step": 471500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7598440331704103e-05, | |
| "loss": 1.151, | |
| "step": 472000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7564116645614805e-05, | |
| "loss": 1.158, | |
| "step": 472500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.752979295952551e-05, | |
| "loss": 1.1673, | |
| "step": 473000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7495469273436212e-05, | |
| "loss": 1.1419, | |
| "step": 473500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7461145587346917e-05, | |
| "loss": 1.1342, | |
| "step": 474000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.742682190125762e-05, | |
| "loss": 1.1691, | |
| "step": 474500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7392498215168325e-05, | |
| "loss": 1.174, | |
| "step": 475000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7358174529079027e-05, | |
| "loss": 1.1565, | |
| "step": 475500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7323850842989732e-05, | |
| "loss": 1.1523, | |
| "step": 476000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7289527156900434e-05, | |
| "loss": 1.1378, | |
| "step": 476500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.7255203470811136e-05, | |
| "loss": 1.1516, | |
| "step": 477000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.722087978472184e-05, | |
| "loss": 1.1512, | |
| "step": 477500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7186556098632544e-05, | |
| "loss": 1.1571, | |
| "step": 478000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.715223241254325e-05, | |
| "loss": 1.1408, | |
| "step": 478500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.711790872645395e-05, | |
| "loss": 1.1404, | |
| "step": 479000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7083585040364656e-05, | |
| "loss": 1.1433, | |
| "step": 479500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.704926135427536e-05, | |
| "loss": 1.1466, | |
| "step": 480000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.7014937668186064e-05, | |
| "loss": 1.1476, | |
| "step": 480500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6980613982096766e-05, | |
| "loss": 1.1366, | |
| "step": 481000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6946290296007468e-05, | |
| "loss": 1.1394, | |
| "step": 481500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6911966609918173e-05, | |
| "loss": 1.1466, | |
| "step": 482000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6877642923828875e-05, | |
| "loss": 1.1489, | |
| "step": 482500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.684331923773958e-05, | |
| "loss": 1.1527, | |
| "step": 483000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6808995551650282e-05, | |
| "loss": 1.13, | |
| "step": 483500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.6774671865560988e-05, | |
| "loss": 1.1549, | |
| "step": 484000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.674034817947169e-05, | |
| "loss": 1.1246, | |
| "step": 484500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6706024493382395e-05, | |
| "loss": 1.1478, | |
| "step": 485000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6671700807293097e-05, | |
| "loss": 1.1496, | |
| "step": 485500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6637377121203803e-05, | |
| "loss": 1.1511, | |
| "step": 486000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6603053435114505e-05, | |
| "loss": 1.1252, | |
| "step": 486500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6568729749025207e-05, | |
| "loss": 1.1503, | |
| "step": 487000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6534406062935912e-05, | |
| "loss": 1.1525, | |
| "step": 487500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6500082376846614e-05, | |
| "loss": 1.1391, | |
| "step": 488000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.646575869075732e-05, | |
| "loss": 1.1636, | |
| "step": 488500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.643143500466802e-05, | |
| "loss": 1.1541, | |
| "step": 489000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6397111318578727e-05, | |
| "loss": 1.1253, | |
| "step": 489500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.636278763248943e-05, | |
| "loss": 1.1437, | |
| "step": 490000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6328463946400134e-05, | |
| "loss": 1.1327, | |
| "step": 490500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6294140260310836e-05, | |
| "loss": 1.1451, | |
| "step": 491000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.6259816574221538e-05, | |
| "loss": 1.1586, | |
| "step": 491500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6225492888132243e-05, | |
| "loss": 1.1416, | |
| "step": 492000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6191169202042945e-05, | |
| "loss": 1.1438, | |
| "step": 492500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.615684551595365e-05, | |
| "loss": 1.1366, | |
| "step": 493000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6122521829864353e-05, | |
| "loss": 1.1383, | |
| "step": 493500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6088198143775058e-05, | |
| "loss": 1.1445, | |
| "step": 494000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.605387445768576e-05, | |
| "loss": 1.1624, | |
| "step": 494500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.6019550771596466e-05, | |
| "loss": 1.1331, | |
| "step": 495000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5985227085507168e-05, | |
| "loss": 1.1451, | |
| "step": 495500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5950903399417873e-05, | |
| "loss": 1.1531, | |
| "step": 496000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5916579713328575e-05, | |
| "loss": 1.1537, | |
| "step": 496500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5882256027239277e-05, | |
| "loss": 1.1514, | |
| "step": 497000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5847932341149982e-05, | |
| "loss": 1.1564, | |
| "step": 497500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.5813608655060684e-05, | |
| "loss": 1.1311, | |
| "step": 498000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.577928496897139e-05, | |
| "loss": 1.1517, | |
| "step": 498500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.574496128288209e-05, | |
| "loss": 1.1388, | |
| "step": 499000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5710637596792797e-05, | |
| "loss": 1.1404, | |
| "step": 499500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.56763139107035e-05, | |
| "loss": 1.1275, | |
| "step": 500000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5641990224614204e-05, | |
| "loss": 1.1455, | |
| "step": 500500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5607666538524906e-05, | |
| "loss": 1.1472, | |
| "step": 501000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5573342852435612e-05, | |
| "loss": 1.1409, | |
| "step": 501500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5539019166346314e-05, | |
| "loss": 1.1419, | |
| "step": 502000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5504695480257016e-05, | |
| "loss": 1.1368, | |
| "step": 502500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.547037179416772e-05, | |
| "loss": 1.1421, | |
| "step": 503000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5436048108078423e-05, | |
| "loss": 1.1307, | |
| "step": 503500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.540172442198913e-05, | |
| "loss": 1.1503, | |
| "step": 504000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.536740073589983e-05, | |
| "loss": 1.1415, | |
| "step": 504500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5333077049810536e-05, | |
| "loss": 1.1348, | |
| "step": 505000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5298753363721238e-05, | |
| "loss": 1.1459, | |
| "step": 505500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.5264429677631943e-05, | |
| "loss": 1.1477, | |
| "step": 506000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5230105991542645e-05, | |
| "loss": 1.1372, | |
| "step": 506500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5195782305453349e-05, | |
| "loss": 1.1253, | |
| "step": 507000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5161458619364053e-05, | |
| "loss": 1.1266, | |
| "step": 507500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5127134933274756e-05, | |
| "loss": 1.1288, | |
| "step": 508000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.509281124718546e-05, | |
| "loss": 1.1388, | |
| "step": 508500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5058487561096164e-05, | |
| "loss": 1.1431, | |
| "step": 509000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.5024163875006866e-05, | |
| "loss": 1.1306, | |
| "step": 509500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.498984018891757e-05, | |
| "loss": 1.1233, | |
| "step": 510000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.4955516502828273e-05, | |
| "loss": 1.1334, | |
| "step": 510500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.4921192816738977e-05, | |
| "loss": 1.1422, | |
| "step": 511000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.488686913064968e-05, | |
| "loss": 1.1234, | |
| "step": 511500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.4852545444560384e-05, | |
| "loss": 1.1309, | |
| "step": 512000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.4818221758471088e-05, | |
| "loss": 1.1296, | |
| "step": 512500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.4783898072381791e-05, | |
| "loss": 1.1503, | |
| "step": 513000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4749574386292495e-05, | |
| "loss": 1.1445, | |
| "step": 513500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4715250700203199e-05, | |
| "loss": 1.1172, | |
| "step": 514000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.46809270141139e-05, | |
| "loss": 1.1197, | |
| "step": 514500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4646603328024604e-05, | |
| "loss": 1.116, | |
| "step": 515000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4612279641935308e-05, | |
| "loss": 1.1234, | |
| "step": 515500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4577955955846012e-05, | |
| "loss": 1.1256, | |
| "step": 516000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4543632269756716e-05, | |
| "loss": 1.1435, | |
| "step": 516500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.450930858366742e-05, | |
| "loss": 1.1163, | |
| "step": 517000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4474984897578123e-05, | |
| "loss": 1.1273, | |
| "step": 517500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4440661211488827e-05, | |
| "loss": 1.1213, | |
| "step": 518000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.440633752539953e-05, | |
| "loss": 1.1277, | |
| "step": 518500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4372013839310234e-05, | |
| "loss": 1.1185, | |
| "step": 519000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4337690153220934e-05, | |
| "loss": 1.1358, | |
| "step": 519500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4303366467131638e-05, | |
| "loss": 1.1315, | |
| "step": 520000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.4269042781042342e-05, | |
| "loss": 1.1348, | |
| "step": 520500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4234719094953045e-05, | |
| "loss": 1.1121, | |
| "step": 521000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4200395408863747e-05, | |
| "loss": 1.132, | |
| "step": 521500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4166071722774451e-05, | |
| "loss": 1.1265, | |
| "step": 522000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4131748036685155e-05, | |
| "loss": 1.1156, | |
| "step": 522500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4097424350595858e-05, | |
| "loss": 1.1258, | |
| "step": 523000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4063100664506562e-05, | |
| "loss": 1.118, | |
| "step": 523500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.4028776978417266e-05, | |
| "loss": 1.121, | |
| "step": 524000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.399445329232797e-05, | |
| "loss": 1.1362, | |
| "step": 524500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.3960129606238673e-05, | |
| "loss": 1.1463, | |
| "step": 525000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.3925805920149377e-05, | |
| "loss": 1.1338, | |
| "step": 525500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.389148223406008e-05, | |
| "loss": 1.1241, | |
| "step": 526000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.3857158547970783e-05, | |
| "loss": 1.1262, | |
| "step": 526500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.3822834861881486e-05, | |
| "loss": 1.1376, | |
| "step": 527000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.378851117579219e-05, | |
| "loss": 1.1267, | |
| "step": 527500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.3754187489702894e-05, | |
| "loss": 1.1305, | |
| "step": 528000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3719863803613597e-05, | |
| "loss": 1.1264, | |
| "step": 528500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3685540117524301e-05, | |
| "loss": 1.1392, | |
| "step": 529000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3651216431435005e-05, | |
| "loss": 1.1286, | |
| "step": 529500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3616892745345708e-05, | |
| "loss": 1.1257, | |
| "step": 530000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3582569059256412e-05, | |
| "loss": 1.1205, | |
| "step": 530500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3548245373167116e-05, | |
| "loss": 1.1129, | |
| "step": 531000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3513921687077818e-05, | |
| "loss": 1.1415, | |
| "step": 531500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3479598000988521e-05, | |
| "loss": 1.1416, | |
| "step": 532000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3445274314899225e-05, | |
| "loss": 1.1123, | |
| "step": 532500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3410950628809929e-05, | |
| "loss": 1.1168, | |
| "step": 533000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3376626942720632e-05, | |
| "loss": 1.1235, | |
| "step": 533500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3342303256631336e-05, | |
| "loss": 1.1275, | |
| "step": 534000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.330797957054204e-05, | |
| "loss": 1.1171, | |
| "step": 534500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.3273655884452743e-05, | |
| "loss": 1.1186, | |
| "step": 535000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3239332198363447e-05, | |
| "loss": 1.1296, | |
| "step": 535500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.320500851227415e-05, | |
| "loss": 1.1191, | |
| "step": 536000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3170684826184853e-05, | |
| "loss": 1.1147, | |
| "step": 536500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3136361140095557e-05, | |
| "loss": 1.1247, | |
| "step": 537000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.310203745400626e-05, | |
| "loss": 1.1293, | |
| "step": 537500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3067713767916964e-05, | |
| "loss": 1.1188, | |
| "step": 538000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.3033390081827668e-05, | |
| "loss": 1.1189, | |
| "step": 538500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2999066395738371e-05, | |
| "loss": 1.1277, | |
| "step": 539000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2964742709649075e-05, | |
| "loss": 1.1154, | |
| "step": 539500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2930419023559779e-05, | |
| "loss": 1.1186, | |
| "step": 540000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2896095337470482e-05, | |
| "loss": 1.1305, | |
| "step": 540500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2861771651381186e-05, | |
| "loss": 1.1133, | |
| "step": 541000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.282744796529189e-05, | |
| "loss": 1.1174, | |
| "step": 541500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2793124279202592e-05, | |
| "loss": 1.1097, | |
| "step": 542000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.2758800593113295e-05, | |
| "loss": 1.1199, | |
| "step": 542500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2724476907023999e-05, | |
| "loss": 1.1168, | |
| "step": 543000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2690153220934703e-05, | |
| "loss": 1.1189, | |
| "step": 543500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2655829534845406e-05, | |
| "loss": 1.1175, | |
| "step": 544000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.262150584875611e-05, | |
| "loss": 1.1258, | |
| "step": 544500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2587182162666814e-05, | |
| "loss": 1.1275, | |
| "step": 545000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2552858476577517e-05, | |
| "loss": 1.1183, | |
| "step": 545500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2518534790488221e-05, | |
| "loss": 1.1264, | |
| "step": 546000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2484211104398925e-05, | |
| "loss": 1.1107, | |
| "step": 546500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2449887418309627e-05, | |
| "loss": 1.1207, | |
| "step": 547000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.241556373222033e-05, | |
| "loss": 1.1197, | |
| "step": 547500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2381240046131034e-05, | |
| "loss": 1.1133, | |
| "step": 548000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2346916360041738e-05, | |
| "loss": 1.1144, | |
| "step": 548500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2312592673952442e-05, | |
| "loss": 1.1179, | |
| "step": 549000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.2278268987863145e-05, | |
| "loss": 1.103, | |
| "step": 549500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2243945301773849e-05, | |
| "loss": 1.1107, | |
| "step": 550000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2209621615684553e-05, | |
| "loss": 1.1112, | |
| "step": 550500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2175297929595256e-05, | |
| "loss": 1.1213, | |
| "step": 551000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.214097424350596e-05, | |
| "loss": 1.1145, | |
| "step": 551500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2106650557416662e-05, | |
| "loss": 1.1192, | |
| "step": 552000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2072326871327366e-05, | |
| "loss": 1.1066, | |
| "step": 552500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.203800318523807e-05, | |
| "loss": 1.1142, | |
| "step": 553000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.2003679499148773e-05, | |
| "loss": 1.1231, | |
| "step": 553500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1969355813059477e-05, | |
| "loss": 1.1043, | |
| "step": 554000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.193503212697018e-05, | |
| "loss": 1.1253, | |
| "step": 554500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1900708440880884e-05, | |
| "loss": 1.1159, | |
| "step": 555000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1866384754791588e-05, | |
| "loss": 1.1332, | |
| "step": 555500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1832061068702292e-05, | |
| "loss": 1.1203, | |
| "step": 556000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1797737382612995e-05, | |
| "loss": 1.1176, | |
| "step": 556500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.1763413696523697e-05, | |
| "loss": 1.1288, | |
| "step": 557000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1729090010434401e-05, | |
| "loss": 1.1123, | |
| "step": 557500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1694766324345105e-05, | |
| "loss": 1.1144, | |
| "step": 558000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1660442638255808e-05, | |
| "loss": 1.1042, | |
| "step": 558500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1626118952166512e-05, | |
| "loss": 1.1046, | |
| "step": 559000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1591795266077216e-05, | |
| "loss": 1.1178, | |
| "step": 559500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.155747157998792e-05, | |
| "loss": 1.1204, | |
| "step": 560000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1523147893898623e-05, | |
| "loss": 1.122, | |
| "step": 560500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1488824207809327e-05, | |
| "loss": 1.1163, | |
| "step": 561000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.145450052172003e-05, | |
| "loss": 1.1089, | |
| "step": 561500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1420176835630732e-05, | |
| "loss": 1.0995, | |
| "step": 562000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1385853149541436e-05, | |
| "loss": 1.1071, | |
| "step": 562500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.135152946345214e-05, | |
| "loss": 1.1279, | |
| "step": 563000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1317205777362843e-05, | |
| "loss": 1.1133, | |
| "step": 563500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.1282882091273547e-05, | |
| "loss": 1.1104, | |
| "step": 564000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.124855840518425e-05, | |
| "loss": 1.1127, | |
| "step": 564500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1214234719094954e-05, | |
| "loss": 1.1086, | |
| "step": 565000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1179911033005656e-05, | |
| "loss": 1.1049, | |
| "step": 565500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.114558734691636e-05, | |
| "loss": 1.1204, | |
| "step": 566000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1111263660827064e-05, | |
| "loss": 1.1175, | |
| "step": 566500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1076939974737768e-05, | |
| "loss": 1.1029, | |
| "step": 567000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1042616288648471e-05, | |
| "loss": 1.1157, | |
| "step": 567500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.1008292602559173e-05, | |
| "loss": 1.1129, | |
| "step": 568000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0973968916469877e-05, | |
| "loss": 1.119, | |
| "step": 568500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.093964523038058e-05, | |
| "loss": 1.1126, | |
| "step": 569000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0905321544291284e-05, | |
| "loss": 1.1201, | |
| "step": 569500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0870997858201988e-05, | |
| "loss": 1.1253, | |
| "step": 570000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0836674172112692e-05, | |
| "loss": 1.1019, | |
| "step": 570500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0802350486023395e-05, | |
| "loss": 1.1002, | |
| "step": 571000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.0768026799934099e-05, | |
| "loss": 1.1029, | |
| "step": 571500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0733703113844803e-05, | |
| "loss": 1.1161, | |
| "step": 572000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0699379427755506e-05, | |
| "loss": 1.0901, | |
| "step": 572500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0665055741666208e-05, | |
| "loss": 1.1245, | |
| "step": 573000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0630732055576912e-05, | |
| "loss": 1.1055, | |
| "step": 573500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0596408369487616e-05, | |
| "loss": 1.1068, | |
| "step": 574000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.056208468339832e-05, | |
| "loss": 1.0998, | |
| "step": 574500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0527760997309023e-05, | |
| "loss": 1.0957, | |
| "step": 575000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0493437311219727e-05, | |
| "loss": 1.0896, | |
| "step": 575500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.045911362513043e-05, | |
| "loss": 1.0978, | |
| "step": 576000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0424789939041134e-05, | |
| "loss": 1.106, | |
| "step": 576500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0390466252951838e-05, | |
| "loss": 1.1152, | |
| "step": 577000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0356142566862542e-05, | |
| "loss": 1.0977, | |
| "step": 577500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0321818880773244e-05, | |
| "loss": 1.1115, | |
| "step": 578000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0287495194683947e-05, | |
| "loss": 1.0966, | |
| "step": 578500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.0253171508594651e-05, | |
| "loss": 1.1201, | |
| "step": 579000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0218847822505355e-05, | |
| "loss": 1.1168, | |
| "step": 579500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0184524136416058e-05, | |
| "loss": 1.1141, | |
| "step": 580000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0150200450326762e-05, | |
| "loss": 1.1115, | |
| "step": 580500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0115876764237466e-05, | |
| "loss": 1.1158, | |
| "step": 581000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.008155307814817e-05, | |
| "loss": 1.1131, | |
| "step": 581500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0047229392058873e-05, | |
| "loss": 1.1007, | |
| "step": 582000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.0012905705969577e-05, | |
| "loss": 1.1066, | |
| "step": 582500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.97858201988028e-06, | |
| "loss": 1.1094, | |
| "step": 583000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.944258333790982e-06, | |
| "loss": 1.0962, | |
| "step": 583500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.909934647701686e-06, | |
| "loss": 1.0962, | |
| "step": 584000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.87561096161239e-06, | |
| "loss": 1.1004, | |
| "step": 584500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.841287275523093e-06, | |
| "loss": 1.1165, | |
| "step": 585000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.806963589433797e-06, | |
| "loss": 1.1079, | |
| "step": 585500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.7726399033445e-06, | |
| "loss": 1.1008, | |
| "step": 586000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.738316217255205e-06, | |
| "loss": 1.0872, | |
| "step": 586500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.703992531165908e-06, | |
| "loss": 1.1128, | |
| "step": 587000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.669668845076612e-06, | |
| "loss": 1.1272, | |
| "step": 587500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.635345158987316e-06, | |
| "loss": 1.1073, | |
| "step": 588000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.601021472898018e-06, | |
| "loss": 1.0932, | |
| "step": 588500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.566697786808721e-06, | |
| "loss": 1.1003, | |
| "step": 589000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.532374100719425e-06, | |
| "loss": 1.1132, | |
| "step": 589500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.498050414630129e-06, | |
| "loss": 1.0932, | |
| "step": 590000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.463726728540832e-06, | |
| "loss": 1.1129, | |
| "step": 590500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.429403042451536e-06, | |
| "loss": 1.1182, | |
| "step": 591000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.39507935636224e-06, | |
| "loss": 1.1044, | |
| "step": 591500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.360755670272943e-06, | |
| "loss": 1.0989, | |
| "step": 592000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.326431984183647e-06, | |
| "loss": 1.1118, | |
| "step": 592500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.29210829809435e-06, | |
| "loss": 1.0983, | |
| "step": 593000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.257784612005053e-06, | |
| "loss": 1.1071, | |
| "step": 593500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.223460925915756e-06, | |
| "loss": 1.091, | |
| "step": 594000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.18913723982646e-06, | |
| "loss": 1.0953, | |
| "step": 594500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.154813553737164e-06, | |
| "loss": 1.1123, | |
| "step": 595000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.120489867647867e-06, | |
| "loss": 1.1069, | |
| "step": 595500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.08616618155857e-06, | |
| "loss": 1.0998, | |
| "step": 596000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.051842495469273e-06, | |
| "loss": 1.0986, | |
| "step": 596500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.017518809379977e-06, | |
| "loss": 1.1044, | |
| "step": 597000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.98319512329068e-06, | |
| "loss": 1.0957, | |
| "step": 597500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.948871437201384e-06, | |
| "loss": 1.0912, | |
| "step": 598000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.914547751112088e-06, | |
| "loss": 1.095, | |
| "step": 598500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.880224065022792e-06, | |
| "loss": 1.0936, | |
| "step": 599000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.845900378933494e-06, | |
| "loss": 1.0951, | |
| "step": 599500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.811576692844197e-06, | |
| "loss": 1.1066, | |
| "step": 600000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.777253006754901e-06, | |
| "loss": 1.0902, | |
| "step": 600500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.742929320665605e-06, | |
| "loss": 1.0947, | |
| "step": 601000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.708605634576308e-06, | |
| "loss": 1.0913, | |
| "step": 601500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.674281948487012e-06, | |
| "loss": 1.089, | |
| "step": 602000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.639958262397716e-06, | |
| "loss": 1.1062, | |
| "step": 602500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.60563457630842e-06, | |
| "loss": 1.1013, | |
| "step": 603000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.571310890219123e-06, | |
| "loss": 1.09, | |
| "step": 603500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.536987204129827e-06, | |
| "loss": 1.1014, | |
| "step": 604000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.502663518040529e-06, | |
| "loss": 1.1057, | |
| "step": 604500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.468339831951232e-06, | |
| "loss": 1.088, | |
| "step": 605000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.434016145861936e-06, | |
| "loss": 1.11, | |
| "step": 605500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.39969245977264e-06, | |
| "loss": 1.1013, | |
| "step": 606000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.365368773683344e-06, | |
| "loss": 1.0868, | |
| "step": 606500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.331045087594047e-06, | |
| "loss": 1.1111, | |
| "step": 607000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.296721401504751e-06, | |
| "loss": 1.1081, | |
| "step": 607500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.262397715415455e-06, | |
| "loss": 1.1064, | |
| "step": 608000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.228074029326158e-06, | |
| "loss": 1.0929, | |
| "step": 608500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.193750343236862e-06, | |
| "loss": 1.0806, | |
| "step": 609000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.159426657147564e-06, | |
| "loss": 1.0952, | |
| "step": 609500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.125102971058268e-06, | |
| "loss": 1.0859, | |
| "step": 610000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.090779284968971e-06, | |
| "loss": 1.1025, | |
| "step": 610500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.056455598879675e-06, | |
| "loss": 1.1, | |
| "step": 611000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.022131912790379e-06, | |
| "loss": 1.0921, | |
| "step": 611500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.987808226701082e-06, | |
| "loss": 1.1135, | |
| "step": 612000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.953484540611786e-06, | |
| "loss": 1.1048, | |
| "step": 612500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.91916085452249e-06, | |
| "loss": 1.0836, | |
| "step": 613000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.884837168433193e-06, | |
| "loss": 1.0704, | |
| "step": 613500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.850513482343897e-06, | |
| "loss": 1.0878, | |
| "step": 614000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.816189796254599e-06, | |
| "loss": 1.1043, | |
| "step": 614500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 7.781866110165303e-06, | |
| "loss": 1.0907, | |
| "step": 615000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.747542424076006e-06, | |
| "loss": 1.0865, | |
| "step": 615500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.71321873798671e-06, | |
| "loss": 1.1002, | |
| "step": 616000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.678895051897414e-06, | |
| "loss": 1.0805, | |
| "step": 616500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.644571365808118e-06, | |
| "loss": 1.0951, | |
| "step": 617000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.610247679718821e-06, | |
| "loss": 1.0911, | |
| "step": 617500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.575923993629525e-06, | |
| "loss": 1.1089, | |
| "step": 618000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.541600307540228e-06, | |
| "loss": 1.0927, | |
| "step": 618500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.5072766214509314e-06, | |
| "loss": 1.0991, | |
| "step": 619000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.472952935361635e-06, | |
| "loss": 1.086, | |
| "step": 619500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.438629249272339e-06, | |
| "loss": 1.0876, | |
| "step": 620000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.4043055631830425e-06, | |
| "loss": 1.0998, | |
| "step": 620500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.369981877093745e-06, | |
| "loss": 1.0976, | |
| "step": 621000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.335658191004449e-06, | |
| "loss": 1.0973, | |
| "step": 621500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.301334504915153e-06, | |
| "loss": 1.0889, | |
| "step": 622000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 7.267010818825856e-06, | |
| "loss": 1.1159, | |
| "step": 622500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.23268713273656e-06, | |
| "loss": 1.1006, | |
| "step": 623000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.198363446647263e-06, | |
| "loss": 1.1102, | |
| "step": 623500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.164039760557967e-06, | |
| "loss": 1.0785, | |
| "step": 624000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.12971607446867e-06, | |
| "loss": 1.0885, | |
| "step": 624500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.095392388379374e-06, | |
| "loss": 1.0783, | |
| "step": 625000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.061068702290078e-06, | |
| "loss": 1.0796, | |
| "step": 625500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 7.0267450162007805e-06, | |
| "loss": 1.0923, | |
| "step": 626000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.992421330111483e-06, | |
| "loss": 1.0983, | |
| "step": 626500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.958097644022186e-06, | |
| "loss": 1.0887, | |
| "step": 627000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.92377395793289e-06, | |
| "loss": 1.0974, | |
| "step": 627500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.8894502718435936e-06, | |
| "loss": 1.0868, | |
| "step": 628000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.855126585754297e-06, | |
| "loss": 1.0818, | |
| "step": 628500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.820802899665001e-06, | |
| "loss": 1.085, | |
| "step": 629000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.786479213575704e-06, | |
| "loss": 1.0858, | |
| "step": 629500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.7521555274864075e-06, | |
| "loss": 1.0874, | |
| "step": 630000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.717831841397111e-06, | |
| "loss": 1.0957, | |
| "step": 630500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.683508155307815e-06, | |
| "loss": 1.0764, | |
| "step": 631000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.6491844692185185e-06, | |
| "loss": 1.1055, | |
| "step": 631500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.614860783129221e-06, | |
| "loss": 1.0771, | |
| "step": 632000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.580537097039925e-06, | |
| "loss": 1.093, | |
| "step": 632500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.546213410950629e-06, | |
| "loss": 1.0797, | |
| "step": 633000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.511889724861332e-06, | |
| "loss": 1.0846, | |
| "step": 633500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.477566038772036e-06, | |
| "loss": 1.0957, | |
| "step": 634000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.443242352682739e-06, | |
| "loss": 1.0838, | |
| "step": 634500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.408918666593443e-06, | |
| "loss": 1.0865, | |
| "step": 635000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.374594980504146e-06, | |
| "loss": 1.0796, | |
| "step": 635500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.34027129441485e-06, | |
| "loss": 1.0867, | |
| "step": 636000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.305947608325554e-06, | |
| "loss": 1.0906, | |
| "step": 636500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 6.2716239222362565e-06, | |
| "loss": 1.0798, | |
| "step": 637000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.23730023614696e-06, | |
| "loss": 1.1051, | |
| "step": 637500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.202976550057664e-06, | |
| "loss": 1.0906, | |
| "step": 638000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.168652863968368e-06, | |
| "loss": 1.0716, | |
| "step": 638500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.134329177879071e-06, | |
| "loss": 1.0777, | |
| "step": 639000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.100005491789774e-06, | |
| "loss": 1.108, | |
| "step": 639500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.065681805700478e-06, | |
| "loss": 1.0964, | |
| "step": 640000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.0313581196111815e-06, | |
| "loss": 1.1027, | |
| "step": 640500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.997034433521885e-06, | |
| "loss": 1.0936, | |
| "step": 641000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.962710747432589e-06, | |
| "loss": 1.0964, | |
| "step": 641500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.928387061343292e-06, | |
| "loss": 1.0898, | |
| "step": 642000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.894063375253995e-06, | |
| "loss": 1.0904, | |
| "step": 642500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.859739689164699e-06, | |
| "loss": 1.0861, | |
| "step": 643000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.825416003075403e-06, | |
| "loss": 1.1034, | |
| "step": 643500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.7910923169861064e-06, | |
| "loss": 1.0817, | |
| "step": 644000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 5.756768630896809e-06, | |
| "loss": 1.0789, | |
| "step": 644500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.722444944807513e-06, | |
| "loss": 1.085, | |
| "step": 645000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.688121258718217e-06, | |
| "loss": 1.088, | |
| "step": 645500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.65379757262892e-06, | |
| "loss": 1.083, | |
| "step": 646000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.619473886539624e-06, | |
| "loss": 1.0947, | |
| "step": 646500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.585150200450327e-06, | |
| "loss": 1.085, | |
| "step": 647000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.5508265143610305e-06, | |
| "loss": 1.0759, | |
| "step": 647500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.516502828271734e-06, | |
| "loss": 1.0871, | |
| "step": 648000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.482179142182438e-06, | |
| "loss": 1.0826, | |
| "step": 648500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.447855456093142e-06, | |
| "loss": 1.0747, | |
| "step": 649000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.4135317700038444e-06, | |
| "loss": 1.0902, | |
| "step": 649500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.379208083914547e-06, | |
| "loss": 1.0796, | |
| "step": 650000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.344884397825251e-06, | |
| "loss": 1.0885, | |
| "step": 650500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.310560711735955e-06, | |
| "loss": 1.0806, | |
| "step": 651000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.276237025646658e-06, | |
| "loss": 1.0777, | |
| "step": 651500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.241913339557362e-06, | |
| "loss": 1.0792, | |
| "step": 652000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.207589653468066e-06, | |
| "loss": 1.0692, | |
| "step": 652500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.1732659673787686e-06, | |
| "loss": 1.086, | |
| "step": 653000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.138942281289472e-06, | |
| "loss": 1.0826, | |
| "step": 653500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.104618595200176e-06, | |
| "loss": 1.0807, | |
| "step": 654000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.07029490911088e-06, | |
| "loss": 1.094, | |
| "step": 654500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.035971223021583e-06, | |
| "loss": 1.0893, | |
| "step": 655000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.001647536932286e-06, | |
| "loss": 1.077, | |
| "step": 655500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.96732385084299e-06, | |
| "loss": 1.0764, | |
| "step": 656000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.9330001647536935e-06, | |
| "loss": 1.083, | |
| "step": 656500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.898676478664397e-06, | |
| "loss": 1.0704, | |
| "step": 657000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.864352792575101e-06, | |
| "loss": 1.0803, | |
| "step": 657500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.830029106485804e-06, | |
| "loss": 1.0949, | |
| "step": 658000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.795705420396507e-06, | |
| "loss": 1.0691, | |
| "step": 658500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.761381734307211e-06, | |
| "loss": 1.067, | |
| "step": 659000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.727058048217915e-06, | |
| "loss": 1.0707, | |
| "step": 659500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.6927343621286185e-06, | |
| "loss": 1.0976, | |
| "step": 660000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.658410676039321e-06, | |
| "loss": 1.0884, | |
| "step": 660500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.624086989950025e-06, | |
| "loss": 1.0792, | |
| "step": 661000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.589763303860729e-06, | |
| "loss": 1.0864, | |
| "step": 661500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.555439617771432e-06, | |
| "loss": 1.0789, | |
| "step": 662000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.521115931682136e-06, | |
| "loss": 1.0789, | |
| "step": 662500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.486792245592839e-06, | |
| "loss": 1.0631, | |
| "step": 663000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.452468559503543e-06, | |
| "loss": 1.0762, | |
| "step": 663500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.418144873414246e-06, | |
| "loss": 1.0925, | |
| "step": 664000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.38382118732495e-06, | |
| "loss": 1.0835, | |
| "step": 664500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.349497501235653e-06, | |
| "loss": 1.0793, | |
| "step": 665000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.3151738151463565e-06, | |
| "loss": 1.0744, | |
| "step": 665500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.280850129057059e-06, | |
| "loss": 1.0862, | |
| "step": 666000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.246526442967763e-06, | |
| "loss": 1.093, | |
| "step": 666500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.212202756878467e-06, | |
| "loss": 1.0831, | |
| "step": 667000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.17787907078917e-06, | |
| "loss": 1.0688, | |
| "step": 667500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.143555384699874e-06, | |
| "loss": 1.083, | |
| "step": 668000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.109231698610577e-06, | |
| "loss": 1.076, | |
| "step": 668500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.074908012521281e-06, | |
| "loss": 1.077, | |
| "step": 669000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.040584326431984e-06, | |
| "loss": 1.0744, | |
| "step": 669500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.006260640342688e-06, | |
| "loss": 1.0867, | |
| "step": 670000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.971936954253392e-06, | |
| "loss": 1.0703, | |
| "step": 670500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.9376132681640945e-06, | |
| "loss": 1.0791, | |
| "step": 671000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.903289582074798e-06, | |
| "loss": 1.0697, | |
| "step": 671500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.868965895985502e-06, | |
| "loss": 1.0797, | |
| "step": 672000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.8346422098962055e-06, | |
| "loss": 1.0853, | |
| "step": 672500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.800318523806909e-06, | |
| "loss": 1.0701, | |
| "step": 673000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.7659948377176125e-06, | |
| "loss": 1.0709, | |
| "step": 673500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.7316711516283158e-06, | |
| "loss": 1.0844, | |
| "step": 674000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.6973474655390194e-06, | |
| "loss": 1.0955, | |
| "step": 674500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.663023779449723e-06, | |
| "loss": 1.0801, | |
| "step": 675000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.6287000933604264e-06, | |
| "loss": 1.0712, | |
| "step": 675500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.59437640727113e-06, | |
| "loss": 1.0934, | |
| "step": 676000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.5600527211818333e-06, | |
| "loss": 1.0748, | |
| "step": 676500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.525729035092537e-06, | |
| "loss": 1.0807, | |
| "step": 677000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.4914053490032407e-06, | |
| "loss": 1.0654, | |
| "step": 677500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.457081662913944e-06, | |
| "loss": 1.0735, | |
| "step": 678000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.4227579768246477e-06, | |
| "loss": 1.0799, | |
| "step": 678500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.388434290735351e-06, | |
| "loss": 1.0725, | |
| "step": 679000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.3541106046460546e-06, | |
| "loss": 1.0753, | |
| "step": 679500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.3197869185567574e-06, | |
| "loss": 1.0753, | |
| "step": 680000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.285463232467461e-06, | |
| "loss": 1.0686, | |
| "step": 680500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.2511395463781644e-06, | |
| "loss": 1.0728, | |
| "step": 681000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.216815860288868e-06, | |
| "loss": 1.0814, | |
| "step": 681500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.1824921741995713e-06, | |
| "loss": 1.0723, | |
| "step": 682000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.148168488110275e-06, | |
| "loss": 1.0686, | |
| "step": 682500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.1138448020209787e-06, | |
| "loss": 1.0765, | |
| "step": 683000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0795211159316824e-06, | |
| "loss": 1.0784, | |
| "step": 683500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0451974298423857e-06, | |
| "loss": 1.0658, | |
| "step": 684000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.010873743753089e-06, | |
| "loss": 1.0706, | |
| "step": 684500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.9765500576637926e-06, | |
| "loss": 1.064, | |
| "step": 685000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.9422263715744963e-06, | |
| "loss": 1.0607, | |
| "step": 685500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.9079026854851996e-06, | |
| "loss": 1.0754, | |
| "step": 686000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.8735789993959033e-06, | |
| "loss": 1.0703, | |
| "step": 686500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.839255313306607e-06, | |
| "loss": 1.075, | |
| "step": 687000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.80493162721731e-06, | |
| "loss": 1.0853, | |
| "step": 687500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.770607941128014e-06, | |
| "loss": 1.0648, | |
| "step": 688000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.736284255038717e-06, | |
| "loss": 1.0669, | |
| "step": 688500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.701960568949421e-06, | |
| "loss": 1.0867, | |
| "step": 689000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.6676368828601245e-06, | |
| "loss": 1.0738, | |
| "step": 689500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.6333131967708278e-06, | |
| "loss": 1.0695, | |
| "step": 690000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.5989895106815315e-06, | |
| "loss": 1.0743, | |
| "step": 690500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.5646658245922347e-06, | |
| "loss": 1.0844, | |
| "step": 691000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.530342138502938e-06, | |
| "loss": 1.0666, | |
| "step": 691500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.4960184524136417e-06, | |
| "loss": 1.0844, | |
| "step": 692000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.461694766324345e-06, | |
| "loss": 1.0772, | |
| "step": 692500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.4273710802350486e-06, | |
| "loss": 1.0795, | |
| "step": 693000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.3930473941457523e-06, | |
| "loss": 1.083, | |
| "step": 693500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.3587237080564556e-06, | |
| "loss": 1.0688, | |
| "step": 694000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.3244000219671593e-06, | |
| "loss": 1.0663, | |
| "step": 694500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.2900763358778625e-06, | |
| "loss": 1.0721, | |
| "step": 695000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.2557526497885662e-06, | |
| "loss": 1.0708, | |
| "step": 695500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.22142896369927e-06, | |
| "loss": 1.0769, | |
| "step": 696000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.187105277609973e-06, | |
| "loss": 1.0713, | |
| "step": 696500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.152781591520677e-06, | |
| "loss": 1.0719, | |
| "step": 697000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.11845790543138e-06, | |
| "loss": 1.0891, | |
| "step": 697500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.084134219342084e-06, | |
| "loss": 1.0772, | |
| "step": 698000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.0498105332527875e-06, | |
| "loss": 1.0737, | |
| "step": 698500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.0154868471634903e-06, | |
| "loss": 1.073, | |
| "step": 699000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.981163161074194e-06, | |
| "loss": 1.0852, | |
| "step": 699500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.9468394749848977e-06, | |
| "loss": 1.0722, | |
| "step": 700000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.912515788895601e-06, | |
| "loss": 1.0753, | |
| "step": 700500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.8781921028063047e-06, | |
| "loss": 1.0826, | |
| "step": 701000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.8438684167170081e-06, | |
| "loss": 1.0649, | |
| "step": 701500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.8095447306277116e-06, | |
| "loss": 1.0634, | |
| "step": 702000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.775221044538415e-06, | |
| "loss": 1.0606, | |
| "step": 702500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.7408973584491188e-06, | |
| "loss": 1.0734, | |
| "step": 703000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.7065736723598222e-06, | |
| "loss": 1.0655, | |
| "step": 703500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.6722499862705257e-06, | |
| "loss": 1.0617, | |
| "step": 704000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.6379263001812292e-06, | |
| "loss": 1.0674, | |
| "step": 704500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.6036026140919327e-06, | |
| "loss": 1.0725, | |
| "step": 705000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5692789280026363e-06, | |
| "loss": 1.0747, | |
| "step": 705500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.5349552419133396e-06, | |
| "loss": 1.0743, | |
| "step": 706000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.500631555824043e-06, | |
| "loss": 1.0667, | |
| "step": 706500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.4663078697347466e-06, | |
| "loss": 1.0732, | |
| "step": 707000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.4319841836454502e-06, | |
| "loss": 1.0785, | |
| "step": 707500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.3976604975561537e-06, | |
| "loss": 1.0849, | |
| "step": 708000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.363336811466857e-06, | |
| "loss": 1.068, | |
| "step": 708500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.3290131253775605e-06, | |
| "loss": 1.0581, | |
| "step": 709000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.2946894392882641e-06, | |
| "loss": 1.0778, | |
| "step": 709500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.2603657531989676e-06, | |
| "loss": 1.0694, | |
| "step": 710000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.226042067109671e-06, | |
| "loss": 1.0748, | |
| "step": 710500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.1917183810203746e-06, | |
| "loss": 1.0765, | |
| "step": 711000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.157394694931078e-06, | |
| "loss": 1.0645, | |
| "step": 711500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.1230710088417817e-06, | |
| "loss": 1.0692, | |
| "step": 712000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.088747322752485e-06, | |
| "loss": 1.0628, | |
| "step": 712500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.0544236366631885e-06, | |
| "loss": 1.0544, | |
| "step": 713000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.0200999505738922e-06, | |
| "loss": 1.068, | |
| "step": 713500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.857762644845956e-07, | |
| "loss": 1.0732, | |
| "step": 714000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.514525783952991e-07, | |
| "loss": 1.0678, | |
| "step": 714500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.171288923060026e-07, | |
| "loss": 1.0572, | |
| "step": 715000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.828052062167062e-07, | |
| "loss": 1.0625, | |
| "step": 715500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.484815201274096e-07, | |
| "loss": 1.0809, | |
| "step": 716000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.14157834038113e-07, | |
| "loss": 1.0675, | |
| "step": 716500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 7.798341479488166e-07, | |
| "loss": 1.0576, | |
| "step": 717000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 7.455104618595201e-07, | |
| "loss": 1.0636, | |
| "step": 717500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 7.111867757702235e-07, | |
| "loss": 1.0804, | |
| "step": 718000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 6.768630896809271e-07, | |
| "loss": 1.0678, | |
| "step": 718500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 6.425394035916305e-07, | |
| "loss": 1.0545, | |
| "step": 719000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 6.082157175023341e-07, | |
| "loss": 1.0657, | |
| "step": 719500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 5.738920314130375e-07, | |
| "loss": 1.0486, | |
| "step": 720000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 5.395683453237411e-07, | |
| "loss": 1.063, | |
| "step": 720500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 5.052446592344445e-07, | |
| "loss": 1.0733, | |
| "step": 721000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.70920973145148e-07, | |
| "loss": 1.0695, | |
| "step": 721500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.3659728705585154e-07, | |
| "loss": 1.0728, | |
| "step": 722000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.0227360096655506e-07, | |
| "loss": 1.0687, | |
| "step": 722500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.6794991487725854e-07, | |
| "loss": 1.0717, | |
| "step": 723000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.33626228787962e-07, | |
| "loss": 1.0643, | |
| "step": 723500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9930254269866554e-07, | |
| "loss": 1.0738, | |
| "step": 724000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.64978856609369e-07, | |
| "loss": 1.0722, | |
| "step": 724500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.3065517052007252e-07, | |
| "loss": 1.0768, | |
| "step": 725000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.96331484430776e-07, | |
| "loss": 1.0689, | |
| "step": 725500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.620077983414795e-07, | |
| "loss": 1.0709, | |
| "step": 726000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.27684112252183e-07, | |
| "loss": 1.0672, | |
| "step": 726500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.336042616288648e-08, | |
| "loss": 1.0662, | |
| "step": 727000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 5.903674007358998e-08, | |
| "loss": 1.0778, | |
| "step": 727500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.4713053984293483e-08, | |
| "loss": 1.0693, | |
| "step": 728000 | |
| } | |
| ], | |
| "max_steps": 728360, | |
| "num_train_epochs": 1, | |
| "total_flos": 3.0658096594944e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |