| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 20.0, |
| "eval_steps": 2.0, |
| "global_step": 40920, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.9987781036168134e-05, |
| "loss": 1.0996, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.9975562072336266e-05, |
| "loss": 1.0049, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.99633431085044e-05, |
| "loss": 0.8837, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9951124144672535e-05, |
| "loss": 0.7945, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9938905180840667e-05, |
| "loss": 0.7541, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.99266862170088e-05, |
| "loss": 0.694, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9914467253176932e-05, |
| "loss": 0.6936, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9902248289345064e-05, |
| "loss": 0.6625, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9890029325513197e-05, |
| "loss": 0.661, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9877810361681332e-05, |
| "loss": 0.6525, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9865591397849465e-05, |
| "loss": 0.6298, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9853372434017597e-05, |
| "loss": 0.6111, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.984115347018573e-05, |
| "loss": 0.62, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9828934506353862e-05, |
| "loss": 0.611, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9816715542521994e-05, |
| "loss": 0.5751, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.980449657869013e-05, |
| "loss": 0.6004, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9792277614858263e-05, |
| "loss": 0.5816, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9780058651026395e-05, |
| "loss": 0.5813, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9767839687194527e-05, |
| "loss": 0.5784, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.975562072336266e-05, |
| "loss": 0.5869, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9743401759530792e-05, |
| "loss": 0.5793, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9731182795698928e-05, |
| "loss": 0.5782, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.971896383186706e-05, |
| "loss": 0.5638, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9706744868035193e-05, |
| "loss": 0.5645, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9694525904203325e-05, |
| "loss": 0.552, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9682306940371458e-05, |
| "loss": 0.539, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.967008797653959e-05, |
| "loss": 0.5464, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9657869012707726e-05, |
| "loss": 0.538, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9645650048875858e-05, |
| "loss": 0.5394, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.963343108504399e-05, |
| "loss": 0.5586, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9621212121212123e-05, |
| "loss": 0.5241, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9608993157380255e-05, |
| "loss": 0.5304, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9596774193548388e-05, |
| "loss": 0.5313, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9584555229716524e-05, |
| "loss": 0.5423, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9572336265884656e-05, |
| "loss": 0.5086, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.956011730205279e-05, |
| "loss": 0.5172, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.954789833822092e-05, |
| "loss": 0.5081, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9535679374389053e-05, |
| "loss": 0.5244, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9523460410557186e-05, |
| "loss": 0.5371, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.951124144672532e-05, |
| "loss": 0.5278, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9499022482893454e-05, |
| "loss": 0.5325, |
| "step": 1025 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9486803519061586e-05, |
| "loss": 0.5019, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.947458455522972e-05, |
| "loss": 0.504, |
| "step": 1075 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.946236559139785e-05, |
| "loss": 0.5225, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.9450146627565983e-05, |
| "loss": 0.5232, |
| "step": 1125 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.9437927663734116e-05, |
| "loss": 0.5282, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.942570869990225e-05, |
| "loss": 0.514, |
| "step": 1175 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.9413489736070384e-05, |
| "loss": 0.4916, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.9401270772238516e-05, |
| "loss": 0.5079, |
| "step": 1225 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.938905180840665e-05, |
| "loss": 0.5221, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.937683284457478e-05, |
| "loss": 0.5097, |
| "step": 1275 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.9364613880742914e-05, |
| "loss": 0.512, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.935239491691105e-05, |
| "loss": 0.5001, |
| "step": 1325 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.9340175953079182e-05, |
| "loss": 0.5021, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.9327956989247314e-05, |
| "loss": 0.5044, |
| "step": 1375 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.9315738025415447e-05, |
| "loss": 0.4943, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.930351906158358e-05, |
| "loss": 0.4935, |
| "step": 1425 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.929130009775171e-05, |
| "loss": 0.5144, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.9279081133919847e-05, |
| "loss": 0.506, |
| "step": 1475 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.926686217008798e-05, |
| "loss": 0.5079, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.9254643206256112e-05, |
| "loss": 0.4939, |
| "step": 1525 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.9242424242424244e-05, |
| "loss": 0.4874, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.9230205278592377e-05, |
| "loss": 0.4994, |
| "step": 1575 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.921798631476051e-05, |
| "loss": 0.4818, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.9205767350928645e-05, |
| "loss": 0.5026, |
| "step": 1625 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.9193548387096777e-05, |
| "loss": 0.4634, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.918132942326491e-05, |
| "loss": 0.4824, |
| "step": 1675 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.9169110459433042e-05, |
| "loss": 0.472, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.9156891495601175e-05, |
| "loss": 0.4811, |
| "step": 1725 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.9144672531769307e-05, |
| "loss": 0.4763, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.9132453567937443e-05, |
| "loss": 0.4809, |
| "step": 1775 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.9120234604105575e-05, |
| "loss": 0.5, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.9108015640273708e-05, |
| "loss": 0.4846, |
| "step": 1825 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.909579667644184e-05, |
| "loss": 0.4624, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.9083577712609972e-05, |
| "loss": 0.4824, |
| "step": 1875 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.9071358748778105e-05, |
| "loss": 0.4708, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.905913978494624e-05, |
| "loss": 0.4884, |
| "step": 1925 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.9046920821114373e-05, |
| "loss": 0.4768, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.9034701857282505e-05, |
| "loss": 0.4583, |
| "step": 1975 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.9022482893450638e-05, |
| "loss": 0.4634, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.901026392961877e-05, |
| "loss": 0.4741, |
| "step": 2025 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.8998044965786903e-05, |
| "loss": 0.458, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.8985826001955038e-05, |
| "loss": 0.3961, |
| "step": 2075 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.897360703812317e-05, |
| "loss": 0.3929, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.8961388074291303e-05, |
| "loss": 0.4265, |
| "step": 2125 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.8949169110459435e-05, |
| "loss": 0.4049, |
| "step": 2150 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.8936950146627568e-05, |
| "loss": 0.4171, |
| "step": 2175 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.89247311827957e-05, |
| "loss": 0.4024, |
| "step": 2200 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.8912512218963833e-05, |
| "loss": 0.403, |
| "step": 2225 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.890029325513197e-05, |
| "loss": 0.4105, |
| "step": 2250 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.88880742913001e-05, |
| "loss": 0.3956, |
| "step": 2275 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.8875855327468233e-05, |
| "loss": 0.397, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.8863636363636366e-05, |
| "loss": 0.3883, |
| "step": 2325 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.8851417399804498e-05, |
| "loss": 0.4056, |
| "step": 2350 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.883919843597263e-05, |
| "loss": 0.3884, |
| "step": 2375 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.8826979472140766e-05, |
| "loss": 0.4129, |
| "step": 2400 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.88147605083089e-05, |
| "loss": 0.4097, |
| "step": 2425 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.880254154447703e-05, |
| "loss": 0.4036, |
| "step": 2450 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.8790322580645163e-05, |
| "loss": 0.3894, |
| "step": 2475 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.8778103616813296e-05, |
| "loss": 0.4147, |
| "step": 2500 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.8765884652981428e-05, |
| "loss": 0.409, |
| "step": 2525 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.8753665689149564e-05, |
| "loss": 0.4106, |
| "step": 2550 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.8741446725317693e-05, |
| "loss": 0.3971, |
| "step": 2575 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.8729227761485825e-05, |
| "loss": 0.3995, |
| "step": 2600 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.8717008797653958e-05, |
| "loss": 0.3947, |
| "step": 2625 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.870527859237537e-05, |
| "loss": 0.4141, |
| "step": 2650 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.8693059628543502e-05, |
| "loss": 0.4024, |
| "step": 2675 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.8680840664711635e-05, |
| "loss": 0.4074, |
| "step": 2700 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.8668621700879767e-05, |
| "loss": 0.3953, |
| "step": 2725 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.86564027370479e-05, |
| "loss": 0.3861, |
| "step": 2750 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.8644183773216032e-05, |
| "loss": 0.3939, |
| "step": 2775 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.8631964809384168e-05, |
| "loss": 0.4088, |
| "step": 2800 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.86197458455523e-05, |
| "loss": 0.4089, |
| "step": 2825 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.8607526881720433e-05, |
| "loss": 0.3987, |
| "step": 2850 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.8595307917888565e-05, |
| "loss": 0.3881, |
| "step": 2875 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.8583088954056697e-05, |
| "loss": 0.3877, |
| "step": 2900 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.857086999022483e-05, |
| "loss": 0.3996, |
| "step": 2925 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.8558651026392966e-05, |
| "loss": 0.3979, |
| "step": 2950 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.8546432062561098e-05, |
| "loss": 0.4009, |
| "step": 2975 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.853421309872923e-05, |
| "loss": 0.4069, |
| "step": 3000 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.8521994134897363e-05, |
| "loss": 0.3725, |
| "step": 3025 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.8509775171065495e-05, |
| "loss": 0.3899, |
| "step": 3050 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.8497556207233628e-05, |
| "loss": 0.3916, |
| "step": 3075 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.8485337243401763e-05, |
| "loss": 0.4114, |
| "step": 3100 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.8473118279569896e-05, |
| "loss": 0.3945, |
| "step": 3125 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.8460899315738028e-05, |
| "loss": 0.3984, |
| "step": 3150 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 1.844868035190616e-05, |
| "loss": 0.3871, |
| "step": 3175 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 1.8436461388074293e-05, |
| "loss": 0.3859, |
| "step": 3200 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 1.8424242424242425e-05, |
| "loss": 0.3926, |
| "step": 3225 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 1.841202346041056e-05, |
| "loss": 0.3941, |
| "step": 3250 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 1.8399804496578694e-05, |
| "loss": 0.3922, |
| "step": 3275 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 1.8387585532746826e-05, |
| "loss": 0.3847, |
| "step": 3300 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 1.837536656891496e-05, |
| "loss": 0.3709, |
| "step": 3325 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 1.836314760508309e-05, |
| "loss": 0.3963, |
| "step": 3350 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 1.8350928641251223e-05, |
| "loss": 0.3639, |
| "step": 3375 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 1.833870967741936e-05, |
| "loss": 0.4072, |
| "step": 3400 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 1.832649071358749e-05, |
| "loss": 0.3918, |
| "step": 3425 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 1.8314271749755624e-05, |
| "loss": 0.3847, |
| "step": 3450 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 1.8302052785923756e-05, |
| "loss": 0.393, |
| "step": 3475 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 1.828983382209189e-05, |
| "loss": 0.4028, |
| "step": 3500 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 1.827761485826002e-05, |
| "loss": 0.3967, |
| "step": 3525 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 1.8265395894428157e-05, |
| "loss": 0.3993, |
| "step": 3550 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 1.825317693059629e-05, |
| "loss": 0.3854, |
| "step": 3575 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 1.8240957966764418e-05, |
| "loss": 0.3644, |
| "step": 3600 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 1.822873900293255e-05, |
| "loss": 0.4081, |
| "step": 3625 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 1.8216520039100686e-05, |
| "loss": 0.3836, |
| "step": 3650 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 1.820430107526882e-05, |
| "loss": 0.3927, |
| "step": 3675 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 1.819208211143695e-05, |
| "loss": 0.4068, |
| "step": 3700 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 1.8179863147605084e-05, |
| "loss": 0.4027, |
| "step": 3725 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 1.8167644183773216e-05, |
| "loss": 0.3736, |
| "step": 3750 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 1.815542521994135e-05, |
| "loss": 0.3916, |
| "step": 3775 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 1.814320625610948e-05, |
| "loss": 0.3872, |
| "step": 3800 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 1.8130987292277617e-05, |
| "loss": 0.3867, |
| "step": 3825 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 1.811876832844575e-05, |
| "loss": 0.3879, |
| "step": 3850 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 1.810654936461388e-05, |
| "loss": 0.4013, |
| "step": 3875 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 1.8094330400782014e-05, |
| "loss": 0.3964, |
| "step": 3900 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 1.8082111436950146e-05, |
| "loss": 0.4086, |
| "step": 3925 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 1.806989247311828e-05, |
| "loss": 0.3772, |
| "step": 3950 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 1.8057673509286414e-05, |
| "loss": 0.3886, |
| "step": 3975 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 1.8045454545454547e-05, |
| "loss": 0.3952, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 1.803323558162268e-05, |
| "loss": 0.3962, |
| "step": 4025 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 1.802101661779081e-05, |
| "loss": 0.3772, |
| "step": 4050 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 1.8008797653958944e-05, |
| "loss": 0.3781, |
| "step": 4075 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 1.7996578690127076e-05, |
| "loss": 0.3505, |
| "step": 4100 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 1.7984359726295212e-05, |
| "loss": 0.3105, |
| "step": 4125 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 1.7972140762463344e-05, |
| "loss": 0.3035, |
| "step": 4150 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 1.7959921798631477e-05, |
| "loss": 0.3081, |
| "step": 4175 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 1.794770283479961e-05, |
| "loss": 0.3007, |
| "step": 4200 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 1.7935483870967742e-05, |
| "loss": 0.3166, |
| "step": 4225 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 1.7923264907135874e-05, |
| "loss": 0.3119, |
| "step": 4250 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 1.791104594330401e-05, |
| "loss": 0.2894, |
| "step": 4275 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 1.7898826979472142e-05, |
| "loss": 0.3103, |
| "step": 4300 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 1.7886608015640275e-05, |
| "loss": 0.3049, |
| "step": 4325 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 1.7874389051808407e-05, |
| "loss": 0.2996, |
| "step": 4350 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 1.786217008797654e-05, |
| "loss": 0.3012, |
| "step": 4375 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 1.7849951124144672e-05, |
| "loss": 0.2915, |
| "step": 4400 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 1.7837732160312808e-05, |
| "loss": 0.3193, |
| "step": 4425 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 1.782551319648094e-05, |
| "loss": 0.3132, |
| "step": 4450 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 1.7813294232649072e-05, |
| "loss": 0.3209, |
| "step": 4475 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 1.7801075268817205e-05, |
| "loss": 0.2968, |
| "step": 4500 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 1.7788856304985337e-05, |
| "loss": 0.3092, |
| "step": 4525 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 1.777663734115347e-05, |
| "loss": 0.2926, |
| "step": 4550 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 1.7764418377321605e-05, |
| "loss": 0.3175, |
| "step": 4575 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 1.7752199413489738e-05, |
| "loss": 0.3062, |
| "step": 4600 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 1.773998044965787e-05, |
| "loss": 0.3107, |
| "step": 4625 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 1.7727761485826003e-05, |
| "loss": 0.3072, |
| "step": 4650 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 1.7715542521994135e-05, |
| "loss": 0.3166, |
| "step": 4675 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 1.7703323558162267e-05, |
| "loss": 0.3181, |
| "step": 4700 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 1.7691104594330403e-05, |
| "loss": 0.2843, |
| "step": 4725 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 1.7678885630498536e-05, |
| "loss": 0.3103, |
| "step": 4750 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 1.7666666666666668e-05, |
| "loss": 0.3039, |
| "step": 4775 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 1.76544477028348e-05, |
| "loss": 0.3099, |
| "step": 4800 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 1.7642228739002933e-05, |
| "loss": 0.2989, |
| "step": 4825 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 1.7630009775171065e-05, |
| "loss": 0.3132, |
| "step": 4850 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 1.7617790811339198e-05, |
| "loss": 0.2937, |
| "step": 4875 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 1.7605571847507333e-05, |
| "loss": 0.3067, |
| "step": 4900 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.7593352883675466e-05, |
| "loss": 0.3163, |
| "step": 4925 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.7581133919843598e-05, |
| "loss": 0.3078, |
| "step": 4950 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.756891495601173e-05, |
| "loss": 0.3114, |
| "step": 4975 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7556695992179863e-05, |
| "loss": 0.3121, |
| "step": 5000 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.7544965786901272e-05, |
| "loss": 0.3018, |
| "step": 5025 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.7532746823069404e-05, |
| "loss": 0.3122, |
| "step": 5050 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.7520527859237537e-05, |
| "loss": 0.3052, |
| "step": 5075 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.750830889540567e-05, |
| "loss": 0.3206, |
| "step": 5100 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.7496089931573805e-05, |
| "loss": 0.3136, |
| "step": 5125 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.7483870967741937e-05, |
| "loss": 0.316, |
| "step": 5150 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.747165200391007e-05, |
| "loss": 0.3142, |
| "step": 5175 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.7459433040078202e-05, |
| "loss": 0.3104, |
| "step": 5200 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.7447214076246334e-05, |
| "loss": 0.3, |
| "step": 5225 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.7434995112414467e-05, |
| "loss": 0.3027, |
| "step": 5250 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.74227761485826e-05, |
| "loss": 0.3097, |
| "step": 5275 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 1.7410557184750735e-05, |
| "loss": 0.314, |
| "step": 5300 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 1.7398338220918867e-05, |
| "loss": 0.3252, |
| "step": 5325 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 1.7386119257087e-05, |
| "loss": 0.3158, |
| "step": 5350 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 1.7373900293255132e-05, |
| "loss": 0.3041, |
| "step": 5375 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 1.7361681329423265e-05, |
| "loss": 0.3214, |
| "step": 5400 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 1.7349462365591397e-05, |
| "loss": 0.3352, |
| "step": 5425 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 1.7337243401759533e-05, |
| "loss": 0.3188, |
| "step": 5450 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 1.7325024437927665e-05, |
| "loss": 0.3087, |
| "step": 5475 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 1.7312805474095798e-05, |
| "loss": 0.3064, |
| "step": 5500 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 1.730058651026393e-05, |
| "loss": 0.3153, |
| "step": 5525 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 1.7288367546432062e-05, |
| "loss": 0.3084, |
| "step": 5550 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 1.7276148582600195e-05, |
| "loss": 0.292, |
| "step": 5575 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 1.726392961876833e-05, |
| "loss": 0.3167, |
| "step": 5600 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 1.7251710654936463e-05, |
| "loss": 0.3141, |
| "step": 5625 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 1.7239491691104595e-05, |
| "loss": 0.2913, |
| "step": 5650 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 1.7227272727272728e-05, |
| "loss": 0.3279, |
| "step": 5675 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 1.721505376344086e-05, |
| "loss": 0.3225, |
| "step": 5700 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 1.7202834799608993e-05, |
| "loss": 0.306, |
| "step": 5725 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 1.719061583577713e-05, |
| "loss": 0.3322, |
| "step": 5750 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.717839687194526e-05, |
| "loss": 0.3102, |
| "step": 5775 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.7166177908113393e-05, |
| "loss": 0.3123, |
| "step": 5800 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.7153958944281526e-05, |
| "loss": 0.3077, |
| "step": 5825 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.7141739980449658e-05, |
| "loss": 0.3223, |
| "step": 5850 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 1.712952101661779e-05, |
| "loss": 0.3178, |
| "step": 5875 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 1.7117302052785926e-05, |
| "loss": 0.3097, |
| "step": 5900 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 1.710508308895406e-05, |
| "loss": 0.3047, |
| "step": 5925 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 1.709286412512219e-05, |
| "loss": 0.3316, |
| "step": 5950 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 1.7080645161290323e-05, |
| "loss": 0.3107, |
| "step": 5975 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 1.7068426197458456e-05, |
| "loss": 0.3037, |
| "step": 6000 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 1.7056207233626588e-05, |
| "loss": 0.3075, |
| "step": 6025 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 1.7043988269794724e-05, |
| "loss": 0.3249, |
| "step": 6050 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 1.7031769305962856e-05, |
| "loss": 0.3078, |
| "step": 6075 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.701955034213099e-05, |
| "loss": 0.3015, |
| "step": 6100 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 1.700733137829912e-05, |
| "loss": 0.3186, |
| "step": 6125 |
| }, |
| { |
| "epoch": 3.01, |
| "learning_rate": 1.6995112414467254e-05, |
| "loss": 0.2948, |
| "step": 6150 |
| }, |
| { |
| "epoch": 3.02, |
| "learning_rate": 1.6982893450635386e-05, |
| "loss": 0.2312, |
| "step": 6175 |
| }, |
| { |
| "epoch": 3.03, |
| "learning_rate": 1.697067448680352e-05, |
| "loss": 0.2379, |
| "step": 6200 |
| }, |
| { |
| "epoch": 3.04, |
| "learning_rate": 1.6958455522971654e-05, |
| "loss": 0.24, |
| "step": 6225 |
| }, |
| { |
| "epoch": 3.05, |
| "learning_rate": 1.6946236559139786e-05, |
| "loss": 0.2301, |
| "step": 6250 |
| }, |
| { |
| "epoch": 3.07, |
| "learning_rate": 1.693401759530792e-05, |
| "loss": 0.2257, |
| "step": 6275 |
| }, |
| { |
| "epoch": 3.08, |
| "learning_rate": 1.692179863147605e-05, |
| "loss": 0.2318, |
| "step": 6300 |
| }, |
| { |
| "epoch": 3.09, |
| "learning_rate": 1.6909579667644184e-05, |
| "loss": 0.2417, |
| "step": 6325 |
| }, |
| { |
| "epoch": 3.1, |
| "learning_rate": 1.6897360703812316e-05, |
| "loss": 0.2334, |
| "step": 6350 |
| }, |
| { |
| "epoch": 3.12, |
| "learning_rate": 1.6885141739980452e-05, |
| "loss": 0.2507, |
| "step": 6375 |
| }, |
| { |
| "epoch": 3.13, |
| "learning_rate": 1.6872922776148584e-05, |
| "loss": 0.2401, |
| "step": 6400 |
| }, |
| { |
| "epoch": 3.14, |
| "learning_rate": 1.6860703812316717e-05, |
| "loss": 0.2264, |
| "step": 6425 |
| }, |
| { |
| "epoch": 3.15, |
| "learning_rate": 1.684848484848485e-05, |
| "loss": 0.2395, |
| "step": 6450 |
| }, |
| { |
| "epoch": 3.16, |
| "learning_rate": 1.683626588465298e-05, |
| "loss": 0.2282, |
| "step": 6475 |
| }, |
| { |
| "epoch": 3.18, |
| "learning_rate": 1.6824046920821114e-05, |
| "loss": 0.237, |
| "step": 6500 |
| }, |
| { |
| "epoch": 3.19, |
| "learning_rate": 1.681182795698925e-05, |
| "loss": 0.2148, |
| "step": 6525 |
| }, |
| { |
| "epoch": 3.2, |
| "learning_rate": 1.6799608993157382e-05, |
| "loss": 0.2392, |
| "step": 6550 |
| }, |
| { |
| "epoch": 3.21, |
| "learning_rate": 1.6787390029325514e-05, |
| "loss": 0.2331, |
| "step": 6575 |
| }, |
| { |
| "epoch": 3.23, |
| "learning_rate": 1.6775171065493647e-05, |
| "loss": 0.2397, |
| "step": 6600 |
| }, |
| { |
| "epoch": 3.24, |
| "learning_rate": 1.676295210166178e-05, |
| "loss": 0.2398, |
| "step": 6625 |
| }, |
| { |
| "epoch": 3.25, |
| "learning_rate": 1.675073313782991e-05, |
| "loss": 0.2326, |
| "step": 6650 |
| }, |
| { |
| "epoch": 3.26, |
| "learning_rate": 1.6738514173998047e-05, |
| "loss": 0.2549, |
| "step": 6675 |
| }, |
| { |
| "epoch": 3.27, |
| "learning_rate": 1.672629521016618e-05, |
| "loss": 0.2261, |
| "step": 6700 |
| }, |
| { |
| "epoch": 3.29, |
| "learning_rate": 1.6714076246334312e-05, |
| "loss": 0.2366, |
| "step": 6725 |
| }, |
| { |
| "epoch": 3.3, |
| "learning_rate": 1.6701857282502445e-05, |
| "loss": 0.2504, |
| "step": 6750 |
| }, |
| { |
| "epoch": 3.31, |
| "learning_rate": 1.6689638318670577e-05, |
| "loss": 0.2381, |
| "step": 6775 |
| }, |
| { |
| "epoch": 3.32, |
| "learning_rate": 1.667741935483871e-05, |
| "loss": 0.2503, |
| "step": 6800 |
| }, |
| { |
| "epoch": 3.34, |
| "learning_rate": 1.6665200391006845e-05, |
| "loss": 0.2169, |
| "step": 6825 |
| }, |
| { |
| "epoch": 3.35, |
| "learning_rate": 1.6652981427174978e-05, |
| "loss": 0.2441, |
| "step": 6850 |
| }, |
| { |
| "epoch": 3.36, |
| "learning_rate": 1.664076246334311e-05, |
| "loss": 0.2348, |
| "step": 6875 |
| }, |
| { |
| "epoch": 3.37, |
| "learning_rate": 1.6628543499511242e-05, |
| "loss": 0.2479, |
| "step": 6900 |
| }, |
| { |
| "epoch": 3.38, |
| "learning_rate": 1.6616324535679375e-05, |
| "loss": 0.2479, |
| "step": 6925 |
| }, |
| { |
| "epoch": 3.4, |
| "learning_rate": 1.6604105571847507e-05, |
| "loss": 0.258, |
| "step": 6950 |
| }, |
| { |
| "epoch": 3.41, |
| "learning_rate": 1.6591886608015643e-05, |
| "loss": 0.2269, |
| "step": 6975 |
| }, |
| { |
| "epoch": 3.42, |
| "learning_rate": 1.6579667644183775e-05, |
| "loss": 0.2448, |
| "step": 7000 |
| }, |
| { |
| "epoch": 3.43, |
| "learning_rate": 1.6567448680351908e-05, |
| "loss": 0.2293, |
| "step": 7025 |
| }, |
| { |
| "epoch": 3.45, |
| "learning_rate": 1.655522971652004e-05, |
| "loss": 0.2407, |
| "step": 7050 |
| }, |
| { |
| "epoch": 3.46, |
| "learning_rate": 1.6543010752688173e-05, |
| "loss": 0.2318, |
| "step": 7075 |
| }, |
| { |
| "epoch": 3.47, |
| "learning_rate": 1.6530791788856305e-05, |
| "loss": 0.2363, |
| "step": 7100 |
| }, |
| { |
| "epoch": 3.48, |
| "learning_rate": 1.651857282502444e-05, |
| "loss": 0.2474, |
| "step": 7125 |
| }, |
| { |
| "epoch": 3.49, |
| "learning_rate": 1.6506353861192573e-05, |
| "loss": 0.2394, |
| "step": 7150 |
| }, |
| { |
| "epoch": 3.51, |
| "learning_rate": 1.6494134897360706e-05, |
| "loss": 0.2434, |
| "step": 7175 |
| }, |
| { |
| "epoch": 3.52, |
| "learning_rate": 1.6481915933528838e-05, |
| "loss": 0.2403, |
| "step": 7200 |
| }, |
| { |
| "epoch": 3.53, |
| "learning_rate": 1.646969696969697e-05, |
| "loss": 0.2456, |
| "step": 7225 |
| }, |
| { |
| "epoch": 3.54, |
| "learning_rate": 1.6457478005865103e-05, |
| "loss": 0.2441, |
| "step": 7250 |
| }, |
| { |
| "epoch": 3.56, |
| "learning_rate": 1.644525904203324e-05, |
| "loss": 0.2332, |
| "step": 7275 |
| }, |
| { |
| "epoch": 3.57, |
| "learning_rate": 1.643304007820137e-05, |
| "loss": 0.23, |
| "step": 7300 |
| }, |
| { |
| "epoch": 3.58, |
| "learning_rate": 1.6420821114369503e-05, |
| "loss": 0.2293, |
| "step": 7325 |
| }, |
| { |
| "epoch": 3.59, |
| "learning_rate": 1.640909090909091e-05, |
| "loss": 0.2224, |
| "step": 7350 |
| }, |
| { |
| "epoch": 3.6, |
| "learning_rate": 1.6396871945259045e-05, |
| "loss": 0.2324, |
| "step": 7375 |
| }, |
| { |
| "epoch": 3.62, |
| "learning_rate": 1.6384652981427177e-05, |
| "loss": 0.2409, |
| "step": 7400 |
| }, |
| { |
| "epoch": 3.63, |
| "learning_rate": 1.637243401759531e-05, |
| "loss": 0.2462, |
| "step": 7425 |
| }, |
| { |
| "epoch": 3.64, |
| "learning_rate": 1.6360215053763442e-05, |
| "loss": 0.2375, |
| "step": 7450 |
| }, |
| { |
| "epoch": 3.65, |
| "learning_rate": 1.6347996089931574e-05, |
| "loss": 0.2445, |
| "step": 7475 |
| }, |
| { |
| "epoch": 3.67, |
| "learning_rate": 1.6335777126099707e-05, |
| "loss": 0.2364, |
| "step": 7500 |
| }, |
| { |
| "epoch": 3.68, |
| "learning_rate": 1.6323558162267842e-05, |
| "loss": 0.2374, |
| "step": 7525 |
| }, |
| { |
| "epoch": 3.69, |
| "learning_rate": 1.6311339198435975e-05, |
| "loss": 0.2383, |
| "step": 7550 |
| }, |
| { |
| "epoch": 3.7, |
| "learning_rate": 1.6299120234604107e-05, |
| "loss": 0.2312, |
| "step": 7575 |
| }, |
| { |
| "epoch": 3.71, |
| "learning_rate": 1.628690127077224e-05, |
| "loss": 0.2475, |
| "step": 7600 |
| }, |
| { |
| "epoch": 3.73, |
| "learning_rate": 1.6274682306940372e-05, |
| "loss": 0.2552, |
| "step": 7625 |
| }, |
| { |
| "epoch": 3.74, |
| "learning_rate": 1.6262463343108504e-05, |
| "loss": 0.2507, |
| "step": 7650 |
| }, |
| { |
| "epoch": 3.75, |
| "learning_rate": 1.625024437927664e-05, |
| "loss": 0.248, |
| "step": 7675 |
| }, |
| { |
| "epoch": 3.76, |
| "learning_rate": 1.6238025415444773e-05, |
| "loss": 0.2333, |
| "step": 7700 |
| }, |
| { |
| "epoch": 3.78, |
| "learning_rate": 1.6225806451612905e-05, |
| "loss": 0.2424, |
| "step": 7725 |
| }, |
| { |
| "epoch": 3.79, |
| "learning_rate": 1.6213587487781037e-05, |
| "loss": 0.2565, |
| "step": 7750 |
| }, |
| { |
| "epoch": 3.8, |
| "learning_rate": 1.620136852394917e-05, |
| "loss": 0.2568, |
| "step": 7775 |
| }, |
| { |
| "epoch": 3.81, |
| "learning_rate": 1.6189149560117302e-05, |
| "loss": 0.2409, |
| "step": 7800 |
| }, |
| { |
| "epoch": 3.82, |
| "learning_rate": 1.6176930596285435e-05, |
| "loss": 0.2444, |
| "step": 7825 |
| }, |
| { |
| "epoch": 3.84, |
| "learning_rate": 1.616471163245357e-05, |
| "loss": 0.243, |
| "step": 7850 |
| }, |
| { |
| "epoch": 3.85, |
| "learning_rate": 1.6152492668621703e-05, |
| "loss": 0.2417, |
| "step": 7875 |
| }, |
| { |
| "epoch": 3.86, |
| "learning_rate": 1.6140273704789835e-05, |
| "loss": 0.2246, |
| "step": 7900 |
| }, |
| { |
| "epoch": 3.87, |
| "learning_rate": 1.6128054740957968e-05, |
| "loss": 0.2533, |
| "step": 7925 |
| }, |
| { |
| "epoch": 3.89, |
| "learning_rate": 1.61158357771261e-05, |
| "loss": 0.2508, |
| "step": 7950 |
| }, |
| { |
| "epoch": 3.9, |
| "learning_rate": 1.6103616813294232e-05, |
| "loss": 0.2384, |
| "step": 7975 |
| }, |
| { |
| "epoch": 3.91, |
| "learning_rate": 1.6091397849462368e-05, |
| "loss": 0.2531, |
| "step": 8000 |
| }, |
| { |
| "epoch": 3.92, |
| "learning_rate": 1.60791788856305e-05, |
| "loss": 0.2448, |
| "step": 8025 |
| }, |
| { |
| "epoch": 3.93, |
| "learning_rate": 1.6066959921798633e-05, |
| "loss": 0.241, |
| "step": 8050 |
| }, |
| { |
| "epoch": 3.95, |
| "learning_rate": 1.6054740957966765e-05, |
| "loss": 0.2566, |
| "step": 8075 |
| }, |
| { |
| "epoch": 3.96, |
| "learning_rate": 1.6042521994134898e-05, |
| "loss": 0.2297, |
| "step": 8100 |
| }, |
| { |
| "epoch": 3.97, |
| "learning_rate": 1.603030303030303e-05, |
| "loss": 0.2446, |
| "step": 8125 |
| }, |
| { |
| "epoch": 3.98, |
| "learning_rate": 1.6018084066471166e-05, |
| "loss": 0.2305, |
| "step": 8150 |
| }, |
| { |
| "epoch": 4.0, |
| "learning_rate": 1.6005865102639298e-05, |
| "loss": 0.2438, |
| "step": 8175 |
| }, |
| { |
| "epoch": 4.01, |
| "learning_rate": 1.599364613880743e-05, |
| "loss": 0.2059, |
| "step": 8200 |
| }, |
| { |
| "epoch": 4.02, |
| "learning_rate": 1.5981427174975563e-05, |
| "loss": 0.1706, |
| "step": 8225 |
| }, |
| { |
| "epoch": 4.03, |
| "learning_rate": 1.5969208211143695e-05, |
| "loss": 0.1822, |
| "step": 8250 |
| }, |
| { |
| "epoch": 4.04, |
| "learning_rate": 1.5956989247311828e-05, |
| "loss": 0.1947, |
| "step": 8275 |
| }, |
| { |
| "epoch": 4.06, |
| "learning_rate": 1.5944770283479964e-05, |
| "loss": 0.1714, |
| "step": 8300 |
| }, |
| { |
| "epoch": 4.07, |
| "learning_rate": 1.5932551319648096e-05, |
| "loss": 0.1776, |
| "step": 8325 |
| }, |
| { |
| "epoch": 4.08, |
| "learning_rate": 1.592033235581623e-05, |
| "loss": 0.1974, |
| "step": 8350 |
| }, |
| { |
| "epoch": 4.09, |
| "learning_rate": 1.590811339198436e-05, |
| "loss": 0.1898, |
| "step": 8375 |
| }, |
| { |
| "epoch": 4.11, |
| "learning_rate": 1.5895894428152493e-05, |
| "loss": 0.1747, |
| "step": 8400 |
| }, |
| { |
| "epoch": 4.12, |
| "learning_rate": 1.5883675464320626e-05, |
| "loss": 0.1692, |
| "step": 8425 |
| }, |
| { |
| "epoch": 4.13, |
| "learning_rate": 1.587145650048876e-05, |
| "loss": 0.1773, |
| "step": 8450 |
| }, |
| { |
| "epoch": 4.14, |
| "learning_rate": 1.5859237536656894e-05, |
| "loss": 0.1605, |
| "step": 8475 |
| }, |
| { |
| "epoch": 4.15, |
| "learning_rate": 1.5847018572825026e-05, |
| "loss": 0.1773, |
| "step": 8500 |
| }, |
| { |
| "epoch": 4.17, |
| "learning_rate": 1.583479960899316e-05, |
| "loss": 0.1778, |
| "step": 8525 |
| }, |
| { |
| "epoch": 4.18, |
| "learning_rate": 1.582258064516129e-05, |
| "loss": 0.1865, |
| "step": 8550 |
| }, |
| { |
| "epoch": 4.19, |
| "learning_rate": 1.5810361681329423e-05, |
| "loss": 0.1844, |
| "step": 8575 |
| }, |
| { |
| "epoch": 4.2, |
| "learning_rate": 1.579814271749756e-05, |
| "loss": 0.1694, |
| "step": 8600 |
| }, |
| { |
| "epoch": 4.22, |
| "learning_rate": 1.578592375366569e-05, |
| "loss": 0.1813, |
| "step": 8625 |
| }, |
| { |
| "epoch": 4.23, |
| "learning_rate": 1.5773704789833824e-05, |
| "loss": 0.1899, |
| "step": 8650 |
| }, |
| { |
| "epoch": 4.24, |
| "learning_rate": 1.5761485826001956e-05, |
| "loss": 0.1691, |
| "step": 8675 |
| }, |
| { |
| "epoch": 4.25, |
| "learning_rate": 1.574926686217009e-05, |
| "loss": 0.1777, |
| "step": 8700 |
| }, |
| { |
| "epoch": 4.26, |
| "learning_rate": 1.573704789833822e-05, |
| "loss": 0.1877, |
| "step": 8725 |
| }, |
| { |
| "epoch": 4.28, |
| "learning_rate": 1.5724828934506357e-05, |
| "loss": 0.1846, |
| "step": 8750 |
| }, |
| { |
| "epoch": 4.29, |
| "learning_rate": 1.571260997067449e-05, |
| "loss": 0.1963, |
| "step": 8775 |
| }, |
| { |
| "epoch": 4.3, |
| "learning_rate": 1.5700391006842622e-05, |
| "loss": 0.1771, |
| "step": 8800 |
| }, |
| { |
| "epoch": 4.31, |
| "learning_rate": 1.5688172043010754e-05, |
| "loss": 0.1865, |
| "step": 8825 |
| }, |
| { |
| "epoch": 4.33, |
| "learning_rate": 1.5675953079178887e-05, |
| "loss": 0.1763, |
| "step": 8850 |
| }, |
| { |
| "epoch": 4.34, |
| "learning_rate": 1.566373411534702e-05, |
| "loss": 0.1768, |
| "step": 8875 |
| }, |
| { |
| "epoch": 4.35, |
| "learning_rate": 1.565151515151515e-05, |
| "loss": 0.1645, |
| "step": 8900 |
| }, |
| { |
| "epoch": 4.36, |
| "learning_rate": 1.5639296187683287e-05, |
| "loss": 0.1816, |
| "step": 8925 |
| }, |
| { |
| "epoch": 4.37, |
| "learning_rate": 1.562707722385142e-05, |
| "loss": 0.1793, |
| "step": 8950 |
| }, |
| { |
| "epoch": 4.39, |
| "learning_rate": 1.5614858260019552e-05, |
| "loss": 0.1888, |
| "step": 8975 |
| }, |
| { |
| "epoch": 4.4, |
| "learning_rate": 1.5602639296187684e-05, |
| "loss": 0.1769, |
| "step": 9000 |
| }, |
| { |
| "epoch": 4.41, |
| "learning_rate": 1.5590420332355817e-05, |
| "loss": 0.1791, |
| "step": 9025 |
| }, |
| { |
| "epoch": 4.42, |
| "learning_rate": 1.557820136852395e-05, |
| "loss": 0.1823, |
| "step": 9050 |
| }, |
| { |
| "epoch": 4.44, |
| "learning_rate": 1.5565982404692085e-05, |
| "loss": 0.1835, |
| "step": 9075 |
| }, |
| { |
| "epoch": 4.45, |
| "learning_rate": 1.5553763440860217e-05, |
| "loss": 0.1875, |
| "step": 9100 |
| }, |
| { |
| "epoch": 4.46, |
| "learning_rate": 1.554154447702835e-05, |
| "loss": 0.1761, |
| "step": 9125 |
| }, |
| { |
| "epoch": 4.47, |
| "learning_rate": 1.5529325513196482e-05, |
| "loss": 0.1838, |
| "step": 9150 |
| }, |
| { |
| "epoch": 4.48, |
| "learning_rate": 1.5517106549364615e-05, |
| "loss": 0.1966, |
| "step": 9175 |
| }, |
| { |
| "epoch": 4.5, |
| "learning_rate": 1.5504887585532747e-05, |
| "loss": 0.1874, |
| "step": 9200 |
| }, |
| { |
| "epoch": 4.51, |
| "learning_rate": 1.5492668621700883e-05, |
| "loss": 0.1841, |
| "step": 9225 |
| }, |
| { |
| "epoch": 4.52, |
| "learning_rate": 1.5480449657869015e-05, |
| "loss": 0.2053, |
| "step": 9250 |
| }, |
| { |
| "epoch": 4.53, |
| "learning_rate": 1.5468230694037148e-05, |
| "loss": 0.1925, |
| "step": 9275 |
| }, |
| { |
| "epoch": 4.55, |
| "learning_rate": 1.545601173020528e-05, |
| "loss": 0.1784, |
| "step": 9300 |
| }, |
| { |
| "epoch": 4.56, |
| "learning_rate": 1.5443792766373412e-05, |
| "loss": 0.1914, |
| "step": 9325 |
| }, |
| { |
| "epoch": 4.57, |
| "learning_rate": 1.5431573802541545e-05, |
| "loss": 0.1875, |
| "step": 9350 |
| }, |
| { |
| "epoch": 4.58, |
| "learning_rate": 1.541935483870968e-05, |
| "loss": 0.183, |
| "step": 9375 |
| }, |
| { |
| "epoch": 4.59, |
| "learning_rate": 1.5407135874877813e-05, |
| "loss": 0.1964, |
| "step": 9400 |
| }, |
| { |
| "epoch": 4.61, |
| "learning_rate": 1.539540566959922e-05, |
| "loss": 0.1753, |
| "step": 9425 |
| }, |
| { |
| "epoch": 4.62, |
| "learning_rate": 1.538318670576735e-05, |
| "loss": 0.1911, |
| "step": 9450 |
| }, |
| { |
| "epoch": 4.63, |
| "learning_rate": 1.5370967741935487e-05, |
| "loss": 0.1957, |
| "step": 9475 |
| }, |
| { |
| "epoch": 4.64, |
| "learning_rate": 1.535874877810362e-05, |
| "loss": 0.1861, |
| "step": 9500 |
| }, |
| { |
| "epoch": 4.66, |
| "learning_rate": 1.534652981427175e-05, |
| "loss": 0.191, |
| "step": 9525 |
| }, |
| { |
| "epoch": 4.67, |
| "learning_rate": 1.5334310850439884e-05, |
| "loss": 0.1844, |
| "step": 9550 |
| }, |
| { |
| "epoch": 4.68, |
| "learning_rate": 1.5322091886608016e-05, |
| "loss": 0.2053, |
| "step": 9575 |
| }, |
| { |
| "epoch": 4.69, |
| "learning_rate": 1.530987292277615e-05, |
| "loss": 0.1857, |
| "step": 9600 |
| }, |
| { |
| "epoch": 4.7, |
| "learning_rate": 1.5297653958944284e-05, |
| "loss": 0.1783, |
| "step": 9625 |
| }, |
| { |
| "epoch": 4.72, |
| "learning_rate": 1.5285434995112417e-05, |
| "loss": 0.1917, |
| "step": 9650 |
| }, |
| { |
| "epoch": 4.73, |
| "learning_rate": 1.527321603128055e-05, |
| "loss": 0.1816, |
| "step": 9675 |
| }, |
| { |
| "epoch": 4.74, |
| "learning_rate": 1.526099706744868e-05, |
| "loss": 0.1925, |
| "step": 9700 |
| }, |
| { |
| "epoch": 4.75, |
| "learning_rate": 1.5248778103616814e-05, |
| "loss": 0.2035, |
| "step": 9725 |
| }, |
| { |
| "epoch": 4.77, |
| "learning_rate": 1.5236559139784948e-05, |
| "loss": 0.1839, |
| "step": 9750 |
| }, |
| { |
| "epoch": 4.78, |
| "learning_rate": 1.522434017595308e-05, |
| "loss": 0.1897, |
| "step": 9775 |
| }, |
| { |
| "epoch": 4.79, |
| "learning_rate": 1.5212121212121213e-05, |
| "loss": 0.188, |
| "step": 9800 |
| }, |
| { |
| "epoch": 4.8, |
| "learning_rate": 1.5199902248289347e-05, |
| "loss": 0.1791, |
| "step": 9825 |
| }, |
| { |
| "epoch": 4.81, |
| "learning_rate": 1.518768328445748e-05, |
| "loss": 0.1894, |
| "step": 9850 |
| }, |
| { |
| "epoch": 4.83, |
| "learning_rate": 1.5175464320625612e-05, |
| "loss": 0.1975, |
| "step": 9875 |
| }, |
| { |
| "epoch": 4.84, |
| "learning_rate": 1.5163245356793746e-05, |
| "loss": 0.181, |
| "step": 9900 |
| }, |
| { |
| "epoch": 4.85, |
| "learning_rate": 1.5151026392961878e-05, |
| "loss": 0.2079, |
| "step": 9925 |
| }, |
| { |
| "epoch": 4.86, |
| "learning_rate": 1.513880742913001e-05, |
| "loss": 0.1866, |
| "step": 9950 |
| }, |
| { |
| "epoch": 4.88, |
| "learning_rate": 1.5126588465298145e-05, |
| "loss": 0.1868, |
| "step": 9975 |
| }, |
| { |
| "epoch": 4.89, |
| "learning_rate": 1.5114369501466277e-05, |
| "loss": 0.1956, |
| "step": 10000 |
| }, |
| { |
| "epoch": 4.9, |
| "learning_rate": 1.510215053763441e-05, |
| "loss": 0.1871, |
| "step": 10025 |
| }, |
| { |
| "epoch": 4.91, |
| "learning_rate": 1.5089931573802544e-05, |
| "loss": 0.1954, |
| "step": 10050 |
| }, |
| { |
| "epoch": 4.92, |
| "learning_rate": 1.5077712609970676e-05, |
| "loss": 0.1872, |
| "step": 10075 |
| }, |
| { |
| "epoch": 4.94, |
| "learning_rate": 1.5065493646138808e-05, |
| "loss": 0.2023, |
| "step": 10100 |
| }, |
| { |
| "epoch": 4.95, |
| "learning_rate": 1.5053274682306943e-05, |
| "loss": 0.2043, |
| "step": 10125 |
| }, |
| { |
| "epoch": 4.96, |
| "learning_rate": 1.5041055718475075e-05, |
| "loss": 0.1863, |
| "step": 10150 |
| }, |
| { |
| "epoch": 4.97, |
| "learning_rate": 1.5028836754643207e-05, |
| "loss": 0.1987, |
| "step": 10175 |
| }, |
| { |
| "epoch": 4.99, |
| "learning_rate": 1.5016617790811341e-05, |
| "loss": 0.2, |
| "step": 10200 |
| }, |
| { |
| "epoch": 5.0, |
| "learning_rate": 1.5004398826979474e-05, |
| "loss": 0.1963, |
| "step": 10225 |
| }, |
| { |
| "epoch": 5.01, |
| "learning_rate": 1.4992179863147606e-05, |
| "loss": 0.1458, |
| "step": 10250 |
| }, |
| { |
| "epoch": 5.02, |
| "learning_rate": 1.497996089931574e-05, |
| "loss": 0.1321, |
| "step": 10275 |
| }, |
| { |
| "epoch": 5.03, |
| "learning_rate": 1.4967741935483873e-05, |
| "loss": 0.1246, |
| "step": 10300 |
| }, |
| { |
| "epoch": 5.05, |
| "learning_rate": 1.4955522971652005e-05, |
| "loss": 0.1298, |
| "step": 10325 |
| }, |
| { |
| "epoch": 5.06, |
| "learning_rate": 1.494330400782014e-05, |
| "loss": 0.1401, |
| "step": 10350 |
| }, |
| { |
| "epoch": 5.07, |
| "learning_rate": 1.4931085043988272e-05, |
| "loss": 0.1299, |
| "step": 10375 |
| }, |
| { |
| "epoch": 5.08, |
| "learning_rate": 1.4918866080156404e-05, |
| "loss": 0.128, |
| "step": 10400 |
| }, |
| { |
| "epoch": 5.1, |
| "learning_rate": 1.4906647116324538e-05, |
| "loss": 0.134, |
| "step": 10425 |
| }, |
| { |
| "epoch": 5.11, |
| "learning_rate": 1.489442815249267e-05, |
| "loss": 0.137, |
| "step": 10450 |
| }, |
| { |
| "epoch": 5.12, |
| "learning_rate": 1.4882209188660803e-05, |
| "loss": 0.1393, |
| "step": 10475 |
| }, |
| { |
| "epoch": 5.13, |
| "learning_rate": 1.4869990224828937e-05, |
| "loss": 0.1303, |
| "step": 10500 |
| }, |
| { |
| "epoch": 5.14, |
| "learning_rate": 1.485777126099707e-05, |
| "loss": 0.1412, |
| "step": 10525 |
| }, |
| { |
| "epoch": 5.16, |
| "learning_rate": 1.4845552297165202e-05, |
| "loss": 0.1454, |
| "step": 10550 |
| }, |
| { |
| "epoch": 5.17, |
| "learning_rate": 1.4833333333333336e-05, |
| "loss": 0.1464, |
| "step": 10575 |
| }, |
| { |
| "epoch": 5.18, |
| "learning_rate": 1.4821114369501468e-05, |
| "loss": 0.138, |
| "step": 10600 |
| }, |
| { |
| "epoch": 5.19, |
| "learning_rate": 1.48088954056696e-05, |
| "loss": 0.1256, |
| "step": 10625 |
| }, |
| { |
| "epoch": 5.21, |
| "learning_rate": 1.4796676441837735e-05, |
| "loss": 0.1352, |
| "step": 10650 |
| }, |
| { |
| "epoch": 5.22, |
| "learning_rate": 1.4784457478005867e-05, |
| "loss": 0.138, |
| "step": 10675 |
| }, |
| { |
| "epoch": 5.23, |
| "learning_rate": 1.4772238514174e-05, |
| "loss": 0.1388, |
| "step": 10700 |
| }, |
| { |
| "epoch": 5.24, |
| "learning_rate": 1.4760019550342134e-05, |
| "loss": 0.1364, |
| "step": 10725 |
| }, |
| { |
| "epoch": 5.25, |
| "learning_rate": 1.4747800586510266e-05, |
| "loss": 0.1414, |
| "step": 10750 |
| }, |
| { |
| "epoch": 5.27, |
| "learning_rate": 1.4735581622678398e-05, |
| "loss": 0.1409, |
| "step": 10775 |
| }, |
| { |
| "epoch": 5.28, |
| "learning_rate": 1.4723362658846531e-05, |
| "loss": 0.1475, |
| "step": 10800 |
| }, |
| { |
| "epoch": 5.29, |
| "learning_rate": 1.4711143695014665e-05, |
| "loss": 0.1465, |
| "step": 10825 |
| }, |
| { |
| "epoch": 5.3, |
| "learning_rate": 1.4698924731182797e-05, |
| "loss": 0.1565, |
| "step": 10850 |
| }, |
| { |
| "epoch": 5.32, |
| "learning_rate": 1.468670576735093e-05, |
| "loss": 0.1534, |
| "step": 10875 |
| }, |
| { |
| "epoch": 5.33, |
| "learning_rate": 1.4674486803519064e-05, |
| "loss": 0.1549, |
| "step": 10900 |
| }, |
| { |
| "epoch": 5.34, |
| "learning_rate": 1.4662267839687196e-05, |
| "loss": 0.1404, |
| "step": 10925 |
| }, |
| { |
| "epoch": 5.35, |
| "learning_rate": 1.4650048875855329e-05, |
| "loss": 0.1352, |
| "step": 10950 |
| }, |
| { |
| "epoch": 5.36, |
| "learning_rate": 1.4637829912023463e-05, |
| "loss": 0.1455, |
| "step": 10975 |
| }, |
| { |
| "epoch": 5.38, |
| "learning_rate": 1.4625610948191595e-05, |
| "loss": 0.1359, |
| "step": 11000 |
| }, |
| { |
| "epoch": 5.39, |
| "learning_rate": 1.4613391984359728e-05, |
| "loss": 0.1395, |
| "step": 11025 |
| }, |
| { |
| "epoch": 5.4, |
| "learning_rate": 1.4601173020527862e-05, |
| "loss": 0.1544, |
| "step": 11050 |
| }, |
| { |
| "epoch": 5.41, |
| "learning_rate": 1.4588954056695994e-05, |
| "loss": 0.1603, |
| "step": 11075 |
| }, |
| { |
| "epoch": 5.43, |
| "learning_rate": 1.4576735092864126e-05, |
| "loss": 0.141, |
| "step": 11100 |
| }, |
| { |
| "epoch": 5.44, |
| "learning_rate": 1.456451612903226e-05, |
| "loss": 0.1459, |
| "step": 11125 |
| }, |
| { |
| "epoch": 5.45, |
| "learning_rate": 1.4552297165200393e-05, |
| "loss": 0.1327, |
| "step": 11150 |
| }, |
| { |
| "epoch": 5.46, |
| "learning_rate": 1.4540078201368525e-05, |
| "loss": 0.1393, |
| "step": 11175 |
| }, |
| { |
| "epoch": 5.47, |
| "learning_rate": 1.452785923753666e-05, |
| "loss": 0.1569, |
| "step": 11200 |
| }, |
| { |
| "epoch": 5.49, |
| "learning_rate": 1.4515640273704792e-05, |
| "loss": 0.1387, |
| "step": 11225 |
| }, |
| { |
| "epoch": 5.5, |
| "learning_rate": 1.4503421309872924e-05, |
| "loss": 0.1332, |
| "step": 11250 |
| }, |
| { |
| "epoch": 5.51, |
| "learning_rate": 1.4491202346041058e-05, |
| "loss": 0.1439, |
| "step": 11275 |
| }, |
| { |
| "epoch": 5.52, |
| "learning_rate": 1.447898338220919e-05, |
| "loss": 0.1252, |
| "step": 11300 |
| }, |
| { |
| "epoch": 5.54, |
| "learning_rate": 1.4466764418377323e-05, |
| "loss": 0.1516, |
| "step": 11325 |
| }, |
| { |
| "epoch": 5.55, |
| "learning_rate": 1.4454545454545457e-05, |
| "loss": 0.1378, |
| "step": 11350 |
| }, |
| { |
| "epoch": 5.56, |
| "learning_rate": 1.4442815249266864e-05, |
| "loss": 0.1571, |
| "step": 11375 |
| }, |
| { |
| "epoch": 5.57, |
| "learning_rate": 1.4430596285434997e-05, |
| "loss": 0.1519, |
| "step": 11400 |
| }, |
| { |
| "epoch": 5.58, |
| "learning_rate": 1.4418377321603129e-05, |
| "loss": 0.1498, |
| "step": 11425 |
| }, |
| { |
| "epoch": 5.6, |
| "learning_rate": 1.4406158357771263e-05, |
| "loss": 0.1408, |
| "step": 11450 |
| }, |
| { |
| "epoch": 5.61, |
| "learning_rate": 1.4393939393939396e-05, |
| "loss": 0.1336, |
| "step": 11475 |
| }, |
| { |
| "epoch": 5.62, |
| "learning_rate": 1.4381720430107528e-05, |
| "loss": 0.1423, |
| "step": 11500 |
| }, |
| { |
| "epoch": 5.63, |
| "learning_rate": 1.4369501466275662e-05, |
| "loss": 0.1498, |
| "step": 11525 |
| }, |
| { |
| "epoch": 5.65, |
| "learning_rate": 1.4357282502443794e-05, |
| "loss": 0.1476, |
| "step": 11550 |
| }, |
| { |
| "epoch": 5.66, |
| "learning_rate": 1.4345063538611927e-05, |
| "loss": 0.1363, |
| "step": 11575 |
| }, |
| { |
| "epoch": 5.67, |
| "learning_rate": 1.4332844574780061e-05, |
| "loss": 0.141, |
| "step": 11600 |
| }, |
| { |
| "epoch": 5.68, |
| "learning_rate": 1.4320625610948193e-05, |
| "loss": 0.1329, |
| "step": 11625 |
| }, |
| { |
| "epoch": 5.69, |
| "learning_rate": 1.4308406647116326e-05, |
| "loss": 0.153, |
| "step": 11650 |
| }, |
| { |
| "epoch": 5.71, |
| "learning_rate": 1.429618768328446e-05, |
| "loss": 0.1587, |
| "step": 11675 |
| }, |
| { |
| "epoch": 5.72, |
| "learning_rate": 1.4283968719452592e-05, |
| "loss": 0.1518, |
| "step": 11700 |
| }, |
| { |
| "epoch": 5.73, |
| "learning_rate": 1.4271749755620725e-05, |
| "loss": 0.1449, |
| "step": 11725 |
| }, |
| { |
| "epoch": 5.74, |
| "learning_rate": 1.4259530791788859e-05, |
| "loss": 0.154, |
| "step": 11750 |
| }, |
| { |
| "epoch": 5.76, |
| "learning_rate": 1.4247311827956991e-05, |
| "loss": 0.1519, |
| "step": 11775 |
| }, |
| { |
| "epoch": 5.77, |
| "learning_rate": 1.4235092864125124e-05, |
| "loss": 0.1417, |
| "step": 11800 |
| }, |
| { |
| "epoch": 5.78, |
| "learning_rate": 1.4222873900293258e-05, |
| "loss": 0.1479, |
| "step": 11825 |
| }, |
| { |
| "epoch": 5.79, |
| "learning_rate": 1.421065493646139e-05, |
| "loss": 0.1499, |
| "step": 11850 |
| }, |
| { |
| "epoch": 5.8, |
| "learning_rate": 1.4198435972629522e-05, |
| "loss": 0.1362, |
| "step": 11875 |
| }, |
| { |
| "epoch": 5.82, |
| "learning_rate": 1.4186217008797657e-05, |
| "loss": 0.1522, |
| "step": 11900 |
| }, |
| { |
| "epoch": 5.83, |
| "learning_rate": 1.4173998044965789e-05, |
| "loss": 0.149, |
| "step": 11925 |
| }, |
| { |
| "epoch": 5.84, |
| "learning_rate": 1.4161779081133921e-05, |
| "loss": 0.1448, |
| "step": 11950 |
| }, |
| { |
| "epoch": 5.85, |
| "learning_rate": 1.4149560117302055e-05, |
| "loss": 0.1602, |
| "step": 11975 |
| }, |
| { |
| "epoch": 5.87, |
| "learning_rate": 1.4137341153470188e-05, |
| "loss": 0.1481, |
| "step": 12000 |
| }, |
| { |
| "epoch": 5.88, |
| "learning_rate": 1.412512218963832e-05, |
| "loss": 0.1631, |
| "step": 12025 |
| }, |
| { |
| "epoch": 5.89, |
| "learning_rate": 1.4112903225806454e-05, |
| "loss": 0.1516, |
| "step": 12050 |
| }, |
| { |
| "epoch": 5.9, |
| "learning_rate": 1.4100684261974587e-05, |
| "loss": 0.1449, |
| "step": 12075 |
| }, |
| { |
| "epoch": 5.91, |
| "learning_rate": 1.4088465298142719e-05, |
| "loss": 0.1549, |
| "step": 12100 |
| }, |
| { |
| "epoch": 5.93, |
| "learning_rate": 1.4076246334310853e-05, |
| "loss": 0.1456, |
| "step": 12125 |
| }, |
| { |
| "epoch": 5.94, |
| "learning_rate": 1.4064027370478986e-05, |
| "loss": 0.1443, |
| "step": 12150 |
| }, |
| { |
| "epoch": 5.95, |
| "learning_rate": 1.4051808406647118e-05, |
| "loss": 0.1445, |
| "step": 12175 |
| }, |
| { |
| "epoch": 5.96, |
| "learning_rate": 1.4039589442815252e-05, |
| "loss": 0.15, |
| "step": 12200 |
| }, |
| { |
| "epoch": 5.98, |
| "learning_rate": 1.4027370478983385e-05, |
| "loss": 0.1459, |
| "step": 12225 |
| }, |
| { |
| "epoch": 5.99, |
| "learning_rate": 1.4015151515151517e-05, |
| "loss": 0.1594, |
| "step": 12250 |
| }, |
| { |
| "epoch": 6.0, |
| "learning_rate": 1.400293255131965e-05, |
| "loss": 0.1511, |
| "step": 12275 |
| }, |
| { |
| "epoch": 6.01, |
| "learning_rate": 1.3990713587487783e-05, |
| "loss": 0.1114, |
| "step": 12300 |
| }, |
| { |
| "epoch": 6.02, |
| "learning_rate": 1.3978494623655916e-05, |
| "loss": 0.1002, |
| "step": 12325 |
| }, |
| { |
| "epoch": 6.04, |
| "learning_rate": 1.3966275659824048e-05, |
| "loss": 0.117, |
| "step": 12350 |
| }, |
| { |
| "epoch": 6.05, |
| "learning_rate": 1.3954056695992182e-05, |
| "loss": 0.1082, |
| "step": 12375 |
| }, |
| { |
| "epoch": 6.06, |
| "learning_rate": 1.3941837732160315e-05, |
| "loss": 0.1215, |
| "step": 12400 |
| }, |
| { |
| "epoch": 6.07, |
| "learning_rate": 1.3929618768328447e-05, |
| "loss": 0.1046, |
| "step": 12425 |
| }, |
| { |
| "epoch": 6.09, |
| "learning_rate": 1.3917399804496581e-05, |
| "loss": 0.1063, |
| "step": 12450 |
| }, |
| { |
| "epoch": 6.1, |
| "learning_rate": 1.3905180840664714e-05, |
| "loss": 0.1178, |
| "step": 12475 |
| }, |
| { |
| "epoch": 6.11, |
| "learning_rate": 1.3892961876832846e-05, |
| "loss": 0.1093, |
| "step": 12500 |
| }, |
| { |
| "epoch": 6.12, |
| "learning_rate": 1.388074291300098e-05, |
| "loss": 0.1085, |
| "step": 12525 |
| }, |
| { |
| "epoch": 6.13, |
| "learning_rate": 1.3868523949169112e-05, |
| "loss": 0.1171, |
| "step": 12550 |
| }, |
| { |
| "epoch": 6.15, |
| "learning_rate": 1.3856304985337245e-05, |
| "loss": 0.1145, |
| "step": 12575 |
| }, |
| { |
| "epoch": 6.16, |
| "learning_rate": 1.3844086021505379e-05, |
| "loss": 0.1042, |
| "step": 12600 |
| }, |
| { |
| "epoch": 6.17, |
| "learning_rate": 1.3831867057673511e-05, |
| "loss": 0.1059, |
| "step": 12625 |
| }, |
| { |
| "epoch": 6.18, |
| "learning_rate": 1.3819648093841644e-05, |
| "loss": 0.1144, |
| "step": 12650 |
| }, |
| { |
| "epoch": 6.2, |
| "learning_rate": 1.3807429130009778e-05, |
| "loss": 0.1023, |
| "step": 12675 |
| }, |
| { |
| "epoch": 6.21, |
| "learning_rate": 1.379521016617791e-05, |
| "loss": 0.1188, |
| "step": 12700 |
| }, |
| { |
| "epoch": 6.22, |
| "learning_rate": 1.3782991202346043e-05, |
| "loss": 0.1028, |
| "step": 12725 |
| }, |
| { |
| "epoch": 6.23, |
| "learning_rate": 1.3770772238514177e-05, |
| "loss": 0.1174, |
| "step": 12750 |
| }, |
| { |
| "epoch": 6.24, |
| "learning_rate": 1.375855327468231e-05, |
| "loss": 0.1113, |
| "step": 12775 |
| }, |
| { |
| "epoch": 6.26, |
| "learning_rate": 1.374633431085044e-05, |
| "loss": 0.1183, |
| "step": 12800 |
| }, |
| { |
| "epoch": 6.27, |
| "learning_rate": 1.3734115347018572e-05, |
| "loss": 0.1145, |
| "step": 12825 |
| }, |
| { |
| "epoch": 6.28, |
| "learning_rate": 1.3721896383186706e-05, |
| "loss": 0.1075, |
| "step": 12850 |
| }, |
| { |
| "epoch": 6.29, |
| "learning_rate": 1.3709677419354839e-05, |
| "loss": 0.1083, |
| "step": 12875 |
| }, |
| { |
| "epoch": 6.3, |
| "learning_rate": 1.3697458455522971e-05, |
| "loss": 0.1123, |
| "step": 12900 |
| }, |
| { |
| "epoch": 6.32, |
| "learning_rate": 1.3685239491691105e-05, |
| "loss": 0.1097, |
| "step": 12925 |
| }, |
| { |
| "epoch": 6.33, |
| "learning_rate": 1.3673020527859238e-05, |
| "loss": 0.1262, |
| "step": 12950 |
| }, |
| { |
| "epoch": 6.34, |
| "learning_rate": 1.366080156402737e-05, |
| "loss": 0.1156, |
| "step": 12975 |
| }, |
| { |
| "epoch": 6.35, |
| "learning_rate": 1.3648582600195504e-05, |
| "loss": 0.1229, |
| "step": 13000 |
| }, |
| { |
| "epoch": 6.37, |
| "learning_rate": 1.3636363636363637e-05, |
| "loss": 0.1173, |
| "step": 13025 |
| }, |
| { |
| "epoch": 6.38, |
| "learning_rate": 1.3624144672531769e-05, |
| "loss": 0.113, |
| "step": 13050 |
| }, |
| { |
| "epoch": 6.39, |
| "learning_rate": 1.3611925708699903e-05, |
| "loss": 0.1119, |
| "step": 13075 |
| }, |
| { |
| "epoch": 6.4, |
| "learning_rate": 1.3599706744868035e-05, |
| "loss": 0.1205, |
| "step": 13100 |
| }, |
| { |
| "epoch": 6.41, |
| "learning_rate": 1.3587487781036168e-05, |
| "loss": 0.1074, |
| "step": 13125 |
| }, |
| { |
| "epoch": 6.43, |
| "learning_rate": 1.3575268817204302e-05, |
| "loss": 0.1134, |
| "step": 13150 |
| }, |
| { |
| "epoch": 6.44, |
| "learning_rate": 1.3563049853372434e-05, |
| "loss": 0.1146, |
| "step": 13175 |
| }, |
| { |
| "epoch": 6.45, |
| "learning_rate": 1.3550830889540567e-05, |
| "loss": 0.1055, |
| "step": 13200 |
| }, |
| { |
| "epoch": 6.46, |
| "learning_rate": 1.35386119257087e-05, |
| "loss": 0.1067, |
| "step": 13225 |
| }, |
| { |
| "epoch": 6.48, |
| "learning_rate": 1.3526392961876833e-05, |
| "loss": 0.1141, |
| "step": 13250 |
| }, |
| { |
| "epoch": 6.49, |
| "learning_rate": 1.3514173998044966e-05, |
| "loss": 0.1143, |
| "step": 13275 |
| }, |
| { |
| "epoch": 6.5, |
| "learning_rate": 1.35019550342131e-05, |
| "loss": 0.115, |
| "step": 13300 |
| }, |
| { |
| "epoch": 6.51, |
| "learning_rate": 1.3489736070381232e-05, |
| "loss": 0.1158, |
| "step": 13325 |
| }, |
| { |
| "epoch": 6.52, |
| "learning_rate": 1.3477517106549365e-05, |
| "loss": 0.1119, |
| "step": 13350 |
| }, |
| { |
| "epoch": 6.54, |
| "learning_rate": 1.3465298142717499e-05, |
| "loss": 0.1225, |
| "step": 13375 |
| }, |
| { |
| "epoch": 6.55, |
| "learning_rate": 1.3453079178885631e-05, |
| "loss": 0.1171, |
| "step": 13400 |
| }, |
| { |
| "epoch": 6.56, |
| "learning_rate": 1.3440860215053763e-05, |
| "loss": 0.1235, |
| "step": 13425 |
| }, |
| { |
| "epoch": 6.57, |
| "learning_rate": 1.3428641251221896e-05, |
| "loss": 0.1156, |
| "step": 13450 |
| }, |
| { |
| "epoch": 6.59, |
| "learning_rate": 1.341642228739003e-05, |
| "loss": 0.1129, |
| "step": 13475 |
| }, |
| { |
| "epoch": 6.6, |
| "learning_rate": 1.3404203323558162e-05, |
| "loss": 0.1219, |
| "step": 13500 |
| }, |
| { |
| "epoch": 6.61, |
| "learning_rate": 1.3391984359726295e-05, |
| "loss": 0.1131, |
| "step": 13525 |
| }, |
| { |
| "epoch": 6.62, |
| "learning_rate": 1.3379765395894429e-05, |
| "loss": 0.1163, |
| "step": 13550 |
| }, |
| { |
| "epoch": 6.63, |
| "learning_rate": 1.3367546432062561e-05, |
| "loss": 0.1073, |
| "step": 13575 |
| }, |
| { |
| "epoch": 6.65, |
| "learning_rate": 1.3355327468230694e-05, |
| "loss": 0.1082, |
| "step": 13600 |
| }, |
| { |
| "epoch": 6.66, |
| "learning_rate": 1.3343108504398828e-05, |
| "loss": 0.1161, |
| "step": 13625 |
| }, |
| { |
| "epoch": 6.67, |
| "learning_rate": 1.333088954056696e-05, |
| "loss": 0.1021, |
| "step": 13650 |
| }, |
| { |
| "epoch": 6.68, |
| "learning_rate": 1.3318670576735093e-05, |
| "loss": 0.117, |
| "step": 13675 |
| }, |
| { |
| "epoch": 6.7, |
| "learning_rate": 1.3306451612903227e-05, |
| "loss": 0.1294, |
| "step": 13700 |
| }, |
| { |
| "epoch": 6.71, |
| "learning_rate": 1.3294232649071359e-05, |
| "loss": 0.1148, |
| "step": 13725 |
| }, |
| { |
| "epoch": 6.72, |
| "learning_rate": 1.3282013685239491e-05, |
| "loss": 0.1037, |
| "step": 13750 |
| }, |
| { |
| "epoch": 6.73, |
| "learning_rate": 1.3269794721407626e-05, |
| "loss": 0.1125, |
| "step": 13775 |
| }, |
| { |
| "epoch": 6.74, |
| "learning_rate": 1.3257575757575758e-05, |
| "loss": 0.119, |
| "step": 13800 |
| }, |
| { |
| "epoch": 6.76, |
| "learning_rate": 1.324535679374389e-05, |
| "loss": 0.1161, |
| "step": 13825 |
| }, |
| { |
| "epoch": 6.77, |
| "learning_rate": 1.3233137829912024e-05, |
| "loss": 0.112, |
| "step": 13850 |
| }, |
| { |
| "epoch": 6.78, |
| "learning_rate": 1.3220918866080157e-05, |
| "loss": 0.1124, |
| "step": 13875 |
| }, |
| { |
| "epoch": 6.79, |
| "learning_rate": 1.320869990224829e-05, |
| "loss": 0.1352, |
| "step": 13900 |
| }, |
| { |
| "epoch": 6.81, |
| "learning_rate": 1.3196480938416423e-05, |
| "loss": 0.1252, |
| "step": 13925 |
| }, |
| { |
| "epoch": 6.82, |
| "learning_rate": 1.3184261974584556e-05, |
| "loss": 0.1054, |
| "step": 13950 |
| }, |
| { |
| "epoch": 6.83, |
| "learning_rate": 1.3172043010752688e-05, |
| "loss": 0.1153, |
| "step": 13975 |
| }, |
| { |
| "epoch": 6.84, |
| "learning_rate": 1.3159824046920822e-05, |
| "loss": 0.1323, |
| "step": 14000 |
| }, |
| { |
| "epoch": 6.85, |
| "learning_rate": 1.3147605083088955e-05, |
| "loss": 0.1227, |
| "step": 14025 |
| }, |
| { |
| "epoch": 6.87, |
| "learning_rate": 1.3135386119257087e-05, |
| "loss": 0.1197, |
| "step": 14050 |
| }, |
| { |
| "epoch": 6.88, |
| "learning_rate": 1.3123167155425221e-05, |
| "loss": 0.1202, |
| "step": 14075 |
| }, |
| { |
| "epoch": 6.89, |
| "learning_rate": 1.3110948191593353e-05, |
| "loss": 0.1119, |
| "step": 14100 |
| }, |
| { |
| "epoch": 6.9, |
| "learning_rate": 1.3098729227761486e-05, |
| "loss": 0.115, |
| "step": 14125 |
| }, |
| { |
| "epoch": 6.92, |
| "learning_rate": 1.308651026392962e-05, |
| "loss": 0.1209, |
| "step": 14150 |
| }, |
| { |
| "epoch": 6.93, |
| "learning_rate": 1.3074291300097752e-05, |
| "loss": 0.1201, |
| "step": 14175 |
| }, |
| { |
| "epoch": 6.94, |
| "learning_rate": 1.3062072336265885e-05, |
| "loss": 0.1227, |
| "step": 14200 |
| }, |
| { |
| "epoch": 6.95, |
| "learning_rate": 1.3049853372434019e-05, |
| "loss": 0.1132, |
| "step": 14225 |
| }, |
| { |
| "epoch": 6.96, |
| "learning_rate": 1.3037634408602151e-05, |
| "loss": 0.1326, |
| "step": 14250 |
| }, |
| { |
| "epoch": 6.98, |
| "learning_rate": 1.3025415444770284e-05, |
| "loss": 0.1098, |
| "step": 14275 |
| }, |
| { |
| "epoch": 6.99, |
| "learning_rate": 1.3013196480938418e-05, |
| "loss": 0.1168, |
| "step": 14300 |
| }, |
| { |
| "epoch": 7.0, |
| "learning_rate": 1.300097751710655e-05, |
| "loss": 0.132, |
| "step": 14325 |
| }, |
| { |
| "epoch": 7.01, |
| "learning_rate": 1.2988758553274683e-05, |
| "loss": 0.0912, |
| "step": 14350 |
| }, |
| { |
| "epoch": 7.03, |
| "learning_rate": 1.2976539589442817e-05, |
| "loss": 0.091, |
| "step": 14375 |
| }, |
| { |
| "epoch": 7.04, |
| "learning_rate": 1.2964320625610949e-05, |
| "loss": 0.0789, |
| "step": 14400 |
| }, |
| { |
| "epoch": 7.05, |
| "learning_rate": 1.2952101661779081e-05, |
| "loss": 0.0815, |
| "step": 14425 |
| }, |
| { |
| "epoch": 7.06, |
| "learning_rate": 1.2939882697947216e-05, |
| "loss": 0.0935, |
| "step": 14450 |
| }, |
| { |
| "epoch": 7.07, |
| "learning_rate": 1.2927663734115348e-05, |
| "loss": 0.0863, |
| "step": 14475 |
| }, |
| { |
| "epoch": 7.09, |
| "learning_rate": 1.291544477028348e-05, |
| "loss": 0.0877, |
| "step": 14500 |
| }, |
| { |
| "epoch": 7.1, |
| "learning_rate": 1.2903225806451613e-05, |
| "loss": 0.0832, |
| "step": 14525 |
| }, |
| { |
| "epoch": 7.11, |
| "learning_rate": 1.2891006842619747e-05, |
| "loss": 0.0888, |
| "step": 14550 |
| }, |
| { |
| "epoch": 7.12, |
| "learning_rate": 1.287878787878788e-05, |
| "loss": 0.0901, |
| "step": 14575 |
| }, |
| { |
| "epoch": 7.14, |
| "learning_rate": 1.2866568914956012e-05, |
| "loss": 0.0891, |
| "step": 14600 |
| }, |
| { |
| "epoch": 7.15, |
| "learning_rate": 1.2854349951124146e-05, |
| "loss": 0.0887, |
| "step": 14625 |
| }, |
| { |
| "epoch": 7.16, |
| "learning_rate": 1.2842130987292278e-05, |
| "loss": 0.0916, |
| "step": 14650 |
| }, |
| { |
| "epoch": 7.17, |
| "learning_rate": 1.282991202346041e-05, |
| "loss": 0.0853, |
| "step": 14675 |
| }, |
| { |
| "epoch": 7.18, |
| "learning_rate": 1.2817693059628545e-05, |
| "loss": 0.0858, |
| "step": 14700 |
| }, |
| { |
| "epoch": 7.2, |
| "learning_rate": 1.2805474095796677e-05, |
| "loss": 0.0876, |
| "step": 14725 |
| }, |
| { |
| "epoch": 7.21, |
| "learning_rate": 1.279325513196481e-05, |
| "loss": 0.1036, |
| "step": 14750 |
| }, |
| { |
| "epoch": 7.22, |
| "learning_rate": 1.2781036168132944e-05, |
| "loss": 0.0959, |
| "step": 14775 |
| }, |
| { |
| "epoch": 7.23, |
| "learning_rate": 1.2768817204301076e-05, |
| "loss": 0.0789, |
| "step": 14800 |
| }, |
| { |
| "epoch": 7.25, |
| "learning_rate": 1.2756598240469208e-05, |
| "loss": 0.0843, |
| "step": 14825 |
| }, |
| { |
| "epoch": 7.26, |
| "learning_rate": 1.2744379276637342e-05, |
| "loss": 0.0895, |
| "step": 14850 |
| }, |
| { |
| "epoch": 7.27, |
| "learning_rate": 1.2732160312805475e-05, |
| "loss": 0.0985, |
| "step": 14875 |
| }, |
| { |
| "epoch": 7.28, |
| "learning_rate": 1.2719941348973607e-05, |
| "loss": 0.0793, |
| "step": 14900 |
| }, |
| { |
| "epoch": 7.29, |
| "learning_rate": 1.2707722385141741e-05, |
| "loss": 0.0808, |
| "step": 14925 |
| }, |
| { |
| "epoch": 7.31, |
| "learning_rate": 1.2695503421309874e-05, |
| "loss": 0.0843, |
| "step": 14950 |
| }, |
| { |
| "epoch": 7.32, |
| "learning_rate": 1.2683284457478006e-05, |
| "loss": 0.0979, |
| "step": 14975 |
| }, |
| { |
| "epoch": 7.33, |
| "learning_rate": 1.267106549364614e-05, |
| "loss": 0.1021, |
| "step": 15000 |
| }, |
| { |
| "epoch": 7.34, |
| "learning_rate": 1.2658846529814273e-05, |
| "loss": 0.0892, |
| "step": 15025 |
| }, |
| { |
| "epoch": 7.36, |
| "learning_rate": 1.2646627565982405e-05, |
| "loss": 0.0842, |
| "step": 15050 |
| }, |
| { |
| "epoch": 7.37, |
| "learning_rate": 1.2634408602150539e-05, |
| "loss": 0.0836, |
| "step": 15075 |
| }, |
| { |
| "epoch": 7.38, |
| "learning_rate": 1.2622189638318672e-05, |
| "loss": 0.0844, |
| "step": 15100 |
| }, |
| { |
| "epoch": 7.39, |
| "learning_rate": 1.2609970674486804e-05, |
| "loss": 0.1078, |
| "step": 15125 |
| }, |
| { |
| "epoch": 7.4, |
| "learning_rate": 1.2597751710654938e-05, |
| "loss": 0.0929, |
| "step": 15150 |
| }, |
| { |
| "epoch": 7.42, |
| "learning_rate": 1.258553274682307e-05, |
| "loss": 0.089, |
| "step": 15175 |
| }, |
| { |
| "epoch": 7.43, |
| "learning_rate": 1.2573313782991203e-05, |
| "loss": 0.1009, |
| "step": 15200 |
| }, |
| { |
| "epoch": 7.44, |
| "learning_rate": 1.2561094819159337e-05, |
| "loss": 0.0925, |
| "step": 15225 |
| }, |
| { |
| "epoch": 7.45, |
| "learning_rate": 1.254887585532747e-05, |
| "loss": 0.0892, |
| "step": 15250 |
| }, |
| { |
| "epoch": 7.47, |
| "learning_rate": 1.2536656891495602e-05, |
| "loss": 0.0887, |
| "step": 15275 |
| }, |
| { |
| "epoch": 7.48, |
| "learning_rate": 1.2524437927663736e-05, |
| "loss": 0.0964, |
| "step": 15300 |
| }, |
| { |
| "epoch": 7.49, |
| "learning_rate": 1.2512218963831868e-05, |
| "loss": 0.0894, |
| "step": 15325 |
| }, |
| { |
| "epoch": 7.5, |
| "learning_rate": 1.25e-05, |
| "loss": 0.0946, |
| "step": 15350 |
| }, |
| { |
| "epoch": 7.51, |
| "learning_rate": 1.2487781036168135e-05, |
| "loss": 0.0921, |
| "step": 15375 |
| }, |
| { |
| "epoch": 7.53, |
| "learning_rate": 1.2475562072336267e-05, |
| "loss": 0.0926, |
| "step": 15400 |
| }, |
| { |
| "epoch": 7.54, |
| "learning_rate": 1.24633431085044e-05, |
| "loss": 0.0959, |
| "step": 15425 |
| }, |
| { |
| "epoch": 7.55, |
| "learning_rate": 1.2451124144672534e-05, |
| "loss": 0.0936, |
| "step": 15450 |
| }, |
| { |
| "epoch": 7.56, |
| "learning_rate": 1.243939393939394e-05, |
| "loss": 0.0971, |
| "step": 15475 |
| }, |
| { |
| "epoch": 7.58, |
| "learning_rate": 1.2427174975562073e-05, |
| "loss": 0.0852, |
| "step": 15500 |
| }, |
| { |
| "epoch": 7.59, |
| "learning_rate": 1.2414956011730205e-05, |
| "loss": 0.0913, |
| "step": 15525 |
| }, |
| { |
| "epoch": 7.6, |
| "learning_rate": 1.240273704789834e-05, |
| "loss": 0.0922, |
| "step": 15550 |
| }, |
| { |
| "epoch": 7.61, |
| "learning_rate": 1.2390518084066472e-05, |
| "loss": 0.0918, |
| "step": 15575 |
| }, |
| { |
| "epoch": 7.62, |
| "learning_rate": 1.2378299120234604e-05, |
| "loss": 0.1045, |
| "step": 15600 |
| }, |
| { |
| "epoch": 7.64, |
| "learning_rate": 1.2366080156402738e-05, |
| "loss": 0.1003, |
| "step": 15625 |
| }, |
| { |
| "epoch": 7.65, |
| "learning_rate": 1.235386119257087e-05, |
| "loss": 0.087, |
| "step": 15650 |
| }, |
| { |
| "epoch": 7.66, |
| "learning_rate": 1.2341642228739003e-05, |
| "loss": 0.0965, |
| "step": 15675 |
| }, |
| { |
| "epoch": 7.67, |
| "learning_rate": 1.232991202346041e-05, |
| "loss": 0.0913, |
| "step": 15700 |
| }, |
| { |
| "epoch": 7.69, |
| "learning_rate": 1.2317693059628544e-05, |
| "loss": 0.0866, |
| "step": 15725 |
| }, |
| { |
| "epoch": 7.7, |
| "learning_rate": 1.2305474095796677e-05, |
| "loss": 0.091, |
| "step": 15750 |
| }, |
| { |
| "epoch": 7.71, |
| "learning_rate": 1.2293255131964809e-05, |
| "loss": 0.101, |
| "step": 15775 |
| }, |
| { |
| "epoch": 7.72, |
| "learning_rate": 1.2281036168132943e-05, |
| "loss": 0.1011, |
| "step": 15800 |
| }, |
| { |
| "epoch": 7.73, |
| "learning_rate": 1.2268817204301076e-05, |
| "loss": 0.0854, |
| "step": 15825 |
| }, |
| { |
| "epoch": 7.75, |
| "learning_rate": 1.2256598240469208e-05, |
| "loss": 0.0872, |
| "step": 15850 |
| }, |
| { |
| "epoch": 7.76, |
| "learning_rate": 1.2244379276637342e-05, |
| "loss": 0.0904, |
| "step": 15875 |
| }, |
| { |
| "epoch": 7.77, |
| "learning_rate": 1.2232160312805475e-05, |
| "loss": 0.1043, |
| "step": 15900 |
| }, |
| { |
| "epoch": 7.78, |
| "learning_rate": 1.2219941348973607e-05, |
| "loss": 0.0815, |
| "step": 15925 |
| }, |
| { |
| "epoch": 7.8, |
| "learning_rate": 1.2207722385141741e-05, |
| "loss": 0.0909, |
| "step": 15950 |
| }, |
| { |
| "epoch": 7.81, |
| "learning_rate": 1.2195503421309873e-05, |
| "loss": 0.0998, |
| "step": 15975 |
| }, |
| { |
| "epoch": 7.82, |
| "learning_rate": 1.2183284457478006e-05, |
| "loss": 0.0947, |
| "step": 16000 |
| }, |
| { |
| "epoch": 7.83, |
| "learning_rate": 1.217106549364614e-05, |
| "loss": 0.1099, |
| "step": 16025 |
| }, |
| { |
| "epoch": 7.84, |
| "learning_rate": 1.2158846529814272e-05, |
| "loss": 0.0971, |
| "step": 16050 |
| }, |
| { |
| "epoch": 7.86, |
| "learning_rate": 1.2146627565982405e-05, |
| "loss": 0.0944, |
| "step": 16075 |
| }, |
| { |
| "epoch": 7.87, |
| "learning_rate": 1.2134408602150539e-05, |
| "loss": 0.1006, |
| "step": 16100 |
| }, |
| { |
| "epoch": 7.88, |
| "learning_rate": 1.2122189638318671e-05, |
| "loss": 0.1007, |
| "step": 16125 |
| }, |
| { |
| "epoch": 7.89, |
| "learning_rate": 1.2109970674486804e-05, |
| "loss": 0.1004, |
| "step": 16150 |
| }, |
| { |
| "epoch": 7.91, |
| "learning_rate": 1.2097751710654938e-05, |
| "loss": 0.0946, |
| "step": 16175 |
| }, |
| { |
| "epoch": 7.92, |
| "learning_rate": 1.208553274682307e-05, |
| "loss": 0.0988, |
| "step": 16200 |
| }, |
| { |
| "epoch": 7.93, |
| "learning_rate": 1.2073313782991203e-05, |
| "loss": 0.0994, |
| "step": 16225 |
| }, |
| { |
| "epoch": 7.94, |
| "learning_rate": 1.2061094819159337e-05, |
| "loss": 0.0912, |
| "step": 16250 |
| }, |
| { |
| "epoch": 7.95, |
| "learning_rate": 1.2048875855327469e-05, |
| "loss": 0.094, |
| "step": 16275 |
| }, |
| { |
| "epoch": 7.97, |
| "learning_rate": 1.2036656891495601e-05, |
| "loss": 0.1037, |
| "step": 16300 |
| }, |
| { |
| "epoch": 7.98, |
| "learning_rate": 1.2024437927663734e-05, |
| "loss": 0.1082, |
| "step": 16325 |
| }, |
| { |
| "epoch": 7.99, |
| "learning_rate": 1.2012218963831868e-05, |
| "loss": 0.0959, |
| "step": 16350 |
| }, |
| { |
| "epoch": 8.0, |
| "learning_rate": 1.2e-05, |
| "loss": 0.0848, |
| "step": 16375 |
| }, |
| { |
| "epoch": 8.02, |
| "learning_rate": 1.1987781036168133e-05, |
| "loss": 0.07, |
| "step": 16400 |
| }, |
| { |
| "epoch": 8.03, |
| "learning_rate": 1.1975562072336267e-05, |
| "loss": 0.0713, |
| "step": 16425 |
| }, |
| { |
| "epoch": 8.04, |
| "learning_rate": 1.19633431085044e-05, |
| "loss": 0.0816, |
| "step": 16450 |
| }, |
| { |
| "epoch": 8.05, |
| "learning_rate": 1.1951124144672532e-05, |
| "loss": 0.0823, |
| "step": 16475 |
| }, |
| { |
| "epoch": 8.06, |
| "learning_rate": 1.1938905180840666e-05, |
| "loss": 0.0624, |
| "step": 16500 |
| }, |
| { |
| "epoch": 8.08, |
| "learning_rate": 1.1926686217008798e-05, |
| "loss": 0.0606, |
| "step": 16525 |
| }, |
| { |
| "epoch": 8.09, |
| "learning_rate": 1.191446725317693e-05, |
| "loss": 0.063, |
| "step": 16550 |
| }, |
| { |
| "epoch": 8.1, |
| "learning_rate": 1.1902248289345065e-05, |
| "loss": 0.0679, |
| "step": 16575 |
| }, |
| { |
| "epoch": 8.11, |
| "learning_rate": 1.1890029325513197e-05, |
| "loss": 0.0748, |
| "step": 16600 |
| }, |
| { |
| "epoch": 8.13, |
| "learning_rate": 1.187781036168133e-05, |
| "loss": 0.0666, |
| "step": 16625 |
| }, |
| { |
| "epoch": 8.14, |
| "learning_rate": 1.1865591397849463e-05, |
| "loss": 0.0703, |
| "step": 16650 |
| }, |
| { |
| "epoch": 8.15, |
| "learning_rate": 1.1853372434017596e-05, |
| "loss": 0.0756, |
| "step": 16675 |
| }, |
| { |
| "epoch": 8.16, |
| "learning_rate": 1.1841153470185728e-05, |
| "loss": 0.0611, |
| "step": 16700 |
| }, |
| { |
| "epoch": 8.17, |
| "learning_rate": 1.1828934506353862e-05, |
| "loss": 0.0766, |
| "step": 16725 |
| }, |
| { |
| "epoch": 8.19, |
| "learning_rate": 1.1816715542521995e-05, |
| "loss": 0.067, |
| "step": 16750 |
| }, |
| { |
| "epoch": 8.2, |
| "learning_rate": 1.1804496578690127e-05, |
| "loss": 0.075, |
| "step": 16775 |
| }, |
| { |
| "epoch": 8.21, |
| "learning_rate": 1.1792277614858261e-05, |
| "loss": 0.0887, |
| "step": 16800 |
| }, |
| { |
| "epoch": 8.22, |
| "learning_rate": 1.1780058651026394e-05, |
| "loss": 0.0652, |
| "step": 16825 |
| }, |
| { |
| "epoch": 8.24, |
| "learning_rate": 1.1767839687194526e-05, |
| "loss": 0.0787, |
| "step": 16850 |
| }, |
| { |
| "epoch": 8.25, |
| "learning_rate": 1.175562072336266e-05, |
| "loss": 0.0753, |
| "step": 16875 |
| }, |
| { |
| "epoch": 8.26, |
| "learning_rate": 1.1743401759530793e-05, |
| "loss": 0.0747, |
| "step": 16900 |
| }, |
| { |
| "epoch": 8.27, |
| "learning_rate": 1.1731182795698925e-05, |
| "loss": 0.0681, |
| "step": 16925 |
| }, |
| { |
| "epoch": 8.28, |
| "learning_rate": 1.1718963831867059e-05, |
| "loss": 0.0716, |
| "step": 16950 |
| }, |
| { |
| "epoch": 8.3, |
| "learning_rate": 1.1706744868035191e-05, |
| "loss": 0.0696, |
| "step": 16975 |
| }, |
| { |
| "epoch": 8.31, |
| "learning_rate": 1.1694525904203324e-05, |
| "loss": 0.0787, |
| "step": 17000 |
| }, |
| { |
| "epoch": 8.32, |
| "learning_rate": 1.1682306940371458e-05, |
| "loss": 0.0835, |
| "step": 17025 |
| }, |
| { |
| "epoch": 8.33, |
| "learning_rate": 1.167008797653959e-05, |
| "loss": 0.0712, |
| "step": 17050 |
| }, |
| { |
| "epoch": 8.35, |
| "learning_rate": 1.1657869012707723e-05, |
| "loss": 0.0728, |
| "step": 17075 |
| }, |
| { |
| "epoch": 8.36, |
| "learning_rate": 1.1645650048875857e-05, |
| "loss": 0.0765, |
| "step": 17100 |
| }, |
| { |
| "epoch": 8.37, |
| "learning_rate": 1.163343108504399e-05, |
| "loss": 0.0742, |
| "step": 17125 |
| }, |
| { |
| "epoch": 8.38, |
| "learning_rate": 1.1621212121212122e-05, |
| "loss": 0.0949, |
| "step": 17150 |
| }, |
| { |
| "epoch": 8.39, |
| "learning_rate": 1.1608993157380256e-05, |
| "loss": 0.0709, |
| "step": 17175 |
| }, |
| { |
| "epoch": 8.41, |
| "learning_rate": 1.1596774193548388e-05, |
| "loss": 0.0712, |
| "step": 17200 |
| }, |
| { |
| "epoch": 8.42, |
| "learning_rate": 1.158455522971652e-05, |
| "loss": 0.082, |
| "step": 17225 |
| }, |
| { |
| "epoch": 8.43, |
| "learning_rate": 1.1572336265884655e-05, |
| "loss": 0.0783, |
| "step": 17250 |
| }, |
| { |
| "epoch": 8.44, |
| "learning_rate": 1.1560117302052787e-05, |
| "loss": 0.068, |
| "step": 17275 |
| }, |
| { |
| "epoch": 8.46, |
| "learning_rate": 1.154789833822092e-05, |
| "loss": 0.0788, |
| "step": 17300 |
| }, |
| { |
| "epoch": 8.47, |
| "learning_rate": 1.1535679374389054e-05, |
| "loss": 0.0783, |
| "step": 17325 |
| }, |
| { |
| "epoch": 8.48, |
| "learning_rate": 1.1523460410557186e-05, |
| "loss": 0.0796, |
| "step": 17350 |
| }, |
| { |
| "epoch": 8.49, |
| "learning_rate": 1.1511241446725318e-05, |
| "loss": 0.0751, |
| "step": 17375 |
| }, |
| { |
| "epoch": 8.5, |
| "learning_rate": 1.149902248289345e-05, |
| "loss": 0.0754, |
| "step": 17400 |
| }, |
| { |
| "epoch": 8.52, |
| "learning_rate": 1.1486803519061585e-05, |
| "loss": 0.0855, |
| "step": 17425 |
| }, |
| { |
| "epoch": 8.53, |
| "learning_rate": 1.1474584555229717e-05, |
| "loss": 0.0773, |
| "step": 17450 |
| }, |
| { |
| "epoch": 8.54, |
| "learning_rate": 1.146236559139785e-05, |
| "loss": 0.0844, |
| "step": 17475 |
| }, |
| { |
| "epoch": 8.55, |
| "learning_rate": 1.1450146627565984e-05, |
| "loss": 0.0797, |
| "step": 17500 |
| }, |
| { |
| "epoch": 8.57, |
| "learning_rate": 1.1437927663734116e-05, |
| "loss": 0.0675, |
| "step": 17525 |
| }, |
| { |
| "epoch": 8.58, |
| "learning_rate": 1.1425708699902249e-05, |
| "loss": 0.0742, |
| "step": 17550 |
| }, |
| { |
| "epoch": 8.59, |
| "learning_rate": 1.1413489736070383e-05, |
| "loss": 0.0761, |
| "step": 17575 |
| }, |
| { |
| "epoch": 8.6, |
| "learning_rate": 1.1401270772238515e-05, |
| "loss": 0.0806, |
| "step": 17600 |
| }, |
| { |
| "epoch": 8.61, |
| "learning_rate": 1.1389051808406647e-05, |
| "loss": 0.071, |
| "step": 17625 |
| }, |
| { |
| "epoch": 8.63, |
| "learning_rate": 1.1376832844574782e-05, |
| "loss": 0.0782, |
| "step": 17650 |
| }, |
| { |
| "epoch": 8.64, |
| "learning_rate": 1.1364613880742914e-05, |
| "loss": 0.067, |
| "step": 17675 |
| }, |
| { |
| "epoch": 8.65, |
| "learning_rate": 1.1352394916911046e-05, |
| "loss": 0.0778, |
| "step": 17700 |
| }, |
| { |
| "epoch": 8.66, |
| "learning_rate": 1.134017595307918e-05, |
| "loss": 0.0729, |
| "step": 17725 |
| }, |
| { |
| "epoch": 8.68, |
| "learning_rate": 1.1327956989247313e-05, |
| "loss": 0.0775, |
| "step": 17750 |
| }, |
| { |
| "epoch": 8.69, |
| "learning_rate": 1.1315738025415445e-05, |
| "loss": 0.0726, |
| "step": 17775 |
| }, |
| { |
| "epoch": 8.7, |
| "learning_rate": 1.130351906158358e-05, |
| "loss": 0.0761, |
| "step": 17800 |
| }, |
| { |
| "epoch": 8.71, |
| "learning_rate": 1.1291300097751712e-05, |
| "loss": 0.0824, |
| "step": 17825 |
| }, |
| { |
| "epoch": 8.72, |
| "learning_rate": 1.1279081133919844e-05, |
| "loss": 0.0753, |
| "step": 17850 |
| }, |
| { |
| "epoch": 8.74, |
| "learning_rate": 1.1266862170087978e-05, |
| "loss": 0.069, |
| "step": 17875 |
| }, |
| { |
| "epoch": 8.75, |
| "learning_rate": 1.125464320625611e-05, |
| "loss": 0.0848, |
| "step": 17900 |
| }, |
| { |
| "epoch": 8.76, |
| "learning_rate": 1.1242424242424243e-05, |
| "loss": 0.0845, |
| "step": 17925 |
| }, |
| { |
| "epoch": 8.77, |
| "learning_rate": 1.1230205278592377e-05, |
| "loss": 0.0823, |
| "step": 17950 |
| }, |
| { |
| "epoch": 8.79, |
| "learning_rate": 1.121798631476051e-05, |
| "loss": 0.0764, |
| "step": 17975 |
| }, |
| { |
| "epoch": 8.8, |
| "learning_rate": 1.1205767350928642e-05, |
| "loss": 0.0899, |
| "step": 18000 |
| }, |
| { |
| "epoch": 8.81, |
| "learning_rate": 1.1193548387096776e-05, |
| "loss": 0.0819, |
| "step": 18025 |
| }, |
| { |
| "epoch": 8.82, |
| "learning_rate": 1.1181329423264908e-05, |
| "loss": 0.0823, |
| "step": 18050 |
| }, |
| { |
| "epoch": 8.83, |
| "learning_rate": 1.1169599217986315e-05, |
| "loss": 0.0717, |
| "step": 18075 |
| }, |
| { |
| "epoch": 8.85, |
| "learning_rate": 1.1157380254154448e-05, |
| "loss": 0.0776, |
| "step": 18100 |
| }, |
| { |
| "epoch": 8.86, |
| "learning_rate": 1.1145161290322582e-05, |
| "loss": 0.0786, |
| "step": 18125 |
| }, |
| { |
| "epoch": 8.87, |
| "learning_rate": 1.1132942326490714e-05, |
| "loss": 0.0759, |
| "step": 18150 |
| }, |
| { |
| "epoch": 8.88, |
| "learning_rate": 1.1120723362658847e-05, |
| "loss": 0.0759, |
| "step": 18175 |
| }, |
| { |
| "epoch": 8.9, |
| "learning_rate": 1.110850439882698e-05, |
| "loss": 0.0659, |
| "step": 18200 |
| }, |
| { |
| "epoch": 8.91, |
| "learning_rate": 1.1096285434995113e-05, |
| "loss": 0.0849, |
| "step": 18225 |
| }, |
| { |
| "epoch": 8.92, |
| "learning_rate": 1.1084066471163246e-05, |
| "loss": 0.0795, |
| "step": 18250 |
| }, |
| { |
| "epoch": 8.93, |
| "learning_rate": 1.107184750733138e-05, |
| "loss": 0.0871, |
| "step": 18275 |
| }, |
| { |
| "epoch": 8.94, |
| "learning_rate": 1.1059628543499512e-05, |
| "loss": 0.0786, |
| "step": 18300 |
| }, |
| { |
| "epoch": 8.96, |
| "learning_rate": 1.1047409579667645e-05, |
| "loss": 0.0857, |
| "step": 18325 |
| }, |
| { |
| "epoch": 8.97, |
| "learning_rate": 1.1035190615835779e-05, |
| "loss": 0.0847, |
| "step": 18350 |
| }, |
| { |
| "epoch": 8.98, |
| "learning_rate": 1.1022971652003911e-05, |
| "loss": 0.0869, |
| "step": 18375 |
| }, |
| { |
| "epoch": 8.99, |
| "learning_rate": 1.1010752688172043e-05, |
| "loss": 0.0711, |
| "step": 18400 |
| }, |
| { |
| "epoch": 9.01, |
| "learning_rate": 1.0998533724340177e-05, |
| "loss": 0.0767, |
| "step": 18425 |
| }, |
| { |
| "epoch": 9.02, |
| "learning_rate": 1.098631476050831e-05, |
| "loss": 0.063, |
| "step": 18450 |
| }, |
| { |
| "epoch": 9.03, |
| "learning_rate": 1.0974095796676442e-05, |
| "loss": 0.0544, |
| "step": 18475 |
| }, |
| { |
| "epoch": 9.04, |
| "learning_rate": 1.0961876832844576e-05, |
| "loss": 0.0664, |
| "step": 18500 |
| }, |
| { |
| "epoch": 9.05, |
| "learning_rate": 1.0949657869012709e-05, |
| "loss": 0.0577, |
| "step": 18525 |
| }, |
| { |
| "epoch": 9.07, |
| "learning_rate": 1.0937438905180841e-05, |
| "loss": 0.0518, |
| "step": 18550 |
| }, |
| { |
| "epoch": 9.08, |
| "learning_rate": 1.0925219941348975e-05, |
| "loss": 0.0689, |
| "step": 18575 |
| }, |
| { |
| "epoch": 9.09, |
| "learning_rate": 1.0913000977517108e-05, |
| "loss": 0.0557, |
| "step": 18600 |
| }, |
| { |
| "epoch": 9.1, |
| "learning_rate": 1.090078201368524e-05, |
| "loss": 0.0646, |
| "step": 18625 |
| }, |
| { |
| "epoch": 9.12, |
| "learning_rate": 1.0888563049853374e-05, |
| "loss": 0.0705, |
| "step": 18650 |
| }, |
| { |
| "epoch": 9.13, |
| "learning_rate": 1.0876344086021507e-05, |
| "loss": 0.049, |
| "step": 18675 |
| }, |
| { |
| "epoch": 9.14, |
| "learning_rate": 1.0864125122189639e-05, |
| "loss": 0.0568, |
| "step": 18700 |
| }, |
| { |
| "epoch": 9.15, |
| "learning_rate": 1.0851906158357773e-05, |
| "loss": 0.0547, |
| "step": 18725 |
| }, |
| { |
| "epoch": 9.16, |
| "learning_rate": 1.0839687194525905e-05, |
| "loss": 0.0629, |
| "step": 18750 |
| }, |
| { |
| "epoch": 9.18, |
| "learning_rate": 1.0827468230694038e-05, |
| "loss": 0.0647, |
| "step": 18775 |
| }, |
| { |
| "epoch": 9.19, |
| "learning_rate": 1.0815249266862172e-05, |
| "loss": 0.0692, |
| "step": 18800 |
| }, |
| { |
| "epoch": 9.2, |
| "learning_rate": 1.0803030303030304e-05, |
| "loss": 0.0495, |
| "step": 18825 |
| }, |
| { |
| "epoch": 9.21, |
| "learning_rate": 1.0790811339198437e-05, |
| "loss": 0.0628, |
| "step": 18850 |
| }, |
| { |
| "epoch": 9.23, |
| "learning_rate": 1.077859237536657e-05, |
| "loss": 0.0563, |
| "step": 18875 |
| }, |
| { |
| "epoch": 9.24, |
| "learning_rate": 1.0766373411534703e-05, |
| "loss": 0.0657, |
| "step": 18900 |
| }, |
| { |
| "epoch": 9.25, |
| "learning_rate": 1.0754154447702836e-05, |
| "loss": 0.0535, |
| "step": 18925 |
| }, |
| { |
| "epoch": 9.26, |
| "learning_rate": 1.0741935483870968e-05, |
| "loss": 0.0667, |
| "step": 18950 |
| }, |
| { |
| "epoch": 9.27, |
| "learning_rate": 1.0729716520039102e-05, |
| "loss": 0.0594, |
| "step": 18975 |
| }, |
| { |
| "epoch": 9.29, |
| "learning_rate": 1.0717497556207235e-05, |
| "loss": 0.0628, |
| "step": 19000 |
| }, |
| { |
| "epoch": 9.3, |
| "learning_rate": 1.0705278592375367e-05, |
| "loss": 0.058, |
| "step": 19025 |
| }, |
| { |
| "epoch": 9.31, |
| "learning_rate": 1.0693059628543501e-05, |
| "loss": 0.0677, |
| "step": 19050 |
| }, |
| { |
| "epoch": 9.32, |
| "learning_rate": 1.0680840664711633e-05, |
| "loss": 0.0623, |
| "step": 19075 |
| }, |
| { |
| "epoch": 9.34, |
| "learning_rate": 1.0668621700879766e-05, |
| "loss": 0.0631, |
| "step": 19100 |
| }, |
| { |
| "epoch": 9.35, |
| "learning_rate": 1.06564027370479e-05, |
| "loss": 0.0616, |
| "step": 19125 |
| }, |
| { |
| "epoch": 9.36, |
| "learning_rate": 1.0644183773216032e-05, |
| "loss": 0.0613, |
| "step": 19150 |
| }, |
| { |
| "epoch": 9.37, |
| "learning_rate": 1.0631964809384165e-05, |
| "loss": 0.066, |
| "step": 19175 |
| }, |
| { |
| "epoch": 9.38, |
| "learning_rate": 1.0619745845552299e-05, |
| "loss": 0.0653, |
| "step": 19200 |
| }, |
| { |
| "epoch": 9.4, |
| "learning_rate": 1.0607526881720431e-05, |
| "loss": 0.0701, |
| "step": 19225 |
| }, |
| { |
| "epoch": 9.41, |
| "learning_rate": 1.0595307917888564e-05, |
| "loss": 0.0612, |
| "step": 19250 |
| }, |
| { |
| "epoch": 9.42, |
| "learning_rate": 1.0583088954056698e-05, |
| "loss": 0.055, |
| "step": 19275 |
| }, |
| { |
| "epoch": 9.43, |
| "learning_rate": 1.057086999022483e-05, |
| "loss": 0.0603, |
| "step": 19300 |
| }, |
| { |
| "epoch": 9.45, |
| "learning_rate": 1.0558651026392963e-05, |
| "loss": 0.0648, |
| "step": 19325 |
| }, |
| { |
| "epoch": 9.46, |
| "learning_rate": 1.0546432062561097e-05, |
| "loss": 0.0659, |
| "step": 19350 |
| }, |
| { |
| "epoch": 9.47, |
| "learning_rate": 1.0534213098729229e-05, |
| "loss": 0.0548, |
| "step": 19375 |
| }, |
| { |
| "epoch": 9.48, |
| "learning_rate": 1.0521994134897361e-05, |
| "loss": 0.0605, |
| "step": 19400 |
| }, |
| { |
| "epoch": 9.49, |
| "learning_rate": 1.0509775171065496e-05, |
| "loss": 0.0542, |
| "step": 19425 |
| }, |
| { |
| "epoch": 9.51, |
| "learning_rate": 1.0497556207233628e-05, |
| "loss": 0.0647, |
| "step": 19450 |
| }, |
| { |
| "epoch": 9.52, |
| "learning_rate": 1.048533724340176e-05, |
| "loss": 0.0677, |
| "step": 19475 |
| }, |
| { |
| "epoch": 9.53, |
| "learning_rate": 1.0473118279569894e-05, |
| "loss": 0.06, |
| "step": 19500 |
| }, |
| { |
| "epoch": 9.54, |
| "learning_rate": 1.0460899315738027e-05, |
| "loss": 0.0664, |
| "step": 19525 |
| }, |
| { |
| "epoch": 9.56, |
| "learning_rate": 1.044868035190616e-05, |
| "loss": 0.0666, |
| "step": 19550 |
| }, |
| { |
| "epoch": 9.57, |
| "learning_rate": 1.0436461388074293e-05, |
| "loss": 0.0576, |
| "step": 19575 |
| }, |
| { |
| "epoch": 9.58, |
| "learning_rate": 1.0424242424242426e-05, |
| "loss": 0.0586, |
| "step": 19600 |
| }, |
| { |
| "epoch": 9.59, |
| "learning_rate": 1.0412023460410558e-05, |
| "loss": 0.0631, |
| "step": 19625 |
| }, |
| { |
| "epoch": 9.6, |
| "learning_rate": 1.0399804496578692e-05, |
| "loss": 0.0688, |
| "step": 19650 |
| }, |
| { |
| "epoch": 9.62, |
| "learning_rate": 1.0387585532746825e-05, |
| "loss": 0.0619, |
| "step": 19675 |
| }, |
| { |
| "epoch": 9.63, |
| "learning_rate": 1.0375366568914957e-05, |
| "loss": 0.0695, |
| "step": 19700 |
| }, |
| { |
| "epoch": 9.64, |
| "learning_rate": 1.0363147605083091e-05, |
| "loss": 0.0694, |
| "step": 19725 |
| }, |
| { |
| "epoch": 9.65, |
| "learning_rate": 1.0350928641251224e-05, |
| "loss": 0.0654, |
| "step": 19750 |
| }, |
| { |
| "epoch": 9.67, |
| "learning_rate": 1.0338709677419356e-05, |
| "loss": 0.0688, |
| "step": 19775 |
| }, |
| { |
| "epoch": 9.68, |
| "learning_rate": 1.032649071358749e-05, |
| "loss": 0.0657, |
| "step": 19800 |
| }, |
| { |
| "epoch": 9.69, |
| "learning_rate": 1.0314271749755622e-05, |
| "loss": 0.0575, |
| "step": 19825 |
| }, |
| { |
| "epoch": 9.7, |
| "learning_rate": 1.0302052785923755e-05, |
| "loss": 0.0545, |
| "step": 19850 |
| }, |
| { |
| "epoch": 9.71, |
| "learning_rate": 1.0289833822091889e-05, |
| "loss": 0.0694, |
| "step": 19875 |
| }, |
| { |
| "epoch": 9.73, |
| "learning_rate": 1.0277614858260021e-05, |
| "loss": 0.0706, |
| "step": 19900 |
| }, |
| { |
| "epoch": 9.74, |
| "learning_rate": 1.0265395894428154e-05, |
| "loss": 0.061, |
| "step": 19925 |
| }, |
| { |
| "epoch": 9.75, |
| "learning_rate": 1.0253176930596286e-05, |
| "loss": 0.064, |
| "step": 19950 |
| }, |
| { |
| "epoch": 9.76, |
| "learning_rate": 1.024095796676442e-05, |
| "loss": 0.0629, |
| "step": 19975 |
| }, |
| { |
| "epoch": 9.78, |
| "learning_rate": 1.0228739002932553e-05, |
| "loss": 0.0654, |
| "step": 20000 |
| }, |
| { |
| "epoch": 9.79, |
| "learning_rate": 1.0216520039100685e-05, |
| "loss": 0.0594, |
| "step": 20025 |
| }, |
| { |
| "epoch": 9.8, |
| "learning_rate": 1.0204301075268819e-05, |
| "loss": 0.0644, |
| "step": 20050 |
| }, |
| { |
| "epoch": 9.81, |
| "learning_rate": 1.0192082111436951e-05, |
| "loss": 0.0781, |
| "step": 20075 |
| }, |
| { |
| "epoch": 9.82, |
| "learning_rate": 1.0179863147605084e-05, |
| "loss": 0.0659, |
| "step": 20100 |
| }, |
| { |
| "epoch": 9.84, |
| "learning_rate": 1.0167644183773218e-05, |
| "loss": 0.072, |
| "step": 20125 |
| }, |
| { |
| "epoch": 9.85, |
| "learning_rate": 1.015542521994135e-05, |
| "loss": 0.0661, |
| "step": 20150 |
| }, |
| { |
| "epoch": 9.86, |
| "learning_rate": 1.0143206256109483e-05, |
| "loss": 0.0615, |
| "step": 20175 |
| }, |
| { |
| "epoch": 9.87, |
| "learning_rate": 1.0130987292277617e-05, |
| "loss": 0.0625, |
| "step": 20200 |
| }, |
| { |
| "epoch": 9.89, |
| "learning_rate": 1.011876832844575e-05, |
| "loss": 0.0634, |
| "step": 20225 |
| }, |
| { |
| "epoch": 9.9, |
| "learning_rate": 1.0106549364613882e-05, |
| "loss": 0.0616, |
| "step": 20250 |
| }, |
| { |
| "epoch": 9.91, |
| "learning_rate": 1.0094330400782016e-05, |
| "loss": 0.0602, |
| "step": 20275 |
| }, |
| { |
| "epoch": 9.92, |
| "learning_rate": 1.0082111436950148e-05, |
| "loss": 0.0739, |
| "step": 20300 |
| }, |
| { |
| "epoch": 9.93, |
| "learning_rate": 1.006989247311828e-05, |
| "loss": 0.0654, |
| "step": 20325 |
| }, |
| { |
| "epoch": 9.95, |
| "learning_rate": 1.0057673509286415e-05, |
| "loss": 0.0724, |
| "step": 20350 |
| }, |
| { |
| "epoch": 9.96, |
| "learning_rate": 1.0045454545454547e-05, |
| "loss": 0.0735, |
| "step": 20375 |
| }, |
| { |
| "epoch": 9.97, |
| "learning_rate": 1.003323558162268e-05, |
| "loss": 0.0652, |
| "step": 20400 |
| }, |
| { |
| "epoch": 9.98, |
| "learning_rate": 1.0021016617790814e-05, |
| "loss": 0.0694, |
| "step": 20425 |
| }, |
| { |
| "epoch": 10.0, |
| "learning_rate": 1.0008797653958946e-05, |
| "loss": 0.071, |
| "step": 20450 |
| }, |
| { |
| "epoch": 10.01, |
| "learning_rate": 9.996578690127078e-06, |
| "loss": 0.0609, |
| "step": 20475 |
| }, |
| { |
| "epoch": 10.02, |
| "learning_rate": 9.98435972629521e-06, |
| "loss": 0.0383, |
| "step": 20500 |
| }, |
| { |
| "epoch": 10.03, |
| "learning_rate": 9.972140762463343e-06, |
| "loss": 0.0541, |
| "step": 20525 |
| }, |
| { |
| "epoch": 10.04, |
| "learning_rate": 9.959921798631477e-06, |
| "loss": 0.0552, |
| "step": 20550 |
| }, |
| { |
| "epoch": 10.06, |
| "learning_rate": 9.94770283479961e-06, |
| "loss": 0.052, |
| "step": 20575 |
| }, |
| { |
| "epoch": 10.07, |
| "learning_rate": 9.935483870967742e-06, |
| "loss": 0.0512, |
| "step": 20600 |
| }, |
| { |
| "epoch": 10.08, |
| "learning_rate": 9.923264907135876e-06, |
| "loss": 0.0453, |
| "step": 20625 |
| }, |
| { |
| "epoch": 10.09, |
| "learning_rate": 9.911045943304009e-06, |
| "loss": 0.0465, |
| "step": 20650 |
| }, |
| { |
| "epoch": 10.11, |
| "learning_rate": 9.898826979472141e-06, |
| "loss": 0.0522, |
| "step": 20675 |
| }, |
| { |
| "epoch": 10.12, |
| "learning_rate": 9.886608015640275e-06, |
| "loss": 0.054, |
| "step": 20700 |
| }, |
| { |
| "epoch": 10.13, |
| "learning_rate": 9.874389051808407e-06, |
| "loss": 0.0575, |
| "step": 20725 |
| }, |
| { |
| "epoch": 10.14, |
| "learning_rate": 9.86217008797654e-06, |
| "loss": 0.0526, |
| "step": 20750 |
| }, |
| { |
| "epoch": 10.15, |
| "learning_rate": 9.849951124144674e-06, |
| "loss": 0.0572, |
| "step": 20775 |
| }, |
| { |
| "epoch": 10.17, |
| "learning_rate": 9.837732160312806e-06, |
| "loss": 0.0539, |
| "step": 20800 |
| }, |
| { |
| "epoch": 10.18, |
| "learning_rate": 9.825513196480939e-06, |
| "loss": 0.0505, |
| "step": 20825 |
| }, |
| { |
| "epoch": 10.19, |
| "learning_rate": 9.813294232649073e-06, |
| "loss": 0.0607, |
| "step": 20850 |
| }, |
| { |
| "epoch": 10.2, |
| "learning_rate": 9.801075268817205e-06, |
| "loss": 0.0487, |
| "step": 20875 |
| }, |
| { |
| "epoch": 10.22, |
| "learning_rate": 9.788856304985338e-06, |
| "loss": 0.0438, |
| "step": 20900 |
| }, |
| { |
| "epoch": 10.23, |
| "learning_rate": 9.776637341153472e-06, |
| "loss": 0.0585, |
| "step": 20925 |
| }, |
| { |
| "epoch": 10.24, |
| "learning_rate": 9.764418377321604e-06, |
| "loss": 0.0608, |
| "step": 20950 |
| }, |
| { |
| "epoch": 10.25, |
| "learning_rate": 9.752199413489737e-06, |
| "loss": 0.0421, |
| "step": 20975 |
| }, |
| { |
| "epoch": 10.26, |
| "learning_rate": 9.73998044965787e-06, |
| "loss": 0.0511, |
| "step": 21000 |
| }, |
| { |
| "epoch": 10.28, |
| "learning_rate": 9.727761485826003e-06, |
| "loss": 0.0517, |
| "step": 21025 |
| }, |
| { |
| "epoch": 10.29, |
| "learning_rate": 9.715542521994135e-06, |
| "loss": 0.0639, |
| "step": 21050 |
| }, |
| { |
| "epoch": 10.3, |
| "learning_rate": 9.703323558162268e-06, |
| "loss": 0.0544, |
| "step": 21075 |
| }, |
| { |
| "epoch": 10.31, |
| "learning_rate": 9.691104594330402e-06, |
| "loss": 0.049, |
| "step": 21100 |
| }, |
| { |
| "epoch": 10.33, |
| "learning_rate": 9.678885630498534e-06, |
| "loss": 0.0522, |
| "step": 21125 |
| }, |
| { |
| "epoch": 10.34, |
| "learning_rate": 9.666666666666667e-06, |
| "loss": 0.0596, |
| "step": 21150 |
| }, |
| { |
| "epoch": 10.35, |
| "learning_rate": 9.6544477028348e-06, |
| "loss": 0.0452, |
| "step": 21175 |
| }, |
| { |
| "epoch": 10.36, |
| "learning_rate": 9.642228739002933e-06, |
| "loss": 0.0555, |
| "step": 21200 |
| }, |
| { |
| "epoch": 10.37, |
| "learning_rate": 9.630009775171066e-06, |
| "loss": 0.0507, |
| "step": 21225 |
| }, |
| { |
| "epoch": 10.39, |
| "learning_rate": 9.6177908113392e-06, |
| "loss": 0.0371, |
| "step": 21250 |
| }, |
| { |
| "epoch": 10.4, |
| "learning_rate": 9.605571847507332e-06, |
| "loss": 0.0534, |
| "step": 21275 |
| }, |
| { |
| "epoch": 10.41, |
| "learning_rate": 9.593352883675465e-06, |
| "loss": 0.0482, |
| "step": 21300 |
| }, |
| { |
| "epoch": 10.42, |
| "learning_rate": 9.581133919843599e-06, |
| "loss": 0.0473, |
| "step": 21325 |
| }, |
| { |
| "epoch": 10.43, |
| "learning_rate": 9.568914956011731e-06, |
| "loss": 0.0538, |
| "step": 21350 |
| }, |
| { |
| "epoch": 10.45, |
| "learning_rate": 9.556695992179863e-06, |
| "loss": 0.0545, |
| "step": 21375 |
| }, |
| { |
| "epoch": 10.46, |
| "learning_rate": 9.544477028347998e-06, |
| "loss": 0.0528, |
| "step": 21400 |
| }, |
| { |
| "epoch": 10.47, |
| "learning_rate": 9.53225806451613e-06, |
| "loss": 0.0476, |
| "step": 21425 |
| }, |
| { |
| "epoch": 10.48, |
| "learning_rate": 9.520039100684262e-06, |
| "loss": 0.0582, |
| "step": 21450 |
| }, |
| { |
| "epoch": 10.5, |
| "learning_rate": 9.507820136852396e-06, |
| "loss": 0.05, |
| "step": 21475 |
| }, |
| { |
| "epoch": 10.51, |
| "learning_rate": 9.495601173020529e-06, |
| "loss": 0.0498, |
| "step": 21500 |
| }, |
| { |
| "epoch": 10.52, |
| "learning_rate": 9.483382209188661e-06, |
| "loss": 0.0537, |
| "step": 21525 |
| }, |
| { |
| "epoch": 10.53, |
| "learning_rate": 9.471163245356795e-06, |
| "loss": 0.0596, |
| "step": 21550 |
| }, |
| { |
| "epoch": 10.54, |
| "learning_rate": 9.458944281524928e-06, |
| "loss": 0.0495, |
| "step": 21575 |
| }, |
| { |
| "epoch": 10.56, |
| "learning_rate": 9.44672531769306e-06, |
| "loss": 0.056, |
| "step": 21600 |
| }, |
| { |
| "epoch": 10.57, |
| "learning_rate": 9.434506353861194e-06, |
| "loss": 0.0493, |
| "step": 21625 |
| }, |
| { |
| "epoch": 10.58, |
| "learning_rate": 9.422776148582601e-06, |
| "loss": 0.0561, |
| "step": 21650 |
| }, |
| { |
| "epoch": 10.59, |
| "learning_rate": 9.410557184750734e-06, |
| "loss": 0.0483, |
| "step": 21675 |
| }, |
| { |
| "epoch": 10.61, |
| "learning_rate": 9.398338220918866e-06, |
| "loss": 0.0545, |
| "step": 21700 |
| }, |
| { |
| "epoch": 10.62, |
| "learning_rate": 9.386119257087e-06, |
| "loss": 0.0519, |
| "step": 21725 |
| }, |
| { |
| "epoch": 10.63, |
| "learning_rate": 9.373900293255133e-06, |
| "loss": 0.051, |
| "step": 21750 |
| }, |
| { |
| "epoch": 10.64, |
| "learning_rate": 9.361681329423265e-06, |
| "loss": 0.054, |
| "step": 21775 |
| }, |
| { |
| "epoch": 10.65, |
| "learning_rate": 9.349462365591399e-06, |
| "loss": 0.0594, |
| "step": 21800 |
| }, |
| { |
| "epoch": 10.67, |
| "learning_rate": 9.337243401759531e-06, |
| "loss": 0.0561, |
| "step": 21825 |
| }, |
| { |
| "epoch": 10.68, |
| "learning_rate": 9.325024437927664e-06, |
| "loss": 0.0607, |
| "step": 21850 |
| }, |
| { |
| "epoch": 10.69, |
| "learning_rate": 9.312805474095798e-06, |
| "loss": 0.0416, |
| "step": 21875 |
| }, |
| { |
| "epoch": 10.7, |
| "learning_rate": 9.30058651026393e-06, |
| "loss": 0.053, |
| "step": 21900 |
| }, |
| { |
| "epoch": 10.72, |
| "learning_rate": 9.288367546432063e-06, |
| "loss": 0.0555, |
| "step": 21925 |
| }, |
| { |
| "epoch": 10.73, |
| "learning_rate": 9.276148582600197e-06, |
| "loss": 0.0633, |
| "step": 21950 |
| }, |
| { |
| "epoch": 10.74, |
| "learning_rate": 9.26392961876833e-06, |
| "loss": 0.0621, |
| "step": 21975 |
| }, |
| { |
| "epoch": 10.75, |
| "learning_rate": 9.251710654936462e-06, |
| "loss": 0.052, |
| "step": 22000 |
| }, |
| { |
| "epoch": 10.76, |
| "learning_rate": 9.239491691104596e-06, |
| "loss": 0.0548, |
| "step": 22025 |
| }, |
| { |
| "epoch": 10.78, |
| "learning_rate": 9.227272727272728e-06, |
| "loss": 0.06, |
| "step": 22050 |
| }, |
| { |
| "epoch": 10.79, |
| "learning_rate": 9.21505376344086e-06, |
| "loss": 0.0528, |
| "step": 22075 |
| }, |
| { |
| "epoch": 10.8, |
| "learning_rate": 9.202834799608995e-06, |
| "loss": 0.06, |
| "step": 22100 |
| }, |
| { |
| "epoch": 10.81, |
| "learning_rate": 9.190615835777127e-06, |
| "loss": 0.0561, |
| "step": 22125 |
| }, |
| { |
| "epoch": 10.83, |
| "learning_rate": 9.17839687194526e-06, |
| "loss": 0.057, |
| "step": 22150 |
| }, |
| { |
| "epoch": 10.84, |
| "learning_rate": 9.166177908113393e-06, |
| "loss": 0.0546, |
| "step": 22175 |
| }, |
| { |
| "epoch": 10.85, |
| "learning_rate": 9.153958944281526e-06, |
| "loss": 0.0549, |
| "step": 22200 |
| }, |
| { |
| "epoch": 10.86, |
| "learning_rate": 9.141739980449658e-06, |
| "loss": 0.0615, |
| "step": 22225 |
| }, |
| { |
| "epoch": 10.87, |
| "learning_rate": 9.129521016617792e-06, |
| "loss": 0.0557, |
| "step": 22250 |
| }, |
| { |
| "epoch": 10.89, |
| "learning_rate": 9.117302052785925e-06, |
| "loss": 0.0605, |
| "step": 22275 |
| }, |
| { |
| "epoch": 10.9, |
| "learning_rate": 9.105083088954057e-06, |
| "loss": 0.0611, |
| "step": 22300 |
| }, |
| { |
| "epoch": 10.91, |
| "learning_rate": 9.092864125122191e-06, |
| "loss": 0.0596, |
| "step": 22325 |
| }, |
| { |
| "epoch": 10.92, |
| "learning_rate": 9.080645161290324e-06, |
| "loss": 0.057, |
| "step": 22350 |
| }, |
| { |
| "epoch": 10.94, |
| "learning_rate": 9.068426197458456e-06, |
| "loss": 0.0543, |
| "step": 22375 |
| }, |
| { |
| "epoch": 10.95, |
| "learning_rate": 9.05620723362659e-06, |
| "loss": 0.0604, |
| "step": 22400 |
| }, |
| { |
| "epoch": 10.96, |
| "learning_rate": 9.043988269794723e-06, |
| "loss": 0.0546, |
| "step": 22425 |
| }, |
| { |
| "epoch": 10.97, |
| "learning_rate": 9.031769305962855e-06, |
| "loss": 0.0552, |
| "step": 22450 |
| }, |
| { |
| "epoch": 10.98, |
| "learning_rate": 9.019550342130989e-06, |
| "loss": 0.065, |
| "step": 22475 |
| }, |
| { |
| "epoch": 11.0, |
| "learning_rate": 9.007331378299121e-06, |
| "loss": 0.0511, |
| "step": 22500 |
| }, |
| { |
| "epoch": 11.01, |
| "learning_rate": 8.995112414467254e-06, |
| "loss": 0.0367, |
| "step": 22525 |
| }, |
| { |
| "epoch": 11.02, |
| "learning_rate": 8.982893450635386e-06, |
| "loss": 0.0531, |
| "step": 22550 |
| }, |
| { |
| "epoch": 11.03, |
| "learning_rate": 8.97067448680352e-06, |
| "loss": 0.0524, |
| "step": 22575 |
| }, |
| { |
| "epoch": 11.05, |
| "learning_rate": 8.958455522971653e-06, |
| "loss": 0.0442, |
| "step": 22600 |
| }, |
| { |
| "epoch": 11.06, |
| "learning_rate": 8.946236559139785e-06, |
| "loss": 0.0443, |
| "step": 22625 |
| }, |
| { |
| "epoch": 11.07, |
| "learning_rate": 8.93401759530792e-06, |
| "loss": 0.0395, |
| "step": 22650 |
| }, |
| { |
| "epoch": 11.08, |
| "learning_rate": 8.921798631476052e-06, |
| "loss": 0.0429, |
| "step": 22675 |
| }, |
| { |
| "epoch": 11.09, |
| "learning_rate": 8.909579667644184e-06, |
| "loss": 0.0411, |
| "step": 22700 |
| }, |
| { |
| "epoch": 11.11, |
| "learning_rate": 8.897360703812318e-06, |
| "loss": 0.0516, |
| "step": 22725 |
| }, |
| { |
| "epoch": 11.12, |
| "learning_rate": 8.88514173998045e-06, |
| "loss": 0.0428, |
| "step": 22750 |
| }, |
| { |
| "epoch": 11.13, |
| "learning_rate": 8.872922776148583e-06, |
| "loss": 0.0367, |
| "step": 22775 |
| }, |
| { |
| "epoch": 11.14, |
| "learning_rate": 8.860703812316717e-06, |
| "loss": 0.045, |
| "step": 22800 |
| }, |
| { |
| "epoch": 11.16, |
| "learning_rate": 8.84848484848485e-06, |
| "loss": 0.0447, |
| "step": 22825 |
| }, |
| { |
| "epoch": 11.17, |
| "learning_rate": 8.836265884652982e-06, |
| "loss": 0.0475, |
| "step": 22850 |
| }, |
| { |
| "epoch": 11.18, |
| "learning_rate": 8.824046920821116e-06, |
| "loss": 0.0441, |
| "step": 22875 |
| }, |
| { |
| "epoch": 11.19, |
| "learning_rate": 8.811827956989248e-06, |
| "loss": 0.0416, |
| "step": 22900 |
| }, |
| { |
| "epoch": 11.2, |
| "learning_rate": 8.79960899315738e-06, |
| "loss": 0.0555, |
| "step": 22925 |
| }, |
| { |
| "epoch": 11.22, |
| "learning_rate": 8.787390029325515e-06, |
| "loss": 0.035, |
| "step": 22950 |
| }, |
| { |
| "epoch": 11.23, |
| "learning_rate": 8.775171065493647e-06, |
| "loss": 0.0486, |
| "step": 22975 |
| }, |
| { |
| "epoch": 11.24, |
| "learning_rate": 8.76295210166178e-06, |
| "loss": 0.0478, |
| "step": 23000 |
| }, |
| { |
| "epoch": 11.25, |
| "learning_rate": 8.750733137829914e-06, |
| "loss": 0.0494, |
| "step": 23025 |
| }, |
| { |
| "epoch": 11.27, |
| "learning_rate": 8.738514173998046e-06, |
| "loss": 0.0452, |
| "step": 23050 |
| }, |
| { |
| "epoch": 11.28, |
| "learning_rate": 8.726295210166179e-06, |
| "loss": 0.0398, |
| "step": 23075 |
| }, |
| { |
| "epoch": 11.29, |
| "learning_rate": 8.714076246334313e-06, |
| "loss": 0.0438, |
| "step": 23100 |
| }, |
| { |
| "epoch": 11.3, |
| "learning_rate": 8.701857282502445e-06, |
| "loss": 0.0479, |
| "step": 23125 |
| }, |
| { |
| "epoch": 11.31, |
| "learning_rate": 8.689638318670577e-06, |
| "loss": 0.0484, |
| "step": 23150 |
| }, |
| { |
| "epoch": 11.33, |
| "learning_rate": 8.677419354838712e-06, |
| "loss": 0.04, |
| "step": 23175 |
| }, |
| { |
| "epoch": 11.34, |
| "learning_rate": 8.665200391006844e-06, |
| "loss": 0.043, |
| "step": 23200 |
| }, |
| { |
| "epoch": 11.35, |
| "learning_rate": 8.652981427174976e-06, |
| "loss": 0.0434, |
| "step": 23225 |
| }, |
| { |
| "epoch": 11.36, |
| "learning_rate": 8.64076246334311e-06, |
| "loss": 0.0462, |
| "step": 23250 |
| }, |
| { |
| "epoch": 11.38, |
| "learning_rate": 8.628543499511243e-06, |
| "loss": 0.0479, |
| "step": 23275 |
| }, |
| { |
| "epoch": 11.39, |
| "learning_rate": 8.616324535679375e-06, |
| "loss": 0.0419, |
| "step": 23300 |
| }, |
| { |
| "epoch": 11.4, |
| "learning_rate": 8.604105571847508e-06, |
| "loss": 0.0459, |
| "step": 23325 |
| }, |
| { |
| "epoch": 11.41, |
| "learning_rate": 8.59188660801564e-06, |
| "loss": 0.0449, |
| "step": 23350 |
| }, |
| { |
| "epoch": 11.42, |
| "learning_rate": 8.579667644183774e-06, |
| "loss": 0.0398, |
| "step": 23375 |
| }, |
| { |
| "epoch": 11.44, |
| "learning_rate": 8.567448680351907e-06, |
| "loss": 0.0443, |
| "step": 23400 |
| }, |
| { |
| "epoch": 11.45, |
| "learning_rate": 8.555229716520039e-06, |
| "loss": 0.0518, |
| "step": 23425 |
| }, |
| { |
| "epoch": 11.46, |
| "learning_rate": 8.543010752688173e-06, |
| "loss": 0.0377, |
| "step": 23450 |
| }, |
| { |
| "epoch": 11.47, |
| "learning_rate": 8.530791788856305e-06, |
| "loss": 0.0547, |
| "step": 23475 |
| }, |
| { |
| "epoch": 11.49, |
| "learning_rate": 8.518572825024438e-06, |
| "loss": 0.045, |
| "step": 23500 |
| }, |
| { |
| "epoch": 11.5, |
| "learning_rate": 8.506353861192572e-06, |
| "loss": 0.0444, |
| "step": 23525 |
| }, |
| { |
| "epoch": 11.51, |
| "learning_rate": 8.494134897360704e-06, |
| "loss": 0.0483, |
| "step": 23550 |
| }, |
| { |
| "epoch": 11.52, |
| "learning_rate": 8.481915933528837e-06, |
| "loss": 0.0436, |
| "step": 23575 |
| }, |
| { |
| "epoch": 11.53, |
| "learning_rate": 8.46969696969697e-06, |
| "loss": 0.0505, |
| "step": 23600 |
| }, |
| { |
| "epoch": 11.55, |
| "learning_rate": 8.457478005865103e-06, |
| "loss": 0.0395, |
| "step": 23625 |
| }, |
| { |
| "epoch": 11.56, |
| "learning_rate": 8.445259042033236e-06, |
| "loss": 0.0417, |
| "step": 23650 |
| }, |
| { |
| "epoch": 11.57, |
| "learning_rate": 8.433040078201368e-06, |
| "loss": 0.0577, |
| "step": 23675 |
| }, |
| { |
| "epoch": 11.58, |
| "learning_rate": 8.420821114369502e-06, |
| "loss": 0.0476, |
| "step": 23700 |
| }, |
| { |
| "epoch": 11.6, |
| "learning_rate": 8.408602150537634e-06, |
| "loss": 0.0415, |
| "step": 23725 |
| }, |
| { |
| "epoch": 11.61, |
| "learning_rate": 8.396383186705767e-06, |
| "loss": 0.0455, |
| "step": 23750 |
| }, |
| { |
| "epoch": 11.62, |
| "learning_rate": 8.384164222873901e-06, |
| "loss": 0.0572, |
| "step": 23775 |
| }, |
| { |
| "epoch": 11.63, |
| "learning_rate": 8.371945259042033e-06, |
| "loss": 0.0442, |
| "step": 23800 |
| }, |
| { |
| "epoch": 11.64, |
| "learning_rate": 8.359726295210166e-06, |
| "loss": 0.0376, |
| "step": 23825 |
| }, |
| { |
| "epoch": 11.66, |
| "learning_rate": 8.3475073313783e-06, |
| "loss": 0.0455, |
| "step": 23850 |
| }, |
| { |
| "epoch": 11.67, |
| "learning_rate": 8.335288367546432e-06, |
| "loss": 0.0514, |
| "step": 23875 |
| }, |
| { |
| "epoch": 11.68, |
| "learning_rate": 8.323069403714565e-06, |
| "loss": 0.0496, |
| "step": 23900 |
| }, |
| { |
| "epoch": 11.69, |
| "learning_rate": 8.310850439882699e-06, |
| "loss": 0.0498, |
| "step": 23925 |
| }, |
| { |
| "epoch": 11.71, |
| "learning_rate": 8.298631476050831e-06, |
| "loss": 0.0479, |
| "step": 23950 |
| }, |
| { |
| "epoch": 11.72, |
| "learning_rate": 8.286412512218964e-06, |
| "loss": 0.0423, |
| "step": 23975 |
| }, |
| { |
| "epoch": 11.73, |
| "learning_rate": 8.274193548387098e-06, |
| "loss": 0.0444, |
| "step": 24000 |
| }, |
| { |
| "epoch": 11.74, |
| "learning_rate": 8.26197458455523e-06, |
| "loss": 0.0482, |
| "step": 24025 |
| }, |
| { |
| "epoch": 11.75, |
| "learning_rate": 8.249755620723362e-06, |
| "loss": 0.0502, |
| "step": 24050 |
| }, |
| { |
| "epoch": 11.77, |
| "learning_rate": 8.237536656891497e-06, |
| "loss": 0.0503, |
| "step": 24075 |
| }, |
| { |
| "epoch": 11.78, |
| "learning_rate": 8.225317693059629e-06, |
| "loss": 0.0503, |
| "step": 24100 |
| }, |
| { |
| "epoch": 11.79, |
| "learning_rate": 8.213098729227761e-06, |
| "loss": 0.0424, |
| "step": 24125 |
| }, |
| { |
| "epoch": 11.8, |
| "learning_rate": 8.200879765395895e-06, |
| "loss": 0.0543, |
| "step": 24150 |
| }, |
| { |
| "epoch": 11.82, |
| "learning_rate": 8.188660801564028e-06, |
| "loss": 0.0505, |
| "step": 24175 |
| }, |
| { |
| "epoch": 11.83, |
| "learning_rate": 8.17644183773216e-06, |
| "loss": 0.0513, |
| "step": 24200 |
| }, |
| { |
| "epoch": 11.84, |
| "learning_rate": 8.164222873900294e-06, |
| "loss": 0.0537, |
| "step": 24225 |
| }, |
| { |
| "epoch": 11.85, |
| "learning_rate": 8.152003910068427e-06, |
| "loss": 0.0449, |
| "step": 24250 |
| }, |
| { |
| "epoch": 11.86, |
| "learning_rate": 8.139784946236559e-06, |
| "loss": 0.0469, |
| "step": 24275 |
| }, |
| { |
| "epoch": 11.88, |
| "learning_rate": 8.127565982404693e-06, |
| "loss": 0.055, |
| "step": 24300 |
| }, |
| { |
| "epoch": 11.89, |
| "learning_rate": 8.115347018572826e-06, |
| "loss": 0.0455, |
| "step": 24325 |
| }, |
| { |
| "epoch": 11.9, |
| "learning_rate": 8.103128054740958e-06, |
| "loss": 0.049, |
| "step": 24350 |
| }, |
| { |
| "epoch": 11.91, |
| "learning_rate": 8.090909090909092e-06, |
| "loss": 0.0457, |
| "step": 24375 |
| }, |
| { |
| "epoch": 11.93, |
| "learning_rate": 8.078690127077225e-06, |
| "loss": 0.0512, |
| "step": 24400 |
| }, |
| { |
| "epoch": 11.94, |
| "learning_rate": 8.066471163245357e-06, |
| "loss": 0.052, |
| "step": 24425 |
| }, |
| { |
| "epoch": 11.95, |
| "learning_rate": 8.054252199413491e-06, |
| "loss": 0.0468, |
| "step": 24450 |
| }, |
| { |
| "epoch": 11.96, |
| "learning_rate": 8.042033235581623e-06, |
| "loss": 0.0451, |
| "step": 24475 |
| }, |
| { |
| "epoch": 11.97, |
| "learning_rate": 8.029814271749756e-06, |
| "loss": 0.0443, |
| "step": 24500 |
| }, |
| { |
| "epoch": 11.99, |
| "learning_rate": 8.01759530791789e-06, |
| "loss": 0.0454, |
| "step": 24525 |
| }, |
| { |
| "epoch": 12.0, |
| "learning_rate": 8.005376344086022e-06, |
| "loss": 0.0415, |
| "step": 24550 |
| }, |
| { |
| "epoch": 12.01, |
| "learning_rate": 7.993157380254155e-06, |
| "loss": 0.0313, |
| "step": 24575 |
| }, |
| { |
| "epoch": 12.02, |
| "learning_rate": 7.980938416422289e-06, |
| "loss": 0.0428, |
| "step": 24600 |
| }, |
| { |
| "epoch": 12.04, |
| "learning_rate": 7.968719452590421e-06, |
| "loss": 0.0386, |
| "step": 24625 |
| }, |
| { |
| "epoch": 12.05, |
| "learning_rate": 7.956500488758554e-06, |
| "loss": 0.0357, |
| "step": 24650 |
| }, |
| { |
| "epoch": 12.06, |
| "learning_rate": 7.944281524926688e-06, |
| "loss": 0.0376, |
| "step": 24675 |
| }, |
| { |
| "epoch": 12.07, |
| "learning_rate": 7.93206256109482e-06, |
| "loss": 0.0398, |
| "step": 24700 |
| }, |
| { |
| "epoch": 12.08, |
| "learning_rate": 7.919843597262953e-06, |
| "loss": 0.0413, |
| "step": 24725 |
| }, |
| { |
| "epoch": 12.1, |
| "learning_rate": 7.907624633431087e-06, |
| "loss": 0.0261, |
| "step": 24750 |
| }, |
| { |
| "epoch": 12.11, |
| "learning_rate": 7.895405669599219e-06, |
| "loss": 0.0375, |
| "step": 24775 |
| }, |
| { |
| "epoch": 12.12, |
| "learning_rate": 7.883186705767351e-06, |
| "loss": 0.0396, |
| "step": 24800 |
| }, |
| { |
| "epoch": 12.13, |
| "learning_rate": 7.870967741935484e-06, |
| "loss": 0.0356, |
| "step": 24825 |
| }, |
| { |
| "epoch": 12.15, |
| "learning_rate": 7.858748778103618e-06, |
| "loss": 0.0284, |
| "step": 24850 |
| }, |
| { |
| "epoch": 12.16, |
| "learning_rate": 7.84652981427175e-06, |
| "loss": 0.0369, |
| "step": 24875 |
| }, |
| { |
| "epoch": 12.17, |
| "learning_rate": 7.834310850439883e-06, |
| "loss": 0.033, |
| "step": 24900 |
| }, |
| { |
| "epoch": 12.18, |
| "learning_rate": 7.822091886608017e-06, |
| "loss": 0.0361, |
| "step": 24925 |
| }, |
| { |
| "epoch": 12.19, |
| "learning_rate": 7.80987292277615e-06, |
| "loss": 0.0384, |
| "step": 24950 |
| }, |
| { |
| "epoch": 12.21, |
| "learning_rate": 7.797653958944282e-06, |
| "loss": 0.0418, |
| "step": 24975 |
| }, |
| { |
| "epoch": 12.22, |
| "learning_rate": 7.785434995112416e-06, |
| "loss": 0.0395, |
| "step": 25000 |
| }, |
| { |
| "epoch": 12.23, |
| "learning_rate": 7.773216031280548e-06, |
| "loss": 0.0356, |
| "step": 25025 |
| }, |
| { |
| "epoch": 12.24, |
| "learning_rate": 7.76099706744868e-06, |
| "loss": 0.0326, |
| "step": 25050 |
| }, |
| { |
| "epoch": 12.26, |
| "learning_rate": 7.748778103616815e-06, |
| "loss": 0.0402, |
| "step": 25075 |
| }, |
| { |
| "epoch": 12.27, |
| "learning_rate": 7.736559139784947e-06, |
| "loss": 0.0372, |
| "step": 25100 |
| }, |
| { |
| "epoch": 12.28, |
| "learning_rate": 7.72434017595308e-06, |
| "loss": 0.0486, |
| "step": 25125 |
| }, |
| { |
| "epoch": 12.29, |
| "learning_rate": 7.712121212121213e-06, |
| "loss": 0.0351, |
| "step": 25150 |
| }, |
| { |
| "epoch": 12.3, |
| "learning_rate": 7.699902248289346e-06, |
| "loss": 0.0388, |
| "step": 25175 |
| }, |
| { |
| "epoch": 12.32, |
| "learning_rate": 7.687683284457478e-06, |
| "loss": 0.05, |
| "step": 25200 |
| }, |
| { |
| "epoch": 12.33, |
| "learning_rate": 7.675464320625612e-06, |
| "loss": 0.0448, |
| "step": 25225 |
| }, |
| { |
| "epoch": 12.34, |
| "learning_rate": 7.663245356793745e-06, |
| "loss": 0.0338, |
| "step": 25250 |
| }, |
| { |
| "epoch": 12.35, |
| "learning_rate": 7.651026392961877e-06, |
| "loss": 0.0393, |
| "step": 25275 |
| }, |
| { |
| "epoch": 12.37, |
| "learning_rate": 7.638807429130011e-06, |
| "loss": 0.0363, |
| "step": 25300 |
| }, |
| { |
| "epoch": 12.38, |
| "learning_rate": 7.626588465298144e-06, |
| "loss": 0.039, |
| "step": 25325 |
| }, |
| { |
| "epoch": 12.39, |
| "learning_rate": 7.614369501466277e-06, |
| "loss": 0.0343, |
| "step": 25350 |
| }, |
| { |
| "epoch": 12.4, |
| "learning_rate": 7.602150537634409e-06, |
| "loss": 0.0416, |
| "step": 25375 |
| }, |
| { |
| "epoch": 12.41, |
| "learning_rate": 7.5899315738025426e-06, |
| "loss": 0.0472, |
| "step": 25400 |
| }, |
| { |
| "epoch": 12.43, |
| "learning_rate": 7.577712609970676e-06, |
| "loss": 0.0384, |
| "step": 25425 |
| }, |
| { |
| "epoch": 12.44, |
| "learning_rate": 7.565493646138808e-06, |
| "loss": 0.0376, |
| "step": 25450 |
| }, |
| { |
| "epoch": 12.45, |
| "learning_rate": 7.5532746823069415e-06, |
| "loss": 0.0386, |
| "step": 25475 |
| }, |
| { |
| "epoch": 12.46, |
| "learning_rate": 7.541055718475075e-06, |
| "loss": 0.0413, |
| "step": 25500 |
| }, |
| { |
| "epoch": 12.48, |
| "learning_rate": 7.528836754643207e-06, |
| "loss": 0.0372, |
| "step": 25525 |
| }, |
| { |
| "epoch": 12.49, |
| "learning_rate": 7.51661779081134e-06, |
| "loss": 0.04, |
| "step": 25550 |
| }, |
| { |
| "epoch": 12.5, |
| "learning_rate": 7.504398826979474e-06, |
| "loss": 0.0416, |
| "step": 25575 |
| }, |
| { |
| "epoch": 12.51, |
| "learning_rate": 7.492179863147606e-06, |
| "loss": 0.0443, |
| "step": 25600 |
| }, |
| { |
| "epoch": 12.52, |
| "learning_rate": 7.479960899315739e-06, |
| "loss": 0.038, |
| "step": 25625 |
| }, |
| { |
| "epoch": 12.54, |
| "learning_rate": 7.467741935483872e-06, |
| "loss": 0.0452, |
| "step": 25650 |
| }, |
| { |
| "epoch": 12.55, |
| "learning_rate": 7.455522971652005e-06, |
| "loss": 0.0366, |
| "step": 25675 |
| }, |
| { |
| "epoch": 12.56, |
| "learning_rate": 7.443304007820138e-06, |
| "loss": 0.0463, |
| "step": 25700 |
| }, |
| { |
| "epoch": 12.57, |
| "learning_rate": 7.4310850439882706e-06, |
| "loss": 0.0447, |
| "step": 25725 |
| }, |
| { |
| "epoch": 12.59, |
| "learning_rate": 7.418866080156404e-06, |
| "loss": 0.0411, |
| "step": 25750 |
| }, |
| { |
| "epoch": 12.6, |
| "learning_rate": 7.406647116324537e-06, |
| "loss": 0.0413, |
| "step": 25775 |
| }, |
| { |
| "epoch": 12.61, |
| "learning_rate": 7.3944281524926694e-06, |
| "loss": 0.0322, |
| "step": 25800 |
| }, |
| { |
| "epoch": 12.62, |
| "learning_rate": 7.382697947214077e-06, |
| "loss": 0.0464, |
| "step": 25825 |
| }, |
| { |
| "epoch": 12.63, |
| "learning_rate": 7.37047898338221e-06, |
| "loss": 0.0337, |
| "step": 25850 |
| }, |
| { |
| "epoch": 12.65, |
| "learning_rate": 7.358260019550343e-06, |
| "loss": 0.0365, |
| "step": 25875 |
| }, |
| { |
| "epoch": 12.66, |
| "learning_rate": 7.346041055718476e-06, |
| "loss": 0.0432, |
| "step": 25900 |
| }, |
| { |
| "epoch": 12.67, |
| "learning_rate": 7.333822091886609e-06, |
| "loss": 0.0457, |
| "step": 25925 |
| }, |
| { |
| "epoch": 12.68, |
| "learning_rate": 7.321603128054742e-06, |
| "loss": 0.0368, |
| "step": 25950 |
| }, |
| { |
| "epoch": 12.7, |
| "learning_rate": 7.309384164222874e-06, |
| "loss": 0.0385, |
| "step": 25975 |
| }, |
| { |
| "epoch": 12.71, |
| "learning_rate": 7.2971652003910075e-06, |
| "loss": 0.044, |
| "step": 26000 |
| }, |
| { |
| "epoch": 12.72, |
| "learning_rate": 7.284946236559141e-06, |
| "loss": 0.0417, |
| "step": 26025 |
| }, |
| { |
| "epoch": 12.73, |
| "learning_rate": 7.272727272727273e-06, |
| "loss": 0.0418, |
| "step": 26050 |
| }, |
| { |
| "epoch": 12.74, |
| "learning_rate": 7.260508308895406e-06, |
| "loss": 0.0328, |
| "step": 26075 |
| }, |
| { |
| "epoch": 12.76, |
| "learning_rate": 7.24828934506354e-06, |
| "loss": 0.0343, |
| "step": 26100 |
| }, |
| { |
| "epoch": 12.77, |
| "learning_rate": 7.236070381231672e-06, |
| "loss": 0.0349, |
| "step": 26125 |
| }, |
| { |
| "epoch": 12.78, |
| "learning_rate": 7.223851417399805e-06, |
| "loss": 0.0364, |
| "step": 26150 |
| }, |
| { |
| "epoch": 12.79, |
| "learning_rate": 7.2116324535679386e-06, |
| "loss": 0.0358, |
| "step": 26175 |
| }, |
| { |
| "epoch": 12.81, |
| "learning_rate": 7.199413489736071e-06, |
| "loss": 0.0399, |
| "step": 26200 |
| }, |
| { |
| "epoch": 12.82, |
| "learning_rate": 7.187194525904204e-06, |
| "loss": 0.0376, |
| "step": 26225 |
| }, |
| { |
| "epoch": 12.83, |
| "learning_rate": 7.1749755620723375e-06, |
| "loss": 0.0396, |
| "step": 26250 |
| }, |
| { |
| "epoch": 12.84, |
| "learning_rate": 7.16275659824047e-06, |
| "loss": 0.0377, |
| "step": 26275 |
| }, |
| { |
| "epoch": 12.85, |
| "learning_rate": 7.150537634408603e-06, |
| "loss": 0.0274, |
| "step": 26300 |
| }, |
| { |
| "epoch": 12.87, |
| "learning_rate": 7.138318670576736e-06, |
| "loss": 0.0409, |
| "step": 26325 |
| }, |
| { |
| "epoch": 12.88, |
| "learning_rate": 7.126099706744869e-06, |
| "loss": 0.0409, |
| "step": 26350 |
| }, |
| { |
| "epoch": 12.89, |
| "learning_rate": 7.113880742913002e-06, |
| "loss": 0.0351, |
| "step": 26375 |
| }, |
| { |
| "epoch": 12.9, |
| "learning_rate": 7.101661779081135e-06, |
| "loss": 0.0372, |
| "step": 26400 |
| }, |
| { |
| "epoch": 12.92, |
| "learning_rate": 7.089442815249268e-06, |
| "loss": 0.0394, |
| "step": 26425 |
| }, |
| { |
| "epoch": 12.93, |
| "learning_rate": 7.077223851417401e-06, |
| "loss": 0.0409, |
| "step": 26450 |
| }, |
| { |
| "epoch": 12.94, |
| "learning_rate": 7.065004887585533e-06, |
| "loss": 0.0328, |
| "step": 26475 |
| }, |
| { |
| "epoch": 12.95, |
| "learning_rate": 7.0527859237536665e-06, |
| "loss": 0.0348, |
| "step": 26500 |
| }, |
| { |
| "epoch": 12.96, |
| "learning_rate": 7.0405669599218e-06, |
| "loss": 0.0355, |
| "step": 26525 |
| }, |
| { |
| "epoch": 12.98, |
| "learning_rate": 7.028347996089932e-06, |
| "loss": 0.043, |
| "step": 26550 |
| }, |
| { |
| "epoch": 12.99, |
| "learning_rate": 7.0161290322580654e-06, |
| "loss": 0.0468, |
| "step": 26575 |
| }, |
| { |
| "epoch": 13.0, |
| "learning_rate": 7.003910068426199e-06, |
| "loss": 0.0427, |
| "step": 26600 |
| }, |
| { |
| "epoch": 13.01, |
| "learning_rate": 6.991691104594331e-06, |
| "loss": 0.0343, |
| "step": 26625 |
| }, |
| { |
| "epoch": 13.03, |
| "learning_rate": 6.979472140762464e-06, |
| "loss": 0.0318, |
| "step": 26650 |
| }, |
| { |
| "epoch": 13.04, |
| "learning_rate": 6.9672531769305976e-06, |
| "loss": 0.0302, |
| "step": 26675 |
| }, |
| { |
| "epoch": 13.05, |
| "learning_rate": 6.95503421309873e-06, |
| "loss": 0.032, |
| "step": 26700 |
| }, |
| { |
| "epoch": 13.06, |
| "learning_rate": 6.942815249266863e-06, |
| "loss": 0.0327, |
| "step": 26725 |
| }, |
| { |
| "epoch": 13.07, |
| "learning_rate": 6.9305962854349965e-06, |
| "loss": 0.0335, |
| "step": 26750 |
| }, |
| { |
| "epoch": 13.09, |
| "learning_rate": 6.918377321603129e-06, |
| "loss": 0.0319, |
| "step": 26775 |
| }, |
| { |
| "epoch": 13.1, |
| "learning_rate": 6.906158357771262e-06, |
| "loss": 0.036, |
| "step": 26800 |
| }, |
| { |
| "epoch": 13.11, |
| "learning_rate": 6.893939393939395e-06, |
| "loss": 0.0346, |
| "step": 26825 |
| }, |
| { |
| "epoch": 13.12, |
| "learning_rate": 6.881720430107528e-06, |
| "loss": 0.0309, |
| "step": 26850 |
| }, |
| { |
| "epoch": 13.14, |
| "learning_rate": 6.86950146627566e-06, |
| "loss": 0.0306, |
| "step": 26875 |
| }, |
| { |
| "epoch": 13.15, |
| "learning_rate": 6.8572825024437926e-06, |
| "loss": 0.0311, |
| "step": 26900 |
| }, |
| { |
| "epoch": 13.16, |
| "learning_rate": 6.845063538611926e-06, |
| "loss": 0.0334, |
| "step": 26925 |
| }, |
| { |
| "epoch": 13.17, |
| "learning_rate": 6.832844574780059e-06, |
| "loss": 0.0331, |
| "step": 26950 |
| }, |
| { |
| "epoch": 13.18, |
| "learning_rate": 6.8206256109481915e-06, |
| "loss": 0.0408, |
| "step": 26975 |
| }, |
| { |
| "epoch": 13.2, |
| "learning_rate": 6.808406647116325e-06, |
| "loss": 0.0338, |
| "step": 27000 |
| }, |
| { |
| "epoch": 13.21, |
| "learning_rate": 6.796187683284458e-06, |
| "loss": 0.0267, |
| "step": 27025 |
| }, |
| { |
| "epoch": 13.22, |
| "learning_rate": 6.78396871945259e-06, |
| "loss": 0.0303, |
| "step": 27050 |
| }, |
| { |
| "epoch": 13.23, |
| "learning_rate": 6.771749755620724e-06, |
| "loss": 0.0434, |
| "step": 27075 |
| }, |
| { |
| "epoch": 13.25, |
| "learning_rate": 6.759530791788856e-06, |
| "loss": 0.029, |
| "step": 27100 |
| }, |
| { |
| "epoch": 13.26, |
| "learning_rate": 6.747311827956989e-06, |
| "loss": 0.032, |
| "step": 27125 |
| }, |
| { |
| "epoch": 13.27, |
| "learning_rate": 6.7350928641251225e-06, |
| "loss": 0.0319, |
| "step": 27150 |
| }, |
| { |
| "epoch": 13.28, |
| "learning_rate": 6.722873900293255e-06, |
| "loss": 0.0304, |
| "step": 27175 |
| }, |
| { |
| "epoch": 13.29, |
| "learning_rate": 6.710654936461388e-06, |
| "loss": 0.0315, |
| "step": 27200 |
| }, |
| { |
| "epoch": 13.31, |
| "learning_rate": 6.698435972629521e-06, |
| "loss": 0.0414, |
| "step": 27225 |
| }, |
| { |
| "epoch": 13.32, |
| "learning_rate": 6.686217008797654e-06, |
| "loss": 0.0367, |
| "step": 27250 |
| }, |
| { |
| "epoch": 13.33, |
| "learning_rate": 6.673998044965787e-06, |
| "loss": 0.0297, |
| "step": 27275 |
| }, |
| { |
| "epoch": 13.34, |
| "learning_rate": 6.66177908113392e-06, |
| "loss": 0.0301, |
| "step": 27300 |
| }, |
| { |
| "epoch": 13.36, |
| "learning_rate": 6.649560117302053e-06, |
| "loss": 0.033, |
| "step": 27325 |
| }, |
| { |
| "epoch": 13.37, |
| "learning_rate": 6.637341153470186e-06, |
| "loss": 0.0324, |
| "step": 27350 |
| }, |
| { |
| "epoch": 13.38, |
| "learning_rate": 6.625122189638319e-06, |
| "loss": 0.0251, |
| "step": 27375 |
| }, |
| { |
| "epoch": 13.39, |
| "learning_rate": 6.612903225806452e-06, |
| "loss": 0.035, |
| "step": 27400 |
| }, |
| { |
| "epoch": 13.4, |
| "learning_rate": 6.600684261974585e-06, |
| "loss": 0.0311, |
| "step": 27425 |
| }, |
| { |
| "epoch": 13.42, |
| "learning_rate": 6.588465298142718e-06, |
| "loss": 0.0347, |
| "step": 27450 |
| }, |
| { |
| "epoch": 13.43, |
| "learning_rate": 6.5762463343108505e-06, |
| "loss": 0.0305, |
| "step": 27475 |
| }, |
| { |
| "epoch": 13.44, |
| "learning_rate": 6.564027370478984e-06, |
| "loss": 0.0298, |
| "step": 27500 |
| }, |
| { |
| "epoch": 13.45, |
| "learning_rate": 6.551808406647117e-06, |
| "loss": 0.0313, |
| "step": 27525 |
| }, |
| { |
| "epoch": 13.47, |
| "learning_rate": 6.539589442815249e-06, |
| "loss": 0.0454, |
| "step": 27550 |
| }, |
| { |
| "epoch": 13.48, |
| "learning_rate": 6.527370478983383e-06, |
| "loss": 0.029, |
| "step": 27575 |
| }, |
| { |
| "epoch": 13.49, |
| "learning_rate": 6.515151515151516e-06, |
| "loss": 0.0318, |
| "step": 27600 |
| }, |
| { |
| "epoch": 13.5, |
| "learning_rate": 6.502932551319648e-06, |
| "loss": 0.0381, |
| "step": 27625 |
| }, |
| { |
| "epoch": 13.51, |
| "learning_rate": 6.4907135874877815e-06, |
| "loss": 0.03, |
| "step": 27650 |
| }, |
| { |
| "epoch": 13.53, |
| "learning_rate": 6.478494623655914e-06, |
| "loss": 0.0414, |
| "step": 27675 |
| }, |
| { |
| "epoch": 13.54, |
| "learning_rate": 6.466275659824047e-06, |
| "loss": 0.035, |
| "step": 27700 |
| }, |
| { |
| "epoch": 13.55, |
| "learning_rate": 6.45405669599218e-06, |
| "loss": 0.0411, |
| "step": 27725 |
| }, |
| { |
| "epoch": 13.56, |
| "learning_rate": 6.441837732160313e-06, |
| "loss": 0.0394, |
| "step": 27750 |
| }, |
| { |
| "epoch": 13.58, |
| "learning_rate": 6.429618768328446e-06, |
| "loss": 0.0287, |
| "step": 27775 |
| }, |
| { |
| "epoch": 13.59, |
| "learning_rate": 6.417399804496579e-06, |
| "loss": 0.029, |
| "step": 27800 |
| }, |
| { |
| "epoch": 13.6, |
| "learning_rate": 6.405180840664712e-06, |
| "loss": 0.0396, |
| "step": 27825 |
| }, |
| { |
| "epoch": 13.61, |
| "learning_rate": 6.392961876832845e-06, |
| "loss": 0.0378, |
| "step": 27850 |
| }, |
| { |
| "epoch": 13.62, |
| "learning_rate": 6.380742913000978e-06, |
| "loss": 0.0365, |
| "step": 27875 |
| }, |
| { |
| "epoch": 13.64, |
| "learning_rate": 6.368523949169111e-06, |
| "loss": 0.0385, |
| "step": 27900 |
| }, |
| { |
| "epoch": 13.65, |
| "learning_rate": 6.356304985337244e-06, |
| "loss": 0.04, |
| "step": 27925 |
| }, |
| { |
| "epoch": 13.66, |
| "learning_rate": 6.344086021505377e-06, |
| "loss": 0.0302, |
| "step": 27950 |
| }, |
| { |
| "epoch": 13.67, |
| "learning_rate": 6.332355816226784e-06, |
| "loss": 0.0293, |
| "step": 27975 |
| }, |
| { |
| "epoch": 13.69, |
| "learning_rate": 6.3201368523949165e-06, |
| "loss": 0.0422, |
| "step": 28000 |
| }, |
| { |
| "epoch": 13.7, |
| "learning_rate": 6.30791788856305e-06, |
| "loss": 0.035, |
| "step": 28025 |
| }, |
| { |
| "epoch": 13.71, |
| "learning_rate": 6.295698924731183e-06, |
| "loss": 0.0323, |
| "step": 28050 |
| }, |
| { |
| "epoch": 13.72, |
| "learning_rate": 6.283479960899315e-06, |
| "loss": 0.0275, |
| "step": 28075 |
| }, |
| { |
| "epoch": 13.73, |
| "learning_rate": 6.271260997067449e-06, |
| "loss": 0.0374, |
| "step": 28100 |
| }, |
| { |
| "epoch": 13.75, |
| "learning_rate": 6.259042033235582e-06, |
| "loss": 0.034, |
| "step": 28125 |
| }, |
| { |
| "epoch": 13.76, |
| "learning_rate": 6.246823069403714e-06, |
| "loss": 0.0326, |
| "step": 28150 |
| }, |
| { |
| "epoch": 13.77, |
| "learning_rate": 6.2346041055718476e-06, |
| "loss": 0.0338, |
| "step": 28175 |
| }, |
| { |
| "epoch": 13.78, |
| "learning_rate": 6.222385141739981e-06, |
| "loss": 0.031, |
| "step": 28200 |
| }, |
| { |
| "epoch": 13.8, |
| "learning_rate": 6.210166177908113e-06, |
| "loss": 0.0345, |
| "step": 28225 |
| }, |
| { |
| "epoch": 13.81, |
| "learning_rate": 6.1979472140762465e-06, |
| "loss": 0.0375, |
| "step": 28250 |
| }, |
| { |
| "epoch": 13.82, |
| "learning_rate": 6.18572825024438e-06, |
| "loss": 0.0296, |
| "step": 28275 |
| }, |
| { |
| "epoch": 13.83, |
| "learning_rate": 6.173509286412512e-06, |
| "loss": 0.0329, |
| "step": 28300 |
| }, |
| { |
| "epoch": 13.84, |
| "learning_rate": 6.161290322580645e-06, |
| "loss": 0.0301, |
| "step": 28325 |
| }, |
| { |
| "epoch": 13.86, |
| "learning_rate": 6.149071358748779e-06, |
| "loss": 0.034, |
| "step": 28350 |
| }, |
| { |
| "epoch": 13.87, |
| "learning_rate": 6.136852394916911e-06, |
| "loss": 0.034, |
| "step": 28375 |
| }, |
| { |
| "epoch": 13.88, |
| "learning_rate": 6.124633431085044e-06, |
| "loss": 0.0303, |
| "step": 28400 |
| }, |
| { |
| "epoch": 13.89, |
| "learning_rate": 6.1124144672531775e-06, |
| "loss": 0.0304, |
| "step": 28425 |
| }, |
| { |
| "epoch": 13.91, |
| "learning_rate": 6.10019550342131e-06, |
| "loss": 0.0281, |
| "step": 28450 |
| }, |
| { |
| "epoch": 13.92, |
| "learning_rate": 6.087976539589443e-06, |
| "loss": 0.0342, |
| "step": 28475 |
| }, |
| { |
| "epoch": 13.93, |
| "learning_rate": 6.0757575757575755e-06, |
| "loss": 0.0438, |
| "step": 28500 |
| }, |
| { |
| "epoch": 13.94, |
| "learning_rate": 6.063538611925709e-06, |
| "loss": 0.0241, |
| "step": 28525 |
| }, |
| { |
| "epoch": 13.95, |
| "learning_rate": 6.051319648093842e-06, |
| "loss": 0.0268, |
| "step": 28550 |
| }, |
| { |
| "epoch": 13.97, |
| "learning_rate": 6.0391006842619744e-06, |
| "loss": 0.0462, |
| "step": 28575 |
| }, |
| { |
| "epoch": 13.98, |
| "learning_rate": 6.026881720430108e-06, |
| "loss": 0.0292, |
| "step": 28600 |
| }, |
| { |
| "epoch": 13.99, |
| "learning_rate": 6.014662756598241e-06, |
| "loss": 0.0365, |
| "step": 28625 |
| }, |
| { |
| "epoch": 14.0, |
| "learning_rate": 6.002443792766373e-06, |
| "loss": 0.0397, |
| "step": 28650 |
| }, |
| { |
| "epoch": 14.02, |
| "learning_rate": 5.990224828934507e-06, |
| "loss": 0.0296, |
| "step": 28675 |
| }, |
| { |
| "epoch": 14.03, |
| "learning_rate": 5.97800586510264e-06, |
| "loss": 0.0288, |
| "step": 28700 |
| }, |
| { |
| "epoch": 14.04, |
| "learning_rate": 5.965786901270772e-06, |
| "loss": 0.0268, |
| "step": 28725 |
| }, |
| { |
| "epoch": 14.05, |
| "learning_rate": 5.9535679374389055e-06, |
| "loss": 0.0244, |
| "step": 28750 |
| }, |
| { |
| "epoch": 14.06, |
| "learning_rate": 5.941348973607039e-06, |
| "loss": 0.0288, |
| "step": 28775 |
| }, |
| { |
| "epoch": 14.08, |
| "learning_rate": 5.929130009775171e-06, |
| "loss": 0.0301, |
| "step": 28800 |
| }, |
| { |
| "epoch": 14.09, |
| "learning_rate": 5.916911045943304e-06, |
| "loss": 0.0244, |
| "step": 28825 |
| }, |
| { |
| "epoch": 14.1, |
| "learning_rate": 5.904692082111438e-06, |
| "loss": 0.0289, |
| "step": 28850 |
| }, |
| { |
| "epoch": 14.11, |
| "learning_rate": 5.89247311827957e-06, |
| "loss": 0.0195, |
| "step": 28875 |
| }, |
| { |
| "epoch": 14.13, |
| "learning_rate": 5.880254154447703e-06, |
| "loss": 0.0257, |
| "step": 28900 |
| }, |
| { |
| "epoch": 14.14, |
| "learning_rate": 5.8680351906158365e-06, |
| "loss": 0.0256, |
| "step": 28925 |
| }, |
| { |
| "epoch": 14.15, |
| "learning_rate": 5.855816226783969e-06, |
| "loss": 0.0317, |
| "step": 28950 |
| }, |
| { |
| "epoch": 14.16, |
| "learning_rate": 5.843597262952102e-06, |
| "loss": 0.0346, |
| "step": 28975 |
| }, |
| { |
| "epoch": 14.17, |
| "learning_rate": 5.831378299120235e-06, |
| "loss": 0.0306, |
| "step": 29000 |
| }, |
| { |
| "epoch": 14.19, |
| "learning_rate": 5.819159335288368e-06, |
| "loss": 0.0262, |
| "step": 29025 |
| }, |
| { |
| "epoch": 14.2, |
| "learning_rate": 5.806940371456501e-06, |
| "loss": 0.0255, |
| "step": 29050 |
| }, |
| { |
| "epoch": 14.21, |
| "learning_rate": 5.7947214076246335e-06, |
| "loss": 0.027, |
| "step": 29075 |
| }, |
| { |
| "epoch": 14.22, |
| "learning_rate": 5.782502443792767e-06, |
| "loss": 0.0419, |
| "step": 29100 |
| }, |
| { |
| "epoch": 14.24, |
| "learning_rate": 5.7702834799609e-06, |
| "loss": 0.0211, |
| "step": 29125 |
| }, |
| { |
| "epoch": 14.25, |
| "learning_rate": 5.758064516129032e-06, |
| "loss": 0.0332, |
| "step": 29150 |
| }, |
| { |
| "epoch": 14.26, |
| "learning_rate": 5.745845552297166e-06, |
| "loss": 0.0282, |
| "step": 29175 |
| }, |
| { |
| "epoch": 14.27, |
| "learning_rate": 5.733626588465299e-06, |
| "loss": 0.0274, |
| "step": 29200 |
| }, |
| { |
| "epoch": 14.28, |
| "learning_rate": 5.721407624633431e-06, |
| "loss": 0.0271, |
| "step": 29225 |
| }, |
| { |
| "epoch": 14.3, |
| "learning_rate": 5.7091886608015645e-06, |
| "loss": 0.0276, |
| "step": 29250 |
| }, |
| { |
| "epoch": 14.31, |
| "learning_rate": 5.696969696969698e-06, |
| "loss": 0.0331, |
| "step": 29275 |
| }, |
| { |
| "epoch": 14.32, |
| "learning_rate": 5.68475073313783e-06, |
| "loss": 0.0321, |
| "step": 29300 |
| }, |
| { |
| "epoch": 14.33, |
| "learning_rate": 5.672531769305963e-06, |
| "loss": 0.0253, |
| "step": 29325 |
| }, |
| { |
| "epoch": 14.35, |
| "learning_rate": 5.660312805474097e-06, |
| "loss": 0.0265, |
| "step": 29350 |
| }, |
| { |
| "epoch": 14.36, |
| "learning_rate": 5.648093841642229e-06, |
| "loss": 0.0305, |
| "step": 29375 |
| }, |
| { |
| "epoch": 14.37, |
| "learning_rate": 5.635874877810362e-06, |
| "loss": 0.0266, |
| "step": 29400 |
| }, |
| { |
| "epoch": 14.38, |
| "learning_rate": 5.6236559139784955e-06, |
| "loss": 0.0336, |
| "step": 29425 |
| }, |
| { |
| "epoch": 14.39, |
| "learning_rate": 5.611436950146628e-06, |
| "loss": 0.0267, |
| "step": 29450 |
| }, |
| { |
| "epoch": 14.41, |
| "learning_rate": 5.599217986314761e-06, |
| "loss": 0.0252, |
| "step": 29475 |
| }, |
| { |
| "epoch": 14.42, |
| "learning_rate": 5.586999022482894e-06, |
| "loss": 0.0246, |
| "step": 29500 |
| }, |
| { |
| "epoch": 14.43, |
| "learning_rate": 5.574780058651027e-06, |
| "loss": 0.0302, |
| "step": 29525 |
| }, |
| { |
| "epoch": 14.44, |
| "learning_rate": 5.56256109481916e-06, |
| "loss": 0.0255, |
| "step": 29550 |
| }, |
| { |
| "epoch": 14.46, |
| "learning_rate": 5.550342130987293e-06, |
| "loss": 0.0313, |
| "step": 29575 |
| }, |
| { |
| "epoch": 14.47, |
| "learning_rate": 5.538123167155426e-06, |
| "loss": 0.0289, |
| "step": 29600 |
| }, |
| { |
| "epoch": 14.48, |
| "learning_rate": 5.525904203323559e-06, |
| "loss": 0.0266, |
| "step": 29625 |
| }, |
| { |
| "epoch": 14.49, |
| "learning_rate": 5.513685239491691e-06, |
| "loss": 0.0326, |
| "step": 29650 |
| }, |
| { |
| "epoch": 14.5, |
| "learning_rate": 5.501466275659825e-06, |
| "loss": 0.0298, |
| "step": 29675 |
| }, |
| { |
| "epoch": 14.52, |
| "learning_rate": 5.489247311827958e-06, |
| "loss": 0.0218, |
| "step": 29700 |
| }, |
| { |
| "epoch": 14.53, |
| "learning_rate": 5.47702834799609e-06, |
| "loss": 0.0281, |
| "step": 29725 |
| }, |
| { |
| "epoch": 14.54, |
| "learning_rate": 5.4648093841642235e-06, |
| "loss": 0.0258, |
| "step": 29750 |
| }, |
| { |
| "epoch": 14.55, |
| "learning_rate": 5.452590420332357e-06, |
| "loss": 0.0259, |
| "step": 29775 |
| }, |
| { |
| "epoch": 14.57, |
| "learning_rate": 5.440371456500489e-06, |
| "loss": 0.0255, |
| "step": 29800 |
| }, |
| { |
| "epoch": 14.58, |
| "learning_rate": 5.428152492668622e-06, |
| "loss": 0.0321, |
| "step": 29825 |
| }, |
| { |
| "epoch": 14.59, |
| "learning_rate": 5.415933528836756e-06, |
| "loss": 0.0337, |
| "step": 29850 |
| }, |
| { |
| "epoch": 14.6, |
| "learning_rate": 5.403714565004888e-06, |
| "loss": 0.032, |
| "step": 29875 |
| }, |
| { |
| "epoch": 14.61, |
| "learning_rate": 5.391495601173021e-06, |
| "loss": 0.0374, |
| "step": 29900 |
| }, |
| { |
| "epoch": 14.63, |
| "learning_rate": 5.3792766373411545e-06, |
| "loss": 0.0285, |
| "step": 29925 |
| }, |
| { |
| "epoch": 14.64, |
| "learning_rate": 5.367057673509287e-06, |
| "loss": 0.0283, |
| "step": 29950 |
| }, |
| { |
| "epoch": 14.65, |
| "learning_rate": 5.35483870967742e-06, |
| "loss": 0.0249, |
| "step": 29975 |
| }, |
| { |
| "epoch": 14.66, |
| "learning_rate": 5.3426197458455534e-06, |
| "loss": 0.0305, |
| "step": 30000 |
| }, |
| { |
| "epoch": 14.67, |
| "learning_rate": 5.330400782013686e-06, |
| "loss": 0.035, |
| "step": 30025 |
| }, |
| { |
| "epoch": 14.69, |
| "learning_rate": 5.318181818181819e-06, |
| "loss": 0.0392, |
| "step": 30050 |
| }, |
| { |
| "epoch": 14.7, |
| "learning_rate": 5.305962854349952e-06, |
| "loss": 0.0269, |
| "step": 30075 |
| }, |
| { |
| "epoch": 14.71, |
| "learning_rate": 5.294232649071359e-06, |
| "loss": 0.0258, |
| "step": 30100 |
| }, |
| { |
| "epoch": 14.72, |
| "learning_rate": 5.282013685239492e-06, |
| "loss": 0.0358, |
| "step": 30125 |
| }, |
| { |
| "epoch": 14.74, |
| "learning_rate": 5.269794721407625e-06, |
| "loss": 0.0332, |
| "step": 30150 |
| }, |
| { |
| "epoch": 14.75, |
| "learning_rate": 5.257575757575758e-06, |
| "loss": 0.0273, |
| "step": 30175 |
| }, |
| { |
| "epoch": 14.76, |
| "learning_rate": 5.245356793743891e-06, |
| "loss": 0.0275, |
| "step": 30200 |
| }, |
| { |
| "epoch": 14.77, |
| "learning_rate": 5.233137829912024e-06, |
| "loss": 0.0282, |
| "step": 30225 |
| }, |
| { |
| "epoch": 14.78, |
| "learning_rate": 5.220918866080157e-06, |
| "loss": 0.0232, |
| "step": 30250 |
| }, |
| { |
| "epoch": 14.8, |
| "learning_rate": 5.2086999022482896e-06, |
| "loss": 0.0323, |
| "step": 30275 |
| }, |
| { |
| "epoch": 14.81, |
| "learning_rate": 5.196480938416423e-06, |
| "loss": 0.0354, |
| "step": 30300 |
| }, |
| { |
| "epoch": 14.82, |
| "learning_rate": 5.184261974584556e-06, |
| "loss": 0.0288, |
| "step": 30325 |
| }, |
| { |
| "epoch": 14.83, |
| "learning_rate": 5.1720430107526885e-06, |
| "loss": 0.032, |
| "step": 30350 |
| }, |
| { |
| "epoch": 14.85, |
| "learning_rate": 5.159824046920822e-06, |
| "loss": 0.0333, |
| "step": 30375 |
| }, |
| { |
| "epoch": 14.86, |
| "learning_rate": 5.147605083088955e-06, |
| "loss": 0.0299, |
| "step": 30400 |
| }, |
| { |
| "epoch": 14.87, |
| "learning_rate": 5.135386119257087e-06, |
| "loss": 0.0395, |
| "step": 30425 |
| }, |
| { |
| "epoch": 14.88, |
| "learning_rate": 5.123167155425221e-06, |
| "loss": 0.0292, |
| "step": 30450 |
| }, |
| { |
| "epoch": 14.89, |
| "learning_rate": 5.110948191593354e-06, |
| "loss": 0.0283, |
| "step": 30475 |
| }, |
| { |
| "epoch": 14.91, |
| "learning_rate": 5.098729227761486e-06, |
| "loss": 0.0315, |
| "step": 30500 |
| }, |
| { |
| "epoch": 14.92, |
| "learning_rate": 5.0865102639296195e-06, |
| "loss": 0.028, |
| "step": 30525 |
| }, |
| { |
| "epoch": 14.93, |
| "learning_rate": 5.074291300097752e-06, |
| "loss": 0.0343, |
| "step": 30550 |
| }, |
| { |
| "epoch": 14.94, |
| "learning_rate": 5.062072336265885e-06, |
| "loss": 0.0243, |
| "step": 30575 |
| }, |
| { |
| "epoch": 14.96, |
| "learning_rate": 5.049853372434018e-06, |
| "loss": 0.0299, |
| "step": 30600 |
| }, |
| { |
| "epoch": 14.97, |
| "learning_rate": 5.037634408602151e-06, |
| "loss": 0.0295, |
| "step": 30625 |
| }, |
| { |
| "epoch": 14.98, |
| "learning_rate": 5.025415444770284e-06, |
| "loss": 0.0278, |
| "step": 30650 |
| }, |
| { |
| "epoch": 14.99, |
| "learning_rate": 5.013196480938417e-06, |
| "loss": 0.038, |
| "step": 30675 |
| }, |
| { |
| "epoch": 15.0, |
| "learning_rate": 5.00097751710655e-06, |
| "loss": 0.038, |
| "step": 30700 |
| }, |
| { |
| "epoch": 15.02, |
| "learning_rate": 4.988758553274683e-06, |
| "loss": 0.0213, |
| "step": 30725 |
| }, |
| { |
| "epoch": 15.03, |
| "learning_rate": 4.976539589442816e-06, |
| "loss": 0.0222, |
| "step": 30750 |
| }, |
| { |
| "epoch": 15.04, |
| "learning_rate": 4.9643206256109486e-06, |
| "loss": 0.0252, |
| "step": 30775 |
| }, |
| { |
| "epoch": 15.05, |
| "learning_rate": 4.952101661779082e-06, |
| "loss": 0.026, |
| "step": 30800 |
| }, |
| { |
| "epoch": 15.07, |
| "learning_rate": 4.939882697947215e-06, |
| "loss": 0.0261, |
| "step": 30825 |
| }, |
| { |
| "epoch": 15.08, |
| "learning_rate": 4.9276637341153475e-06, |
| "loss": 0.0296, |
| "step": 30850 |
| }, |
| { |
| "epoch": 15.09, |
| "learning_rate": 4.91544477028348e-06, |
| "loss": 0.0262, |
| "step": 30875 |
| }, |
| { |
| "epoch": 15.1, |
| "learning_rate": 4.903225806451613e-06, |
| "loss": 0.0173, |
| "step": 30900 |
| }, |
| { |
| "epoch": 15.11, |
| "learning_rate": 4.891006842619746e-06, |
| "loss": 0.0303, |
| "step": 30925 |
| }, |
| { |
| "epoch": 15.13, |
| "learning_rate": 4.878787878787879e-06, |
| "loss": 0.0275, |
| "step": 30950 |
| }, |
| { |
| "epoch": 15.14, |
| "learning_rate": 4.866568914956012e-06, |
| "loss": 0.0257, |
| "step": 30975 |
| }, |
| { |
| "epoch": 15.15, |
| "learning_rate": 4.854349951124145e-06, |
| "loss": 0.0221, |
| "step": 31000 |
| }, |
| { |
| "epoch": 15.16, |
| "learning_rate": 4.842130987292278e-06, |
| "loss": 0.0224, |
| "step": 31025 |
| }, |
| { |
| "epoch": 15.18, |
| "learning_rate": 4.829912023460411e-06, |
| "loss": 0.0215, |
| "step": 31050 |
| }, |
| { |
| "epoch": 15.19, |
| "learning_rate": 4.817693059628543e-06, |
| "loss": 0.0303, |
| "step": 31075 |
| }, |
| { |
| "epoch": 15.2, |
| "learning_rate": 4.8054740957966766e-06, |
| "loss": 0.0222, |
| "step": 31100 |
| }, |
| { |
| "epoch": 15.21, |
| "learning_rate": 4.79325513196481e-06, |
| "loss": 0.0268, |
| "step": 31125 |
| }, |
| { |
| "epoch": 15.22, |
| "learning_rate": 4.781036168132942e-06, |
| "loss": 0.0276, |
| "step": 31150 |
| }, |
| { |
| "epoch": 15.24, |
| "learning_rate": 4.7688172043010755e-06, |
| "loss": 0.0245, |
| "step": 31175 |
| }, |
| { |
| "epoch": 15.25, |
| "learning_rate": 4.757086999022483e-06, |
| "loss": 0.0185, |
| "step": 31200 |
| }, |
| { |
| "epoch": 15.26, |
| "learning_rate": 4.744868035190617e-06, |
| "loss": 0.0214, |
| "step": 31225 |
| }, |
| { |
| "epoch": 15.27, |
| "learning_rate": 4.732649071358749e-06, |
| "loss": 0.0236, |
| "step": 31250 |
| }, |
| { |
| "epoch": 15.29, |
| "learning_rate": 4.720430107526882e-06, |
| "loss": 0.0289, |
| "step": 31275 |
| }, |
| { |
| "epoch": 15.3, |
| "learning_rate": 4.7082111436950155e-06, |
| "loss": 0.0335, |
| "step": 31300 |
| }, |
| { |
| "epoch": 15.31, |
| "learning_rate": 4.695992179863148e-06, |
| "loss": 0.0229, |
| "step": 31325 |
| }, |
| { |
| "epoch": 15.32, |
| "learning_rate": 4.683773216031281e-06, |
| "loss": 0.0237, |
| "step": 31350 |
| }, |
| { |
| "epoch": 15.33, |
| "learning_rate": 4.6715542521994135e-06, |
| "loss": 0.0186, |
| "step": 31375 |
| }, |
| { |
| "epoch": 15.35, |
| "learning_rate": 4.659335288367547e-06, |
| "loss": 0.0219, |
| "step": 31400 |
| }, |
| { |
| "epoch": 15.36, |
| "learning_rate": 4.64711632453568e-06, |
| "loss": 0.0292, |
| "step": 31425 |
| }, |
| { |
| "epoch": 15.37, |
| "learning_rate": 4.634897360703812e-06, |
| "loss": 0.0261, |
| "step": 31450 |
| }, |
| { |
| "epoch": 15.38, |
| "learning_rate": 4.622678396871946e-06, |
| "loss": 0.0301, |
| "step": 31475 |
| }, |
| { |
| "epoch": 15.4, |
| "learning_rate": 4.610459433040079e-06, |
| "loss": 0.0243, |
| "step": 31500 |
| }, |
| { |
| "epoch": 15.41, |
| "learning_rate": 4.598240469208211e-06, |
| "loss": 0.034, |
| "step": 31525 |
| }, |
| { |
| "epoch": 15.42, |
| "learning_rate": 4.5860215053763446e-06, |
| "loss": 0.0253, |
| "step": 31550 |
| }, |
| { |
| "epoch": 15.43, |
| "learning_rate": 4.573802541544478e-06, |
| "loss": 0.0265, |
| "step": 31575 |
| }, |
| { |
| "epoch": 15.44, |
| "learning_rate": 4.56158357771261e-06, |
| "loss": 0.0178, |
| "step": 31600 |
| }, |
| { |
| "epoch": 15.46, |
| "learning_rate": 4.5493646138807435e-06, |
| "loss": 0.0208, |
| "step": 31625 |
| }, |
| { |
| "epoch": 15.47, |
| "learning_rate": 4.537145650048877e-06, |
| "loss": 0.0264, |
| "step": 31650 |
| }, |
| { |
| "epoch": 15.48, |
| "learning_rate": 4.524926686217009e-06, |
| "loss": 0.0266, |
| "step": 31675 |
| }, |
| { |
| "epoch": 15.49, |
| "learning_rate": 4.512707722385142e-06, |
| "loss": 0.0267, |
| "step": 31700 |
| }, |
| { |
| "epoch": 15.51, |
| "learning_rate": 4.500488758553276e-06, |
| "loss": 0.0218, |
| "step": 31725 |
| }, |
| { |
| "epoch": 15.52, |
| "learning_rate": 4.488269794721408e-06, |
| "loss": 0.0252, |
| "step": 31750 |
| }, |
| { |
| "epoch": 15.53, |
| "learning_rate": 4.476050830889541e-06, |
| "loss": 0.023, |
| "step": 31775 |
| }, |
| { |
| "epoch": 15.54, |
| "learning_rate": 4.4638318670576745e-06, |
| "loss": 0.0216, |
| "step": 31800 |
| }, |
| { |
| "epoch": 15.55, |
| "learning_rate": 4.451612903225807e-06, |
| "loss": 0.0278, |
| "step": 31825 |
| }, |
| { |
| "epoch": 15.57, |
| "learning_rate": 4.43939393939394e-06, |
| "loss": 0.0224, |
| "step": 31850 |
| }, |
| { |
| "epoch": 15.58, |
| "learning_rate": 4.4271749755620725e-06, |
| "loss": 0.0302, |
| "step": 31875 |
| }, |
| { |
| "epoch": 15.59, |
| "learning_rate": 4.414956011730206e-06, |
| "loss": 0.0213, |
| "step": 31900 |
| }, |
| { |
| "epoch": 15.6, |
| "learning_rate": 4.402737047898338e-06, |
| "loss": 0.0243, |
| "step": 31925 |
| }, |
| { |
| "epoch": 15.62, |
| "learning_rate": 4.3905180840664714e-06, |
| "loss": 0.0263, |
| "step": 31950 |
| }, |
| { |
| "epoch": 15.63, |
| "learning_rate": 4.378299120234604e-06, |
| "loss": 0.018, |
| "step": 31975 |
| }, |
| { |
| "epoch": 15.64, |
| "learning_rate": 4.366080156402737e-06, |
| "loss": 0.0155, |
| "step": 32000 |
| }, |
| { |
| "epoch": 15.65, |
| "learning_rate": 4.35386119257087e-06, |
| "loss": 0.0368, |
| "step": 32025 |
| }, |
| { |
| "epoch": 15.66, |
| "learning_rate": 4.341642228739003e-06, |
| "loss": 0.0271, |
| "step": 32050 |
| }, |
| { |
| "epoch": 15.68, |
| "learning_rate": 4.329423264907136e-06, |
| "loss": 0.0309, |
| "step": 32075 |
| }, |
| { |
| "epoch": 15.69, |
| "learning_rate": 4.317204301075269e-06, |
| "loss": 0.0252, |
| "step": 32100 |
| }, |
| { |
| "epoch": 15.7, |
| "learning_rate": 4.304985337243402e-06, |
| "loss": 0.0353, |
| "step": 32125 |
| }, |
| { |
| "epoch": 15.71, |
| "learning_rate": 4.292766373411535e-06, |
| "loss": 0.0246, |
| "step": 32150 |
| }, |
| { |
| "epoch": 15.73, |
| "learning_rate": 4.280547409579668e-06, |
| "loss": 0.0315, |
| "step": 32175 |
| }, |
| { |
| "epoch": 15.74, |
| "learning_rate": 4.2683284457478005e-06, |
| "loss": 0.0209, |
| "step": 32200 |
| }, |
| { |
| "epoch": 15.75, |
| "learning_rate": 4.256109481915934e-06, |
| "loss": 0.0209, |
| "step": 32225 |
| }, |
| { |
| "epoch": 15.76, |
| "learning_rate": 4.243890518084067e-06, |
| "loss": 0.0264, |
| "step": 32250 |
| }, |
| { |
| "epoch": 15.77, |
| "learning_rate": 4.231671554252199e-06, |
| "loss": 0.0232, |
| "step": 32275 |
| }, |
| { |
| "epoch": 15.79, |
| "learning_rate": 4.219452590420333e-06, |
| "loss": 0.0259, |
| "step": 32300 |
| }, |
| { |
| "epoch": 15.8, |
| "learning_rate": 4.207233626588466e-06, |
| "loss": 0.0284, |
| "step": 32325 |
| }, |
| { |
| "epoch": 15.81, |
| "learning_rate": 4.195014662756598e-06, |
| "loss": 0.0319, |
| "step": 32350 |
| }, |
| { |
| "epoch": 15.82, |
| "learning_rate": 4.1827956989247316e-06, |
| "loss": 0.0208, |
| "step": 32375 |
| }, |
| { |
| "epoch": 15.84, |
| "learning_rate": 4.170576735092865e-06, |
| "loss": 0.0265, |
| "step": 32400 |
| }, |
| { |
| "epoch": 15.85, |
| "learning_rate": 4.158357771260997e-06, |
| "loss": 0.0241, |
| "step": 32425 |
| }, |
| { |
| "epoch": 15.86, |
| "learning_rate": 4.1461388074291305e-06, |
| "loss": 0.023, |
| "step": 32450 |
| }, |
| { |
| "epoch": 15.87, |
| "learning_rate": 4.133919843597264e-06, |
| "loss": 0.0263, |
| "step": 32475 |
| }, |
| { |
| "epoch": 15.88, |
| "learning_rate": 4.121700879765396e-06, |
| "loss": 0.0181, |
| "step": 32500 |
| }, |
| { |
| "epoch": 15.9, |
| "learning_rate": 4.109481915933529e-06, |
| "loss": 0.0224, |
| "step": 32525 |
| }, |
| { |
| "epoch": 15.91, |
| "learning_rate": 4.097262952101662e-06, |
| "loss": 0.0259, |
| "step": 32550 |
| }, |
| { |
| "epoch": 15.92, |
| "learning_rate": 4.085043988269795e-06, |
| "loss": 0.0242, |
| "step": 32575 |
| }, |
| { |
| "epoch": 15.93, |
| "learning_rate": 4.072825024437928e-06, |
| "loss": 0.0293, |
| "step": 32600 |
| }, |
| { |
| "epoch": 15.95, |
| "learning_rate": 4.060606060606061e-06, |
| "loss": 0.0252, |
| "step": 32625 |
| }, |
| { |
| "epoch": 15.96, |
| "learning_rate": 4.048387096774194e-06, |
| "loss": 0.0234, |
| "step": 32650 |
| }, |
| { |
| "epoch": 15.97, |
| "learning_rate": 4.036168132942327e-06, |
| "loss": 0.0224, |
| "step": 32675 |
| }, |
| { |
| "epoch": 15.98, |
| "learning_rate": 4.0239491691104595e-06, |
| "loss": 0.025, |
| "step": 32700 |
| }, |
| { |
| "epoch": 15.99, |
| "learning_rate": 4.011730205278593e-06, |
| "loss": 0.0226, |
| "step": 32725 |
| }, |
| { |
| "epoch": 16.01, |
| "learning_rate": 3.999511241446726e-06, |
| "loss": 0.0239, |
| "step": 32750 |
| }, |
| { |
| "epoch": 16.02, |
| "learning_rate": 3.9872922776148584e-06, |
| "loss": 0.0246, |
| "step": 32775 |
| }, |
| { |
| "epoch": 16.03, |
| "learning_rate": 3.975073313782992e-06, |
| "loss": 0.0157, |
| "step": 32800 |
| }, |
| { |
| "epoch": 16.04, |
| "learning_rate": 3.962854349951125e-06, |
| "loss": 0.022, |
| "step": 32825 |
| }, |
| { |
| "epoch": 16.06, |
| "learning_rate": 3.950635386119257e-06, |
| "loss": 0.0236, |
| "step": 32850 |
| }, |
| { |
| "epoch": 16.07, |
| "learning_rate": 3.9384164222873906e-06, |
| "loss": 0.023, |
| "step": 32875 |
| }, |
| { |
| "epoch": 16.08, |
| "learning_rate": 3.926197458455524e-06, |
| "loss": 0.034, |
| "step": 32900 |
| }, |
| { |
| "epoch": 16.09, |
| "learning_rate": 3.913978494623656e-06, |
| "loss": 0.0249, |
| "step": 32925 |
| }, |
| { |
| "epoch": 16.1, |
| "learning_rate": 3.9017595307917895e-06, |
| "loss": 0.0224, |
| "step": 32950 |
| }, |
| { |
| "epoch": 16.12, |
| "learning_rate": 3.889540566959923e-06, |
| "loss": 0.0261, |
| "step": 32975 |
| }, |
| { |
| "epoch": 16.13, |
| "learning_rate": 3.877321603128055e-06, |
| "loss": 0.0153, |
| "step": 33000 |
| }, |
| { |
| "epoch": 16.14, |
| "learning_rate": 3.865102639296188e-06, |
| "loss": 0.0267, |
| "step": 33025 |
| }, |
| { |
| "epoch": 16.15, |
| "learning_rate": 3.852883675464321e-06, |
| "loss": 0.0158, |
| "step": 33050 |
| }, |
| { |
| "epoch": 16.17, |
| "learning_rate": 3.840664711632454e-06, |
| "loss": 0.02, |
| "step": 33075 |
| }, |
| { |
| "epoch": 16.18, |
| "learning_rate": 3.828445747800587e-06, |
| "loss": 0.0227, |
| "step": 33100 |
| }, |
| { |
| "epoch": 16.19, |
| "learning_rate": 3.81622678396872e-06, |
| "loss": 0.0202, |
| "step": 33125 |
| }, |
| { |
| "epoch": 16.2, |
| "learning_rate": 3.8040078201368525e-06, |
| "loss": 0.024, |
| "step": 33150 |
| }, |
| { |
| "epoch": 16.21, |
| "learning_rate": 3.7917888563049853e-06, |
| "loss": 0.0255, |
| "step": 33175 |
| }, |
| { |
| "epoch": 16.23, |
| "learning_rate": 3.7795698924731186e-06, |
| "loss": 0.0202, |
| "step": 33200 |
| }, |
| { |
| "epoch": 16.24, |
| "learning_rate": 3.7673509286412514e-06, |
| "loss": 0.0299, |
| "step": 33225 |
| }, |
| { |
| "epoch": 16.25, |
| "learning_rate": 3.755131964809384e-06, |
| "loss": 0.0266, |
| "step": 33250 |
| }, |
| { |
| "epoch": 16.26, |
| "learning_rate": 3.742913000977517e-06, |
| "loss": 0.0232, |
| "step": 33275 |
| }, |
| { |
| "epoch": 16.28, |
| "learning_rate": 3.7306940371456503e-06, |
| "loss": 0.0193, |
| "step": 33300 |
| }, |
| { |
| "epoch": 16.29, |
| "learning_rate": 3.718475073313783e-06, |
| "loss": 0.0192, |
| "step": 33325 |
| }, |
| { |
| "epoch": 16.3, |
| "learning_rate": 3.706256109481916e-06, |
| "loss": 0.0195, |
| "step": 33350 |
| }, |
| { |
| "epoch": 16.31, |
| "learning_rate": 3.694037145650049e-06, |
| "loss": 0.0182, |
| "step": 33375 |
| }, |
| { |
| "epoch": 16.32, |
| "learning_rate": 3.6823069403714566e-06, |
| "loss": 0.0261, |
| "step": 33400 |
| }, |
| { |
| "epoch": 16.34, |
| "learning_rate": 3.67008797653959e-06, |
| "loss": 0.0219, |
| "step": 33425 |
| }, |
| { |
| "epoch": 16.35, |
| "learning_rate": 3.6578690127077227e-06, |
| "loss": 0.0293, |
| "step": 33450 |
| }, |
| { |
| "epoch": 16.36, |
| "learning_rate": 3.6456500488758555e-06, |
| "loss": 0.0318, |
| "step": 33475 |
| }, |
| { |
| "epoch": 16.37, |
| "learning_rate": 3.6334310850439888e-06, |
| "loss": 0.0198, |
| "step": 33500 |
| }, |
| { |
| "epoch": 16.39, |
| "learning_rate": 3.6212121212121216e-06, |
| "loss": 0.0257, |
| "step": 33525 |
| }, |
| { |
| "epoch": 16.4, |
| "learning_rate": 3.6089931573802544e-06, |
| "loss": 0.0192, |
| "step": 33550 |
| }, |
| { |
| "epoch": 16.41, |
| "learning_rate": 3.5967741935483872e-06, |
| "loss": 0.0204, |
| "step": 33575 |
| }, |
| { |
| "epoch": 16.42, |
| "learning_rate": 3.5845552297165205e-06, |
| "loss": 0.0221, |
| "step": 33600 |
| }, |
| { |
| "epoch": 16.43, |
| "learning_rate": 3.5723362658846533e-06, |
| "loss": 0.0229, |
| "step": 33625 |
| }, |
| { |
| "epoch": 16.45, |
| "learning_rate": 3.560117302052786e-06, |
| "loss": 0.0234, |
| "step": 33650 |
| }, |
| { |
| "epoch": 16.46, |
| "learning_rate": 3.5478983382209194e-06, |
| "loss": 0.0233, |
| "step": 33675 |
| }, |
| { |
| "epoch": 16.47, |
| "learning_rate": 3.535679374389052e-06, |
| "loss": 0.0248, |
| "step": 33700 |
| }, |
| { |
| "epoch": 16.48, |
| "learning_rate": 3.523460410557185e-06, |
| "loss": 0.0251, |
| "step": 33725 |
| }, |
| { |
| "epoch": 16.5, |
| "learning_rate": 3.5112414467253183e-06, |
| "loss": 0.0297, |
| "step": 33750 |
| }, |
| { |
| "epoch": 16.51, |
| "learning_rate": 3.499022482893451e-06, |
| "loss": 0.0256, |
| "step": 33775 |
| }, |
| { |
| "epoch": 16.52, |
| "learning_rate": 3.486803519061584e-06, |
| "loss": 0.0194, |
| "step": 33800 |
| }, |
| { |
| "epoch": 16.53, |
| "learning_rate": 3.4745845552297167e-06, |
| "loss": 0.026, |
| "step": 33825 |
| }, |
| { |
| "epoch": 16.54, |
| "learning_rate": 3.46236559139785e-06, |
| "loss": 0.0165, |
| "step": 33850 |
| }, |
| { |
| "epoch": 16.56, |
| "learning_rate": 3.450146627565983e-06, |
| "loss": 0.021, |
| "step": 33875 |
| }, |
| { |
| "epoch": 16.57, |
| "learning_rate": 3.4379276637341156e-06, |
| "loss": 0.0226, |
| "step": 33900 |
| }, |
| { |
| "epoch": 16.58, |
| "learning_rate": 3.425708699902249e-06, |
| "loss": 0.0273, |
| "step": 33925 |
| }, |
| { |
| "epoch": 16.59, |
| "learning_rate": 3.4134897360703817e-06, |
| "loss": 0.0209, |
| "step": 33950 |
| }, |
| { |
| "epoch": 16.61, |
| "learning_rate": 3.4012707722385145e-06, |
| "loss": 0.0228, |
| "step": 33975 |
| }, |
| { |
| "epoch": 16.62, |
| "learning_rate": 3.3890518084066478e-06, |
| "loss": 0.026, |
| "step": 34000 |
| }, |
| { |
| "epoch": 16.63, |
| "learning_rate": 3.3768328445747806e-06, |
| "loss": 0.0192, |
| "step": 34025 |
| }, |
| { |
| "epoch": 16.64, |
| "learning_rate": 3.3646138807429134e-06, |
| "loss": 0.0226, |
| "step": 34050 |
| }, |
| { |
| "epoch": 16.65, |
| "learning_rate": 3.3523949169110463e-06, |
| "loss": 0.021, |
| "step": 34075 |
| }, |
| { |
| "epoch": 16.67, |
| "learning_rate": 3.3401759530791795e-06, |
| "loss": 0.0185, |
| "step": 34100 |
| }, |
| { |
| "epoch": 16.68, |
| "learning_rate": 3.3279569892473123e-06, |
| "loss": 0.0279, |
| "step": 34125 |
| }, |
| { |
| "epoch": 16.69, |
| "learning_rate": 3.315738025415445e-06, |
| "loss": 0.0306, |
| "step": 34150 |
| }, |
| { |
| "epoch": 16.7, |
| "learning_rate": 3.3035190615835775e-06, |
| "loss": 0.033, |
| "step": 34175 |
| }, |
| { |
| "epoch": 16.72, |
| "learning_rate": 3.291300097751711e-06, |
| "loss": 0.0196, |
| "step": 34200 |
| }, |
| { |
| "epoch": 16.73, |
| "learning_rate": 3.2790811339198436e-06, |
| "loss": 0.0293, |
| "step": 34225 |
| }, |
| { |
| "epoch": 16.74, |
| "learning_rate": 3.2668621700879764e-06, |
| "loss": 0.0224, |
| "step": 34250 |
| }, |
| { |
| "epoch": 16.75, |
| "learning_rate": 3.2546432062561097e-06, |
| "loss": 0.0271, |
| "step": 34275 |
| }, |
| { |
| "epoch": 16.76, |
| "learning_rate": 3.2424242424242425e-06, |
| "loss": 0.0323, |
| "step": 34300 |
| }, |
| { |
| "epoch": 16.78, |
| "learning_rate": 3.2302052785923753e-06, |
| "loss": 0.0242, |
| "step": 34325 |
| }, |
| { |
| "epoch": 16.79, |
| "learning_rate": 3.2179863147605086e-06, |
| "loss": 0.0176, |
| "step": 34350 |
| }, |
| { |
| "epoch": 16.8, |
| "learning_rate": 3.2057673509286414e-06, |
| "loss": 0.0229, |
| "step": 34375 |
| }, |
| { |
| "epoch": 16.81, |
| "learning_rate": 3.1935483870967742e-06, |
| "loss": 0.022, |
| "step": 34400 |
| }, |
| { |
| "epoch": 16.83, |
| "learning_rate": 3.181329423264907e-06, |
| "loss": 0.0181, |
| "step": 34425 |
| }, |
| { |
| "epoch": 16.84, |
| "learning_rate": 3.1691104594330403e-06, |
| "loss": 0.0219, |
| "step": 34450 |
| }, |
| { |
| "epoch": 16.85, |
| "learning_rate": 3.156891495601173e-06, |
| "loss": 0.0192, |
| "step": 34475 |
| }, |
| { |
| "epoch": 16.86, |
| "learning_rate": 3.144672531769306e-06, |
| "loss": 0.0228, |
| "step": 34500 |
| }, |
| { |
| "epoch": 16.87, |
| "learning_rate": 3.132453567937439e-06, |
| "loss": 0.026, |
| "step": 34525 |
| }, |
| { |
| "epoch": 16.89, |
| "learning_rate": 3.120234604105572e-06, |
| "loss": 0.0213, |
| "step": 34550 |
| }, |
| { |
| "epoch": 16.9, |
| "learning_rate": 3.108015640273705e-06, |
| "loss": 0.0219, |
| "step": 34575 |
| }, |
| { |
| "epoch": 16.91, |
| "learning_rate": 3.095796676441838e-06, |
| "loss": 0.0241, |
| "step": 34600 |
| }, |
| { |
| "epoch": 16.92, |
| "learning_rate": 3.083577712609971e-06, |
| "loss": 0.0218, |
| "step": 34625 |
| }, |
| { |
| "epoch": 16.94, |
| "learning_rate": 3.0713587487781037e-06, |
| "loss": 0.0175, |
| "step": 34650 |
| }, |
| { |
| "epoch": 16.95, |
| "learning_rate": 3.0591397849462366e-06, |
| "loss": 0.0182, |
| "step": 34675 |
| }, |
| { |
| "epoch": 16.96, |
| "learning_rate": 3.04692082111437e-06, |
| "loss": 0.0235, |
| "step": 34700 |
| }, |
| { |
| "epoch": 16.97, |
| "learning_rate": 3.0347018572825026e-06, |
| "loss": 0.0176, |
| "step": 34725 |
| }, |
| { |
| "epoch": 16.98, |
| "learning_rate": 3.0224828934506355e-06, |
| "loss": 0.0187, |
| "step": 34750 |
| }, |
| { |
| "epoch": 17.0, |
| "learning_rate": 3.0102639296187687e-06, |
| "loss": 0.0206, |
| "step": 34775 |
| }, |
| { |
| "epoch": 17.01, |
| "learning_rate": 2.9980449657869015e-06, |
| "loss": 0.0201, |
| "step": 34800 |
| }, |
| { |
| "epoch": 17.02, |
| "learning_rate": 2.9858260019550344e-06, |
| "loss": 0.0172, |
| "step": 34825 |
| }, |
| { |
| "epoch": 17.03, |
| "learning_rate": 2.9736070381231676e-06, |
| "loss": 0.0219, |
| "step": 34850 |
| }, |
| { |
| "epoch": 17.05, |
| "learning_rate": 2.9613880742913004e-06, |
| "loss": 0.0164, |
| "step": 34875 |
| }, |
| { |
| "epoch": 17.06, |
| "learning_rate": 2.9491691104594332e-06, |
| "loss": 0.0212, |
| "step": 34900 |
| }, |
| { |
| "epoch": 17.07, |
| "learning_rate": 2.936950146627566e-06, |
| "loss": 0.0187, |
| "step": 34925 |
| }, |
| { |
| "epoch": 17.08, |
| "learning_rate": 2.9247311827956993e-06, |
| "loss": 0.0248, |
| "step": 34950 |
| }, |
| { |
| "epoch": 17.09, |
| "learning_rate": 2.912512218963832e-06, |
| "loss": 0.0203, |
| "step": 34975 |
| }, |
| { |
| "epoch": 17.11, |
| "learning_rate": 2.900293255131965e-06, |
| "loss": 0.0255, |
| "step": 35000 |
| }, |
| { |
| "epoch": 17.12, |
| "learning_rate": 2.8880742913000982e-06, |
| "loss": 0.0222, |
| "step": 35025 |
| }, |
| { |
| "epoch": 17.13, |
| "learning_rate": 2.875855327468231e-06, |
| "loss": 0.0234, |
| "step": 35050 |
| }, |
| { |
| "epoch": 17.14, |
| "learning_rate": 2.863636363636364e-06, |
| "loss": 0.0234, |
| "step": 35075 |
| }, |
| { |
| "epoch": 17.16, |
| "learning_rate": 2.851417399804497e-06, |
| "loss": 0.0174, |
| "step": 35100 |
| }, |
| { |
| "epoch": 17.17, |
| "learning_rate": 2.83919843597263e-06, |
| "loss": 0.0167, |
| "step": 35125 |
| }, |
| { |
| "epoch": 17.18, |
| "learning_rate": 2.8269794721407628e-06, |
| "loss": 0.0182, |
| "step": 35150 |
| }, |
| { |
| "epoch": 17.19, |
| "learning_rate": 2.814760508308896e-06, |
| "loss": 0.0146, |
| "step": 35175 |
| }, |
| { |
| "epoch": 17.2, |
| "learning_rate": 2.802541544477029e-06, |
| "loss": 0.0199, |
| "step": 35200 |
| }, |
| { |
| "epoch": 17.22, |
| "learning_rate": 2.7903225806451617e-06, |
| "loss": 0.021, |
| "step": 35225 |
| }, |
| { |
| "epoch": 17.23, |
| "learning_rate": 2.7781036168132945e-06, |
| "loss": 0.0173, |
| "step": 35250 |
| }, |
| { |
| "epoch": 17.24, |
| "learning_rate": 2.7658846529814277e-06, |
| "loss": 0.0164, |
| "step": 35275 |
| }, |
| { |
| "epoch": 17.25, |
| "learning_rate": 2.7536656891495605e-06, |
| "loss": 0.0172, |
| "step": 35300 |
| }, |
| { |
| "epoch": 17.27, |
| "learning_rate": 2.7414467253176934e-06, |
| "loss": 0.0202, |
| "step": 35325 |
| }, |
| { |
| "epoch": 17.28, |
| "learning_rate": 2.7292277614858266e-06, |
| "loss": 0.019, |
| "step": 35350 |
| }, |
| { |
| "epoch": 17.29, |
| "learning_rate": 2.7170087976539594e-06, |
| "loss": 0.0211, |
| "step": 35375 |
| }, |
| { |
| "epoch": 17.3, |
| "learning_rate": 2.7047898338220923e-06, |
| "loss": 0.0238, |
| "step": 35400 |
| }, |
| { |
| "epoch": 17.31, |
| "learning_rate": 2.6925708699902255e-06, |
| "loss": 0.0187, |
| "step": 35425 |
| }, |
| { |
| "epoch": 17.33, |
| "learning_rate": 2.680351906158358e-06, |
| "loss": 0.0218, |
| "step": 35450 |
| }, |
| { |
| "epoch": 17.34, |
| "learning_rate": 2.6681329423264907e-06, |
| "loss": 0.0138, |
| "step": 35475 |
| }, |
| { |
| "epoch": 17.35, |
| "learning_rate": 2.6559139784946236e-06, |
| "loss": 0.0146, |
| "step": 35500 |
| }, |
| { |
| "epoch": 17.36, |
| "learning_rate": 2.6436950146627564e-06, |
| "loss": 0.0183, |
| "step": 35525 |
| }, |
| { |
| "epoch": 17.38, |
| "learning_rate": 2.6314760508308896e-06, |
| "loss": 0.0126, |
| "step": 35550 |
| }, |
| { |
| "epoch": 17.39, |
| "learning_rate": 2.6192570869990225e-06, |
| "loss": 0.0127, |
| "step": 35575 |
| }, |
| { |
| "epoch": 17.4, |
| "learning_rate": 2.6070381231671553e-06, |
| "loss": 0.0194, |
| "step": 35600 |
| }, |
| { |
| "epoch": 17.41, |
| "learning_rate": 2.5948191593352885e-06, |
| "loss": 0.0206, |
| "step": 35625 |
| }, |
| { |
| "epoch": 17.42, |
| "learning_rate": 2.5826001955034214e-06, |
| "loss": 0.0194, |
| "step": 35650 |
| }, |
| { |
| "epoch": 17.44, |
| "learning_rate": 2.570381231671554e-06, |
| "loss": 0.0186, |
| "step": 35675 |
| }, |
| { |
| "epoch": 17.45, |
| "learning_rate": 2.5581622678396874e-06, |
| "loss": 0.031, |
| "step": 35700 |
| }, |
| { |
| "epoch": 17.46, |
| "learning_rate": 2.5459433040078202e-06, |
| "loss": 0.0235, |
| "step": 35725 |
| }, |
| { |
| "epoch": 17.47, |
| "learning_rate": 2.533724340175953e-06, |
| "loss": 0.02, |
| "step": 35750 |
| }, |
| { |
| "epoch": 17.49, |
| "learning_rate": 2.521505376344086e-06, |
| "loss": 0.0157, |
| "step": 35775 |
| }, |
| { |
| "epoch": 17.5, |
| "learning_rate": 2.509286412512219e-06, |
| "loss": 0.018, |
| "step": 35800 |
| }, |
| { |
| "epoch": 17.51, |
| "learning_rate": 2.497067448680352e-06, |
| "loss": 0.0209, |
| "step": 35825 |
| }, |
| { |
| "epoch": 17.52, |
| "learning_rate": 2.4848484848484848e-06, |
| "loss": 0.0201, |
| "step": 35850 |
| }, |
| { |
| "epoch": 17.53, |
| "learning_rate": 2.472629521016618e-06, |
| "loss": 0.0174, |
| "step": 35875 |
| }, |
| { |
| "epoch": 17.55, |
| "learning_rate": 2.460410557184751e-06, |
| "loss": 0.0223, |
| "step": 35900 |
| }, |
| { |
| "epoch": 17.56, |
| "learning_rate": 2.4481915933528837e-06, |
| "loss": 0.0249, |
| "step": 35925 |
| }, |
| { |
| "epoch": 17.57, |
| "learning_rate": 2.435972629521017e-06, |
| "loss": 0.0181, |
| "step": 35950 |
| }, |
| { |
| "epoch": 17.58, |
| "learning_rate": 2.4237536656891498e-06, |
| "loss": 0.0239, |
| "step": 35975 |
| }, |
| { |
| "epoch": 17.6, |
| "learning_rate": 2.4115347018572826e-06, |
| "loss": 0.0204, |
| "step": 36000 |
| }, |
| { |
| "epoch": 17.61, |
| "learning_rate": 2.399315738025416e-06, |
| "loss": 0.028, |
| "step": 36025 |
| }, |
| { |
| "epoch": 17.62, |
| "learning_rate": 2.3870967741935486e-06, |
| "loss": 0.0179, |
| "step": 36050 |
| }, |
| { |
| "epoch": 17.63, |
| "learning_rate": 2.3748778103616815e-06, |
| "loss": 0.0211, |
| "step": 36075 |
| }, |
| { |
| "epoch": 17.64, |
| "learning_rate": 2.3626588465298143e-06, |
| "loss": 0.016, |
| "step": 36100 |
| }, |
| { |
| "epoch": 17.66, |
| "learning_rate": 2.3504398826979475e-06, |
| "loss": 0.0202, |
| "step": 36125 |
| }, |
| { |
| "epoch": 17.67, |
| "learning_rate": 2.3382209188660804e-06, |
| "loss": 0.0211, |
| "step": 36150 |
| }, |
| { |
| "epoch": 17.68, |
| "learning_rate": 2.326001955034213e-06, |
| "loss": 0.0198, |
| "step": 36175 |
| }, |
| { |
| "epoch": 17.69, |
| "learning_rate": 2.3137829912023464e-06, |
| "loss": 0.0241, |
| "step": 36200 |
| }, |
| { |
| "epoch": 17.71, |
| "learning_rate": 2.3015640273704793e-06, |
| "loss": 0.0189, |
| "step": 36225 |
| }, |
| { |
| "epoch": 17.72, |
| "learning_rate": 2.289345063538612e-06, |
| "loss": 0.0179, |
| "step": 36250 |
| }, |
| { |
| "epoch": 17.73, |
| "learning_rate": 2.2771260997067453e-06, |
| "loss": 0.0252, |
| "step": 36275 |
| }, |
| { |
| "epoch": 17.74, |
| "learning_rate": 2.264907135874878e-06, |
| "loss": 0.019, |
| "step": 36300 |
| }, |
| { |
| "epoch": 17.75, |
| "learning_rate": 2.252688172043011e-06, |
| "loss": 0.019, |
| "step": 36325 |
| }, |
| { |
| "epoch": 17.77, |
| "learning_rate": 2.240469208211144e-06, |
| "loss": 0.0178, |
| "step": 36350 |
| }, |
| { |
| "epoch": 17.78, |
| "learning_rate": 2.228250244379277e-06, |
| "loss": 0.0174, |
| "step": 36375 |
| }, |
| { |
| "epoch": 17.79, |
| "learning_rate": 2.2160312805474095e-06, |
| "loss": 0.0172, |
| "step": 36400 |
| }, |
| { |
| "epoch": 17.8, |
| "learning_rate": 2.2038123167155427e-06, |
| "loss": 0.0157, |
| "step": 36425 |
| }, |
| { |
| "epoch": 17.82, |
| "learning_rate": 2.1915933528836755e-06, |
| "loss": 0.0212, |
| "step": 36450 |
| }, |
| { |
| "epoch": 17.83, |
| "learning_rate": 2.1793743890518083e-06, |
| "loss": 0.0204, |
| "step": 36475 |
| }, |
| { |
| "epoch": 17.84, |
| "learning_rate": 2.1671554252199416e-06, |
| "loss": 0.0279, |
| "step": 36500 |
| }, |
| { |
| "epoch": 17.85, |
| "learning_rate": 2.1549364613880744e-06, |
| "loss": 0.0197, |
| "step": 36525 |
| }, |
| { |
| "epoch": 17.86, |
| "learning_rate": 2.1427174975562072e-06, |
| "loss": 0.0188, |
| "step": 36550 |
| }, |
| { |
| "epoch": 17.88, |
| "learning_rate": 2.1304985337243405e-06, |
| "loss": 0.0241, |
| "step": 36575 |
| }, |
| { |
| "epoch": 17.89, |
| "learning_rate": 2.1182795698924733e-06, |
| "loss": 0.0207, |
| "step": 36600 |
| }, |
| { |
| "epoch": 17.9, |
| "learning_rate": 2.106060606060606e-06, |
| "loss": 0.0158, |
| "step": 36625 |
| }, |
| { |
| "epoch": 17.91, |
| "learning_rate": 2.093841642228739e-06, |
| "loss": 0.0214, |
| "step": 36650 |
| }, |
| { |
| "epoch": 17.93, |
| "learning_rate": 2.081622678396872e-06, |
| "loss": 0.0178, |
| "step": 36675 |
| }, |
| { |
| "epoch": 17.94, |
| "learning_rate": 2.069403714565005e-06, |
| "loss": 0.0242, |
| "step": 36700 |
| }, |
| { |
| "epoch": 17.95, |
| "learning_rate": 2.057184750733138e-06, |
| "loss": 0.0229, |
| "step": 36725 |
| }, |
| { |
| "epoch": 17.96, |
| "learning_rate": 2.044965786901271e-06, |
| "loss": 0.0168, |
| "step": 36750 |
| }, |
| { |
| "epoch": 17.97, |
| "learning_rate": 2.032746823069404e-06, |
| "loss": 0.0209, |
| "step": 36775 |
| }, |
| { |
| "epoch": 17.99, |
| "learning_rate": 2.0205278592375367e-06, |
| "loss": 0.021, |
| "step": 36800 |
| }, |
| { |
| "epoch": 18.0, |
| "learning_rate": 2.00830889540567e-06, |
| "loss": 0.0219, |
| "step": 36825 |
| }, |
| { |
| "epoch": 18.01, |
| "learning_rate": 1.996089931573803e-06, |
| "loss": 0.0174, |
| "step": 36850 |
| }, |
| { |
| "epoch": 18.02, |
| "learning_rate": 1.9838709677419356e-06, |
| "loss": 0.0169, |
| "step": 36875 |
| }, |
| { |
| "epoch": 18.04, |
| "learning_rate": 1.9716520039100685e-06, |
| "loss": 0.0171, |
| "step": 36900 |
| }, |
| { |
| "epoch": 18.05, |
| "learning_rate": 1.9594330400782017e-06, |
| "loss": 0.0203, |
| "step": 36925 |
| }, |
| { |
| "epoch": 18.06, |
| "learning_rate": 1.9472140762463345e-06, |
| "loss": 0.0165, |
| "step": 36950 |
| }, |
| { |
| "epoch": 18.07, |
| "learning_rate": 1.9349951124144674e-06, |
| "loss": 0.012, |
| "step": 36975 |
| }, |
| { |
| "epoch": 18.08, |
| "learning_rate": 1.9227761485826006e-06, |
| "loss": 0.0178, |
| "step": 37000 |
| }, |
| { |
| "epoch": 18.1, |
| "learning_rate": 1.9105571847507334e-06, |
| "loss": 0.0165, |
| "step": 37025 |
| }, |
| { |
| "epoch": 18.11, |
| "learning_rate": 1.898338220918866e-06, |
| "loss": 0.0157, |
| "step": 37050 |
| }, |
| { |
| "epoch": 18.12, |
| "learning_rate": 1.886119257086999e-06, |
| "loss": 0.0202, |
| "step": 37075 |
| }, |
| { |
| "epoch": 18.13, |
| "learning_rate": 1.873900293255132e-06, |
| "loss": 0.0149, |
| "step": 37100 |
| }, |
| { |
| "epoch": 18.15, |
| "learning_rate": 1.861681329423265e-06, |
| "loss": 0.0185, |
| "step": 37125 |
| }, |
| { |
| "epoch": 18.16, |
| "learning_rate": 1.849462365591398e-06, |
| "loss": 0.0226, |
| "step": 37150 |
| }, |
| { |
| "epoch": 18.17, |
| "learning_rate": 1.8372434017595308e-06, |
| "loss": 0.0152, |
| "step": 37175 |
| }, |
| { |
| "epoch": 18.18, |
| "learning_rate": 1.8250244379276638e-06, |
| "loss": 0.0126, |
| "step": 37200 |
| }, |
| { |
| "epoch": 18.19, |
| "learning_rate": 1.8128054740957969e-06, |
| "loss": 0.0211, |
| "step": 37225 |
| }, |
| { |
| "epoch": 18.21, |
| "learning_rate": 1.8005865102639297e-06, |
| "loss": 0.017, |
| "step": 37250 |
| }, |
| { |
| "epoch": 18.22, |
| "learning_rate": 1.7883675464320627e-06, |
| "loss": 0.0217, |
| "step": 37275 |
| }, |
| { |
| "epoch": 18.23, |
| "learning_rate": 1.7761485826001956e-06, |
| "loss": 0.0195, |
| "step": 37300 |
| }, |
| { |
| "epoch": 18.24, |
| "learning_rate": 1.7639296187683286e-06, |
| "loss": 0.015, |
| "step": 37325 |
| }, |
| { |
| "epoch": 18.26, |
| "learning_rate": 1.7517106549364616e-06, |
| "loss": 0.016, |
| "step": 37350 |
| }, |
| { |
| "epoch": 18.27, |
| "learning_rate": 1.7394916911045944e-06, |
| "loss": 0.019, |
| "step": 37375 |
| }, |
| { |
| "epoch": 18.28, |
| "learning_rate": 1.7272727272727275e-06, |
| "loss": 0.0206, |
| "step": 37400 |
| }, |
| { |
| "epoch": 18.29, |
| "learning_rate": 1.715542521994135e-06, |
| "loss": 0.0207, |
| "step": 37425 |
| }, |
| { |
| "epoch": 18.3, |
| "learning_rate": 1.7033235581622678e-06, |
| "loss": 0.0232, |
| "step": 37450 |
| }, |
| { |
| "epoch": 18.32, |
| "learning_rate": 1.6911045943304008e-06, |
| "loss": 0.0165, |
| "step": 37475 |
| }, |
| { |
| "epoch": 18.33, |
| "learning_rate": 1.6788856304985338e-06, |
| "loss": 0.0152, |
| "step": 37500 |
| }, |
| { |
| "epoch": 18.34, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": 0.0144, |
| "step": 37525 |
| }, |
| { |
| "epoch": 18.35, |
| "learning_rate": 1.6544477028347997e-06, |
| "loss": 0.0149, |
| "step": 37550 |
| }, |
| { |
| "epoch": 18.37, |
| "learning_rate": 1.6422287390029325e-06, |
| "loss": 0.0158, |
| "step": 37575 |
| }, |
| { |
| "epoch": 18.38, |
| "learning_rate": 1.6300097751710656e-06, |
| "loss": 0.0144, |
| "step": 37600 |
| }, |
| { |
| "epoch": 18.39, |
| "learning_rate": 1.6177908113391986e-06, |
| "loss": 0.0224, |
| "step": 37625 |
| }, |
| { |
| "epoch": 18.4, |
| "learning_rate": 1.6055718475073314e-06, |
| "loss": 0.0198, |
| "step": 37650 |
| }, |
| { |
| "epoch": 18.41, |
| "learning_rate": 1.5933528836754645e-06, |
| "loss": 0.0116, |
| "step": 37675 |
| }, |
| { |
| "epoch": 18.43, |
| "learning_rate": 1.5811339198435973e-06, |
| "loss": 0.0169, |
| "step": 37700 |
| }, |
| { |
| "epoch": 18.44, |
| "learning_rate": 1.5689149560117303e-06, |
| "loss": 0.0183, |
| "step": 37725 |
| }, |
| { |
| "epoch": 18.45, |
| "learning_rate": 1.5566959921798633e-06, |
| "loss": 0.0183, |
| "step": 37750 |
| }, |
| { |
| "epoch": 18.46, |
| "learning_rate": 1.5444770283479962e-06, |
| "loss": 0.0264, |
| "step": 37775 |
| }, |
| { |
| "epoch": 18.48, |
| "learning_rate": 1.5322580645161292e-06, |
| "loss": 0.0174, |
| "step": 37800 |
| }, |
| { |
| "epoch": 18.49, |
| "learning_rate": 1.5200391006842622e-06, |
| "loss": 0.0193, |
| "step": 37825 |
| }, |
| { |
| "epoch": 18.5, |
| "learning_rate": 1.507820136852395e-06, |
| "loss": 0.0169, |
| "step": 37850 |
| }, |
| { |
| "epoch": 18.51, |
| "learning_rate": 1.495601173020528e-06, |
| "loss": 0.0188, |
| "step": 37875 |
| }, |
| { |
| "epoch": 18.52, |
| "learning_rate": 1.483382209188661e-06, |
| "loss": 0.0185, |
| "step": 37900 |
| }, |
| { |
| "epoch": 18.54, |
| "learning_rate": 1.471163245356794e-06, |
| "loss": 0.0167, |
| "step": 37925 |
| }, |
| { |
| "epoch": 18.55, |
| "learning_rate": 1.458944281524927e-06, |
| "loss": 0.0199, |
| "step": 37950 |
| }, |
| { |
| "epoch": 18.56, |
| "learning_rate": 1.4467253176930598e-06, |
| "loss": 0.02, |
| "step": 37975 |
| }, |
| { |
| "epoch": 18.57, |
| "learning_rate": 1.4345063538611929e-06, |
| "loss": 0.0184, |
| "step": 38000 |
| }, |
| { |
| "epoch": 18.59, |
| "learning_rate": 1.4222873900293257e-06, |
| "loss": 0.0182, |
| "step": 38025 |
| }, |
| { |
| "epoch": 18.6, |
| "learning_rate": 1.4100684261974587e-06, |
| "loss": 0.0142, |
| "step": 38050 |
| }, |
| { |
| "epoch": 18.61, |
| "learning_rate": 1.3978494623655913e-06, |
| "loss": 0.0195, |
| "step": 38075 |
| }, |
| { |
| "epoch": 18.62, |
| "learning_rate": 1.3856304985337244e-06, |
| "loss": 0.0112, |
| "step": 38100 |
| }, |
| { |
| "epoch": 18.63, |
| "learning_rate": 1.3734115347018572e-06, |
| "loss": 0.0183, |
| "step": 38125 |
| }, |
| { |
| "epoch": 18.65, |
| "learning_rate": 1.3611925708699902e-06, |
| "loss": 0.0187, |
| "step": 38150 |
| }, |
| { |
| "epoch": 18.66, |
| "learning_rate": 1.3489736070381233e-06, |
| "loss": 0.0149, |
| "step": 38175 |
| }, |
| { |
| "epoch": 18.67, |
| "learning_rate": 1.336754643206256e-06, |
| "loss": 0.0212, |
| "step": 38200 |
| }, |
| { |
| "epoch": 18.68, |
| "learning_rate": 1.3245356793743891e-06, |
| "loss": 0.0166, |
| "step": 38225 |
| }, |
| { |
| "epoch": 18.7, |
| "learning_rate": 1.3123167155425222e-06, |
| "loss": 0.0141, |
| "step": 38250 |
| }, |
| { |
| "epoch": 18.71, |
| "learning_rate": 1.300097751710655e-06, |
| "loss": 0.0147, |
| "step": 38275 |
| }, |
| { |
| "epoch": 18.72, |
| "learning_rate": 1.287878787878788e-06, |
| "loss": 0.0214, |
| "step": 38300 |
| }, |
| { |
| "epoch": 18.73, |
| "learning_rate": 1.2756598240469208e-06, |
| "loss": 0.0193, |
| "step": 38325 |
| }, |
| { |
| "epoch": 18.74, |
| "learning_rate": 1.2634408602150539e-06, |
| "loss": 0.0189, |
| "step": 38350 |
| }, |
| { |
| "epoch": 18.76, |
| "learning_rate": 1.251221896383187e-06, |
| "loss": 0.0193, |
| "step": 38375 |
| }, |
| { |
| "epoch": 18.77, |
| "learning_rate": 1.2390029325513197e-06, |
| "loss": 0.0208, |
| "step": 38400 |
| }, |
| { |
| "epoch": 18.78, |
| "learning_rate": 1.2267839687194528e-06, |
| "loss": 0.0179, |
| "step": 38425 |
| }, |
| { |
| "epoch": 18.79, |
| "learning_rate": 1.2145650048875856e-06, |
| "loss": 0.0149, |
| "step": 38450 |
| }, |
| { |
| "epoch": 18.8, |
| "learning_rate": 1.2023460410557186e-06, |
| "loss": 0.0181, |
| "step": 38475 |
| }, |
| { |
| "epoch": 18.82, |
| "learning_rate": 1.1901270772238517e-06, |
| "loss": 0.0233, |
| "step": 38500 |
| }, |
| { |
| "epoch": 18.83, |
| "learning_rate": 1.1779081133919845e-06, |
| "loss": 0.0138, |
| "step": 38525 |
| }, |
| { |
| "epoch": 18.84, |
| "learning_rate": 1.1656891495601173e-06, |
| "loss": 0.0209, |
| "step": 38550 |
| }, |
| { |
| "epoch": 18.85, |
| "learning_rate": 1.1534701857282503e-06, |
| "loss": 0.0222, |
| "step": 38575 |
| }, |
| { |
| "epoch": 18.87, |
| "learning_rate": 1.1412512218963832e-06, |
| "loss": 0.0217, |
| "step": 38600 |
| }, |
| { |
| "epoch": 18.88, |
| "learning_rate": 1.1290322580645162e-06, |
| "loss": 0.0217, |
| "step": 38625 |
| }, |
| { |
| "epoch": 18.89, |
| "learning_rate": 1.1168132942326492e-06, |
| "loss": 0.0175, |
| "step": 38650 |
| }, |
| { |
| "epoch": 18.9, |
| "learning_rate": 1.104594330400782e-06, |
| "loss": 0.0215, |
| "step": 38675 |
| }, |
| { |
| "epoch": 18.91, |
| "learning_rate": 1.092375366568915e-06, |
| "loss": 0.0189, |
| "step": 38700 |
| }, |
| { |
| "epoch": 18.93, |
| "learning_rate": 1.080156402737048e-06, |
| "loss": 0.0179, |
| "step": 38725 |
| }, |
| { |
| "epoch": 18.94, |
| "learning_rate": 1.067937438905181e-06, |
| "loss": 0.0128, |
| "step": 38750 |
| }, |
| { |
| "epoch": 18.95, |
| "learning_rate": 1.055718475073314e-06, |
| "loss": 0.0214, |
| "step": 38775 |
| }, |
| { |
| "epoch": 18.96, |
| "learning_rate": 1.0434995112414468e-06, |
| "loss": 0.0192, |
| "step": 38800 |
| }, |
| { |
| "epoch": 18.98, |
| "learning_rate": 1.0312805474095798e-06, |
| "loss": 0.016, |
| "step": 38825 |
| }, |
| { |
| "epoch": 18.99, |
| "learning_rate": 1.0190615835777127e-06, |
| "loss": 0.0184, |
| "step": 38850 |
| }, |
| { |
| "epoch": 19.0, |
| "learning_rate": 1.0068426197458455e-06, |
| "loss": 0.0146, |
| "step": 38875 |
| }, |
| { |
| "epoch": 19.01, |
| "learning_rate": 9.946236559139785e-07, |
| "loss": 0.0122, |
| "step": 38900 |
| }, |
| { |
| "epoch": 19.02, |
| "learning_rate": 9.824046920821116e-07, |
| "loss": 0.0165, |
| "step": 38925 |
| }, |
| { |
| "epoch": 19.04, |
| "learning_rate": 9.701857282502444e-07, |
| "loss": 0.0165, |
| "step": 38950 |
| }, |
| { |
| "epoch": 19.05, |
| "learning_rate": 9.579667644183774e-07, |
| "loss": 0.0188, |
| "step": 38975 |
| }, |
| { |
| "epoch": 19.06, |
| "learning_rate": 9.457478005865104e-07, |
| "loss": 0.0208, |
| "step": 39000 |
| }, |
| { |
| "epoch": 19.07, |
| "learning_rate": 9.335288367546433e-07, |
| "loss": 0.0188, |
| "step": 39025 |
| }, |
| { |
| "epoch": 19.09, |
| "learning_rate": 9.213098729227762e-07, |
| "loss": 0.0133, |
| "step": 39050 |
| }, |
| { |
| "epoch": 19.1, |
| "learning_rate": 9.090909090909091e-07, |
| "loss": 0.0088, |
| "step": 39075 |
| }, |
| { |
| "epoch": 19.11, |
| "learning_rate": 8.968719452590422e-07, |
| "loss": 0.0206, |
| "step": 39100 |
| }, |
| { |
| "epoch": 19.12, |
| "learning_rate": 8.846529814271751e-07, |
| "loss": 0.0149, |
| "step": 39125 |
| }, |
| { |
| "epoch": 19.13, |
| "learning_rate": 8.72434017595308e-07, |
| "loss": 0.0158, |
| "step": 39150 |
| }, |
| { |
| "epoch": 19.15, |
| "learning_rate": 8.60215053763441e-07, |
| "loss": 0.0162, |
| "step": 39175 |
| }, |
| { |
| "epoch": 19.16, |
| "learning_rate": 8.479960899315738e-07, |
| "loss": 0.0144, |
| "step": 39200 |
| }, |
| { |
| "epoch": 19.17, |
| "learning_rate": 8.357771260997067e-07, |
| "loss": 0.0171, |
| "step": 39225 |
| }, |
| { |
| "epoch": 19.18, |
| "learning_rate": 8.235581622678398e-07, |
| "loss": 0.0192, |
| "step": 39250 |
| }, |
| { |
| "epoch": 19.2, |
| "learning_rate": 8.113391984359727e-07, |
| "loss": 0.0156, |
| "step": 39275 |
| }, |
| { |
| "epoch": 19.21, |
| "learning_rate": 7.991202346041056e-07, |
| "loss": 0.017, |
| "step": 39300 |
| }, |
| { |
| "epoch": 19.22, |
| "learning_rate": 7.869012707722385e-07, |
| "loss": 0.0192, |
| "step": 39325 |
| }, |
| { |
| "epoch": 19.23, |
| "learning_rate": 7.746823069403715e-07, |
| "loss": 0.021, |
| "step": 39350 |
| }, |
| { |
| "epoch": 19.24, |
| "learning_rate": 7.624633431085045e-07, |
| "loss": 0.0211, |
| "step": 39375 |
| }, |
| { |
| "epoch": 19.26, |
| "learning_rate": 7.502443792766374e-07, |
| "loss": 0.0158, |
| "step": 39400 |
| }, |
| { |
| "epoch": 19.27, |
| "learning_rate": 7.38514173998045e-07, |
| "loss": 0.0166, |
| "step": 39425 |
| }, |
| { |
| "epoch": 19.28, |
| "learning_rate": 7.262952101661779e-07, |
| "loss": 0.0116, |
| "step": 39450 |
| }, |
| { |
| "epoch": 19.29, |
| "learning_rate": 7.14076246334311e-07, |
| "loss": 0.0179, |
| "step": 39475 |
| }, |
| { |
| "epoch": 19.31, |
| "learning_rate": 7.018572825024439e-07, |
| "loss": 0.0195, |
| "step": 39500 |
| }, |
| { |
| "epoch": 19.32, |
| "learning_rate": 6.896383186705768e-07, |
| "loss": 0.0198, |
| "step": 39525 |
| }, |
| { |
| "epoch": 19.33, |
| "learning_rate": 6.774193548387098e-07, |
| "loss": 0.0191, |
| "step": 39550 |
| }, |
| { |
| "epoch": 19.34, |
| "learning_rate": 6.652003910068426e-07, |
| "loss": 0.0181, |
| "step": 39575 |
| }, |
| { |
| "epoch": 19.35, |
| "learning_rate": 6.529814271749755e-07, |
| "loss": 0.0167, |
| "step": 39600 |
| }, |
| { |
| "epoch": 19.37, |
| "learning_rate": 6.407624633431086e-07, |
| "loss": 0.0167, |
| "step": 39625 |
| }, |
| { |
| "epoch": 19.38, |
| "learning_rate": 6.285434995112415e-07, |
| "loss": 0.0111, |
| "step": 39650 |
| }, |
| { |
| "epoch": 19.39, |
| "learning_rate": 6.163245356793744e-07, |
| "loss": 0.0152, |
| "step": 39675 |
| }, |
| { |
| "epoch": 19.4, |
| "learning_rate": 6.041055718475073e-07, |
| "loss": 0.0189, |
| "step": 39700 |
| }, |
| { |
| "epoch": 19.42, |
| "learning_rate": 5.92375366568915e-07, |
| "loss": 0.0158, |
| "step": 39725 |
| }, |
| { |
| "epoch": 19.43, |
| "learning_rate": 5.801564027370479e-07, |
| "loss": 0.0207, |
| "step": 39750 |
| }, |
| { |
| "epoch": 19.44, |
| "learning_rate": 5.679374389051809e-07, |
| "loss": 0.0189, |
| "step": 39775 |
| }, |
| { |
| "epoch": 19.45, |
| "learning_rate": 5.557184750733138e-07, |
| "loss": 0.0172, |
| "step": 39800 |
| }, |
| { |
| "epoch": 19.46, |
| "learning_rate": 5.434995112414468e-07, |
| "loss": 0.0173, |
| "step": 39825 |
| }, |
| { |
| "epoch": 19.48, |
| "learning_rate": 5.312805474095798e-07, |
| "loss": 0.0135, |
| "step": 39850 |
| }, |
| { |
| "epoch": 19.49, |
| "learning_rate": 5.190615835777126e-07, |
| "loss": 0.0126, |
| "step": 39875 |
| }, |
| { |
| "epoch": 19.5, |
| "learning_rate": 5.068426197458456e-07, |
| "loss": 0.0183, |
| "step": 39900 |
| }, |
| { |
| "epoch": 19.51, |
| "learning_rate": 4.946236559139786e-07, |
| "loss": 0.0121, |
| "step": 39925 |
| }, |
| { |
| "epoch": 19.53, |
| "learning_rate": 4.824046920821115e-07, |
| "loss": 0.0164, |
| "step": 39950 |
| }, |
| { |
| "epoch": 19.54, |
| "learning_rate": 4.701857282502444e-07, |
| "loss": 0.0227, |
| "step": 39975 |
| }, |
| { |
| "epoch": 19.55, |
| "learning_rate": 4.579667644183774e-07, |
| "loss": 0.0172, |
| "step": 40000 |
| }, |
| { |
| "epoch": 19.56, |
| "learning_rate": 4.4574780058651033e-07, |
| "loss": 0.0156, |
| "step": 40025 |
| }, |
| { |
| "epoch": 19.57, |
| "learning_rate": 4.335288367546432e-07, |
| "loss": 0.0212, |
| "step": 40050 |
| }, |
| { |
| "epoch": 19.59, |
| "learning_rate": 4.213098729227762e-07, |
| "loss": 0.0138, |
| "step": 40075 |
| }, |
| { |
| "epoch": 19.6, |
| "learning_rate": 4.090909090909091e-07, |
| "loss": 0.0179, |
| "step": 40100 |
| }, |
| { |
| "epoch": 19.61, |
| "learning_rate": 3.968719452590421e-07, |
| "loss": 0.0121, |
| "step": 40125 |
| }, |
| { |
| "epoch": 19.62, |
| "learning_rate": 3.8465298142717503e-07, |
| "loss": 0.0175, |
| "step": 40150 |
| }, |
| { |
| "epoch": 19.64, |
| "learning_rate": 3.7243401759530796e-07, |
| "loss": 0.0194, |
| "step": 40175 |
| }, |
| { |
| "epoch": 19.65, |
| "learning_rate": 3.602150537634409e-07, |
| "loss": 0.0119, |
| "step": 40200 |
| }, |
| { |
| "epoch": 19.66, |
| "learning_rate": 3.479960899315738e-07, |
| "loss": 0.0121, |
| "step": 40225 |
| }, |
| { |
| "epoch": 19.67, |
| "learning_rate": 3.3577712609970675e-07, |
| "loss": 0.0207, |
| "step": 40250 |
| }, |
| { |
| "epoch": 19.68, |
| "learning_rate": 3.2355816226783973e-07, |
| "loss": 0.0118, |
| "step": 40275 |
| }, |
| { |
| "epoch": 19.7, |
| "learning_rate": 3.1133919843597266e-07, |
| "loss": 0.0142, |
| "step": 40300 |
| }, |
| { |
| "epoch": 19.71, |
| "learning_rate": 2.991202346041056e-07, |
| "loss": 0.0228, |
| "step": 40325 |
| }, |
| { |
| "epoch": 19.72, |
| "learning_rate": 2.869012707722385e-07, |
| "loss": 0.0185, |
| "step": 40350 |
| }, |
| { |
| "epoch": 19.73, |
| "learning_rate": 2.746823069403715e-07, |
| "loss": 0.017, |
| "step": 40375 |
| }, |
| { |
| "epoch": 19.75, |
| "learning_rate": 2.6246334310850443e-07, |
| "loss": 0.0229, |
| "step": 40400 |
| }, |
| { |
| "epoch": 19.76, |
| "learning_rate": 2.5024437927663736e-07, |
| "loss": 0.0141, |
| "step": 40425 |
| }, |
| { |
| "epoch": 19.77, |
| "learning_rate": 2.3802541544477032e-07, |
| "loss": 0.0139, |
| "step": 40450 |
| }, |
| { |
| "epoch": 19.78, |
| "learning_rate": 2.2580645161290322e-07, |
| "loss": 0.0188, |
| "step": 40475 |
| }, |
| { |
| "epoch": 19.79, |
| "learning_rate": 2.1358748778103618e-07, |
| "loss": 0.0168, |
| "step": 40500 |
| }, |
| { |
| "epoch": 19.81, |
| "learning_rate": 2.0136852394916913e-07, |
| "loss": 0.0158, |
| "step": 40525 |
| }, |
| { |
| "epoch": 19.82, |
| "learning_rate": 1.8914956011730206e-07, |
| "loss": 0.0168, |
| "step": 40550 |
| }, |
| { |
| "epoch": 19.83, |
| "learning_rate": 1.7693059628543502e-07, |
| "loss": 0.0183, |
| "step": 40575 |
| }, |
| { |
| "epoch": 19.84, |
| "learning_rate": 1.6471163245356795e-07, |
| "loss": 0.019, |
| "step": 40600 |
| }, |
| { |
| "epoch": 19.86, |
| "learning_rate": 1.524926686217009e-07, |
| "loss": 0.0162, |
| "step": 40625 |
| }, |
| { |
| "epoch": 19.87, |
| "learning_rate": 1.4027370478983383e-07, |
| "loss": 0.0227, |
| "step": 40650 |
| }, |
| { |
| "epoch": 19.88, |
| "learning_rate": 1.280547409579668e-07, |
| "loss": 0.0169, |
| "step": 40675 |
| }, |
| { |
| "epoch": 19.89, |
| "learning_rate": 1.1583577712609972e-07, |
| "loss": 0.0106, |
| "step": 40700 |
| }, |
| { |
| "epoch": 19.9, |
| "learning_rate": 1.0361681329423266e-07, |
| "loss": 0.0175, |
| "step": 40725 |
| }, |
| { |
| "epoch": 19.92, |
| "learning_rate": 9.13978494623656e-08, |
| "loss": 0.0197, |
| "step": 40750 |
| }, |
| { |
| "epoch": 19.93, |
| "learning_rate": 7.917888563049853e-08, |
| "loss": 0.0176, |
| "step": 40775 |
| }, |
| { |
| "epoch": 19.94, |
| "learning_rate": 6.695992179863148e-08, |
| "loss": 0.0144, |
| "step": 40800 |
| }, |
| { |
| "epoch": 19.95, |
| "learning_rate": 5.4740957966764426e-08, |
| "loss": 0.0128, |
| "step": 40825 |
| }, |
| { |
| "epoch": 19.97, |
| "learning_rate": 4.252199413489737e-08, |
| "loss": 0.0136, |
| "step": 40850 |
| }, |
| { |
| "epoch": 19.98, |
| "learning_rate": 3.0303030303030305e-08, |
| "loss": 0.011, |
| "step": 40875 |
| }, |
| { |
| "epoch": 19.99, |
| "learning_rate": 1.8084066471163248e-08, |
| "loss": 0.0185, |
| "step": 40900 |
| }, |
| { |
| "epoch": 20.0, |
| "step": 40920, |
| "total_flos": 5.166258268431053e+17, |
| "train_loss": 0.12400934287872191, |
| "train_runtime": 16044.2941, |
| "train_samples_per_second": 489.522, |
| "train_steps_per_second": 2.55 |
| } |
| ], |
| "logging_steps": 25, |
| "max_steps": 40920, |
| "num_train_epochs": 20, |
| "save_steps": 500, |
| "total_flos": 5.166258268431053e+17, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|