| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.3852502500985238, | |
| "global_step": 944187, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9973522194226354e-05, | |
| "loss": 1.5613, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.994704438845271e-05, | |
| "loss": 1.4492, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9920566582679065e-05, | |
| "loss": 1.3946, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9894088776905424e-05, | |
| "loss": 1.3436, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9867610971131776e-05, | |
| "loss": 1.3033, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9841133165358134e-05, | |
| "loss": 1.2823, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9814655359584486e-05, | |
| "loss": 1.2504, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.978817755381085e-05, | |
| "loss": 1.2289, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9761699748037204e-05, | |
| "loss": 1.2179, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.973522194226356e-05, | |
| "loss": 1.1966, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9708744136489914e-05, | |
| "loss": 1.1812, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.968226633071627e-05, | |
| "loss": 1.1723, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9655788524942625e-05, | |
| "loss": 1.1612, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9629310719168984e-05, | |
| "loss": 1.1464, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9602832913395336e-05, | |
| "loss": 1.1321, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9576355107621694e-05, | |
| "loss": 1.1153, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9549877301848046e-05, | |
| "loss": 1.113, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9523399496074405e-05, | |
| "loss": 1.1049, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.949692169030076e-05, | |
| "loss": 1.0948, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9470443884527116e-05, | |
| "loss": 1.0879, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.944396607875347e-05, | |
| "loss": 1.0756, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.941748827297982e-05, | |
| "loss": 1.0727, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9391010467206185e-05, | |
| "loss": 1.0657, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.936453266143254e-05, | |
| "loss": 1.0599, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9338054855658896e-05, | |
| "loss": 1.0534, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.931157704988525e-05, | |
| "loss": 1.0443, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9285099244111606e-05, | |
| "loss": 1.0386, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.925862143833796e-05, | |
| "loss": 1.0342, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.923214363256432e-05, | |
| "loss": 1.0284, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.920566582679067e-05, | |
| "loss": 1.026, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.917918802101703e-05, | |
| "loss": 1.0239, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.915271021524338e-05, | |
| "loss": 1.0114, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.912623240946974e-05, | |
| "loss": 1.0092, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.909975460369609e-05, | |
| "loss": 1.0023, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.907327679792245e-05, | |
| "loss": 1.0013, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.90467989921488e-05, | |
| "loss": 0.9939, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.902032118637516e-05, | |
| "loss": 0.9892, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.899384338060151e-05, | |
| "loss": 0.9848, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.896736557482787e-05, | |
| "loss": 0.9805, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.894088776905423e-05, | |
| "loss": 0.9771, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.891440996328058e-05, | |
| "loss": 0.975, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.888793215750694e-05, | |
| "loss": 0.9714, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.886145435173329e-05, | |
| "loss": 0.9691, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.883497654595965e-05, | |
| "loss": 0.966, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.8808498740186e-05, | |
| "loss": 0.9589, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.878202093441236e-05, | |
| "loss": 0.9565, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.875554312863871e-05, | |
| "loss": 0.9557, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.872906532286507e-05, | |
| "loss": 0.9497, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.870258751709142e-05, | |
| "loss": 0.9474, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.867610971131778e-05, | |
| "loss": 0.9499, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.8649631905544134e-05, | |
| "loss": 0.9435, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.862315409977049e-05, | |
| "loss": 0.9402, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.8596676293996845e-05, | |
| "loss": 0.9373, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.85701984882232e-05, | |
| "loss": 0.9322, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.854372068244956e-05, | |
| "loss": 0.9354, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.8517242876675914e-05, | |
| "loss": 0.9274, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.849076507090227e-05, | |
| "loss": 0.9258, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.8464287265128625e-05, | |
| "loss": 0.9212, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.843780945935498e-05, | |
| "loss": 0.9163, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8411331653581335e-05, | |
| "loss": 0.9171, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8384853847807694e-05, | |
| "loss": 0.9116, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8358376042034046e-05, | |
| "loss": 0.9136, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8331898236260405e-05, | |
| "loss": 0.912, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8305420430486757e-05, | |
| "loss": 0.9071, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8278942624713115e-05, | |
| "loss": 0.9079, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.825246481893947e-05, | |
| "loss": 0.9026, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.8225987013165826e-05, | |
| "loss": 0.9038, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.819950920739218e-05, | |
| "loss": 0.8992, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.817303140161854e-05, | |
| "loss": 0.9037, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8146553595844895e-05, | |
| "loss": 0.9005, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8120075790071254e-05, | |
| "loss": 0.8974, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8093597984297606e-05, | |
| "loss": 0.8969, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.806712017852396e-05, | |
| "loss": 0.8885, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8040642372750317e-05, | |
| "loss": 0.8903, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.801416456697667e-05, | |
| "loss": 0.8825, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.798768676120303e-05, | |
| "loss": 0.8896, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.796120895542938e-05, | |
| "loss": 0.8818, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.793473114965574e-05, | |
| "loss": 0.8829, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.790825334388209e-05, | |
| "loss": 0.8741, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.788177553810845e-05, | |
| "loss": 0.8772, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.78552977323348e-05, | |
| "loss": 0.8818, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.782881992656116e-05, | |
| "loss": 0.8697, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.780234212078751e-05, | |
| "loss": 0.8705, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.7775864315013877e-05, | |
| "loss": 0.8708, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.774938650924023e-05, | |
| "loss": 0.8636, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.772290870346659e-05, | |
| "loss": 0.8616, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.769643089769294e-05, | |
| "loss": 0.8609, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.76699530919193e-05, | |
| "loss": 0.8722, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.764347528614565e-05, | |
| "loss": 0.8605, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.761699748037201e-05, | |
| "loss": 0.8605, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.759051967459836e-05, | |
| "loss": 0.8557, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.756404186882471e-05, | |
| "loss": 0.8568, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.753756406305107e-05, | |
| "loss": 0.8608, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.751108625727742e-05, | |
| "loss": 0.8514, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.748460845150378e-05, | |
| "loss": 0.8536, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7458130645730134e-05, | |
| "loss": 0.8505, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.743165283995649e-05, | |
| "loss": 0.8505, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7405175034182844e-05, | |
| "loss": 0.8528, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.737869722840921e-05, | |
| "loss": 0.8524, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.735221942263556e-05, | |
| "loss": 0.8457, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.732574161686192e-05, | |
| "loss": 0.8511, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.729926381108827e-05, | |
| "loss": 0.8439, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.727278600531463e-05, | |
| "loss": 0.8418, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.724630819954098e-05, | |
| "loss": 0.8436, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.721983039376734e-05, | |
| "loss": 0.8379, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.7193352587993694e-05, | |
| "loss": 0.8408, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.716687478222005e-05, | |
| "loss": 0.8397, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7140396976446404e-05, | |
| "loss": 0.8347, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.711391917067276e-05, | |
| "loss": 0.8355, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7087441364899115e-05, | |
| "loss": 0.8323, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7060963559125474e-05, | |
| "loss": 0.8268, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.7034485753351826e-05, | |
| "loss": 0.8309, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.700800794757818e-05, | |
| "loss": 0.829, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.6981530141804536e-05, | |
| "loss": 0.8281, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6955052336030895e-05, | |
| "loss": 0.8321, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6928574530257254e-05, | |
| "loss": 0.826, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6902096724483606e-05, | |
| "loss": 0.8275, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6875618918709964e-05, | |
| "loss": 0.8241, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6849141112936316e-05, | |
| "loss": 0.8237, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6822663307162675e-05, | |
| "loss": 0.8237, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.679618550138903e-05, | |
| "loss": 0.8212, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.6769707695615386e-05, | |
| "loss": 0.8168, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.674322988984174e-05, | |
| "loss": 0.8216, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.6716752084068096e-05, | |
| "loss": 0.8214, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.669027427829445e-05, | |
| "loss": 0.8148, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.666379647252081e-05, | |
| "loss": 0.8115, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.663731866674716e-05, | |
| "loss": 0.8154, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.661084086097352e-05, | |
| "loss": 0.821, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.658436305519987e-05, | |
| "loss": 0.8143, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.655788524942623e-05, | |
| "loss": 0.8134, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.653140744365259e-05, | |
| "loss": 0.8152, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.650492963787894e-05, | |
| "loss": 0.8131, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.64784518321053e-05, | |
| "loss": 0.8077, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.645197402633165e-05, | |
| "loss": 0.812, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.642549622055801e-05, | |
| "loss": 0.8074, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.639901841478436e-05, | |
| "loss": 0.8065, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.637254060901072e-05, | |
| "loss": 0.8081, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.634606280323707e-05, | |
| "loss": 0.8055, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.631958499746343e-05, | |
| "loss": 0.8022, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.629310719168978e-05, | |
| "loss": 0.8049, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.626662938591614e-05, | |
| "loss": 0.7996, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.624015158014249e-05, | |
| "loss": 0.8031, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.621367377436885e-05, | |
| "loss": 0.8045, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.61871959685952e-05, | |
| "loss": 0.8032, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.616071816282156e-05, | |
| "loss": 0.8005, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.613424035704792e-05, | |
| "loss": 0.799, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.610776255127427e-05, | |
| "loss": 0.7956, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.608128474550063e-05, | |
| "loss": 0.7955, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.605480693972698e-05, | |
| "loss": 0.7941, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.602832913395334e-05, | |
| "loss": 0.796, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.6001851328179693e-05, | |
| "loss": 0.7914, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.597537352240605e-05, | |
| "loss": 0.7878, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.5948895716632404e-05, | |
| "loss": 0.7877, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.592241791085876e-05, | |
| "loss": 0.7877, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5895940105085115e-05, | |
| "loss": 0.7898, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5869462299311473e-05, | |
| "loss": 0.7861, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5842984493537825e-05, | |
| "loss": 0.7952, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5816506687764184e-05, | |
| "loss": 0.7836, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5790028881990536e-05, | |
| "loss": 0.7911, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.57635510762169e-05, | |
| "loss": 0.7919, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.5737073270443253e-05, | |
| "loss": 0.7861, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.571059546466961e-05, | |
| "loss": 0.7786, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.5684117658895964e-05, | |
| "loss": 0.7834, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.5657639853122316e-05, | |
| "loss": 0.7799, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.5631162047348675e-05, | |
| "loss": 0.7845, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.560468424157503e-05, | |
| "loss": 0.7788, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.5578206435801385e-05, | |
| "loss": 0.7782, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.555172863002774e-05, | |
| "loss": 0.7783, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.5525250824254096e-05, | |
| "loss": 0.7754, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.549877301848045e-05, | |
| "loss": 0.7779, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.547229521270681e-05, | |
| "loss": 0.7782, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.544581740693316e-05, | |
| "loss": 0.7783, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.541933960115952e-05, | |
| "loss": 0.7792, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.539286179538587e-05, | |
| "loss": 0.7744, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.536638398961223e-05, | |
| "loss": 0.777, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.533990618383859e-05, | |
| "loss": 0.7714, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.5313428378064945e-05, | |
| "loss": 0.7735, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.52869505722913e-05, | |
| "loss": 0.778, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.5260472766517656e-05, | |
| "loss": 0.7746, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.523399496074401e-05, | |
| "loss": 0.7673, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.520751715497037e-05, | |
| "loss": 0.7761, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.518103934919672e-05, | |
| "loss": 0.7713, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.515456154342308e-05, | |
| "loss": 0.7754, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.512808373764943e-05, | |
| "loss": 0.7736, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.510160593187578e-05, | |
| "loss": 0.7704, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.507512812610214e-05, | |
| "loss": 0.772, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.504865032032849e-05, | |
| "loss": 0.7697, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.502217251455485e-05, | |
| "loss": 0.7672, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.49956947087812e-05, | |
| "loss": 0.7625, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.496921690300756e-05, | |
| "loss": 0.7607, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.494273909723392e-05, | |
| "loss": 0.7696, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.491626129146028e-05, | |
| "loss": 0.7667, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.488978348568663e-05, | |
| "loss": 0.7635, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.486330567991299e-05, | |
| "loss": 0.7677, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.483682787413934e-05, | |
| "loss": 0.7661, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.48103500683657e-05, | |
| "loss": 0.7646, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.478387226259205e-05, | |
| "loss": 0.7609, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.475739445681841e-05, | |
| "loss": 0.7573, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.473091665104476e-05, | |
| "loss": 0.7591, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.470443884527112e-05, | |
| "loss": 0.7611, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.467796103949747e-05, | |
| "loss": 0.759, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.465148323372383e-05, | |
| "loss": 0.7588, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.4625005427950184e-05, | |
| "loss": 0.7623, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.459852762217654e-05, | |
| "loss": 0.7534, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.4572049816402894e-05, | |
| "loss": 0.7567, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.454557201062925e-05, | |
| "loss": 0.7572, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.451909420485561e-05, | |
| "loss": 0.7594, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.4492616399081964e-05, | |
| "loss": 0.7563, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.446613859330832e-05, | |
| "loss": 0.7543, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4439660787534674e-05, | |
| "loss": 0.7544, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.441318298176103e-05, | |
| "loss": 0.751, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4386705175987385e-05, | |
| "loss": 0.7468, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4360227370213744e-05, | |
| "loss": 0.7523, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4333749564440096e-05, | |
| "loss": 0.7504, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4307271758666454e-05, | |
| "loss": 0.7521, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4280793952892806e-05, | |
| "loss": 0.7492, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.4254316147119165e-05, | |
| "loss": 0.7509, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.422783834134552e-05, | |
| "loss": 0.7493, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.4201360535571876e-05, | |
| "loss": 0.7501, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.417488272979823e-05, | |
| "loss": 0.7436, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.4148404924024586e-05, | |
| "loss": 0.7475, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.4121927118250945e-05, | |
| "loss": 0.7529, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.40954493124773e-05, | |
| "loss": 0.7447, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.4068971506703656e-05, | |
| "loss": 0.7478, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.404249370093001e-05, | |
| "loss": 0.7503, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.4016015895156366e-05, | |
| "loss": 0.74, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.398953808938272e-05, | |
| "loss": 0.7462, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.396306028360908e-05, | |
| "loss": 0.7462, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.393658247783543e-05, | |
| "loss": 0.7377, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.391010467206179e-05, | |
| "loss": 0.7421, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.388362686628814e-05, | |
| "loss": 0.7369, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.38571490605145e-05, | |
| "loss": 0.7377, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.383067125474085e-05, | |
| "loss": 0.7496, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.380419344896721e-05, | |
| "loss": 0.7403, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.377771564319356e-05, | |
| "loss": 0.7437, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.375123783741992e-05, | |
| "loss": 0.7372, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.372476003164628e-05, | |
| "loss": 0.7421, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.369828222587263e-05, | |
| "loss": 0.7404, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.367180442009899e-05, | |
| "loss": 0.7363, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.364532661432534e-05, | |
| "loss": 0.7413, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.36188488085517e-05, | |
| "loss": 0.7382, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.359237100277805e-05, | |
| "loss": 0.7383, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.356589319700441e-05, | |
| "loss": 0.7362, | |
| "step": 121500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.353941539123076e-05, | |
| "loss": 0.738, | |
| "step": 122000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.351293758545712e-05, | |
| "loss": 0.735, | |
| "step": 122500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.348645977968347e-05, | |
| "loss": 0.7411, | |
| "step": 123000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.345998197390983e-05, | |
| "loss": 0.7347, | |
| "step": 123500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.3433504168136183e-05, | |
| "loss": 0.7334, | |
| "step": 124000 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.340702636236254e-05, | |
| "loss": 0.7376, | |
| "step": 124500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.3380548556588894e-05, | |
| "loss": 0.733, | |
| "step": 125000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.335407075081525e-05, | |
| "loss": 0.7299, | |
| "step": 125500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.332759294504161e-05, | |
| "loss": 0.7312, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.330111513926797e-05, | |
| "loss": 0.7382, | |
| "step": 126500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.327463733349432e-05, | |
| "loss": 0.7322, | |
| "step": 127000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.324815952772068e-05, | |
| "loss": 0.7336, | |
| "step": 127500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.322168172194703e-05, | |
| "loss": 0.734, | |
| "step": 128000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.3195203916173385e-05, | |
| "loss": 0.7317, | |
| "step": 128500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3168726110399743e-05, | |
| "loss": 0.7299, | |
| "step": 129000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3142248304626095e-05, | |
| "loss": 0.7284, | |
| "step": 129500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3115770498852454e-05, | |
| "loss": 0.728, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3089292693078806e-05, | |
| "loss": 0.7302, | |
| "step": 130500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3062814887305165e-05, | |
| "loss": 0.7237, | |
| "step": 131000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.303633708153152e-05, | |
| "loss": 0.7277, | |
| "step": 131500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.3009859275757875e-05, | |
| "loss": 0.7312, | |
| "step": 132000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.298338146998423e-05, | |
| "loss": 0.728, | |
| "step": 132500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2956903664210586e-05, | |
| "loss": 0.7214, | |
| "step": 133000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2930425858436945e-05, | |
| "loss": 0.7316, | |
| "step": 133500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2903948052663303e-05, | |
| "loss": 0.7207, | |
| "step": 134000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2877470246889655e-05, | |
| "loss": 0.7275, | |
| "step": 134500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2850992441116014e-05, | |
| "loss": 0.7262, | |
| "step": 135000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2824514635342366e-05, | |
| "loss": 0.726, | |
| "step": 135500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.2798036829568725e-05, | |
| "loss": 0.7259, | |
| "step": 136000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.277155902379508e-05, | |
| "loss": 0.7253, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.2745081218021435e-05, | |
| "loss": 0.7272, | |
| "step": 137000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.271860341224779e-05, | |
| "loss": 0.7213, | |
| "step": 137500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.269212560647414e-05, | |
| "loss": 0.7217, | |
| "step": 138000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.26656478007005e-05, | |
| "loss": 0.7207, | |
| "step": 138500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.263916999492685e-05, | |
| "loss": 0.7229, | |
| "step": 139000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.261269218915321e-05, | |
| "loss": 0.7226, | |
| "step": 139500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.258621438337956e-05, | |
| "loss": 0.7214, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.255973657760592e-05, | |
| "loss": 0.7215, | |
| "step": 140500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.253325877183228e-05, | |
| "loss": 0.7194, | |
| "step": 141000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.250678096605864e-05, | |
| "loss": 0.7158, | |
| "step": 141500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.248030316028499e-05, | |
| "loss": 0.7244, | |
| "step": 142000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.245382535451135e-05, | |
| "loss": 0.7237, | |
| "step": 142500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.24273475487377e-05, | |
| "loss": 0.7193, | |
| "step": 143000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.240086974296406e-05, | |
| "loss": 0.7188, | |
| "step": 143500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.237439193719041e-05, | |
| "loss": 0.7218, | |
| "step": 144000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.234791413141677e-05, | |
| "loss": 0.7189, | |
| "step": 144500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.232143632564312e-05, | |
| "loss": 0.7147, | |
| "step": 145000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.229495851986948e-05, | |
| "loss": 0.7149, | |
| "step": 145500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.226848071409583e-05, | |
| "loss": 0.7128, | |
| "step": 146000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.224200290832219e-05, | |
| "loss": 0.7194, | |
| "step": 146500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.221552510254854e-05, | |
| "loss": 0.7153, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.21890472967749e-05, | |
| "loss": 0.7172, | |
| "step": 147500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.216256949100125e-05, | |
| "loss": 0.7108, | |
| "step": 148000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.213609168522761e-05, | |
| "loss": 0.7188, | |
| "step": 148500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.210961387945397e-05, | |
| "loss": 0.714, | |
| "step": 149000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.208313607368032e-05, | |
| "loss": 0.7138, | |
| "step": 149500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.205665826790668e-05, | |
| "loss": 0.7153, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.203018046213303e-05, | |
| "loss": 0.7167, | |
| "step": 150500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.200370265635939e-05, | |
| "loss": 0.7145, | |
| "step": 151000 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.197722485058574e-05, | |
| "loss": 0.7173, | |
| "step": 151500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.19507470448121e-05, | |
| "loss": 0.7087, | |
| "step": 152000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1924269239038454e-05, | |
| "loss": 0.7164, | |
| "step": 152500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.189779143326481e-05, | |
| "loss": 0.7086, | |
| "step": 153000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1871313627491164e-05, | |
| "loss": 0.7116, | |
| "step": 153500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.184483582171752e-05, | |
| "loss": 0.7073, | |
| "step": 154000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1818358015943875e-05, | |
| "loss": 0.7081, | |
| "step": 154500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1791880210170234e-05, | |
| "loss": 0.7111, | |
| "step": 155000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1765402404396586e-05, | |
| "loss": 0.711, | |
| "step": 155500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.1738924598622944e-05, | |
| "loss": 0.7133, | |
| "step": 156000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.17124467928493e-05, | |
| "loss": 0.7138, | |
| "step": 156500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1685968987075655e-05, | |
| "loss": 0.713, | |
| "step": 157000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1659491181302014e-05, | |
| "loss": 0.7082, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1633013375528366e-05, | |
| "loss": 0.7103, | |
| "step": 158000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1606535569754724e-05, | |
| "loss": 0.7111, | |
| "step": 158500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1580057763981076e-05, | |
| "loss": 0.7078, | |
| "step": 159000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.1553579958207435e-05, | |
| "loss": 0.7029, | |
| "step": 159500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.152710215243379e-05, | |
| "loss": 0.7089, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.1500624346660146e-05, | |
| "loss": 0.7093, | |
| "step": 160500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.14741465408865e-05, | |
| "loss": 0.7079, | |
| "step": 161000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.1447668735112856e-05, | |
| "loss": 0.7076, | |
| "step": 161500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.142119092933921e-05, | |
| "loss": 0.7083, | |
| "step": 162000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.139471312356557e-05, | |
| "loss": 0.7057, | |
| "step": 162500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.136823531779192e-05, | |
| "loss": 0.6995, | |
| "step": 163000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.134175751201828e-05, | |
| "loss": 0.709, | |
| "step": 163500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.1315279706244636e-05, | |
| "loss": 0.701, | |
| "step": 164000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.128880190047099e-05, | |
| "loss": 0.7098, | |
| "step": 164500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.126232409469735e-05, | |
| "loss": 0.7056, | |
| "step": 165000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.12358462889237e-05, | |
| "loss": 0.7036, | |
| "step": 165500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.120936848315006e-05, | |
| "loss": 0.6969, | |
| "step": 166000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.118289067737641e-05, | |
| "loss": 0.7036, | |
| "step": 166500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.115641287160277e-05, | |
| "loss": 0.7025, | |
| "step": 167000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.112993506582912e-05, | |
| "loss": 0.7058, | |
| "step": 167500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.110345726005548e-05, | |
| "loss": 0.7021, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.107697945428183e-05, | |
| "loss": 0.7027, | |
| "step": 168500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.105050164850819e-05, | |
| "loss": 0.7013, | |
| "step": 169000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.102402384273454e-05, | |
| "loss": 0.6999, | |
| "step": 169500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.09975460369609e-05, | |
| "loss": 0.7052, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.097106823118725e-05, | |
| "loss": 0.7026, | |
| "step": 170500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.094459042541361e-05, | |
| "loss": 0.6993, | |
| "step": 171000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.091811261963997e-05, | |
| "loss": 0.7016, | |
| "step": 171500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.089163481386633e-05, | |
| "loss": 0.6992, | |
| "step": 172000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.086515700809268e-05, | |
| "loss": 0.6954, | |
| "step": 172500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.083867920231904e-05, | |
| "loss": 0.7, | |
| "step": 173000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.081220139654539e-05, | |
| "loss": 0.698, | |
| "step": 173500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.078572359077174e-05, | |
| "loss": 0.7005, | |
| "step": 174000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.07592457849981e-05, | |
| "loss": 0.7035, | |
| "step": 174500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.0732767979224453e-05, | |
| "loss": 0.6945, | |
| "step": 175000 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.070629017345081e-05, | |
| "loss": 0.6994, | |
| "step": 175500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.0679812367677164e-05, | |
| "loss": 0.6907, | |
| "step": 176000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.065333456190352e-05, | |
| "loss": 0.6926, | |
| "step": 176500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.0626856756129875e-05, | |
| "loss": 0.6953, | |
| "step": 177000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.0600378950356233e-05, | |
| "loss": 0.6941, | |
| "step": 177500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.0573901144582585e-05, | |
| "loss": 0.6961, | |
| "step": 178000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.0547423338808944e-05, | |
| "loss": 0.6959, | |
| "step": 178500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.05209455330353e-05, | |
| "loss": 0.6917, | |
| "step": 179000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.049446772726166e-05, | |
| "loss": 0.6976, | |
| "step": 179500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.0467989921488013e-05, | |
| "loss": 0.7003, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.044151211571437e-05, | |
| "loss": 0.6969, | |
| "step": 180500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0415034309940724e-05, | |
| "loss": 0.6943, | |
| "step": 181000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.038855650416708e-05, | |
| "loss": 0.6892, | |
| "step": 181500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0362078698393435e-05, | |
| "loss": 0.6923, | |
| "step": 182000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0335600892619793e-05, | |
| "loss": 0.6938, | |
| "step": 182500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0309123086846145e-05, | |
| "loss": 0.6954, | |
| "step": 183000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0282645281072504e-05, | |
| "loss": 0.6936, | |
| "step": 183500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.0256167475298856e-05, | |
| "loss": 0.6885, | |
| "step": 184000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.022968966952521e-05, | |
| "loss": 0.6981, | |
| "step": 184500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.020321186375157e-05, | |
| "loss": 0.6915, | |
| "step": 185000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.017673405797792e-05, | |
| "loss": 0.6966, | |
| "step": 185500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.015025625220428e-05, | |
| "loss": 0.6962, | |
| "step": 186000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.012377844643063e-05, | |
| "loss": 0.6944, | |
| "step": 186500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.0097300640656995e-05, | |
| "loss": 0.6967, | |
| "step": 187000 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.007082283488335e-05, | |
| "loss": 0.6902, | |
| "step": 187500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.0044345029109705e-05, | |
| "loss": 0.6914, | |
| "step": 188000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.001786722333606e-05, | |
| "loss": 0.6925, | |
| "step": 188500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.9991389417562416e-05, | |
| "loss": 0.6867, | |
| "step": 189000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.996491161178877e-05, | |
| "loss": 0.6862, | |
| "step": 189500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.993843380601513e-05, | |
| "loss": 0.6912, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.991195600024148e-05, | |
| "loss": 0.685, | |
| "step": 190500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.988547819446784e-05, | |
| "loss": 0.6885, | |
| "step": 191000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 3.985900038869419e-05, | |
| "loss": 0.6867, | |
| "step": 191500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.983252258292055e-05, | |
| "loss": 0.6952, | |
| "step": 192000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.98060447771469e-05, | |
| "loss": 0.6912, | |
| "step": 192500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.977956697137326e-05, | |
| "loss": 0.6835, | |
| "step": 193000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.975308916559961e-05, | |
| "loss": 0.692, | |
| "step": 193500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.972661135982597e-05, | |
| "loss": 0.6914, | |
| "step": 194000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.970013355405233e-05, | |
| "loss": 0.6888, | |
| "step": 194500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.967365574827868e-05, | |
| "loss": 0.6865, | |
| "step": 195000 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.964717794250504e-05, | |
| "loss": 0.681, | |
| "step": 195500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.962070013673139e-05, | |
| "loss": 0.6906, | |
| "step": 196000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.959422233095775e-05, | |
| "loss": 0.6831, | |
| "step": 196500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.95677445251841e-05, | |
| "loss": 0.6859, | |
| "step": 197000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.954126671941046e-05, | |
| "loss": 0.6807, | |
| "step": 197500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.951478891363681e-05, | |
| "loss": 0.691, | |
| "step": 198000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.948831110786317e-05, | |
| "loss": 0.6868, | |
| "step": 198500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.946183330208952e-05, | |
| "loss": 0.6896, | |
| "step": 199000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 3.943535549631588e-05, | |
| "loss": 0.6883, | |
| "step": 199500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.940887769054223e-05, | |
| "loss": 0.6808, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.938239988476859e-05, | |
| "loss": 0.6841, | |
| "step": 200500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.9355922078994944e-05, | |
| "loss": 0.6923, | |
| "step": 201000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.93294442732213e-05, | |
| "loss": 0.6827, | |
| "step": 201500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.930296646744766e-05, | |
| "loss": 0.6836, | |
| "step": 202000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.927648866167401e-05, | |
| "loss": 0.6862, | |
| "step": 202500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.925001085590037e-05, | |
| "loss": 0.6866, | |
| "step": 203000 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 3.9223533050126724e-05, | |
| "loss": 0.6792, | |
| "step": 203500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.919705524435308e-05, | |
| "loss": 0.6779, | |
| "step": 204000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9170577438579434e-05, | |
| "loss": 0.6828, | |
| "step": 204500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.914409963280579e-05, | |
| "loss": 0.6816, | |
| "step": 205000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9117621827032145e-05, | |
| "loss": 0.6771, | |
| "step": 205500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9091144021258504e-05, | |
| "loss": 0.6762, | |
| "step": 206000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9064666215484856e-05, | |
| "loss": 0.6833, | |
| "step": 206500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9038188409711214e-05, | |
| "loss": 0.6803, | |
| "step": 207000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 3.9011710603937566e-05, | |
| "loss": 0.6786, | |
| "step": 207500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8985232798163925e-05, | |
| "loss": 0.6817, | |
| "step": 208000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.895875499239028e-05, | |
| "loss": 0.6825, | |
| "step": 208500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8932277186616636e-05, | |
| "loss": 0.6771, | |
| "step": 209000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8905799380842994e-05, | |
| "loss": 0.6813, | |
| "step": 209500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8879321575069346e-05, | |
| "loss": 0.677, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8852843769295705e-05, | |
| "loss": 0.6758, | |
| "step": 210500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.882636596352206e-05, | |
| "loss": 0.6777, | |
| "step": 211000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 3.8799888157748416e-05, | |
| "loss": 0.6762, | |
| "step": 211500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.877341035197477e-05, | |
| "loss": 0.6762, | |
| "step": 212000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.8746932546201126e-05, | |
| "loss": 0.6727, | |
| "step": 212500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.872045474042748e-05, | |
| "loss": 0.6781, | |
| "step": 213000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.869397693465384e-05, | |
| "loss": 0.6818, | |
| "step": 213500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.866749912888019e-05, | |
| "loss": 0.6759, | |
| "step": 214000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.864102132310655e-05, | |
| "loss": 0.6812, | |
| "step": 214500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.86145435173329e-05, | |
| "loss": 0.6798, | |
| "step": 215000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 3.858806571155926e-05, | |
| "loss": 0.6729, | |
| "step": 215500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.856158790578561e-05, | |
| "loss": 0.6784, | |
| "step": 216000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.853511010001197e-05, | |
| "loss": 0.672, | |
| "step": 216500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.850863229423832e-05, | |
| "loss": 0.6746, | |
| "step": 217000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.8482154488464686e-05, | |
| "loss": 0.6705, | |
| "step": 217500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.845567668269104e-05, | |
| "loss": 0.6749, | |
| "step": 218000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.84291988769174e-05, | |
| "loss": 0.6753, | |
| "step": 218500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.840272107114375e-05, | |
| "loss": 0.6714, | |
| "step": 219000 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 3.837624326537011e-05, | |
| "loss": 0.6713, | |
| "step": 219500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.834976545959646e-05, | |
| "loss": 0.6733, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.832328765382281e-05, | |
| "loss": 0.6699, | |
| "step": 220500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.829680984804917e-05, | |
| "loss": 0.6716, | |
| "step": 221000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.827033204227552e-05, | |
| "loss": 0.6733, | |
| "step": 221500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.824385423650188e-05, | |
| "loss": 0.6723, | |
| "step": 222000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.821737643072823e-05, | |
| "loss": 0.6676, | |
| "step": 222500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.819089862495459e-05, | |
| "loss": 0.6735, | |
| "step": 223000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 3.8164420819180943e-05, | |
| "loss": 0.6809, | |
| "step": 223500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.81379430134073e-05, | |
| "loss": 0.6669, | |
| "step": 224000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.8111465207633654e-05, | |
| "loss": 0.674, | |
| "step": 224500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.808498740186002e-05, | |
| "loss": 0.6718, | |
| "step": 225000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.805850959608637e-05, | |
| "loss": 0.6681, | |
| "step": 225500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.803203179031273e-05, | |
| "loss": 0.668, | |
| "step": 226000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.800555398453908e-05, | |
| "loss": 0.6679, | |
| "step": 226500 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.797907617876544e-05, | |
| "loss": 0.6695, | |
| "step": 227000 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 3.795259837299179e-05, | |
| "loss": 0.6729, | |
| "step": 227500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.792612056721815e-05, | |
| "loss": 0.6697, | |
| "step": 228000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.7899642761444503e-05, | |
| "loss": 0.6706, | |
| "step": 228500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.787316495567086e-05, | |
| "loss": 0.6642, | |
| "step": 229000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.7846687149897214e-05, | |
| "loss": 0.6682, | |
| "step": 229500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.7820209344123566e-05, | |
| "loss": 0.6683, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.7793731538349925e-05, | |
| "loss": 0.6693, | |
| "step": 230500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.776725373257628e-05, | |
| "loss": 0.6729, | |
| "step": 231000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 3.7740775926802635e-05, | |
| "loss": 0.6633, | |
| "step": 231500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.771429812102899e-05, | |
| "loss": 0.6644, | |
| "step": 232000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.768782031525535e-05, | |
| "loss": 0.6716, | |
| "step": 232500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7661342509481705e-05, | |
| "loss": 0.6682, | |
| "step": 233000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7634864703708063e-05, | |
| "loss": 0.6678, | |
| "step": 233500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7608386897934415e-05, | |
| "loss": 0.6689, | |
| "step": 234000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7581909092160774e-05, | |
| "loss": 0.666, | |
| "step": 234500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7555431286387126e-05, | |
| "loss": 0.6672, | |
| "step": 235000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 3.7528953480613485e-05, | |
| "loss": 0.6612, | |
| "step": 235500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.750247567483984e-05, | |
| "loss": 0.6624, | |
| "step": 236000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.7475997869066195e-05, | |
| "loss": 0.6631, | |
| "step": 236500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.744952006329255e-05, | |
| "loss": 0.6633, | |
| "step": 237000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.7423042257518906e-05, | |
| "loss": 0.664, | |
| "step": 237500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.739656445174526e-05, | |
| "loss": 0.667, | |
| "step": 238000 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.737008664597162e-05, | |
| "loss": 0.6677, | |
| "step": 238500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 3.734360884019797e-05, | |
| "loss": 0.6607, | |
| "step": 239000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.731713103442433e-05, | |
| "loss": 0.667, | |
| "step": 239500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.7290653228650686e-05, | |
| "loss": 0.6628, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.726417542287704e-05, | |
| "loss": 0.6624, | |
| "step": 240500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.72376976171034e-05, | |
| "loss": 0.6644, | |
| "step": 241000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.721121981132975e-05, | |
| "loss": 0.6624, | |
| "step": 241500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.718474200555611e-05, | |
| "loss": 0.6623, | |
| "step": 242000 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.715826419978246e-05, | |
| "loss": 0.6625, | |
| "step": 242500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 3.713178639400882e-05, | |
| "loss": 0.6621, | |
| "step": 243000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.710530858823517e-05, | |
| "loss": 0.6638, | |
| "step": 243500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.707883078246153e-05, | |
| "loss": 0.6632, | |
| "step": 244000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.705235297668788e-05, | |
| "loss": 0.6606, | |
| "step": 244500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.702587517091424e-05, | |
| "loss": 0.6627, | |
| "step": 245000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.699939736514059e-05, | |
| "loss": 0.6651, | |
| "step": 245500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.697291955936695e-05, | |
| "loss": 0.659, | |
| "step": 246000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.69464417535933e-05, | |
| "loss": 0.6661, | |
| "step": 246500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 3.691996394781966e-05, | |
| "loss": 0.6609, | |
| "step": 247000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.689348614204602e-05, | |
| "loss": 0.665, | |
| "step": 247500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.686700833627237e-05, | |
| "loss": 0.6575, | |
| "step": 248000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.684053053049873e-05, | |
| "loss": 0.6592, | |
| "step": 248500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.681405272472508e-05, | |
| "loss": 0.6576, | |
| "step": 249000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.678757491895144e-05, | |
| "loss": 0.66, | |
| "step": 249500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.676109711317779e-05, | |
| "loss": 0.6617, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.673461930740415e-05, | |
| "loss": 0.6626, | |
| "step": 250500 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 3.67081415016305e-05, | |
| "loss": 0.6551, | |
| "step": 251000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.668166369585686e-05, | |
| "loss": 0.6619, | |
| "step": 251500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.6655185890083214e-05, | |
| "loss": 0.6563, | |
| "step": 252000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.662870808430957e-05, | |
| "loss": 0.6573, | |
| "step": 252500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.6602230278535924e-05, | |
| "loss": 0.6632, | |
| "step": 253000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.657575247276228e-05, | |
| "loss": 0.6533, | |
| "step": 253500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.6549274666988635e-05, | |
| "loss": 0.6571, | |
| "step": 254000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.6522796861214994e-05, | |
| "loss": 0.6555, | |
| "step": 254500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.6496319055441346e-05, | |
| "loss": 0.6573, | |
| "step": 255000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6469841249667704e-05, | |
| "loss": 0.6567, | |
| "step": 255500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.644336344389406e-05, | |
| "loss": 0.6582, | |
| "step": 256000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6416885638120415e-05, | |
| "loss": 0.6583, | |
| "step": 256500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6390407832346774e-05, | |
| "loss": 0.6536, | |
| "step": 257000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6363930026573126e-05, | |
| "loss": 0.6553, | |
| "step": 257500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6337452220799484e-05, | |
| "loss": 0.654, | |
| "step": 258000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6310974415025836e-05, | |
| "loss": 0.6556, | |
| "step": 258500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6284496609252195e-05, | |
| "loss": 0.6546, | |
| "step": 259000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.625801880347855e-05, | |
| "loss": 0.6506, | |
| "step": 259500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.6231540997704906e-05, | |
| "loss": 0.6512, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.620506319193126e-05, | |
| "loss": 0.6571, | |
| "step": 260500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.6178585386157616e-05, | |
| "loss": 0.6576, | |
| "step": 261000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.615210758038397e-05, | |
| "loss": 0.6563, | |
| "step": 261500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.612562977461033e-05, | |
| "loss": 0.6572, | |
| "step": 262000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.609915196883668e-05, | |
| "loss": 0.6532, | |
| "step": 262500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 3.6072674163063044e-05, | |
| "loss": 0.6566, | |
| "step": 263000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.6046196357289396e-05, | |
| "loss": 0.6573, | |
| "step": 263500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.6019718551515755e-05, | |
| "loss": 0.6513, | |
| "step": 264000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.599324074574211e-05, | |
| "loss": 0.6572, | |
| "step": 264500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.5966762939968466e-05, | |
| "loss": 0.6626, | |
| "step": 265000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.594028513419482e-05, | |
| "loss": 0.6502, | |
| "step": 265500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.591380732842117e-05, | |
| "loss": 0.6545, | |
| "step": 266000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.588732952264753e-05, | |
| "loss": 0.6543, | |
| "step": 266500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.586085171687388e-05, | |
| "loss": 0.6482, | |
| "step": 267000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.583437391110024e-05, | |
| "loss": 0.6517, | |
| "step": 267500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.580789610532659e-05, | |
| "loss": 0.6532, | |
| "step": 268000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.578141829955295e-05, | |
| "loss": 0.6519, | |
| "step": 268500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.57549404937793e-05, | |
| "loss": 0.6527, | |
| "step": 269000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.572846268800566e-05, | |
| "loss": 0.6505, | |
| "step": 269500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.570198488223201e-05, | |
| "loss": 0.6575, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.567550707645838e-05, | |
| "loss": 0.6522, | |
| "step": 270500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.564902927068473e-05, | |
| "loss": 0.6535, | |
| "step": 271000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.562255146491109e-05, | |
| "loss": 0.6472, | |
| "step": 271500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.559607365913744e-05, | |
| "loss": 0.6515, | |
| "step": 272000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.55695958533638e-05, | |
| "loss": 0.6493, | |
| "step": 272500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.554311804759015e-05, | |
| "loss": 0.649, | |
| "step": 273000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.551664024181651e-05, | |
| "loss": 0.6535, | |
| "step": 273500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.549016243604286e-05, | |
| "loss": 0.6499, | |
| "step": 274000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.546368463026922e-05, | |
| "loss": 0.6472, | |
| "step": 274500 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.543720682449557e-05, | |
| "loss": 0.6503, | |
| "step": 275000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.541072901872193e-05, | |
| "loss": 0.6494, | |
| "step": 275500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.538425121294828e-05, | |
| "loss": 0.6498, | |
| "step": 276000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.5357773407174635e-05, | |
| "loss": 0.6516, | |
| "step": 276500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.5331295601400993e-05, | |
| "loss": 0.6455, | |
| "step": 277000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.5304817795627345e-05, | |
| "loss": 0.6484, | |
| "step": 277500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.527833998985371e-05, | |
| "loss": 0.654, | |
| "step": 278000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.525186218408006e-05, | |
| "loss": 0.6492, | |
| "step": 278500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.522538437830642e-05, | |
| "loss": 0.6458, | |
| "step": 279000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5198906572532773e-05, | |
| "loss": 0.6462, | |
| "step": 279500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.517242876675913e-05, | |
| "loss": 0.6488, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5145950960985484e-05, | |
| "loss": 0.6457, | |
| "step": 280500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.511947315521184e-05, | |
| "loss": 0.6505, | |
| "step": 281000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5092995349438195e-05, | |
| "loss": 0.6479, | |
| "step": 281500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5066517543664553e-05, | |
| "loss": 0.6458, | |
| "step": 282000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5040039737890905e-05, | |
| "loss": 0.6473, | |
| "step": 282500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.5013561932117264e-05, | |
| "loss": 0.6451, | |
| "step": 283000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.4987084126343616e-05, | |
| "loss": 0.6517, | |
| "step": 283500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.4960606320569975e-05, | |
| "loss": 0.6471, | |
| "step": 284000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.493412851479633e-05, | |
| "loss": 0.6405, | |
| "step": 284500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.4907650709022685e-05, | |
| "loss": 0.6461, | |
| "step": 285000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.488117290324904e-05, | |
| "loss": 0.6462, | |
| "step": 285500 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.4854695097475396e-05, | |
| "loss": 0.648, | |
| "step": 286000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.4828217291701755e-05, | |
| "loss": 0.6424, | |
| "step": 286500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.480173948592811e-05, | |
| "loss": 0.6446, | |
| "step": 287000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.4775261680154465e-05, | |
| "loss": 0.6406, | |
| "step": 287500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.474878387438082e-05, | |
| "loss": 0.6417, | |
| "step": 288000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.4722306068607176e-05, | |
| "loss": 0.6477, | |
| "step": 288500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.469582826283353e-05, | |
| "loss": 0.6456, | |
| "step": 289000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.466935045705989e-05, | |
| "loss": 0.6446, | |
| "step": 289500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.464287265128624e-05, | |
| "loss": 0.646, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.46163948455126e-05, | |
| "loss": 0.6449, | |
| "step": 290500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.458991703973895e-05, | |
| "loss": 0.6468, | |
| "step": 291000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.456343923396531e-05, | |
| "loss": 0.6382, | |
| "step": 291500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.453696142819166e-05, | |
| "loss": 0.6443, | |
| "step": 292000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.451048362241802e-05, | |
| "loss": 0.6438, | |
| "step": 292500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.448400581664437e-05, | |
| "loss": 0.6392, | |
| "step": 293000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.445752801087073e-05, | |
| "loss": 0.6396, | |
| "step": 293500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.443105020509709e-05, | |
| "loss": 0.6399, | |
| "step": 294000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.440457239932344e-05, | |
| "loss": 0.6436, | |
| "step": 294500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.43780945935498e-05, | |
| "loss": 0.641, | |
| "step": 295000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.435161678777615e-05, | |
| "loss": 0.6433, | |
| "step": 295500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.432513898200251e-05, | |
| "loss": 0.6351, | |
| "step": 296000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.429866117622886e-05, | |
| "loss": 0.6442, | |
| "step": 296500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.427218337045522e-05, | |
| "loss": 0.6457, | |
| "step": 297000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.424570556468157e-05, | |
| "loss": 0.6443, | |
| "step": 297500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.421922775890793e-05, | |
| "loss": 0.6426, | |
| "step": 298000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.419274995313428e-05, | |
| "loss": 0.6419, | |
| "step": 298500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.416627214736064e-05, | |
| "loss": 0.6422, | |
| "step": 299000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.413979434158699e-05, | |
| "loss": 0.6418, | |
| "step": 299500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.411331653581335e-05, | |
| "loss": 0.643, | |
| "step": 300000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.4086838730039704e-05, | |
| "loss": 0.6437, | |
| "step": 300500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.406036092426607e-05, | |
| "loss": 0.6392, | |
| "step": 301000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.403388311849242e-05, | |
| "loss": 0.635, | |
| "step": 301500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.400740531271877e-05, | |
| "loss": 0.6384, | |
| "step": 302000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.398092750694513e-05, | |
| "loss": 0.6444, | |
| "step": 302500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3954449701171484e-05, | |
| "loss": 0.6387, | |
| "step": 303000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.392797189539784e-05, | |
| "loss": 0.6447, | |
| "step": 303500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3901494089624194e-05, | |
| "loss": 0.6412, | |
| "step": 304000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.387501628385055e-05, | |
| "loss": 0.636, | |
| "step": 304500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3848538478076905e-05, | |
| "loss": 0.6405, | |
| "step": 305000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3822060672303264e-05, | |
| "loss": 0.6451, | |
| "step": 305500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3795582866529616e-05, | |
| "loss": 0.6381, | |
| "step": 306000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.3769105060755974e-05, | |
| "loss": 0.6407, | |
| "step": 306500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.3742627254982326e-05, | |
| "loss": 0.6357, | |
| "step": 307000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.3716149449208685e-05, | |
| "loss": 0.6451, | |
| "step": 307500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.368967164343504e-05, | |
| "loss": 0.6372, | |
| "step": 308000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.36631938376614e-05, | |
| "loss": 0.6352, | |
| "step": 308500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.3636716031887754e-05, | |
| "loss": 0.6383, | |
| "step": 309000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.361023822611411e-05, | |
| "loss": 0.6344, | |
| "step": 309500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.3583760420340465e-05, | |
| "loss": 0.6385, | |
| "step": 310000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.3557282614566824e-05, | |
| "loss": 0.6331, | |
| "step": 310500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.3530804808793176e-05, | |
| "loss": 0.6398, | |
| "step": 311000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.3504327003019534e-05, | |
| "loss": 0.6387, | |
| "step": 311500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.3477849197245886e-05, | |
| "loss": 0.6377, | |
| "step": 312000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.345137139147224e-05, | |
| "loss": 0.6363, | |
| "step": 312500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.34248935856986e-05, | |
| "loss": 0.6354, | |
| "step": 313000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.339841577992495e-05, | |
| "loss": 0.6374, | |
| "step": 313500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.337193797415131e-05, | |
| "loss": 0.6366, | |
| "step": 314000 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.334546016837766e-05, | |
| "loss": 0.6389, | |
| "step": 314500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.331898236260402e-05, | |
| "loss": 0.637, | |
| "step": 315000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.329250455683037e-05, | |
| "loss": 0.6339, | |
| "step": 315500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.326602675105673e-05, | |
| "loss": 0.6368, | |
| "step": 316000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.323954894528309e-05, | |
| "loss": 0.6369, | |
| "step": 316500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.3213071139509446e-05, | |
| "loss": 0.6386, | |
| "step": 317000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.31865933337358e-05, | |
| "loss": 0.6356, | |
| "step": 317500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.316011552796216e-05, | |
| "loss": 0.6323, | |
| "step": 318000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.313363772218851e-05, | |
| "loss": 0.6399, | |
| "step": 318500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.310715991641487e-05, | |
| "loss": 0.6367, | |
| "step": 319000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.308068211064122e-05, | |
| "loss": 0.6316, | |
| "step": 319500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.305420430486758e-05, | |
| "loss": 0.63, | |
| "step": 320000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.302772649909393e-05, | |
| "loss": 0.6303, | |
| "step": 320500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.300124869332029e-05, | |
| "loss": 0.6345, | |
| "step": 321000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.297477088754664e-05, | |
| "loss": 0.6303, | |
| "step": 321500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.294829308177299e-05, | |
| "loss": 0.6329, | |
| "step": 322000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.292181527599935e-05, | |
| "loss": 0.6289, | |
| "step": 322500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.2895337470225704e-05, | |
| "loss": 0.6342, | |
| "step": 323000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.286885966445206e-05, | |
| "loss": 0.6359, | |
| "step": 323500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.284238185867842e-05, | |
| "loss": 0.6355, | |
| "step": 324000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.281590405290478e-05, | |
| "loss": 0.6314, | |
| "step": 324500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.278942624713113e-05, | |
| "loss": 0.6349, | |
| "step": 325000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.276294844135749e-05, | |
| "loss": 0.6372, | |
| "step": 325500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.273647063558384e-05, | |
| "loss": 0.6297, | |
| "step": 326000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.27099928298102e-05, | |
| "loss": 0.633, | |
| "step": 326500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.268351502403655e-05, | |
| "loss": 0.6287, | |
| "step": 327000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.265703721826291e-05, | |
| "loss": 0.6299, | |
| "step": 327500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.2630559412489264e-05, | |
| "loss": 0.6317, | |
| "step": 328000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.260408160671562e-05, | |
| "loss": 0.6361, | |
| "step": 328500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.2577603800941974e-05, | |
| "loss": 0.6314, | |
| "step": 329000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.255112599516833e-05, | |
| "loss": 0.6316, | |
| "step": 329500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.2524648189394685e-05, | |
| "loss": 0.6294, | |
| "step": 330000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.2498170383621044e-05, | |
| "loss": 0.634, | |
| "step": 330500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2471692577847395e-05, | |
| "loss": 0.6298, | |
| "step": 331000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2445214772073754e-05, | |
| "loss": 0.6284, | |
| "step": 331500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.241873696630011e-05, | |
| "loss": 0.6324, | |
| "step": 332000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2392259160526465e-05, | |
| "loss": 0.6292, | |
| "step": 332500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2365781354752824e-05, | |
| "loss": 0.6262, | |
| "step": 333000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2339303548979175e-05, | |
| "loss": 0.6319, | |
| "step": 333500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.2312825743205534e-05, | |
| "loss": 0.6317, | |
| "step": 334000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.2286347937431886e-05, | |
| "loss": 0.6333, | |
| "step": 334500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.2259870131658245e-05, | |
| "loss": 0.6313, | |
| "step": 335000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.22333923258846e-05, | |
| "loss": 0.6293, | |
| "step": 335500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.2206914520110955e-05, | |
| "loss": 0.6315, | |
| "step": 336000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.218043671433731e-05, | |
| "loss": 0.6341, | |
| "step": 336500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.2153958908563666e-05, | |
| "loss": 0.6317, | |
| "step": 337000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.212748110279002e-05, | |
| "loss": 0.626, | |
| "step": 337500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.210100329701638e-05, | |
| "loss": 0.6295, | |
| "step": 338000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.207452549124273e-05, | |
| "loss": 0.6319, | |
| "step": 338500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.204804768546909e-05, | |
| "loss": 0.6238, | |
| "step": 339000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.2021569879695446e-05, | |
| "loss": 0.6304, | |
| "step": 339500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.19950920739218e-05, | |
| "loss": 0.6243, | |
| "step": 340000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.196861426814816e-05, | |
| "loss": 0.6253, | |
| "step": 340500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.194213646237451e-05, | |
| "loss": 0.6294, | |
| "step": 341000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.191565865660087e-05, | |
| "loss": 0.6222, | |
| "step": 341500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.188918085082722e-05, | |
| "loss": 0.6298, | |
| "step": 342000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.186270304505358e-05, | |
| "loss": 0.6301, | |
| "step": 342500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.183622523927993e-05, | |
| "loss": 0.6256, | |
| "step": 343000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.180974743350629e-05, | |
| "loss": 0.6282, | |
| "step": 343500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.178326962773264e-05, | |
| "loss": 0.6293, | |
| "step": 344000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.1756791821959e-05, | |
| "loss": 0.6321, | |
| "step": 344500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.173031401618535e-05, | |
| "loss": 0.6244, | |
| "step": 345000 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.170383621041171e-05, | |
| "loss": 0.6284, | |
| "step": 345500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.167735840463806e-05, | |
| "loss": 0.6268, | |
| "step": 346000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.165088059886443e-05, | |
| "loss": 0.6197, | |
| "step": 346500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.162440279309078e-05, | |
| "loss": 0.6298, | |
| "step": 347000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.159792498731713e-05, | |
| "loss": 0.6275, | |
| "step": 347500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.157144718154349e-05, | |
| "loss": 0.6229, | |
| "step": 348000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.154496937576984e-05, | |
| "loss": 0.6223, | |
| "step": 348500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.15184915699962e-05, | |
| "loss": 0.6255, | |
| "step": 349000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.149201376422255e-05, | |
| "loss": 0.6278, | |
| "step": 349500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.146553595844891e-05, | |
| "loss": 0.6231, | |
| "step": 350000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.143905815267526e-05, | |
| "loss": 0.6195, | |
| "step": 350500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.141258034690162e-05, | |
| "loss": 0.6228, | |
| "step": 351000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.1386102541127974e-05, | |
| "loss": 0.6243, | |
| "step": 351500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.135962473535433e-05, | |
| "loss": 0.6219, | |
| "step": 352000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.1333146929580684e-05, | |
| "loss": 0.6215, | |
| "step": 352500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.130666912380704e-05, | |
| "loss": 0.6262, | |
| "step": 353000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.1280191318033395e-05, | |
| "loss": 0.6233, | |
| "step": 353500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.1253713512259754e-05, | |
| "loss": 0.6235, | |
| "step": 354000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.122723570648611e-05, | |
| "loss": 0.6228, | |
| "step": 354500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.120075790071247e-05, | |
| "loss": 0.6194, | |
| "step": 355000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.117428009493882e-05, | |
| "loss": 0.6233, | |
| "step": 355500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.114780228916518e-05, | |
| "loss": 0.6213, | |
| "step": 356000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.1121324483391534e-05, | |
| "loss": 0.6274, | |
| "step": 356500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.109484667761789e-05, | |
| "loss": 0.6203, | |
| "step": 357000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.1068368871844244e-05, | |
| "loss": 0.6166, | |
| "step": 357500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.1041891066070596e-05, | |
| "loss": 0.6228, | |
| "step": 358000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.1015413260296955e-05, | |
| "loss": 0.6243, | |
| "step": 358500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.098893545452331e-05, | |
| "loss": 0.6234, | |
| "step": 359000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.0962457648749666e-05, | |
| "loss": 0.6251, | |
| "step": 359500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.093597984297602e-05, | |
| "loss": 0.6198, | |
| "step": 360000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.0909502037202376e-05, | |
| "loss": 0.6241, | |
| "step": 360500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.088302423142873e-05, | |
| "loss": 0.6218, | |
| "step": 361000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.085654642565509e-05, | |
| "loss": 0.6226, | |
| "step": 361500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.0830068619881446e-05, | |
| "loss": 0.6181, | |
| "step": 362000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.0803590814107804e-05, | |
| "loss": 0.6188, | |
| "step": 362500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.0777113008334156e-05, | |
| "loss": 0.6256, | |
| "step": 363000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.0750635202560515e-05, | |
| "loss": 0.6223, | |
| "step": 363500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.072415739678687e-05, | |
| "loss": 0.6228, | |
| "step": 364000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.0697679591013226e-05, | |
| "loss": 0.6206, | |
| "step": 364500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.067120178523958e-05, | |
| "loss": 0.6245, | |
| "step": 365000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.0644723979465936e-05, | |
| "loss": 0.6208, | |
| "step": 365500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.061824617369229e-05, | |
| "loss": 0.6219, | |
| "step": 366000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.059176836791865e-05, | |
| "loss": 0.6143, | |
| "step": 366500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.0565290562145e-05, | |
| "loss": 0.6223, | |
| "step": 367000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.053881275637136e-05, | |
| "loss": 0.6274, | |
| "step": 367500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.051233495059771e-05, | |
| "loss": 0.6226, | |
| "step": 368000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.0485857144824065e-05, | |
| "loss": 0.6221, | |
| "step": 368500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.045937933905042e-05, | |
| "loss": 0.6207, | |
| "step": 369000 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.0432901533276782e-05, | |
| "loss": 0.6215, | |
| "step": 369500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.0406423727503138e-05, | |
| "loss": 0.6201, | |
| "step": 370000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.037994592172949e-05, | |
| "loss": 0.6174, | |
| "step": 370500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0353468115955845e-05, | |
| "loss": 0.6154, | |
| "step": 371000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.03269903101822e-05, | |
| "loss": 0.6158, | |
| "step": 371500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0300512504408556e-05, | |
| "loss": 0.6176, | |
| "step": 372000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.027403469863491e-05, | |
| "loss": 0.6213, | |
| "step": 372500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0247556892861266e-05, | |
| "loss": 0.6169, | |
| "step": 373000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.022107908708762e-05, | |
| "loss": 0.615, | |
| "step": 373500 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.0194601281313977e-05, | |
| "loss": 0.6233, | |
| "step": 374000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0168123475540332e-05, | |
| "loss": 0.6206, | |
| "step": 374500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0141645669766688e-05, | |
| "loss": 0.6201, | |
| "step": 375000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0115167863993043e-05, | |
| "loss": 0.6168, | |
| "step": 375500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0088690058219398e-05, | |
| "loss": 0.6154, | |
| "step": 376000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0062212252445754e-05, | |
| "loss": 0.618, | |
| "step": 376500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.0035734446672116e-05, | |
| "loss": 0.6164, | |
| "step": 377000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.000925664089847e-05, | |
| "loss": 0.6211, | |
| "step": 377500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 2.9982778835124826e-05, | |
| "loss": 0.622, | |
| "step": 378000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.995630102935118e-05, | |
| "loss": 0.6186, | |
| "step": 378500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.9929823223577537e-05, | |
| "loss": 0.6156, | |
| "step": 379000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.9903345417803892e-05, | |
| "loss": 0.6119, | |
| "step": 379500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.9876867612030248e-05, | |
| "loss": 0.6159, | |
| "step": 380000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.9850389806256603e-05, | |
| "loss": 0.6159, | |
| "step": 380500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.9823912000482955e-05, | |
| "loss": 0.6215, | |
| "step": 381000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 2.979743419470931e-05, | |
| "loss": 0.6145, | |
| "step": 381500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.9770956388935665e-05, | |
| "loss": 0.6208, | |
| "step": 382000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.974447858316202e-05, | |
| "loss": 0.6141, | |
| "step": 382500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.9718000777388376e-05, | |
| "loss": 0.6142, | |
| "step": 383000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.969152297161473e-05, | |
| "loss": 0.6165, | |
| "step": 383500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.9665045165841087e-05, | |
| "loss": 0.6127, | |
| "step": 384000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.9638567360067442e-05, | |
| "loss": 0.615, | |
| "step": 384500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.9612089554293804e-05, | |
| "loss": 0.6173, | |
| "step": 385000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 2.958561174852016e-05, | |
| "loss": 0.6126, | |
| "step": 385500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.9559133942746515e-05, | |
| "loss": 0.6118, | |
| "step": 386000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.953265613697287e-05, | |
| "loss": 0.613, | |
| "step": 386500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.9506178331199225e-05, | |
| "loss": 0.6098, | |
| "step": 387000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.947970052542558e-05, | |
| "loss": 0.6177, | |
| "step": 387500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.9453222719651936e-05, | |
| "loss": 0.6104, | |
| "step": 388000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.942674491387829e-05, | |
| "loss": 0.6086, | |
| "step": 388500 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.9400267108104647e-05, | |
| "loss": 0.6113, | |
| "step": 389000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.9373789302331002e-05, | |
| "loss": 0.6173, | |
| "step": 389500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9347311496557357e-05, | |
| "loss": 0.6148, | |
| "step": 390000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9320833690783713e-05, | |
| "loss": 0.6118, | |
| "step": 390500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9294355885010065e-05, | |
| "loss": 0.6088, | |
| "step": 391000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.926787807923642e-05, | |
| "loss": 0.6099, | |
| "step": 391500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9241400273462775e-05, | |
| "loss": 0.6069, | |
| "step": 392000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9214922467689137e-05, | |
| "loss": 0.6129, | |
| "step": 392500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9188444661915493e-05, | |
| "loss": 0.6115, | |
| "step": 393000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.9161966856141848e-05, | |
| "loss": 0.6156, | |
| "step": 393500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.9135489050368203e-05, | |
| "loss": 0.6095, | |
| "step": 394000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.910901124459456e-05, | |
| "loss": 0.6134, | |
| "step": 394500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.9082533438820914e-05, | |
| "loss": 0.6091, | |
| "step": 395000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.905605563304727e-05, | |
| "loss": 0.6101, | |
| "step": 395500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.9029577827273625e-05, | |
| "loss": 0.6125, | |
| "step": 396000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.900310002149998e-05, | |
| "loss": 0.6049, | |
| "step": 396500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.8976622215726335e-05, | |
| "loss": 0.6145, | |
| "step": 397000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.895014440995269e-05, | |
| "loss": 0.6185, | |
| "step": 397500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.8923666604179046e-05, | |
| "loss": 0.6135, | |
| "step": 398000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.88971887984054e-05, | |
| "loss": 0.6101, | |
| "step": 398500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.8870710992631757e-05, | |
| "loss": 0.613, | |
| "step": 399000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.8844233186858112e-05, | |
| "loss": 0.6118, | |
| "step": 399500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.881775538108447e-05, | |
| "loss": 0.6111, | |
| "step": 400000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.8791277575310826e-05, | |
| "loss": 0.6134, | |
| "step": 400500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.876479976953718e-05, | |
| "loss": 0.614, | |
| "step": 401000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.8738321963763537e-05, | |
| "loss": 0.6042, | |
| "step": 401500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8711844157989892e-05, | |
| "loss": 0.6085, | |
| "step": 402000 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8685366352216247e-05, | |
| "loss": 0.6092, | |
| "step": 402500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8658888546442603e-05, | |
| "loss": 0.6076, | |
| "step": 403000 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8632410740668958e-05, | |
| "loss": 0.6074, | |
| "step": 403500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8605932934895313e-05, | |
| "loss": 0.6084, | |
| "step": 404000 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.857945512912167e-05, | |
| "loss": 0.6071, | |
| "step": 404500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.8552977323348024e-05, | |
| "loss": 0.609, | |
| "step": 405000 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.852649951757438e-05, | |
| "loss": 0.6099, | |
| "step": 405500 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.8500021711800735e-05, | |
| "loss": 0.6101, | |
| "step": 406000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.847354390602709e-05, | |
| "loss": 0.6005, | |
| "step": 406500 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.8447066100253445e-05, | |
| "loss": 0.6058, | |
| "step": 407000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.8420588294479804e-05, | |
| "loss": 0.6116, | |
| "step": 407500 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.839411048870616e-05, | |
| "loss": 0.6069, | |
| "step": 408000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.8367632682932515e-05, | |
| "loss": 0.6041, | |
| "step": 408500 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.834115487715887e-05, | |
| "loss": 0.6058, | |
| "step": 409000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 2.8314677071385225e-05, | |
| "loss": 0.6083, | |
| "step": 409500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.828819926561158e-05, | |
| "loss": 0.6033, | |
| "step": 410000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8261721459837936e-05, | |
| "loss": 0.6105, | |
| "step": 410500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.823524365406429e-05, | |
| "loss": 0.6061, | |
| "step": 411000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8208765848290646e-05, | |
| "loss": 0.605, | |
| "step": 411500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8182288042517002e-05, | |
| "loss": 0.61, | |
| "step": 412000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8155810236743357e-05, | |
| "loss": 0.602, | |
| "step": 412500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8129332430969712e-05, | |
| "loss": 0.6048, | |
| "step": 413000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.8102854625196068e-05, | |
| "loss": 0.6079, | |
| "step": 413500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.8076376819422423e-05, | |
| "loss": 0.6105, | |
| "step": 414000 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.804989901364878e-05, | |
| "loss": 0.6081, | |
| "step": 414500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.8023421207875134e-05, | |
| "loss": 0.6075, | |
| "step": 415000 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.7996943402101496e-05, | |
| "loss": 0.609, | |
| "step": 415500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.797046559632785e-05, | |
| "loss": 0.6046, | |
| "step": 416000 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.7943987790554203e-05, | |
| "loss": 0.6072, | |
| "step": 416500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.791750998478056e-05, | |
| "loss": 0.6071, | |
| "step": 417000 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.7891032179006914e-05, | |
| "loss": 0.6039, | |
| "step": 417500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.786455437323327e-05, | |
| "loss": 0.6036, | |
| "step": 418000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.7838076567459624e-05, | |
| "loss": 0.6051, | |
| "step": 418500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.781159876168598e-05, | |
| "loss": 0.6015, | |
| "step": 419000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.7785120955912335e-05, | |
| "loss": 0.6073, | |
| "step": 419500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.775864315013869e-05, | |
| "loss": 0.6024, | |
| "step": 420000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.7732165344365046e-05, | |
| "loss": 0.6108, | |
| "step": 420500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.77056875385914e-05, | |
| "loss": 0.6054, | |
| "step": 421000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.7679209732817756e-05, | |
| "loss": 0.6016, | |
| "step": 421500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.765273192704411e-05, | |
| "loss": 0.6036, | |
| "step": 422000 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.7626254121270467e-05, | |
| "loss": 0.604, | |
| "step": 422500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.759977631549683e-05, | |
| "loss": 0.6023, | |
| "step": 423000 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.7573298509723184e-05, | |
| "loss": 0.6107, | |
| "step": 423500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.754682070394954e-05, | |
| "loss": 0.6052, | |
| "step": 424000 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.7520342898175895e-05, | |
| "loss": 0.6061, | |
| "step": 424500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.749386509240225e-05, | |
| "loss": 0.6025, | |
| "step": 425000 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.7467387286628606e-05, | |
| "loss": 0.604, | |
| "step": 425500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.744090948085496e-05, | |
| "loss": 0.6043, | |
| "step": 426000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.7414431675081313e-05, | |
| "loss": 0.6043, | |
| "step": 426500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.7387953869307668e-05, | |
| "loss": 0.6052, | |
| "step": 427000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.7361476063534024e-05, | |
| "loss": 0.6025, | |
| "step": 427500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.733499825776038e-05, | |
| "loss": 0.6032, | |
| "step": 428000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.7308520451986734e-05, | |
| "loss": 0.6035, | |
| "step": 428500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.728204264621309e-05, | |
| "loss": 0.6035, | |
| "step": 429000 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7255564840439445e-05, | |
| "loss": 0.6036, | |
| "step": 429500 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.72290870346658e-05, | |
| "loss": 0.6018, | |
| "step": 430000 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7202609228892162e-05, | |
| "loss": 0.6001, | |
| "step": 430500 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7176131423118518e-05, | |
| "loss": 0.6036, | |
| "step": 431000 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7149653617344873e-05, | |
| "loss": 0.5989, | |
| "step": 431500 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7123175811571228e-05, | |
| "loss": 0.5996, | |
| "step": 432000 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.7096698005797584e-05, | |
| "loss": 0.5998, | |
| "step": 432500 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.707022020002394e-05, | |
| "loss": 0.6035, | |
| "step": 433000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.7043742394250294e-05, | |
| "loss": 0.5994, | |
| "step": 433500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.701726458847665e-05, | |
| "loss": 0.6047, | |
| "step": 434000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.6990786782703005e-05, | |
| "loss": 0.602, | |
| "step": 434500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.696430897692936e-05, | |
| "loss": 0.6006, | |
| "step": 435000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.6937831171155715e-05, | |
| "loss": 0.6006, | |
| "step": 435500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.691135336538207e-05, | |
| "loss": 0.6006, | |
| "step": 436000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.6884875559608426e-05, | |
| "loss": 0.5997, | |
| "step": 436500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.6858397753834778e-05, | |
| "loss": 0.6002, | |
| "step": 437000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6831919948061133e-05, | |
| "loss": 0.5982, | |
| "step": 437500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6805442142287495e-05, | |
| "loss": 0.5986, | |
| "step": 438000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.677896433651385e-05, | |
| "loss": 0.603, | |
| "step": 438500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6752486530740206e-05, | |
| "loss": 0.6032, | |
| "step": 439000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.672600872496656e-05, | |
| "loss": 0.6048, | |
| "step": 439500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6699530919192917e-05, | |
| "loss": 0.5979, | |
| "step": 440000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6673053113419272e-05, | |
| "loss": 0.6085, | |
| "step": 440500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6646575307645627e-05, | |
| "loss": 0.5976, | |
| "step": 441000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.6620097501871983e-05, | |
| "loss": 0.5967, | |
| "step": 441500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.6593619696098338e-05, | |
| "loss": 0.597, | |
| "step": 442000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.6567141890324693e-05, | |
| "loss": 0.5997, | |
| "step": 442500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.654066408455105e-05, | |
| "loss": 0.6015, | |
| "step": 443000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.6514186278777404e-05, | |
| "loss": 0.5997, | |
| "step": 443500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.648770847300376e-05, | |
| "loss": 0.5977, | |
| "step": 444000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.6461230667230115e-05, | |
| "loss": 0.5953, | |
| "step": 444500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 2.643475286145647e-05, | |
| "loss": 0.6038, | |
| "step": 445000 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.640827505568283e-05, | |
| "loss": 0.602, | |
| "step": 445500 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.6381797249909184e-05, | |
| "loss": 0.5974, | |
| "step": 446000 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.635531944413554e-05, | |
| "loss": 0.593, | |
| "step": 446500 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.6328841638361895e-05, | |
| "loss": 0.6019, | |
| "step": 447000 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.630236383258825e-05, | |
| "loss": 0.6006, | |
| "step": 447500 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.6275886026814605e-05, | |
| "loss": 0.5966, | |
| "step": 448000 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.624940822104096e-05, | |
| "loss": 0.5981, | |
| "step": 448500 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 2.6222930415267316e-05, | |
| "loss": 0.5962, | |
| "step": 449000 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.619645260949367e-05, | |
| "loss": 0.6012, | |
| "step": 449500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6169974803720027e-05, | |
| "loss": 0.5944, | |
| "step": 450000 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6143496997946382e-05, | |
| "loss": 0.5964, | |
| "step": 450500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6117019192172737e-05, | |
| "loss": 0.5973, | |
| "step": 451000 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6090541386399093e-05, | |
| "loss": 0.597, | |
| "step": 451500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6064063580625448e-05, | |
| "loss": 0.5939, | |
| "step": 452000 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.6037585774851803e-05, | |
| "loss": 0.5934, | |
| "step": 452500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 2.601110796907816e-05, | |
| "loss": 0.5988, | |
| "step": 453000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5984630163304517e-05, | |
| "loss": 0.5912, | |
| "step": 453500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5958152357530873e-05, | |
| "loss": 0.5939, | |
| "step": 454000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5931674551757228e-05, | |
| "loss": 0.6025, | |
| "step": 454500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5905196745983583e-05, | |
| "loss": 0.5967, | |
| "step": 455000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.587871894020994e-05, | |
| "loss": 0.5968, | |
| "step": 455500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5852241134436294e-05, | |
| "loss": 0.5958, | |
| "step": 456000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.582576332866265e-05, | |
| "loss": 0.5946, | |
| "step": 456500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 2.5799285522889005e-05, | |
| "loss": 0.5971, | |
| "step": 457000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.577280771711536e-05, | |
| "loss": 0.5957, | |
| "step": 457500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.5746329911341715e-05, | |
| "loss": 0.5961, | |
| "step": 458000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.571985210556807e-05, | |
| "loss": 0.5969, | |
| "step": 458500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.5693374299794426e-05, | |
| "loss": 0.5987, | |
| "step": 459000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.566689649402078e-05, | |
| "loss": 0.5959, | |
| "step": 459500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.5640418688247136e-05, | |
| "loss": 0.6012, | |
| "step": 460000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.5613940882473492e-05, | |
| "loss": 0.5954, | |
| "step": 460500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 2.5587463076699854e-05, | |
| "loss": 0.5913, | |
| "step": 461000 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.556098527092621e-05, | |
| "loss": 0.595, | |
| "step": 461500 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5534507465152565e-05, | |
| "loss": 0.5886, | |
| "step": 462000 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5508029659378916e-05, | |
| "loss": 0.5908, | |
| "step": 462500 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5481551853605272e-05, | |
| "loss": 0.5915, | |
| "step": 463000 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5455074047831627e-05, | |
| "loss": 0.594, | |
| "step": 463500 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5428596242057982e-05, | |
| "loss": 0.5894, | |
| "step": 464000 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5402118436284338e-05, | |
| "loss": 0.5969, | |
| "step": 464500 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 2.5375640630510693e-05, | |
| "loss": 0.5965, | |
| "step": 465000 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.534916282473705e-05, | |
| "loss": 0.5921, | |
| "step": 465500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.5322685018963404e-05, | |
| "loss": 0.5959, | |
| "step": 466000 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.529620721318976e-05, | |
| "loss": 0.5926, | |
| "step": 466500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.5269729407416114e-05, | |
| "loss": 0.593, | |
| "step": 467000 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.524325160164247e-05, | |
| "loss": 0.5947, | |
| "step": 467500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.5216773795868825e-05, | |
| "loss": 0.5991, | |
| "step": 468000 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.5190295990095187e-05, | |
| "loss": 0.5909, | |
| "step": 468500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.5163818184321542e-05, | |
| "loss": 0.5925, | |
| "step": 469000 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.5137340378547898e-05, | |
| "loss": 0.5974, | |
| "step": 469500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.5110862572774253e-05, | |
| "loss": 0.5944, | |
| "step": 470000 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.508438476700061e-05, | |
| "loss": 0.5948, | |
| "step": 470500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.5057906961226964e-05, | |
| "loss": 0.5934, | |
| "step": 471000 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.503142915545332e-05, | |
| "loss": 0.5932, | |
| "step": 471500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.5004951349679674e-05, | |
| "loss": 0.5952, | |
| "step": 472000 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.4978473543906026e-05, | |
| "loss": 0.5927, | |
| "step": 472500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 2.495199573813238e-05, | |
| "loss": 0.5927, | |
| "step": 473000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.4925517932358737e-05, | |
| "loss": 0.5893, | |
| "step": 473500 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.4899040126585092e-05, | |
| "loss": 0.5922, | |
| "step": 474000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.487256232081145e-05, | |
| "loss": 0.5896, | |
| "step": 474500 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.4846084515037806e-05, | |
| "loss": 0.5914, | |
| "step": 475000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.481960670926416e-05, | |
| "loss": 0.5909, | |
| "step": 475500 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.4793128903490517e-05, | |
| "loss": 0.5915, | |
| "step": 476000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 2.4766651097716872e-05, | |
| "loss": 0.599, | |
| "step": 476500 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4740173291943228e-05, | |
| "loss": 0.5945, | |
| "step": 477000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4713695486169583e-05, | |
| "loss": 0.5887, | |
| "step": 477500 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.468721768039594e-05, | |
| "loss": 0.5881, | |
| "step": 478000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4660739874622297e-05, | |
| "loss": 0.5897, | |
| "step": 478500 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4634262068848652e-05, | |
| "loss": 0.5913, | |
| "step": 479000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4607784263075008e-05, | |
| "loss": 0.5909, | |
| "step": 479500 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4581306457301363e-05, | |
| "loss": 0.589, | |
| "step": 480000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 2.4554828651527718e-05, | |
| "loss": 0.5906, | |
| "step": 480500 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.4528350845754074e-05, | |
| "loss": 0.5884, | |
| "step": 481000 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.450187303998043e-05, | |
| "loss": 0.5906, | |
| "step": 481500 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.4475395234206784e-05, | |
| "loss": 0.5922, | |
| "step": 482000 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.444891742843314e-05, | |
| "loss": 0.5936, | |
| "step": 482500 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.4422439622659495e-05, | |
| "loss": 0.59, | |
| "step": 483000 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.439596181688585e-05, | |
| "loss": 0.5907, | |
| "step": 483500 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.4369484011112206e-05, | |
| "loss": 0.5951, | |
| "step": 484000 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 2.434300620533856e-05, | |
| "loss": 0.5893, | |
| "step": 484500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.4316528399564916e-05, | |
| "loss": 0.5909, | |
| "step": 485000 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.4290050593791275e-05, | |
| "loss": 0.594, | |
| "step": 485500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.426357278801763e-05, | |
| "loss": 0.5944, | |
| "step": 486000 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.4237094982243986e-05, | |
| "loss": 0.5903, | |
| "step": 486500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.421061717647034e-05, | |
| "loss": 0.5907, | |
| "step": 487000 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.4184139370696696e-05, | |
| "loss": 0.5931, | |
| "step": 487500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.415766156492305e-05, | |
| "loss": 0.588, | |
| "step": 488000 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 2.4131183759149407e-05, | |
| "loss": 0.5895, | |
| "step": 488500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.4104705953375762e-05, | |
| "loss": 0.5857, | |
| "step": 489000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.407822814760212e-05, | |
| "loss": 0.5864, | |
| "step": 489500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.4051750341828476e-05, | |
| "loss": 0.5892, | |
| "step": 490000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.4025272536054828e-05, | |
| "loss": 0.5883, | |
| "step": 490500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.3998794730281183e-05, | |
| "loss": 0.5883, | |
| "step": 491000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.397231692450754e-05, | |
| "loss": 0.5869, | |
| "step": 491500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.3945839118733894e-05, | |
| "loss": 0.5862, | |
| "step": 492000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 2.391936131296025e-05, | |
| "loss": 0.5851, | |
| "step": 492500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.3892883507186605e-05, | |
| "loss": 0.5905, | |
| "step": 493000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.3866405701412963e-05, | |
| "loss": 0.5869, | |
| "step": 493500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.383992789563932e-05, | |
| "loss": 0.5931, | |
| "step": 494000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.3813450089865674e-05, | |
| "loss": 0.5871, | |
| "step": 494500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.378697228409203e-05, | |
| "loss": 0.5949, | |
| "step": 495000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.3760494478318385e-05, | |
| "loss": 0.5875, | |
| "step": 495500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.373401667254474e-05, | |
| "loss": 0.5892, | |
| "step": 496000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.3707538866771095e-05, | |
| "loss": 0.5849, | |
| "step": 496500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.3681061060997454e-05, | |
| "loss": 0.5862, | |
| "step": 497000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.365458325522381e-05, | |
| "loss": 0.588, | |
| "step": 497500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.3628105449450165e-05, | |
| "loss": 0.5876, | |
| "step": 498000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.360162764367652e-05, | |
| "loss": 0.5886, | |
| "step": 498500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.3575149837902875e-05, | |
| "loss": 0.5902, | |
| "step": 499000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.354867203212923e-05, | |
| "loss": 0.5871, | |
| "step": 499500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.3522194226355586e-05, | |
| "loss": 0.5889, | |
| "step": 500000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.349571642058194e-05, | |
| "loss": 0.5859, | |
| "step": 500500 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3469238614808297e-05, | |
| "loss": 0.5826, | |
| "step": 501000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3442760809034652e-05, | |
| "loss": 0.5859, | |
| "step": 501500 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3416283003261007e-05, | |
| "loss": 0.5891, | |
| "step": 502000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3389805197487363e-05, | |
| "loss": 0.5839, | |
| "step": 502500 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3363327391713718e-05, | |
| "loss": 0.5904, | |
| "step": 503000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3336849585940073e-05, | |
| "loss": 0.5866, | |
| "step": 503500 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.331037178016643e-05, | |
| "loss": 0.578, | |
| "step": 504000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3283893974392787e-05, | |
| "loss": 0.5838, | |
| "step": 504500 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.3257416168619143e-05, | |
| "loss": 0.5816, | |
| "step": 505000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.3230938362845498e-05, | |
| "loss": 0.5837, | |
| "step": 505500 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.3204460557071853e-05, | |
| "loss": 0.5833, | |
| "step": 506000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.317798275129821e-05, | |
| "loss": 0.5828, | |
| "step": 506500 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.3151504945524564e-05, | |
| "loss": 0.5856, | |
| "step": 507000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.312502713975092e-05, | |
| "loss": 0.586, | |
| "step": 507500 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.3098549333977275e-05, | |
| "loss": 0.5883, | |
| "step": 508000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.307207152820363e-05, | |
| "loss": 0.5856, | |
| "step": 508500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.3045593722429985e-05, | |
| "loss": 0.5889, | |
| "step": 509000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.301911591665634e-05, | |
| "loss": 0.5835, | |
| "step": 509500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.2992638110882696e-05, | |
| "loss": 0.5876, | |
| "step": 510000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.296616030510905e-05, | |
| "loss": 0.5803, | |
| "step": 510500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.2939682499335406e-05, | |
| "loss": 0.5893, | |
| "step": 511000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.2913204693561762e-05, | |
| "loss": 0.578, | |
| "step": 511500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.2886726887788117e-05, | |
| "loss": 0.585, | |
| "step": 512000 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.2860249082014476e-05, | |
| "loss": 0.5867, | |
| "step": 512500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.283377127624083e-05, | |
| "loss": 0.5836, | |
| "step": 513000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2807293470467186e-05, | |
| "loss": 0.5853, | |
| "step": 513500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2780815664693542e-05, | |
| "loss": 0.5846, | |
| "step": 514000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2754337858919897e-05, | |
| "loss": 0.5816, | |
| "step": 514500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2727860053146252e-05, | |
| "loss": 0.5798, | |
| "step": 515000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2701382247372608e-05, | |
| "loss": 0.5871, | |
| "step": 515500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2674904441598966e-05, | |
| "loss": 0.5841, | |
| "step": 516000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.2648426635825322e-05, | |
| "loss": 0.5825, | |
| "step": 516500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.2621948830051677e-05, | |
| "loss": 0.5799, | |
| "step": 517000 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.2595471024278032e-05, | |
| "loss": 0.5875, | |
| "step": 517500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.2568993218504388e-05, | |
| "loss": 0.5861, | |
| "step": 518000 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.254251541273074e-05, | |
| "loss": 0.5825, | |
| "step": 518500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.2516037606957095e-05, | |
| "loss": 0.5794, | |
| "step": 519000 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.248955980118345e-05, | |
| "loss": 0.5805, | |
| "step": 519500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.246308199540981e-05, | |
| "loss": 0.5836, | |
| "step": 520000 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.2436604189636164e-05, | |
| "loss": 0.5897, | |
| "step": 520500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.241012638386252e-05, | |
| "loss": 0.581, | |
| "step": 521000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.2383648578088875e-05, | |
| "loss": 0.579, | |
| "step": 521500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.235717077231523e-05, | |
| "loss": 0.5821, | |
| "step": 522000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.2330692966541586e-05, | |
| "loss": 0.581, | |
| "step": 522500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.230421516076794e-05, | |
| "loss": 0.5831, | |
| "step": 523000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.2277737354994296e-05, | |
| "loss": 0.5807, | |
| "step": 523500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.2251259549220655e-05, | |
| "loss": 0.5849, | |
| "step": 524000 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.222478174344701e-05, | |
| "loss": 0.5836, | |
| "step": 524500 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.2198303937673366e-05, | |
| "loss": 0.5847, | |
| "step": 525000 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.217182613189972e-05, | |
| "loss": 0.585, | |
| "step": 525500 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.2145348326126076e-05, | |
| "loss": 0.583, | |
| "step": 526000 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.211887052035243e-05, | |
| "loss": 0.5813, | |
| "step": 526500 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.2092392714578787e-05, | |
| "loss": 0.5837, | |
| "step": 527000 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.2065914908805142e-05, | |
| "loss": 0.5813, | |
| "step": 527500 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.2039437103031498e-05, | |
| "loss": 0.5787, | |
| "step": 528000 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.2012959297257853e-05, | |
| "loss": 0.5816, | |
| "step": 528500 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.1986481491484208e-05, | |
| "loss": 0.5795, | |
| "step": 529000 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.1960003685710564e-05, | |
| "loss": 0.5805, | |
| "step": 529500 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.193352587993692e-05, | |
| "loss": 0.5786, | |
| "step": 530000 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.1907048074163274e-05, | |
| "loss": 0.5814, | |
| "step": 530500 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.188057026838963e-05, | |
| "loss": 0.5865, | |
| "step": 531000 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.1854092462615988e-05, | |
| "loss": 0.5761, | |
| "step": 531500 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.1827614656842344e-05, | |
| "loss": 0.5805, | |
| "step": 532000 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.18011368510687e-05, | |
| "loss": 0.5801, | |
| "step": 532500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.1774659045295054e-05, | |
| "loss": 0.5805, | |
| "step": 533000 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.174818123952141e-05, | |
| "loss": 0.5795, | |
| "step": 533500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.1721703433747765e-05, | |
| "loss": 0.5815, | |
| "step": 534000 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.169522562797412e-05, | |
| "loss": 0.583, | |
| "step": 534500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.166874782220048e-05, | |
| "loss": 0.5818, | |
| "step": 535000 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.1642270016426834e-05, | |
| "loss": 0.5769, | |
| "step": 535500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.161579221065319e-05, | |
| "loss": 0.5784, | |
| "step": 536000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.158931440487954e-05, | |
| "loss": 0.581, | |
| "step": 536500 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1562836599105897e-05, | |
| "loss": 0.5805, | |
| "step": 537000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1536358793332252e-05, | |
| "loss": 0.5778, | |
| "step": 537500 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1509880987558607e-05, | |
| "loss": 0.5819, | |
| "step": 538000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1483403181784963e-05, | |
| "loss": 0.5794, | |
| "step": 538500 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.145692537601132e-05, | |
| "loss": 0.582, | |
| "step": 539000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1430447570237677e-05, | |
| "loss": 0.5819, | |
| "step": 539500 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.1403969764464032e-05, | |
| "loss": 0.5817, | |
| "step": 540000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1377491958690387e-05, | |
| "loss": 0.578, | |
| "step": 540500 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1351014152916743e-05, | |
| "loss": 0.5801, | |
| "step": 541000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1324536347143098e-05, | |
| "loss": 0.5753, | |
| "step": 541500 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1298058541369453e-05, | |
| "loss": 0.5801, | |
| "step": 542000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.127158073559581e-05, | |
| "loss": 0.5788, | |
| "step": 542500 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1245102929822167e-05, | |
| "loss": 0.5759, | |
| "step": 543000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1218625124048523e-05, | |
| "loss": 0.5772, | |
| "step": 543500 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.1192147318274878e-05, | |
| "loss": 0.5729, | |
| "step": 544000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.1165669512501233e-05, | |
| "loss": 0.5782, | |
| "step": 544500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.113919170672759e-05, | |
| "loss": 0.5785, | |
| "step": 545000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.1112713900953944e-05, | |
| "loss": 0.5808, | |
| "step": 545500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.10862360951803e-05, | |
| "loss": 0.5786, | |
| "step": 546000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.1059758289406655e-05, | |
| "loss": 0.5755, | |
| "step": 546500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.103328048363301e-05, | |
| "loss": 0.5789, | |
| "step": 547000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.1006802677859365e-05, | |
| "loss": 0.5748, | |
| "step": 547500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.098032487208572e-05, | |
| "loss": 0.5809, | |
| "step": 548000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.0953847066312076e-05, | |
| "loss": 0.5816, | |
| "step": 548500 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.092736926053843e-05, | |
| "loss": 0.5775, | |
| "step": 549000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.0900891454764787e-05, | |
| "loss": 0.5807, | |
| "step": 549500 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.0874413648991142e-05, | |
| "loss": 0.5781, | |
| "step": 550000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.08479358432175e-05, | |
| "loss": 0.5793, | |
| "step": 550500 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.0821458037443856e-05, | |
| "loss": 0.5775, | |
| "step": 551000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.079498023167021e-05, | |
| "loss": 0.5786, | |
| "step": 551500 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.0768502425896567e-05, | |
| "loss": 0.5781, | |
| "step": 552000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.0742024620122922e-05, | |
| "loss": 0.5768, | |
| "step": 552500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.0715546814349277e-05, | |
| "loss": 0.5747, | |
| "step": 553000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.0689069008575633e-05, | |
| "loss": 0.5767, | |
| "step": 553500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.066259120280199e-05, | |
| "loss": 0.5761, | |
| "step": 554000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.0636113397028343e-05, | |
| "loss": 0.5774, | |
| "step": 554500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.06096355912547e-05, | |
| "loss": 0.5756, | |
| "step": 555000 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.0583157785481054e-05, | |
| "loss": 0.5727, | |
| "step": 555500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.055667997970741e-05, | |
| "loss": 0.5704, | |
| "step": 556000 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.0530202173933765e-05, | |
| "loss": 0.5749, | |
| "step": 556500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.050372436816012e-05, | |
| "loss": 0.5735, | |
| "step": 557000 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.0477246562386475e-05, | |
| "loss": 0.5775, | |
| "step": 557500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.0450768756612834e-05, | |
| "loss": 0.5769, | |
| "step": 558000 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.042429095083919e-05, | |
| "loss": 0.5775, | |
| "step": 558500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.0397813145065545e-05, | |
| "loss": 0.573, | |
| "step": 559000 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.03713353392919e-05, | |
| "loss": 0.5752, | |
| "step": 559500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.0344857533518255e-05, | |
| "loss": 0.5757, | |
| "step": 560000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.031837972774461e-05, | |
| "loss": 0.5747, | |
| "step": 560500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.0291901921970966e-05, | |
| "loss": 0.5785, | |
| "step": 561000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.026542411619732e-05, | |
| "loss": 0.5763, | |
| "step": 561500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.023894631042368e-05, | |
| "loss": 0.575, | |
| "step": 562000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.0212468504650035e-05, | |
| "loss": 0.5777, | |
| "step": 562500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.018599069887639e-05, | |
| "loss": 0.5763, | |
| "step": 563000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.0159512893102746e-05, | |
| "loss": 0.5783, | |
| "step": 563500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.01330350873291e-05, | |
| "loss": 0.5718, | |
| "step": 564000 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.0106557281555453e-05, | |
| "loss": 0.5712, | |
| "step": 564500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.008007947578181e-05, | |
| "loss": 0.5775, | |
| "step": 565000 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.0053601670008167e-05, | |
| "loss": 0.5743, | |
| "step": 565500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.0027123864234522e-05, | |
| "loss": 0.5714, | |
| "step": 566000 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.0000646058460878e-05, | |
| "loss": 0.5717, | |
| "step": 566500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.9974168252687233e-05, | |
| "loss": 0.5769, | |
| "step": 567000 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.994769044691359e-05, | |
| "loss": 0.5683, | |
| "step": 567500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.9921212641139944e-05, | |
| "loss": 0.5731, | |
| "step": 568000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.98947348353663e-05, | |
| "loss": 0.574, | |
| "step": 568500 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.9868257029592654e-05, | |
| "loss": 0.5785, | |
| "step": 569000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.9841779223819013e-05, | |
| "loss": 0.5732, | |
| "step": 569500 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.981530141804537e-05, | |
| "loss": 0.5727, | |
| "step": 570000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.9788823612271724e-05, | |
| "loss": 0.5695, | |
| "step": 570500 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.976234580649808e-05, | |
| "loss": 0.5782, | |
| "step": 571000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.9735868000724434e-05, | |
| "loss": 0.5703, | |
| "step": 571500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.970939019495079e-05, | |
| "loss": 0.5739, | |
| "step": 572000 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.9682912389177145e-05, | |
| "loss": 0.5737, | |
| "step": 572500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.96564345834035e-05, | |
| "loss": 0.5714, | |
| "step": 573000 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.9629956777629856e-05, | |
| "loss": 0.57, | |
| "step": 573500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.960347897185621e-05, | |
| "loss": 0.5755, | |
| "step": 574000 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.9577001166082566e-05, | |
| "loss": 0.5706, | |
| "step": 574500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.955052336030892e-05, | |
| "loss": 0.57, | |
| "step": 575000 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.9524045554535277e-05, | |
| "loss": 0.572, | |
| "step": 575500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9497567748761632e-05, | |
| "loss": 0.5716, | |
| "step": 576000 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9471089942987988e-05, | |
| "loss": 0.5717, | |
| "step": 576500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9444612137214346e-05, | |
| "loss": 0.573, | |
| "step": 577000 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.94181343314407e-05, | |
| "loss": 0.5715, | |
| "step": 577500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9391656525667057e-05, | |
| "loss": 0.5766, | |
| "step": 578000 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9365178719893412e-05, | |
| "loss": 0.5711, | |
| "step": 578500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9338700914119768e-05, | |
| "loss": 0.5716, | |
| "step": 579000 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.9312223108346123e-05, | |
| "loss": 0.5777, | |
| "step": 579500 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9285745302572478e-05, | |
| "loss": 0.5709, | |
| "step": 580000 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9259267496798834e-05, | |
| "loss": 0.5714, | |
| "step": 580500 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9232789691025192e-05, | |
| "loss": 0.5722, | |
| "step": 581000 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9206311885251548e-05, | |
| "loss": 0.569, | |
| "step": 581500 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9179834079477903e-05, | |
| "loss": 0.5739, | |
| "step": 582000 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9153356273704255e-05, | |
| "loss": 0.5739, | |
| "step": 582500 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.912687846793061e-05, | |
| "loss": 0.5661, | |
| "step": 583000 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.9100400662156966e-05, | |
| "loss": 0.5715, | |
| "step": 583500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.907392285638332e-05, | |
| "loss": 0.5682, | |
| "step": 584000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.904744505060968e-05, | |
| "loss": 0.5717, | |
| "step": 584500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.9020967244836035e-05, | |
| "loss": 0.5687, | |
| "step": 585000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.899448943906239e-05, | |
| "loss": 0.5697, | |
| "step": 585500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.8968011633288746e-05, | |
| "loss": 0.5679, | |
| "step": 586000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.89415338275151e-05, | |
| "loss": 0.5668, | |
| "step": 586500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.8915056021741456e-05, | |
| "loss": 0.5682, | |
| "step": 587000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.888857821596781e-05, | |
| "loss": 0.573, | |
| "step": 587500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8862100410194167e-05, | |
| "loss": 0.5703, | |
| "step": 588000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8835622604420526e-05, | |
| "loss": 0.5722, | |
| "step": 588500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.880914479864688e-05, | |
| "loss": 0.5706, | |
| "step": 589000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8782666992873236e-05, | |
| "loss": 0.5679, | |
| "step": 589500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.875618918709959e-05, | |
| "loss": 0.5719, | |
| "step": 590000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8729711381325947e-05, | |
| "loss": 0.5686, | |
| "step": 590500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8703233575552302e-05, | |
| "loss": 0.5746, | |
| "step": 591000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.8676755769778658e-05, | |
| "loss": 0.5719, | |
| "step": 591500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.8650277964005013e-05, | |
| "loss": 0.5677, | |
| "step": 592000 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.8623800158231368e-05, | |
| "loss": 0.5696, | |
| "step": 592500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.8597322352457723e-05, | |
| "loss": 0.5689, | |
| "step": 593000 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.857084454668408e-05, | |
| "loss": 0.5667, | |
| "step": 593500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.8544366740910434e-05, | |
| "loss": 0.5659, | |
| "step": 594000 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.851788893513679e-05, | |
| "loss": 0.569, | |
| "step": 594500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.8491411129363145e-05, | |
| "loss": 0.5647, | |
| "step": 595000 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.84649333235895e-05, | |
| "loss": 0.5662, | |
| "step": 595500 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.843845551781586e-05, | |
| "loss": 0.5669, | |
| "step": 596000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.8411977712042214e-05, | |
| "loss": 0.5704, | |
| "step": 596500 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.838549990626857e-05, | |
| "loss": 0.5686, | |
| "step": 597000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.8359022100494925e-05, | |
| "loss": 0.5727, | |
| "step": 597500 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.833254429472128e-05, | |
| "loss": 0.5691, | |
| "step": 598000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.8306066488947635e-05, | |
| "loss": 0.5654, | |
| "step": 598500 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.827958868317399e-05, | |
| "loss": 0.5671, | |
| "step": 599000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.8253110877400346e-05, | |
| "loss": 0.5701, | |
| "step": 599500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8226633071626705e-05, | |
| "loss": 0.5689, | |
| "step": 600000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8200155265853057e-05, | |
| "loss": 0.5712, | |
| "step": 600500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8173677460079412e-05, | |
| "loss": 0.5655, | |
| "step": 601000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8147199654305767e-05, | |
| "loss": 0.573, | |
| "step": 601500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8120721848532123e-05, | |
| "loss": 0.5672, | |
| "step": 602000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8094244042758478e-05, | |
| "loss": 0.5665, | |
| "step": 602500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8067766236984833e-05, | |
| "loss": 0.5681, | |
| "step": 603000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.8041288431211192e-05, | |
| "loss": 0.5646, | |
| "step": 603500 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.8014810625437547e-05, | |
| "loss": 0.5682, | |
| "step": 604000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.7988332819663903e-05, | |
| "loss": 0.569, | |
| "step": 604500 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.7961855013890258e-05, | |
| "loss": 0.5652, | |
| "step": 605000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.7935377208116613e-05, | |
| "loss": 0.5667, | |
| "step": 605500 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.790889940234297e-05, | |
| "loss": 0.5643, | |
| "step": 606000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.7882421596569324e-05, | |
| "loss": 0.5627, | |
| "step": 606500 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.785594379079568e-05, | |
| "loss": 0.5665, | |
| "step": 607000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.7829465985022038e-05, | |
| "loss": 0.5711, | |
| "step": 607500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.7802988179248393e-05, | |
| "loss": 0.5639, | |
| "step": 608000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.777651037347475e-05, | |
| "loss": 0.5642, | |
| "step": 608500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.7750032567701104e-05, | |
| "loss": 0.564, | |
| "step": 609000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.772355476192746e-05, | |
| "loss": 0.5708, | |
| "step": 609500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.7697076956153815e-05, | |
| "loss": 0.5644, | |
| "step": 610000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.7670599150380167e-05, | |
| "loss": 0.566, | |
| "step": 610500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.7644121344606522e-05, | |
| "loss": 0.5601, | |
| "step": 611000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.761764353883288e-05, | |
| "loss": 0.5652, | |
| "step": 611500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7591165733059236e-05, | |
| "loss": 0.5684, | |
| "step": 612000 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.756468792728559e-05, | |
| "loss": 0.5669, | |
| "step": 612500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7538210121511947e-05, | |
| "loss": 0.5658, | |
| "step": 613000 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7511732315738302e-05, | |
| "loss": 0.5623, | |
| "step": 613500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7485254509964657e-05, | |
| "loss": 0.5668, | |
| "step": 614000 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7458776704191013e-05, | |
| "loss": 0.5657, | |
| "step": 614500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.743229889841737e-05, | |
| "loss": 0.5659, | |
| "step": 615000 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.7405821092643727e-05, | |
| "loss": 0.5667, | |
| "step": 615500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7379343286870082e-05, | |
| "loss": 0.5657, | |
| "step": 616000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7352865481096437e-05, | |
| "loss": 0.5605, | |
| "step": 616500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7326387675322793e-05, | |
| "loss": 0.5649, | |
| "step": 617000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7299909869549148e-05, | |
| "loss": 0.5654, | |
| "step": 617500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7273432063775503e-05, | |
| "loss": 0.5631, | |
| "step": 618000 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.724695425800186e-05, | |
| "loss": 0.5619, | |
| "step": 618500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.7220476452228214e-05, | |
| "loss": 0.5614, | |
| "step": 619000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.719399864645457e-05, | |
| "loss": 0.5613, | |
| "step": 619500 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.7167520840680924e-05, | |
| "loss": 0.5606, | |
| "step": 620000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.714104303490728e-05, | |
| "loss": 0.5588, | |
| "step": 620500 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.7114565229133635e-05, | |
| "loss": 0.5665, | |
| "step": 621000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.708808742335999e-05, | |
| "loss": 0.5611, | |
| "step": 621500 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.7061609617586346e-05, | |
| "loss": 0.564, | |
| "step": 622000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.70351318118127e-05, | |
| "loss": 0.5585, | |
| "step": 622500 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.700865400603906e-05, | |
| "loss": 0.5659, | |
| "step": 623000 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.6982176200265415e-05, | |
| "loss": 0.5662, | |
| "step": 623500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.695569839449177e-05, | |
| "loss": 0.5646, | |
| "step": 624000 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.6929220588718126e-05, | |
| "loss": 0.5638, | |
| "step": 624500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.690274278294448e-05, | |
| "loss": 0.5618, | |
| "step": 625000 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.6876264977170836e-05, | |
| "loss": 0.5625, | |
| "step": 625500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.6849787171397192e-05, | |
| "loss": 0.5656, | |
| "step": 626000 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.682330936562355e-05, | |
| "loss": 0.5616, | |
| "step": 626500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.6796831559849906e-05, | |
| "loss": 0.5572, | |
| "step": 627000 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.677035375407626e-05, | |
| "loss": 0.5632, | |
| "step": 627500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.6743875948302616e-05, | |
| "loss": 0.564, | |
| "step": 628000 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.671739814252897e-05, | |
| "loss": 0.5632, | |
| "step": 628500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.6690920336755324e-05, | |
| "loss": 0.5658, | |
| "step": 629000 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.666444253098168e-05, | |
| "loss": 0.5656, | |
| "step": 629500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.6637964725208034e-05, | |
| "loss": 0.5622, | |
| "step": 630000 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.6611486919434393e-05, | |
| "loss": 0.5622, | |
| "step": 630500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.658500911366075e-05, | |
| "loss": 0.5633, | |
| "step": 631000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.6558531307887104e-05, | |
| "loss": 0.5629, | |
| "step": 631500 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.653205350211346e-05, | |
| "loss": 0.5577, | |
| "step": 632000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.6505575696339814e-05, | |
| "loss": 0.566, | |
| "step": 632500 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.647909789056617e-05, | |
| "loss": 0.5634, | |
| "step": 633000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.6452620084792525e-05, | |
| "loss": 0.5608, | |
| "step": 633500 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.6426142279018884e-05, | |
| "loss": 0.5586, | |
| "step": 634000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.639966447324524e-05, | |
| "loss": 0.5645, | |
| "step": 634500 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.6373186667471594e-05, | |
| "loss": 0.5657, | |
| "step": 635000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.634670886169795e-05, | |
| "loss": 0.553, | |
| "step": 635500 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.6320231055924305e-05, | |
| "loss": 0.5645, | |
| "step": 636000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.629375325015066e-05, | |
| "loss": 0.5609, | |
| "step": 636500 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.6267275444377016e-05, | |
| "loss": 0.5627, | |
| "step": 637000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.624079763860337e-05, | |
| "loss": 0.563, | |
| "step": 637500 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.6214319832829726e-05, | |
| "loss": 0.5653, | |
| "step": 638000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.618784202705608e-05, | |
| "loss": 0.558, | |
| "step": 638500 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.6161364221282437e-05, | |
| "loss": 0.5572, | |
| "step": 639000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6134886415508792e-05, | |
| "loss": 0.5609, | |
| "step": 639500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6108408609735148e-05, | |
| "loss": 0.5633, | |
| "step": 640000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6081930803961503e-05, | |
| "loss": 0.5591, | |
| "step": 640500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6055452998187858e-05, | |
| "loss": 0.5611, | |
| "step": 641000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6028975192414213e-05, | |
| "loss": 0.5578, | |
| "step": 641500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.6002497386640572e-05, | |
| "loss": 0.5627, | |
| "step": 642000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.5976019580866928e-05, | |
| "loss": 0.5584, | |
| "step": 642500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.5949541775093283e-05, | |
| "loss": 0.5583, | |
| "step": 643000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5923063969319638e-05, | |
| "loss": 0.5632, | |
| "step": 643500 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5896586163545993e-05, | |
| "loss": 0.5574, | |
| "step": 644000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.587010835777235e-05, | |
| "loss": 0.5579, | |
| "step": 644500 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5843630551998704e-05, | |
| "loss": 0.5581, | |
| "step": 645000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5817152746225063e-05, | |
| "loss": 0.5565, | |
| "step": 645500 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5790674940451418e-05, | |
| "loss": 0.5598, | |
| "step": 646000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.576419713467777e-05, | |
| "loss": 0.5566, | |
| "step": 646500 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.5737719328904125e-05, | |
| "loss": 0.5579, | |
| "step": 647000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.571124152313048e-05, | |
| "loss": 0.5644, | |
| "step": 647500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.5684763717356836e-05, | |
| "loss": 0.5597, | |
| "step": 648000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.565828591158319e-05, | |
| "loss": 0.5611, | |
| "step": 648500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.5631808105809547e-05, | |
| "loss": 0.5579, | |
| "step": 649000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.5605330300035905e-05, | |
| "loss": 0.5597, | |
| "step": 649500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.557885249426226e-05, | |
| "loss": 0.561, | |
| "step": 650000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.5552374688488616e-05, | |
| "loss": 0.5577, | |
| "step": 650500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.552589688271497e-05, | |
| "loss": 0.5584, | |
| "step": 651000 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5499419076941327e-05, | |
| "loss": 0.5572, | |
| "step": 651500 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5472941271167682e-05, | |
| "loss": 0.5565, | |
| "step": 652000 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5446463465394037e-05, | |
| "loss": 0.5589, | |
| "step": 652500 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5419985659620396e-05, | |
| "loss": 0.5564, | |
| "step": 653000 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.539350785384675e-05, | |
| "loss": 0.5652, | |
| "step": 653500 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5367030048073107e-05, | |
| "loss": 0.5583, | |
| "step": 654000 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5340552242299462e-05, | |
| "loss": 0.5613, | |
| "step": 654500 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.5314074436525817e-05, | |
| "loss": 0.5545, | |
| "step": 655000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.5287596630752173e-05, | |
| "loss": 0.5531, | |
| "step": 655500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.5261118824978528e-05, | |
| "loss": 0.5605, | |
| "step": 656000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.5234641019204882e-05, | |
| "loss": 0.553, | |
| "step": 656500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.520816321343124e-05, | |
| "loss": 0.558, | |
| "step": 657000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.5181685407657596e-05, | |
| "loss": 0.5544, | |
| "step": 657500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.515520760188395e-05, | |
| "loss": 0.5541, | |
| "step": 658000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.5128729796110305e-05, | |
| "loss": 0.5563, | |
| "step": 658500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 1.510225199033666e-05, | |
| "loss": 0.562, | |
| "step": 659000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.5075774184563015e-05, | |
| "loss": 0.5554, | |
| "step": 659500 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.504929637878937e-05, | |
| "loss": 0.5589, | |
| "step": 660000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.5022818573015726e-05, | |
| "loss": 0.5571, | |
| "step": 660500 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.4996340767242085e-05, | |
| "loss": 0.5588, | |
| "step": 661000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.496986296146844e-05, | |
| "loss": 0.5546, | |
| "step": 661500 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.4943385155694795e-05, | |
| "loss": 0.56, | |
| "step": 662000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.491690734992115e-05, | |
| "loss": 0.5577, | |
| "step": 662500 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.4890429544147504e-05, | |
| "loss": 0.5526, | |
| "step": 663000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.486395173837386e-05, | |
| "loss": 0.5612, | |
| "step": 663500 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.4837473932600215e-05, | |
| "loss": 0.5623, | |
| "step": 664000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.4810996126826574e-05, | |
| "loss": 0.5586, | |
| "step": 664500 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.4784518321052929e-05, | |
| "loss": 0.5524, | |
| "step": 665000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.4758040515279284e-05, | |
| "loss": 0.5546, | |
| "step": 665500 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.473156270950564e-05, | |
| "loss": 0.5567, | |
| "step": 666000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 1.4705084903731995e-05, | |
| "loss": 0.5599, | |
| "step": 666500 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.467860709795835e-05, | |
| "loss": 0.5567, | |
| "step": 667000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4652129292184706e-05, | |
| "loss": 0.5533, | |
| "step": 667500 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4625651486411059e-05, | |
| "loss": 0.5578, | |
| "step": 668000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4599173680637418e-05, | |
| "loss": 0.5544, | |
| "step": 668500 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4572695874863773e-05, | |
| "loss": 0.5513, | |
| "step": 669000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4546218069090129e-05, | |
| "loss": 0.5589, | |
| "step": 669500 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4519740263316484e-05, | |
| "loss": 0.557, | |
| "step": 670000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 1.4493262457542839e-05, | |
| "loss": 0.5499, | |
| "step": 670500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4466784651769194e-05, | |
| "loss": 0.552, | |
| "step": 671000 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.444030684599555e-05, | |
| "loss": 0.5528, | |
| "step": 671500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4413829040221905e-05, | |
| "loss": 0.5528, | |
| "step": 672000 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4387351234448262e-05, | |
| "loss": 0.5578, | |
| "step": 672500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4360873428674617e-05, | |
| "loss": 0.554, | |
| "step": 673000 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4334395622900973e-05, | |
| "loss": 0.5486, | |
| "step": 673500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4307917817127328e-05, | |
| "loss": 0.5574, | |
| "step": 674000 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 1.4281440011353683e-05, | |
| "loss": 0.5551, | |
| "step": 674500 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4254962205580039e-05, | |
| "loss": 0.5538, | |
| "step": 675000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4228484399806394e-05, | |
| "loss": 0.5582, | |
| "step": 675500 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4202006594032751e-05, | |
| "loss": 0.5542, | |
| "step": 676000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4175528788259106e-05, | |
| "loss": 0.5556, | |
| "step": 676500 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4149050982485462e-05, | |
| "loss": 0.5616, | |
| "step": 677000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4122573176711817e-05, | |
| "loss": 0.561, | |
| "step": 677500 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4096095370938172e-05, | |
| "loss": 0.5497, | |
| "step": 678000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 1.4069617565164528e-05, | |
| "loss": 0.5564, | |
| "step": 678500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.4043139759390883e-05, | |
| "loss": 0.5495, | |
| "step": 679000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.4016661953617238e-05, | |
| "loss": 0.5572, | |
| "step": 679500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3990184147843597e-05, | |
| "loss": 0.5552, | |
| "step": 680000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3963706342069952e-05, | |
| "loss": 0.5523, | |
| "step": 680500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3937228536296306e-05, | |
| "loss": 0.5539, | |
| "step": 681000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3910750730522661e-05, | |
| "loss": 0.5592, | |
| "step": 681500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3884272924749017e-05, | |
| "loss": 0.553, | |
| "step": 682000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 1.3857795118975372e-05, | |
| "loss": 0.5577, | |
| "step": 682500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3831317313201727e-05, | |
| "loss": 0.5545, | |
| "step": 683000 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3804839507428086e-05, | |
| "loss": 0.5541, | |
| "step": 683500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3778361701654441e-05, | |
| "loss": 0.5544, | |
| "step": 684000 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3751883895880797e-05, | |
| "loss": 0.5529, | |
| "step": 684500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3725406090107152e-05, | |
| "loss": 0.5557, | |
| "step": 685000 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3698928284333507e-05, | |
| "loss": 0.5524, | |
| "step": 685500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3672450478559861e-05, | |
| "loss": 0.5506, | |
| "step": 686000 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 1.3645972672786216e-05, | |
| "loss": 0.5491, | |
| "step": 686500 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3619494867012572e-05, | |
| "loss": 0.561, | |
| "step": 687000 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.359301706123893e-05, | |
| "loss": 0.5497, | |
| "step": 687500 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3566539255465286e-05, | |
| "loss": 0.552, | |
| "step": 688000 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3540061449691641e-05, | |
| "loss": 0.5577, | |
| "step": 688500 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3513583643917996e-05, | |
| "loss": 0.5575, | |
| "step": 689000 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3487105838144352e-05, | |
| "loss": 0.5505, | |
| "step": 689500 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3460628032370707e-05, | |
| "loss": 0.5511, | |
| "step": 690000 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 1.3434150226597062e-05, | |
| "loss": 0.5471, | |
| "step": 690500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3407672420823416e-05, | |
| "loss": 0.5508, | |
| "step": 691000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3381194615049775e-05, | |
| "loss": 0.5515, | |
| "step": 691500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.335471680927613e-05, | |
| "loss": 0.5499, | |
| "step": 692000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3328239003502485e-05, | |
| "loss": 0.5451, | |
| "step": 692500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.330176119772884e-05, | |
| "loss": 0.5491, | |
| "step": 693000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3275283391955196e-05, | |
| "loss": 0.5528, | |
| "step": 693500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3248805586181551e-05, | |
| "loss": 0.5485, | |
| "step": 694000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 1.3222327780407906e-05, | |
| "loss": 0.5512, | |
| "step": 694500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.3195849974634264e-05, | |
| "loss": 0.5496, | |
| "step": 695000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.3169372168860619e-05, | |
| "loss": 0.5523, | |
| "step": 695500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.3142894363086974e-05, | |
| "loss": 0.5502, | |
| "step": 696000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.311641655731333e-05, | |
| "loss": 0.5502, | |
| "step": 696500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.3089938751539685e-05, | |
| "loss": 0.5528, | |
| "step": 697000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.306346094576604e-05, | |
| "loss": 0.5526, | |
| "step": 697500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.3036983139992395e-05, | |
| "loss": 0.5486, | |
| "step": 698000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 1.301050533421875e-05, | |
| "loss": 0.5493, | |
| "step": 698500 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2984027528445108e-05, | |
| "loss": 0.5534, | |
| "step": 699000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2957549722671463e-05, | |
| "loss": 0.5511, | |
| "step": 699500 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2931071916897818e-05, | |
| "loss": 0.5539, | |
| "step": 700000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2904594111124174e-05, | |
| "loss": 0.5474, | |
| "step": 700500 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2878116305350529e-05, | |
| "loss": 0.5462, | |
| "step": 701000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2851638499576884e-05, | |
| "loss": 0.5526, | |
| "step": 701500 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.282516069380324e-05, | |
| "loss": 0.5478, | |
| "step": 702000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 1.2798682888029598e-05, | |
| "loss": 0.5509, | |
| "step": 702500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2772205082255954e-05, | |
| "loss": 0.5458, | |
| "step": 703000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2745727276482309e-05, | |
| "loss": 0.5503, | |
| "step": 703500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2719249470708663e-05, | |
| "loss": 0.55, | |
| "step": 704000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2692771664935018e-05, | |
| "loss": 0.5545, | |
| "step": 704500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2666293859161373e-05, | |
| "loss": 0.5509, | |
| "step": 705000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2639816053387729e-05, | |
| "loss": 0.5508, | |
| "step": 705500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2613338247614084e-05, | |
| "loss": 0.5497, | |
| "step": 706000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 1.2586860441840443e-05, | |
| "loss": 0.5545, | |
| "step": 706500 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2560382636066798e-05, | |
| "loss": 0.5498, | |
| "step": 707000 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2533904830293153e-05, | |
| "loss": 0.5493, | |
| "step": 707500 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2507427024519509e-05, | |
| "loss": 0.5467, | |
| "step": 708000 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2480949218745864e-05, | |
| "loss": 0.5498, | |
| "step": 708500 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2454471412972218e-05, | |
| "loss": 0.5496, | |
| "step": 709000 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2427993607198575e-05, | |
| "loss": 0.5517, | |
| "step": 709500 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.240151580142493e-05, | |
| "loss": 0.5457, | |
| "step": 710000 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 1.2375037995651285e-05, | |
| "loss": 0.5519, | |
| "step": 710500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.234856018987764e-05, | |
| "loss": 0.5498, | |
| "step": 711000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2322082384103998e-05, | |
| "loss": 0.543, | |
| "step": 711500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2295604578330353e-05, | |
| "loss": 0.5444, | |
| "step": 712000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2269126772556708e-05, | |
| "loss": 0.5496, | |
| "step": 712500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2242648966783064e-05, | |
| "loss": 0.548, | |
| "step": 713000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2216171161009419e-05, | |
| "loss": 0.5512, | |
| "step": 713500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 1.2189693355235774e-05, | |
| "loss": 0.5511, | |
| "step": 714000 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.216321554946213e-05, | |
| "loss": 0.5439, | |
| "step": 714500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.2136737743688487e-05, | |
| "loss": 0.5518, | |
| "step": 715000 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.2110259937914842e-05, | |
| "loss": 0.5514, | |
| "step": 715500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.2083782132141197e-05, | |
| "loss": 0.551, | |
| "step": 716000 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.2057304326367553e-05, | |
| "loss": 0.5535, | |
| "step": 716500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.203082652059391e-05, | |
| "loss": 0.5518, | |
| "step": 717000 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.2004348714820265e-05, | |
| "loss": 0.5464, | |
| "step": 717500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 1.1977870909046619e-05, | |
| "loss": 0.5461, | |
| "step": 718000 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1951393103272974e-05, | |
| "loss": 0.5504, | |
| "step": 718500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1924915297499331e-05, | |
| "loss": 0.5448, | |
| "step": 719000 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1898437491725686e-05, | |
| "loss": 0.5457, | |
| "step": 719500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1871959685952042e-05, | |
| "loss": 0.5493, | |
| "step": 720000 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1845481880178397e-05, | |
| "loss": 0.5498, | |
| "step": 720500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1819004074404754e-05, | |
| "loss": 0.5481, | |
| "step": 721000 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.179252626863111e-05, | |
| "loss": 0.5449, | |
| "step": 721500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 1.1766048462857464e-05, | |
| "loss": 0.5509, | |
| "step": 722000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.173957065708382e-05, | |
| "loss": 0.5436, | |
| "step": 722500 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1713092851310175e-05, | |
| "loss": 0.5438, | |
| "step": 723000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.168661504553653e-05, | |
| "loss": 0.5462, | |
| "step": 723500 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1660137239762886e-05, | |
| "loss": 0.5486, | |
| "step": 724000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1633659433989243e-05, | |
| "loss": 0.5455, | |
| "step": 724500 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1607181628215598e-05, | |
| "loss": 0.5478, | |
| "step": 725000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1580703822441953e-05, | |
| "loss": 0.5508, | |
| "step": 725500 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 1.1554226016668309e-05, | |
| "loss": 0.5456, | |
| "step": 726000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1527748210894666e-05, | |
| "loss": 0.542, | |
| "step": 726500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.150127040512102e-05, | |
| "loss": 0.5475, | |
| "step": 727000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1474792599347375e-05, | |
| "loss": 0.5489, | |
| "step": 727500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.144831479357373e-05, | |
| "loss": 0.5475, | |
| "step": 728000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1421836987800087e-05, | |
| "loss": 0.5472, | |
| "step": 728500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1395359182026442e-05, | |
| "loss": 0.5441, | |
| "step": 729000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1368881376252798e-05, | |
| "loss": 0.5465, | |
| "step": 729500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 1.1342403570479153e-05, | |
| "loss": 0.5483, | |
| "step": 730000 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.131592576470551e-05, | |
| "loss": 0.5486, | |
| "step": 730500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1289447958931865e-05, | |
| "loss": 0.5408, | |
| "step": 731000 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.126297015315822e-05, | |
| "loss": 0.5465, | |
| "step": 731500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1236492347384576e-05, | |
| "loss": 0.5458, | |
| "step": 732000 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1210014541610931e-05, | |
| "loss": 0.5419, | |
| "step": 732500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1183536735837287e-05, | |
| "loss": 0.5447, | |
| "step": 733000 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1157058930063642e-05, | |
| "loss": 0.5427, | |
| "step": 733500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.1130581124289999e-05, | |
| "loss": 0.5423, | |
| "step": 734000 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.1104103318516354e-05, | |
| "loss": 0.5465, | |
| "step": 734500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.107762551274271e-05, | |
| "loss": 0.5433, | |
| "step": 735000 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.1051147706969065e-05, | |
| "loss": 0.546, | |
| "step": 735500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.102466990119542e-05, | |
| "loss": 0.5453, | |
| "step": 736000 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.0998192095421776e-05, | |
| "loss": 0.5457, | |
| "step": 736500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.0971714289648131e-05, | |
| "loss": 0.5469, | |
| "step": 737000 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.0945236483874486e-05, | |
| "loss": 0.5459, | |
| "step": 737500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.0918758678100843e-05, | |
| "loss": 0.544, | |
| "step": 738000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0892280872327199e-05, | |
| "loss": 0.5437, | |
| "step": 738500 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0865803066553554e-05, | |
| "loss": 0.5433, | |
| "step": 739000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.083932526077991e-05, | |
| "loss": 0.5484, | |
| "step": 739500 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0812847455006266e-05, | |
| "loss": 0.5436, | |
| "step": 740000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0786369649232622e-05, | |
| "loss": 0.5438, | |
| "step": 740500 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0759891843458975e-05, | |
| "loss": 0.5453, | |
| "step": 741000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0733414037685332e-05, | |
| "loss": 0.5471, | |
| "step": 741500 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.0706936231911688e-05, | |
| "loss": 0.547, | |
| "step": 742000 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0680458426138043e-05, | |
| "loss": 0.5425, | |
| "step": 742500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0653980620364398e-05, | |
| "loss": 0.5461, | |
| "step": 743000 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0627502814590755e-05, | |
| "loss": 0.5505, | |
| "step": 743500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.060102500881711e-05, | |
| "loss": 0.5408, | |
| "step": 744000 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0574547203043466e-05, | |
| "loss": 0.5447, | |
| "step": 744500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0548069397269821e-05, | |
| "loss": 0.5443, | |
| "step": 745000 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0521591591496177e-05, | |
| "loss": 0.5456, | |
| "step": 745500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.0495113785722532e-05, | |
| "loss": 0.5416, | |
| "step": 746000 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0468635979948887e-05, | |
| "loss": 0.5471, | |
| "step": 746500 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0442158174175242e-05, | |
| "loss": 0.5433, | |
| "step": 747000 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.04156803684016e-05, | |
| "loss": 0.542, | |
| "step": 747500 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0389202562627955e-05, | |
| "loss": 0.5436, | |
| "step": 748000 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.036272475685431e-05, | |
| "loss": 0.5407, | |
| "step": 748500 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0336246951080665e-05, | |
| "loss": 0.548, | |
| "step": 749000 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0309769145307022e-05, | |
| "loss": 0.5429, | |
| "step": 749500 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.0283291339533376e-05, | |
| "loss": 0.5446, | |
| "step": 750000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0256813533759731e-05, | |
| "loss": 0.5498, | |
| "step": 750500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0230335727986087e-05, | |
| "loss": 0.5436, | |
| "step": 751000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0203857922212444e-05, | |
| "loss": 0.5399, | |
| "step": 751500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0177380116438799e-05, | |
| "loss": 0.5411, | |
| "step": 752000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0150902310665154e-05, | |
| "loss": 0.5437, | |
| "step": 752500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0124424504891511e-05, | |
| "loss": 0.5415, | |
| "step": 753000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0097946699117867e-05, | |
| "loss": 0.5467, | |
| "step": 753500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.0071468893344222e-05, | |
| "loss": 0.544, | |
| "step": 754000 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.0044991087570577e-05, | |
| "loss": 0.5422, | |
| "step": 754500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.0018513281796933e-05, | |
| "loss": 0.5385, | |
| "step": 755000 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.992035476023288e-06, | |
| "loss": 0.5418, | |
| "step": 755500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.965557670249643e-06, | |
| "loss": 0.5444, | |
| "step": 756000 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.939079864475999e-06, | |
| "loss": 0.5416, | |
| "step": 756500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.912602058702356e-06, | |
| "loss": 0.54, | |
| "step": 757000 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.886124252928711e-06, | |
| "loss": 0.5406, | |
| "step": 757500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 9.859646447155066e-06, | |
| "loss": 0.5452, | |
| "step": 758000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.833168641381422e-06, | |
| "loss": 0.5398, | |
| "step": 758500 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.806690835607777e-06, | |
| "loss": 0.5415, | |
| "step": 759000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.780213029834132e-06, | |
| "loss": 0.542, | |
| "step": 759500 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.753735224060488e-06, | |
| "loss": 0.5399, | |
| "step": 760000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.727257418286843e-06, | |
| "loss": 0.5469, | |
| "step": 760500 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.7007796125132e-06, | |
| "loss": 0.5392, | |
| "step": 761000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 9.674301806739555e-06, | |
| "loss": 0.542, | |
| "step": 761500 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.64782400096591e-06, | |
| "loss": 0.541, | |
| "step": 762000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.621346195192268e-06, | |
| "loss": 0.5416, | |
| "step": 762500 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.594868389418623e-06, | |
| "loss": 0.5419, | |
| "step": 763000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.568390583644978e-06, | |
| "loss": 0.5395, | |
| "step": 763500 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.541912777871332e-06, | |
| "loss": 0.5359, | |
| "step": 764000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.515434972097689e-06, | |
| "loss": 0.5432, | |
| "step": 764500 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.488957166324044e-06, | |
| "loss": 0.546, | |
| "step": 765000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 9.4624793605504e-06, | |
| "loss": 0.5418, | |
| "step": 765500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.436001554776755e-06, | |
| "loss": 0.5397, | |
| "step": 766000 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.409523749003112e-06, | |
| "loss": 0.5403, | |
| "step": 766500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.383045943229467e-06, | |
| "loss": 0.5385, | |
| "step": 767000 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.356568137455823e-06, | |
| "loss": 0.5402, | |
| "step": 767500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.330090331682178e-06, | |
| "loss": 0.538, | |
| "step": 768000 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.303612525908533e-06, | |
| "loss": 0.5412, | |
| "step": 768500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.277134720134889e-06, | |
| "loss": 0.5401, | |
| "step": 769000 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 9.250656914361244e-06, | |
| "loss": 0.5396, | |
| "step": 769500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.2241791085876e-06, | |
| "loss": 0.5385, | |
| "step": 770000 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.197701302813956e-06, | |
| "loss": 0.5426, | |
| "step": 770500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.171223497040312e-06, | |
| "loss": 0.5437, | |
| "step": 771000 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.144745691266667e-06, | |
| "loss": 0.5411, | |
| "step": 771500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.118267885493024e-06, | |
| "loss": 0.5378, | |
| "step": 772000 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.09179007971938e-06, | |
| "loss": 0.5419, | |
| "step": 772500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.065312273945733e-06, | |
| "loss": 0.5415, | |
| "step": 773000 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 9.038834468172088e-06, | |
| "loss": 0.537, | |
| "step": 773500 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 9.012356662398445e-06, | |
| "loss": 0.539, | |
| "step": 774000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.9858788566248e-06, | |
| "loss": 0.5408, | |
| "step": 774500 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.959401050851156e-06, | |
| "loss": 0.5422, | |
| "step": 775000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.932923245077511e-06, | |
| "loss": 0.5367, | |
| "step": 775500 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.906445439303868e-06, | |
| "loss": 0.5436, | |
| "step": 776000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.879967633530223e-06, | |
| "loss": 0.5392, | |
| "step": 776500 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.853489827756579e-06, | |
| "loss": 0.5412, | |
| "step": 777000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 8.827012021982934e-06, | |
| "loss": 0.5354, | |
| "step": 777500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.80053421620929e-06, | |
| "loss": 0.5394, | |
| "step": 778000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.774056410435645e-06, | |
| "loss": 0.542, | |
| "step": 778500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.747578604662e-06, | |
| "loss": 0.538, | |
| "step": 779000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.721100798888355e-06, | |
| "loss": 0.5364, | |
| "step": 779500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.694622993114712e-06, | |
| "loss": 0.5419, | |
| "step": 780000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.668145187341068e-06, | |
| "loss": 0.5388, | |
| "step": 780500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.641667381567423e-06, | |
| "loss": 0.5412, | |
| "step": 781000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.615189575793778e-06, | |
| "loss": 0.5416, | |
| "step": 781500 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.588711770020134e-06, | |
| "loss": 0.5415, | |
| "step": 782000 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.562233964246489e-06, | |
| "loss": 0.5357, | |
| "step": 782500 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.535756158472844e-06, | |
| "loss": 0.5387, | |
| "step": 783000 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.509278352699201e-06, | |
| "loss": 0.5395, | |
| "step": 783500 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.482800546925557e-06, | |
| "loss": 0.5369, | |
| "step": 784000 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.456322741151912e-06, | |
| "loss": 0.54, | |
| "step": 784500 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.429844935378267e-06, | |
| "loss": 0.5347, | |
| "step": 785000 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 8.403367129604624e-06, | |
| "loss": 0.5396, | |
| "step": 785500 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.37688932383098e-06, | |
| "loss": 0.5392, | |
| "step": 786000 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.350411518057335e-06, | |
| "loss": 0.5407, | |
| "step": 786500 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.323933712283689e-06, | |
| "loss": 0.5349, | |
| "step": 787000 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.297455906510046e-06, | |
| "loss": 0.5358, | |
| "step": 787500 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.270978100736401e-06, | |
| "loss": 0.5424, | |
| "step": 788000 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.244500294962756e-06, | |
| "loss": 0.5378, | |
| "step": 788500 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.218022489189112e-06, | |
| "loss": 0.5406, | |
| "step": 789000 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.191544683415469e-06, | |
| "loss": 0.5438, | |
| "step": 789500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.165066877641824e-06, | |
| "loss": 0.5389, | |
| "step": 790000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.13858907186818e-06, | |
| "loss": 0.5383, | |
| "step": 790500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.112111266094535e-06, | |
| "loss": 0.5362, | |
| "step": 791000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.08563346032089e-06, | |
| "loss": 0.5397, | |
| "step": 791500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.059155654547245e-06, | |
| "loss": 0.5342, | |
| "step": 792000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.0326778487736e-06, | |
| "loss": 0.538, | |
| "step": 792500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.006200042999958e-06, | |
| "loss": 0.5374, | |
| "step": 793000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 7.979722237226313e-06, | |
| "loss": 0.5387, | |
| "step": 793500 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.953244431452668e-06, | |
| "loss": 0.5363, | |
| "step": 794000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.926766625679024e-06, | |
| "loss": 0.5362, | |
| "step": 794500 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.90028881990538e-06, | |
| "loss": 0.5386, | |
| "step": 795000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.873811014131736e-06, | |
| "loss": 0.5355, | |
| "step": 795500 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.84733320835809e-06, | |
| "loss": 0.5376, | |
| "step": 796000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.820855402584445e-06, | |
| "loss": 0.535, | |
| "step": 796500 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.794377596810802e-06, | |
| "loss": 0.5376, | |
| "step": 797000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 7.767899791037157e-06, | |
| "loss": 0.5393, | |
| "step": 797500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.741421985263513e-06, | |
| "loss": 0.5398, | |
| "step": 798000 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.714944179489868e-06, | |
| "loss": 0.5351, | |
| "step": 798500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.688466373716225e-06, | |
| "loss": 0.5312, | |
| "step": 799000 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.66198856794258e-06, | |
| "loss": 0.5378, | |
| "step": 799500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.635510762168935e-06, | |
| "loss": 0.5353, | |
| "step": 800000 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.60903295639529e-06, | |
| "loss": 0.5319, | |
| "step": 800500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.582555150621647e-06, | |
| "loss": 0.5363, | |
| "step": 801000 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.5560773448480015e-06, | |
| "loss": 0.5334, | |
| "step": 801500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.529599539074357e-06, | |
| "loss": 0.5369, | |
| "step": 802000 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.503121733300714e-06, | |
| "loss": 0.5373, | |
| "step": 802500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.476643927527069e-06, | |
| "loss": 0.5404, | |
| "step": 803000 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.4501661217534244e-06, | |
| "loss": 0.5346, | |
| "step": 803500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.423688315979779e-06, | |
| "loss": 0.537, | |
| "step": 804000 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.397210510206136e-06, | |
| "loss": 0.5378, | |
| "step": 804500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.370732704432491e-06, | |
| "loss": 0.5339, | |
| "step": 805000 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 7.344254898658847e-06, | |
| "loss": 0.5361, | |
| "step": 805500 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.317777092885202e-06, | |
| "loss": 0.5341, | |
| "step": 806000 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.291299287111558e-06, | |
| "loss": 0.5341, | |
| "step": 806500 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.264821481337913e-06, | |
| "loss": 0.5327, | |
| "step": 807000 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.238343675564269e-06, | |
| "loss": 0.536, | |
| "step": 807500 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.211865869790624e-06, | |
| "loss": 0.5364, | |
| "step": 808000 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.18538806401698e-06, | |
| "loss": 0.5332, | |
| "step": 808500 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.1589102582433355e-06, | |
| "loss": 0.5342, | |
| "step": 809000 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.132432452469691e-06, | |
| "loss": 0.5355, | |
| "step": 809500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 7.105954646696046e-06, | |
| "loss": 0.5348, | |
| "step": 810000 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 7.079476840922402e-06, | |
| "loss": 0.5369, | |
| "step": 810500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 7.052999035148758e-06, | |
| "loss": 0.5368, | |
| "step": 811000 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 7.026521229375113e-06, | |
| "loss": 0.5352, | |
| "step": 811500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 7.00004342360147e-06, | |
| "loss": 0.5329, | |
| "step": 812000 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.973565617827825e-06, | |
| "loss": 0.535, | |
| "step": 812500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.94708781205418e-06, | |
| "loss": 0.5306, | |
| "step": 813000 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.920610006280535e-06, | |
| "loss": 0.5332, | |
| "step": 813500 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.894132200506892e-06, | |
| "loss": 0.5364, | |
| "step": 814000 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.8676543947332475e-06, | |
| "loss": 0.5366, | |
| "step": 814500 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.841176588959603e-06, | |
| "loss": 0.5364, | |
| "step": 815000 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.814698783185957e-06, | |
| "loss": 0.5319, | |
| "step": 815500 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.788220977412314e-06, | |
| "loss": 0.5386, | |
| "step": 816000 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.76174317163867e-06, | |
| "loss": 0.5388, | |
| "step": 816500 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.735265365865025e-06, | |
| "loss": 0.5316, | |
| "step": 817000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.70878756009138e-06, | |
| "loss": 0.5298, | |
| "step": 817500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.6823097543177364e-06, | |
| "loss": 0.5332, | |
| "step": 818000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.655831948544092e-06, | |
| "loss": 0.5317, | |
| "step": 818500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.629354142770447e-06, | |
| "loss": 0.5371, | |
| "step": 819000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.602876336996802e-06, | |
| "loss": 0.5316, | |
| "step": 819500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.5763985312231586e-06, | |
| "loss": 0.5337, | |
| "step": 820000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.549920725449514e-06, | |
| "loss": 0.5354, | |
| "step": 820500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.523442919675869e-06, | |
| "loss": 0.534, | |
| "step": 821000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.496965113902226e-06, | |
| "loss": 0.5305, | |
| "step": 821500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.470487308128581e-06, | |
| "loss": 0.5344, | |
| "step": 822000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.444009502354936e-06, | |
| "loss": 0.5338, | |
| "step": 822500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.417531696581291e-06, | |
| "loss": 0.5323, | |
| "step": 823000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.391053890807648e-06, | |
| "loss": 0.5344, | |
| "step": 823500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.364576085034004e-06, | |
| "loss": 0.5358, | |
| "step": 824000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.338098279260358e-06, | |
| "loss": 0.5334, | |
| "step": 824500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.3116204734867135e-06, | |
| "loss": 0.5333, | |
| "step": 825000 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.2851426677130705e-06, | |
| "loss": 0.5296, | |
| "step": 825500 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.258664861939426e-06, | |
| "loss": 0.5341, | |
| "step": 826000 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.232187056165781e-06, | |
| "loss": 0.5318, | |
| "step": 826500 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.2057092503921365e-06, | |
| "loss": 0.5348, | |
| "step": 827000 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.179231444618492e-06, | |
| "loss": 0.5319, | |
| "step": 827500 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.152753638844848e-06, | |
| "loss": 0.5347, | |
| "step": 828000 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.126275833071203e-06, | |
| "loss": 0.5283, | |
| "step": 828500 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.099798027297559e-06, | |
| "loss": 0.5296, | |
| "step": 829000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.073320221523914e-06, | |
| "loss": 0.5333, | |
| "step": 829500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.04684241575027e-06, | |
| "loss": 0.5316, | |
| "step": 830000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.020364609976626e-06, | |
| "loss": 0.5328, | |
| "step": 830500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.993886804202982e-06, | |
| "loss": 0.5337, | |
| "step": 831000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.967408998429337e-06, | |
| "loss": 0.5339, | |
| "step": 831500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.940931192655692e-06, | |
| "loss": 0.5316, | |
| "step": 832000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.914453386882048e-06, | |
| "loss": 0.5298, | |
| "step": 832500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.887975581108404e-06, | |
| "loss": 0.5301, | |
| "step": 833000 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.861497775334759e-06, | |
| "loss": 0.5299, | |
| "step": 833500 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.835019969561114e-06, | |
| "loss": 0.5338, | |
| "step": 834000 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.8085421637874706e-06, | |
| "loss": 0.5329, | |
| "step": 834500 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.782064358013826e-06, | |
| "loss": 0.5301, | |
| "step": 835000 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.755586552240182e-06, | |
| "loss": 0.5305, | |
| "step": 835500 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.7291087464665365e-06, | |
| "loss": 0.5338, | |
| "step": 836000 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.702630940692893e-06, | |
| "loss": 0.5278, | |
| "step": 836500 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.676153134919248e-06, | |
| "loss": 0.5297, | |
| "step": 837000 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.649675329145604e-06, | |
| "loss": 0.5318, | |
| "step": 837500 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.6231975233719595e-06, | |
| "loss": 0.5349, | |
| "step": 838000 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.596719717598315e-06, | |
| "loss": 0.5302, | |
| "step": 838500 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.57024191182467e-06, | |
| "loss": 0.532, | |
| "step": 839000 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.543764106051026e-06, | |
| "loss": 0.5302, | |
| "step": 839500 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.517286300277382e-06, | |
| "loss": 0.5297, | |
| "step": 840000 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.490808494503737e-06, | |
| "loss": 0.5306, | |
| "step": 840500 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.464330688730093e-06, | |
| "loss": 0.5298, | |
| "step": 841000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.4378528829564485e-06, | |
| "loss": 0.5311, | |
| "step": 841500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.411375077182805e-06, | |
| "loss": 0.5278, | |
| "step": 842000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.38489727140916e-06, | |
| "loss": 0.5333, | |
| "step": 842500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.358419465635515e-06, | |
| "loss": 0.5287, | |
| "step": 843000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.331941659861871e-06, | |
| "loss": 0.5331, | |
| "step": 843500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.305463854088227e-06, | |
| "loss": 0.5292, | |
| "step": 844000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.278986048314582e-06, | |
| "loss": 0.531, | |
| "step": 844500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.252508242540937e-06, | |
| "loss": 0.5256, | |
| "step": 845000 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.226030436767293e-06, | |
| "loss": 0.5318, | |
| "step": 845500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.199552630993649e-06, | |
| "loss": 0.5339, | |
| "step": 846000 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.173074825220004e-06, | |
| "loss": 0.5295, | |
| "step": 846500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.14659701944636e-06, | |
| "loss": 0.5283, | |
| "step": 847000 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.120119213672715e-06, | |
| "loss": 0.5308, | |
| "step": 847500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.093641407899071e-06, | |
| "loss": 0.5325, | |
| "step": 848000 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.067163602125426e-06, | |
| "loss": 0.5287, | |
| "step": 848500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.0406857963517825e-06, | |
| "loss": 0.5324, | |
| "step": 849000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.014207990578138e-06, | |
| "loss": 0.5285, | |
| "step": 849500 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.987730184804493e-06, | |
| "loss": 0.5337, | |
| "step": 850000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.961252379030849e-06, | |
| "loss": 0.5261, | |
| "step": 850500 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.934774573257205e-06, | |
| "loss": 0.5285, | |
| "step": 851000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.908296767483561e-06, | |
| "loss": 0.5289, | |
| "step": 851500 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.881818961709915e-06, | |
| "loss": 0.5215, | |
| "step": 852000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.8553411559362715e-06, | |
| "loss": 0.5288, | |
| "step": 852500 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.828863350162627e-06, | |
| "loss": 0.5304, | |
| "step": 853000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.802385544388983e-06, | |
| "loss": 0.5312, | |
| "step": 853500 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.775907738615338e-06, | |
| "loss": 0.5289, | |
| "step": 854000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.749429932841694e-06, | |
| "loss": 0.5315, | |
| "step": 854500 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.722952127068049e-06, | |
| "loss": 0.5301, | |
| "step": 855000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.696474321294405e-06, | |
| "loss": 0.5289, | |
| "step": 855500 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.6699965155207604e-06, | |
| "loss": 0.5297, | |
| "step": 856000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.643518709747116e-06, | |
| "loss": 0.5294, | |
| "step": 856500 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.617040903973471e-06, | |
| "loss": 0.5273, | |
| "step": 857000 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.590563098199827e-06, | |
| "loss": 0.53, | |
| "step": 857500 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.564085292426183e-06, | |
| "loss": 0.5303, | |
| "step": 858000 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.537607486652539e-06, | |
| "loss": 0.5275, | |
| "step": 858500 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.511129680878893e-06, | |
| "loss": 0.5259, | |
| "step": 859000 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.484651875105249e-06, | |
| "loss": 0.5306, | |
| "step": 859500 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.4581740693316056e-06, | |
| "loss": 0.5284, | |
| "step": 860000 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.431696263557961e-06, | |
| "loss": 0.5297, | |
| "step": 860500 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.405218457784316e-06, | |
| "loss": 0.5275, | |
| "step": 861000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.3787406520106715e-06, | |
| "loss": 0.524, | |
| "step": 861500 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.352262846237028e-06, | |
| "loss": 0.532, | |
| "step": 862000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.325785040463383e-06, | |
| "loss": 0.5268, | |
| "step": 862500 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.299307234689739e-06, | |
| "loss": 0.5293, | |
| "step": 863000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.272829428916094e-06, | |
| "loss": 0.5275, | |
| "step": 863500 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.24635162314245e-06, | |
| "loss": 0.5286, | |
| "step": 864000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.219873817368805e-06, | |
| "loss": 0.5276, | |
| "step": 864500 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.193396011595161e-06, | |
| "loss": 0.5244, | |
| "step": 865000 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.166918205821517e-06, | |
| "loss": 0.5311, | |
| "step": 865500 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.140440400047872e-06, | |
| "loss": 0.5311, | |
| "step": 866000 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.113962594274227e-06, | |
| "loss": 0.5275, | |
| "step": 866500 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.0874847885005835e-06, | |
| "loss": 0.5301, | |
| "step": 867000 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.061006982726939e-06, | |
| "loss": 0.5317, | |
| "step": 867500 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.034529176953294e-06, | |
| "loss": 0.5317, | |
| "step": 868000 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.0080513711796494e-06, | |
| "loss": 0.527, | |
| "step": 868500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.981573565406006e-06, | |
| "loss": 0.5273, | |
| "step": 869000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.955095759632361e-06, | |
| "loss": 0.5317, | |
| "step": 869500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.928617953858717e-06, | |
| "loss": 0.53, | |
| "step": 870000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.902140148085072e-06, | |
| "loss": 0.5331, | |
| "step": 870500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.875662342311428e-06, | |
| "loss": 0.5264, | |
| "step": 871000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.849184536537784e-06, | |
| "loss": 0.5248, | |
| "step": 871500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.822706730764139e-06, | |
| "loss": 0.5299, | |
| "step": 872000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 3.796228924990495e-06, | |
| "loss": 0.5311, | |
| "step": 872500 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.7697511192168503e-06, | |
| "loss": 0.5259, | |
| "step": 873000 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.743273313443206e-06, | |
| "loss": 0.5307, | |
| "step": 873500 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.7167955076695614e-06, | |
| "loss": 0.531, | |
| "step": 874000 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.690317701895917e-06, | |
| "loss": 0.5289, | |
| "step": 874500 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.6638398961222724e-06, | |
| "loss": 0.5258, | |
| "step": 875000 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.637362090348628e-06, | |
| "loss": 0.5234, | |
| "step": 875500 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.6108842845749835e-06, | |
| "loss": 0.5307, | |
| "step": 876000 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.5844064788013393e-06, | |
| "loss": 0.5306, | |
| "step": 876500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.5579286730276946e-06, | |
| "loss": 0.5274, | |
| "step": 877000 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.5314508672540508e-06, | |
| "loss": 0.5302, | |
| "step": 877500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.5049730614804057e-06, | |
| "loss": 0.5353, | |
| "step": 878000 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.478495255706762e-06, | |
| "loss": 0.5291, | |
| "step": 878500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.4520174499331167e-06, | |
| "loss": 0.5345, | |
| "step": 879000 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.425539644159473e-06, | |
| "loss": 0.5257, | |
| "step": 879500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.3990618383858286e-06, | |
| "loss": 0.5291, | |
| "step": 880000 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.372584032612184e-06, | |
| "loss": 0.5278, | |
| "step": 880500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.3461062268385397e-06, | |
| "loss": 0.526, | |
| "step": 881000 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.319628421064895e-06, | |
| "loss": 0.5259, | |
| "step": 881500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.293150615291251e-06, | |
| "loss": 0.5239, | |
| "step": 882000 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.266672809517606e-06, | |
| "loss": 0.527, | |
| "step": 882500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.2401950037439623e-06, | |
| "loss": 0.5272, | |
| "step": 883000 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.213717197970317e-06, | |
| "loss": 0.5271, | |
| "step": 883500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.1872393921966733e-06, | |
| "loss": 0.524, | |
| "step": 884000 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.1607615864230287e-06, | |
| "loss": 0.5264, | |
| "step": 884500 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.1342837806493844e-06, | |
| "loss": 0.5268, | |
| "step": 885000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.10780597487574e-06, | |
| "loss": 0.5227, | |
| "step": 885500 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0813281691020955e-06, | |
| "loss": 0.5305, | |
| "step": 886000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0548503633284512e-06, | |
| "loss": 0.529, | |
| "step": 886500 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0283725575548065e-06, | |
| "loss": 0.5294, | |
| "step": 887000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.0018947517811623e-06, | |
| "loss": 0.5251, | |
| "step": 887500 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 2.9754169460075176e-06, | |
| "loss": 0.5289, | |
| "step": 888000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 2.9489391402338734e-06, | |
| "loss": 0.5228, | |
| "step": 888500 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.922461334460229e-06, | |
| "loss": 0.5237, | |
| "step": 889000 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.8959835286865844e-06, | |
| "loss": 0.5238, | |
| "step": 889500 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.86950572291294e-06, | |
| "loss": 0.5252, | |
| "step": 890000 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.8430279171392955e-06, | |
| "loss": 0.5237, | |
| "step": 890500 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.8165501113656512e-06, | |
| "loss": 0.5305, | |
| "step": 891000 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.7900723055920066e-06, | |
| "loss": 0.5237, | |
| "step": 891500 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.7635944998183623e-06, | |
| "loss": 0.5284, | |
| "step": 892000 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 2.737116694044718e-06, | |
| "loss": 0.5248, | |
| "step": 892500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.7106388882710734e-06, | |
| "loss": 0.5307, | |
| "step": 893000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.684161082497429e-06, | |
| "loss": 0.5235, | |
| "step": 893500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.6576832767237845e-06, | |
| "loss": 0.5262, | |
| "step": 894000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.6312054709501406e-06, | |
| "loss": 0.5263, | |
| "step": 894500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.604727665176496e-06, | |
| "loss": 0.5275, | |
| "step": 895000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.5782498594028517e-06, | |
| "loss": 0.521, | |
| "step": 895500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.551772053629207e-06, | |
| "loss": 0.5218, | |
| "step": 896000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 2.5252942478555628e-06, | |
| "loss": 0.5237, | |
| "step": 896500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.4988164420819185e-06, | |
| "loss": 0.5211, | |
| "step": 897000 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.472338636308274e-06, | |
| "loss": 0.5294, | |
| "step": 897500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.4458608305346296e-06, | |
| "loss": 0.5284, | |
| "step": 898000 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.419383024760985e-06, | |
| "loss": 0.526, | |
| "step": 898500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.3929052189873406e-06, | |
| "loss": 0.5201, | |
| "step": 899000 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.366427413213696e-06, | |
| "loss": 0.5281, | |
| "step": 899500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.3399496074400517e-06, | |
| "loss": 0.5218, | |
| "step": 900000 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 2.3134718016664075e-06, | |
| "loss": 0.5236, | |
| "step": 900500 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.2869939958927628e-06, | |
| "loss": 0.5235, | |
| "step": 901000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.2605161901191185e-06, | |
| "loss": 0.5283, | |
| "step": 901500 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.234038384345474e-06, | |
| "loss": 0.5301, | |
| "step": 902000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.2075605785718296e-06, | |
| "loss": 0.5253, | |
| "step": 902500 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.181082772798185e-06, | |
| "loss": 0.5238, | |
| "step": 903000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.1546049670245407e-06, | |
| "loss": 0.5297, | |
| "step": 903500 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.1281271612508964e-06, | |
| "loss": 0.5215, | |
| "step": 904000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 2.101649355477252e-06, | |
| "loss": 0.5215, | |
| "step": 904500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.075171549703608e-06, | |
| "loss": 0.5296, | |
| "step": 905000 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.0486937439299632e-06, | |
| "loss": 0.5266, | |
| "step": 905500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.022215938156319e-06, | |
| "loss": 0.5266, | |
| "step": 906000 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.9957381323826743e-06, | |
| "loss": 0.5243, | |
| "step": 906500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.96926032660903e-06, | |
| "loss": 0.5292, | |
| "step": 907000 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.9427825208353854e-06, | |
| "loss": 0.5277, | |
| "step": 907500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.916304715061741e-06, | |
| "loss": 0.5244, | |
| "step": 908000 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8898269092880966e-06, | |
| "loss": 0.5246, | |
| "step": 908500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8633491035144522e-06, | |
| "loss": 0.5239, | |
| "step": 909000 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8368712977408077e-06, | |
| "loss": 0.528, | |
| "step": 909500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.8103934919671632e-06, | |
| "loss": 0.5221, | |
| "step": 910000 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.783915686193519e-06, | |
| "loss": 0.5269, | |
| "step": 910500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.7574378804198745e-06, | |
| "loss": 0.5294, | |
| "step": 911000 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.73096007464623e-06, | |
| "loss": 0.5273, | |
| "step": 911500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.7044822688725856e-06, | |
| "loss": 0.5216, | |
| "step": 912000 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.6780044630989411e-06, | |
| "loss": 0.5272, | |
| "step": 912500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.6515266573252967e-06, | |
| "loss": 0.5192, | |
| "step": 913000 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.6250488515516522e-06, | |
| "loss": 0.5241, | |
| "step": 913500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.598571045778008e-06, | |
| "loss": 0.5295, | |
| "step": 914000 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.5720932400043637e-06, | |
| "loss": 0.5249, | |
| "step": 914500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.5456154342307192e-06, | |
| "loss": 0.524, | |
| "step": 915000 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.5191376284570748e-06, | |
| "loss": 0.5244, | |
| "step": 915500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.4926598226834303e-06, | |
| "loss": 0.5259, | |
| "step": 916000 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.4661820169097858e-06, | |
| "loss": 0.5267, | |
| "step": 916500 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.4397042111361416e-06, | |
| "loss": 0.5299, | |
| "step": 917000 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.4132264053624971e-06, | |
| "loss": 0.5209, | |
| "step": 917500 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3867485995888526e-06, | |
| "loss": 0.5188, | |
| "step": 918000 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3602707938152084e-06, | |
| "loss": 0.5218, | |
| "step": 918500 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.333792988041564e-06, | |
| "loss": 0.5257, | |
| "step": 919000 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3073151822679195e-06, | |
| "loss": 0.5247, | |
| "step": 919500 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.280837376494275e-06, | |
| "loss": 0.5266, | |
| "step": 920000 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.2543595707206305e-06, | |
| "loss": 0.5174, | |
| "step": 920500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.227881764946986e-06, | |
| "loss": 0.5266, | |
| "step": 921000 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.2014039591733416e-06, | |
| "loss": 0.5224, | |
| "step": 921500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.1749261533996973e-06, | |
| "loss": 0.5277, | |
| "step": 922000 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.148448347626053e-06, | |
| "loss": 0.5233, | |
| "step": 922500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.1219705418524086e-06, | |
| "loss": 0.5268, | |
| "step": 923000 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.0954927360787642e-06, | |
| "loss": 0.5271, | |
| "step": 923500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.0690149303051197e-06, | |
| "loss": 0.519, | |
| "step": 924000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.0425371245314752e-06, | |
| "loss": 0.5221, | |
| "step": 924500 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.0160593187578308e-06, | |
| "loss": 0.5267, | |
| "step": 925000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 9.895815129841863e-07, | |
| "loss": 0.5229, | |
| "step": 925500 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 9.63103707210542e-07, | |
| "loss": 0.5221, | |
| "step": 926000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 9.366259014368975e-07, | |
| "loss": 0.524, | |
| "step": 926500 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 9.101480956632532e-07, | |
| "loss": 0.522, | |
| "step": 927000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 8.836702898896089e-07, | |
| "loss": 0.5223, | |
| "step": 927500 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 8.571924841159644e-07, | |
| "loss": 0.5236, | |
| "step": 928000 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 8.307146783423199e-07, | |
| "loss": 0.5259, | |
| "step": 928500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 8.042368725686756e-07, | |
| "loss": 0.5233, | |
| "step": 929000 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 7.777590667950311e-07, | |
| "loss": 0.5299, | |
| "step": 929500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 7.512812610213866e-07, | |
| "loss": 0.5265, | |
| "step": 930000 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 7.248034552477423e-07, | |
| "loss": 0.5208, | |
| "step": 930500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 6.983256494740979e-07, | |
| "loss": 0.5226, | |
| "step": 931000 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 6.718478437004535e-07, | |
| "loss": 0.5257, | |
| "step": 931500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 6.45370037926809e-07, | |
| "loss": 0.525, | |
| "step": 932000 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 6.188922321531645e-07, | |
| "loss": 0.5288, | |
| "step": 932500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.924144263795203e-07, | |
| "loss": 0.5266, | |
| "step": 933000 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.659366206058758e-07, | |
| "loss": 0.5248, | |
| "step": 933500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.394588148322313e-07, | |
| "loss": 0.5185, | |
| "step": 934000 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.129810090585869e-07, | |
| "loss": 0.5248, | |
| "step": 934500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.865032032849425e-07, | |
| "loss": 0.5213, | |
| "step": 935000 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.6002539751129815e-07, | |
| "loss": 0.5222, | |
| "step": 935500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.335475917376537e-07, | |
| "loss": 0.5279, | |
| "step": 936000 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.070697859640093e-07, | |
| "loss": 0.5254, | |
| "step": 936500 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 3.805919801903648e-07, | |
| "loss": 0.5261, | |
| "step": 937000 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 3.5411417441672045e-07, | |
| "loss": 0.519, | |
| "step": 937500 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 3.27636368643076e-07, | |
| "loss": 0.5252, | |
| "step": 938000 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 3.011585628694316e-07, | |
| "loss": 0.5254, | |
| "step": 938500 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.7468075709578716e-07, | |
| "loss": 0.5232, | |
| "step": 939000 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.4820295132214275e-07, | |
| "loss": 0.523, | |
| "step": 939500 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.2172514554849836e-07, | |
| "loss": 0.5235, | |
| "step": 940000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.9524733977485395e-07, | |
| "loss": 0.5209, | |
| "step": 940500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.687695340012095e-07, | |
| "loss": 0.5203, | |
| "step": 941000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.422917282275651e-07, | |
| "loss": 0.5236, | |
| "step": 941500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.1581392245392068e-07, | |
| "loss": 0.5236, | |
| "step": 942000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 8.933611668027626e-08, | |
| "loss": 0.5225, | |
| "step": 942500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 6.285831090663185e-08, | |
| "loss": 0.5259, | |
| "step": 943000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 3.6380505132987426e-08, | |
| "loss": 0.5201, | |
| "step": 943500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 9.902699359343012e-09, | |
| "loss": 0.5231, | |
| "step": 944000 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "step": 944187, | |
| "total_flos": 7.959444107321715e+18, | |
| "train_loss": 0.6286116640401175, | |
| "train_runtime": 410211.7129, | |
| "train_samples_per_second": 73.655, | |
| "train_steps_per_second": 2.302 | |
| } | |
| ], | |
| "max_steps": 944187, | |
| "num_train_epochs": 3, | |
| "total_flos": 7.959444107321715e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |