| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.979591836734694, | |
| "eval_steps": 500, | |
| "global_step": 915, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.7555, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.8402, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3e-06, | |
| "loss": 1.6331, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.9207, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5e-06, | |
| "loss": 1.6091, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6e-06, | |
| "loss": 1.7484, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 7e-06, | |
| "loss": 1.7995, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.7021, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9e-06, | |
| "loss": 1.7542, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1e-05, | |
| "loss": 1.69, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.99996987395229e-06, | |
| "loss": 1.6178, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.999879496172195e-06, | |
| "loss": 1.8358, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.999728867748802e-06, | |
| "loss": 1.8087, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.999517990497248e-06, | |
| "loss": 1.7274, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.999246866958693e-06, | |
| "loss": 1.9479, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.998915500400287e-06, | |
| "loss": 1.6799, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.998523894815137e-06, | |
| "loss": 1.657, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.998072054922257e-06, | |
| "loss": 1.6667, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.997559986166503e-06, | |
| "loss": 1.7878, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.99698769471852e-06, | |
| "loss": 1.6456, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.99635518747466e-06, | |
| "loss": 1.6963, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.9956624720569e-06, | |
| "loss": 1.4927, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.99490955681275e-06, | |
| "loss": 1.7288, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.994096450815157e-06, | |
| "loss": 1.8489, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.993223163862385e-06, | |
| "loss": 1.6431, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.992289706477912e-06, | |
| "loss": 1.5962, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.991296089910289e-06, | |
| "loss": 1.6551, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.99024232613301e-06, | |
| "loss": 1.7767, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.989128427844374e-06, | |
| "loss": 1.6875, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.98795440846732e-06, | |
| "loss": 1.7293, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.986720282149272e-06, | |
| "loss": 1.8094, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.985426063761973e-06, | |
| "loss": 1.7926, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.984071768901294e-06, | |
| "loss": 1.5949, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.982657413887055e-06, | |
| "loss": 1.4929, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.981183015762831e-06, | |
| "loss": 1.4097, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.979648592295735e-06, | |
| "loss": 1.6776, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.978054161976209e-06, | |
| "loss": 1.5844, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.976399744017812e-06, | |
| "loss": 1.6987, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.974685358356972e-06, | |
| "loss": 1.5576, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.972911025652754e-06, | |
| "loss": 1.6193, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.97107676728661e-06, | |
| "loss": 1.6233, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.969182605362125e-06, | |
| "loss": 1.6317, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.96722856270474e-06, | |
| "loss": 1.7902, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.965214662861492e-06, | |
| "loss": 1.4772, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.963140930100713e-06, | |
| "loss": 1.5038, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.961007389411758e-06, | |
| "loss": 1.5661, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.958814066504684e-06, | |
| "loss": 1.4406, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.956560987809948e-06, | |
| "loss": 1.6141, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.954248180478096e-06, | |
| "loss": 1.467, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.951875672379424e-06, | |
| "loss": 1.5145, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.94944349210365e-06, | |
| "loss": 1.4695, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.946951668959567e-06, | |
| "loss": 1.6004, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.944400232974683e-06, | |
| "loss": 1.4118, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.941789214894876e-06, | |
| "loss": 1.6423, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.939118646184007e-06, | |
| "loss": 1.5948, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.936388559023547e-06, | |
| "loss": 1.5026, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.93359898631219e-06, | |
| "loss": 1.3926, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.930749961665459e-06, | |
| "loss": 1.2179, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.92784151941529e-06, | |
| "loss": 1.5675, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.924873694609636e-06, | |
| "loss": 1.5331, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.921846523012028e-06, | |
| "loss": 1.4585, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.918760041101152e-06, | |
| "loss": 1.3743, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.915614286070408e-06, | |
| "loss": 1.5507, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.912409295827463e-06, | |
| "loss": 1.1745, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.909145108993794e-06, | |
| "loss": 1.4218, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.905821764904218e-06, | |
| "loss": 1.295, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.902439303606424e-06, | |
| "loss": 1.3678, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.898997765860492e-06, | |
| "loss": 1.4748, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.89549719313839e-06, | |
| "loss": 1.4412, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.891937627623486e-06, | |
| "loss": 1.3173, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.888319112210041e-06, | |
| "loss": 1.371, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.884641690502677e-06, | |
| "loss": 1.2869, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.88090540681587e-06, | |
| "loss": 1.3332, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.877110306173403e-06, | |
| "loss": 1.4246, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.873256434307828e-06, | |
| "loss": 1.4935, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.86934383765992e-06, | |
| "loss": 1.2236, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.865372563378102e-06, | |
| "loss": 1.3858, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.8613426593179e-06, | |
| "loss": 1.3957, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.857254174041342e-06, | |
| "loss": 1.295, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.853107156816393e-06, | |
| "loss": 1.3521, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.848901657616344e-06, | |
| "loss": 1.3231, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.844637727119227e-06, | |
| "loss": 1.1221, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.840315416707188e-06, | |
| "loss": 1.2744, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.83593477846588e-06, | |
| "loss": 1.46, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.831495865183832e-06, | |
| "loss": 1.3859, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.826998730351806e-06, | |
| "loss": 1.2469, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.822443428162165e-06, | |
| "loss": 1.1787, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.817830013508207e-06, | |
| "loss": 1.2879, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.81315854198351e-06, | |
| "loss": 1.2606, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.808429069881267e-06, | |
| "loss": 1.1404, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.803641654193599e-06, | |
| "loss": 1.3141, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.798796352610869e-06, | |
| "loss": 1.276, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.793893223520992e-06, | |
| "loss": 1.1235, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.78893232600873e-06, | |
| "loss": 1.25, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.783913719854977e-06, | |
| "loss": 1.3262, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.77883746553604e-06, | |
| "loss": 1.136, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.773703624222908e-06, | |
| "loss": 1.3941, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.768512257780524e-06, | |
| "loss": 1.2803, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.76326342876703e-06, | |
| "loss": 1.3667, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.757957200433011e-06, | |
| "loss": 1.2535, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 9.752593636720744e-06, | |
| "loss": 1.1887, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.747172802263423e-06, | |
| "loss": 1.2551, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.74169476238437e-06, | |
| "loss": 1.3562, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.736159583096262e-06, | |
| "loss": 1.264, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 9.730567331100333e-06, | |
| "loss": 1.2811, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.72491807378556e-06, | |
| "loss": 1.2777, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.719211879227863e-06, | |
| "loss": 1.1766, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.713448816189275e-06, | |
| "loss": 1.2182, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 9.707628954117122e-06, | |
| "loss": 1.2595, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.701752363143183e-06, | |
| "loss": 1.1055, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.69581911408284e-06, | |
| "loss": 1.1307, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.68982927843423e-06, | |
| "loss": 1.2346, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 9.683782928377385e-06, | |
| "loss": 1.1779, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.677680136773357e-06, | |
| "loss": 1.0882, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.67152097716334e-06, | |
| "loss": 1.1589, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 9.665305523767792e-06, | |
| "loss": 1.2117, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.659033851485527e-06, | |
| "loss": 1.234, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.652706035892828e-06, | |
| "loss": 1.1928, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.64632215324252e-06, | |
| "loss": 1.1914, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.639882280463071e-06, | |
| "loss": 1.2158, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.633386495157643e-06, | |
| "loss": 1.1415, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.62683487560317e-06, | |
| "loss": 1.1528, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.620227500749418e-06, | |
| "loss": 0.9916, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 9.613564450218019e-06, | |
| "loss": 1.118, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.606845804301523e-06, | |
| "loss": 1.2559, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.600071643962433e-06, | |
| "loss": 1.147, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 9.593242050832219e-06, | |
| "loss": 1.1697, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.586357107210338e-06, | |
| "loss": 1.2459, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.57941689606325e-06, | |
| "loss": 1.1382, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.572421501023403e-06, | |
| "loss": 1.3151, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.565371006388242e-06, | |
| "loss": 1.0931, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.558265497119182e-06, | |
| "loss": 1.0589, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.551105058840588e-06, | |
| "loss": 1.1522, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.54388977783874e-06, | |
| "loss": 1.0964, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 9.536619741060799e-06, | |
| "loss": 1.0978, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.529295036113755e-06, | |
| "loss": 1.1552, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.521915751263373e-06, | |
| "loss": 1.1139, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 9.514481975433125e-06, | |
| "loss": 1.1417, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.50699379820313e-06, | |
| "loss": 1.0902, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.499451309809058e-06, | |
| "loss": 1.1083, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.491854601141057e-06, | |
| "loss": 1.0415, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 9.484203763742647e-06, | |
| "loss": 1.1145, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.476498889809628e-06, | |
| "loss": 1.0296, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.468740072188961e-06, | |
| "loss": 1.1118, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.460927404377647e-06, | |
| "loss": 0.9663, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 9.45306098052161e-06, | |
| "loss": 1.0643, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.44514089541455e-06, | |
| "loss": 1.1021, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.437167244496814e-06, | |
| "loss": 1.091, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 9.429140123854239e-06, | |
| "loss": 1.0772, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.421059630216992e-06, | |
| "loss": 1.1406, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.412925860958404e-06, | |
| "loss": 1.0528, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.404738914093808e-06, | |
| "loss": 1.0789, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 9.396498888279344e-06, | |
| "loss": 1.0843, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.388205882810773e-06, | |
| "loss": 1.1047, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.37985999762229e-06, | |
| "loss": 1.058, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.371461333285308e-06, | |
| "loss": 1.0475, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 9.363009991007252e-06, | |
| "loss": 1.0543, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.354506072630337e-06, | |
| "loss": 0.9872, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.345949680630347e-06, | |
| "loss": 1.1849, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.337340918115385e-06, | |
| "loss": 1.0618, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.328679888824654e-06, | |
| "loss": 1.0944, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.319966697127181e-06, | |
| "loss": 1.0915, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.311201448020582e-06, | |
| "loss": 0.9994, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 9.30238424712978e-06, | |
| "loss": 0.9176, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.29351520070574e-06, | |
| "loss": 1.0784, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.28459441562419e-06, | |
| "loss": 1.062, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.275621999384327e-06, | |
| "loss": 1.0744, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.26659806010753e-06, | |
| "loss": 1.0254, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.257522706536046e-06, | |
| "loss": 0.9448, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.24839604803169e-06, | |
| "loss": 1.1357, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 9.239218194574522e-06, | |
| "loss": 1.0526, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.229989256761522e-06, | |
| "loss": 1.0148, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.220709345805259e-06, | |
| "loss": 1.0845, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.211378573532549e-06, | |
| "loss": 1.1066, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 9.201997052383107e-06, | |
| "loss": 1.011, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.192564895408195e-06, | |
| "loss": 1.0926, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.183082216269259e-06, | |
| "loss": 0.9629, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.173549129236554e-06, | |
| "loss": 1.0738, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.163965749187777e-06, | |
| "loss": 1.0841, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.154332191606671e-06, | |
| "loss": 1.0281, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.144648572581645e-06, | |
| "loss": 1.1063, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.134915008804364e-06, | |
| "loss": 1.0057, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.12513161756835e-06, | |
| "loss": 1.0011, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.11529851676757e-06, | |
| "loss": 1.1683, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.105415824895008e-06, | |
| "loss": 0.9943, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 9.095483661041244e-06, | |
| "loss": 1.0805, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.085502144893014e-06, | |
| "loss": 1.005, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.075471396731771e-06, | |
| "loss": 1.1173, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.065391537432234e-06, | |
| "loss": 1.0501, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 9.055262688460931e-06, | |
| "loss": 0.9866, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.045084971874738e-06, | |
| "loss": 1.1156, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 9.034858510319404e-06, | |
| "loss": 1.0603, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.02458342702808e-06, | |
| "loss": 0.9592, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.014259845819819e-06, | |
| "loss": 0.9943, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 9.003887891098108e-06, | |
| "loss": 0.8909, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 8.993467687849345e-06, | |
| "loss": 0.9349, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 8.982999361641344e-06, | |
| "loss": 1.1811, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 8.972483038621827e-06, | |
| "loss": 1.0453, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 8.961918845516893e-06, | |
| "loss": 0.9181, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 8.951306909629492e-06, | |
| "loss": 1.1174, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 8.940647358837904e-06, | |
| "loss": 0.752, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 8.92994032159418e-06, | |
| "loss": 0.97, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 8.919185926922608e-06, | |
| "loss": 0.9996, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 8.908384304418148e-06, | |
| "loss": 0.9168, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 8.89753558424488e-06, | |
| "loss": 0.9961, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 8.88663989713443e-06, | |
| "loss": 1.0325, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 8.87569737438439e-06, | |
| "loss": 0.9891, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 8.864708147856748e-06, | |
| "loss": 0.7389, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 8.85367234997629e-06, | |
| "loss": 0.9575, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 8.842590113729001e-06, | |
| "loss": 0.9032, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 8.831461572660476e-06, | |
| "loss": 0.9017, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 8.820286860874298e-06, | |
| "loss": 0.9411, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 8.809066113030427e-06, | |
| "loss": 0.9945, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 8.797799464343575e-06, | |
| "loss": 1.0241, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 8.786487050581583e-06, | |
| "loss": 1.0131, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 8.775129008063773e-06, | |
| "loss": 1.0196, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 8.763725473659325e-06, | |
| "loss": 1.0947, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 8.752276584785603e-06, | |
| "loss": 1.0206, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 8.740782479406515e-06, | |
| "loss": 0.9768, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 8.729243296030851e-06, | |
| "loss": 0.9138, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 8.717659173710603e-06, | |
| "loss": 0.9582, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 8.706030252039302e-06, | |
| "loss": 0.8377, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 8.694356671150327e-06, | |
| "loss": 0.9705, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 8.682638571715221e-06, | |
| "loss": 0.9781, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 8.670876094941991e-06, | |
| "loss": 0.9982, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 8.659069382573413e-06, | |
| "loss": 0.9935, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 8.64721857688532e-06, | |
| "loss": 0.9274, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 8.635323820684884e-06, | |
| "loss": 0.989, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 8.623385257308906e-06, | |
| "loss": 0.8472, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 8.611403030622074e-06, | |
| "loss": 0.9318, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 8.599377285015243e-06, | |
| "loss": 0.9046, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 8.587308165403686e-06, | |
| "loss": 0.812, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 8.575195817225357e-06, | |
| "loss": 0.9221, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 8.563040386439123e-06, | |
| "loss": 0.9355, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 8.55084201952302e-06, | |
| "loss": 1.0636, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 8.538600863472481e-06, | |
| "loss": 0.9307, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 8.526317065798568e-06, | |
| "loss": 0.9033, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 8.51399077452619e-06, | |
| "loss": 1.0066, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 8.501622138192323e-06, | |
| "loss": 0.9879, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 8.489211305844216e-06, | |
| "loss": 0.9377, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 8.476758427037607e-06, | |
| "loss": 0.9003, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 8.464263651834894e-06, | |
| "loss": 0.9243, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 8.451727130803362e-06, | |
| "loss": 0.9654, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 8.439149015013343e-06, | |
| "loss": 0.9913, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 8.4265294560364e-06, | |
| "loss": 0.7543, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 8.41386860594351e-06, | |
| "loss": 0.9519, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 8.401166617303221e-06, | |
| "loss": 1.0114, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 8.388423643179822e-06, | |
| "loss": 0.8913, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 8.375639837131488e-06, | |
| "loss": 0.8743, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 8.362815353208441e-06, | |
| "loss": 0.9434, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 8.349950345951086e-06, | |
| "loss": 1.0303, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 8.337044970388155e-06, | |
| "loss": 0.9072, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 8.324099382034828e-06, | |
| "loss": 0.9545, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 8.311113736890873e-06, | |
| "loss": 0.8508, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 8.298088191438753e-06, | |
| "loss": 0.9524, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 8.285022902641753e-06, | |
| "loss": 1.0097, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 8.271918027942076e-06, | |
| "loss": 0.9031, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 8.258773725258955e-06, | |
| "loss": 0.9991, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 8.245590152986746e-06, | |
| "loss": 0.8694, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 8.23236746999302e-06, | |
| "loss": 1.0646, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 8.219105835616645e-06, | |
| "loss": 1.0111, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 8.20580540966588e-06, | |
| "loss": 0.9125, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 8.192466352416425e-06, | |
| "loss": 0.9076, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 8.179088824609515e-06, | |
| "loss": 0.937, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 8.165672987449962e-06, | |
| "loss": 0.905, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 8.15221900260423e-06, | |
| "loss": 0.9514, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 8.138727032198473e-06, | |
| "loss": 1.1232, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 8.125197238816589e-06, | |
| "loss": 0.9331, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 8.111629785498256e-06, | |
| "loss": 0.8878, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 8.098024835736977e-06, | |
| "loss": 0.9899, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 8.084382553478095e-06, | |
| "loss": 0.9031, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 8.070703103116827e-06, | |
| "loss": 0.973, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 8.056986649496288e-06, | |
| "loss": 0.9401, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 8.043233357905488e-06, | |
| "loss": 0.844, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 8.029443394077356e-06, | |
| "loss": 0.9523, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 8.015616924186736e-06, | |
| "loss": 0.9861, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 8.001754114848382e-06, | |
| "loss": 1.0519, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 7.987855133114958e-06, | |
| "loss": 0.8462, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 7.97392014647502e-06, | |
| "loss": 0.8905, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 7.959949322850994e-06, | |
| "loss": 0.9227, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 7.945942830597163e-06, | |
| "loss": 0.9107, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 7.931900838497626e-06, | |
| "loss": 0.9799, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 7.917823515764272e-06, | |
| "loss": 0.9709, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 7.903711032034742e-06, | |
| "loss": 0.8537, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 7.889563557370378e-06, | |
| "loss": 0.9063, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 7.87538126225418e-06, | |
| "loss": 0.9615, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 7.861164317588746e-06, | |
| "loss": 0.9486, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 7.846912894694216e-06, | |
| "loss": 0.8401, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 7.83262716530621e-06, | |
| "loss": 1.0363, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 7.818307301573757e-06, | |
| "loss": 1.0342, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 7.80395347605721e-06, | |
| "loss": 0.9348, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 7.789565861726189e-06, | |
| "loss": 0.8611, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 7.77514463195747e-06, | |
| "loss": 0.989, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 7.76068996053292e-06, | |
| "loss": 0.955, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 7.746202021637385e-06, | |
| "loss": 0.9545, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 7.7316809898566e-06, | |
| "loss": 0.9366, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 7.717127040175084e-06, | |
| "loss": 0.8091, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 7.70254034797403e-06, | |
| "loss": 0.9747, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 7.687921089029194e-06, | |
| "loss": 1.0307, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 7.67326943950877e-06, | |
| "loss": 1.0158, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 7.658585575971278e-06, | |
| "loss": 0.8534, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 7.643869675363425e-06, | |
| "loss": 0.8246, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 7.62912191501798e-06, | |
| "loss": 0.8728, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 7.614342472651639e-06, | |
| "loss": 0.9968, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 7.599531526362873e-06, | |
| "loss": 0.9028, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 7.5846892546297925e-06, | |
| "loss": 0.8097, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 7.569815836307994e-06, | |
| "loss": 0.9532, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 7.5549114506283995e-06, | |
| "loss": 1.0243, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 7.539976277195102e-06, | |
| "loss": 0.9622, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.525010495983202e-06, | |
| "loss": 0.9323, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.510014287336638e-06, | |
| "loss": 0.9608, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.494987831966003e-06, | |
| "loss": 0.9235, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.4799313109463844e-06, | |
| "loss": 0.9056, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.464844905715169e-06, | |
| "loss": 0.988, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.449728798069864e-06, | |
| "loss": 0.9618, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.434583170165903e-06, | |
| "loss": 0.9671, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.419408204514445e-06, | |
| "loss": 0.9739, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 7.40420408398019e-06, | |
| "loss": 0.8321, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.388970991779159e-06, | |
| "loss": 0.9047, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 7.373709111476498e-06, | |
| "loss": 1.0659, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.358418626984262e-06, | |
| "loss": 0.899, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 7.343099722559195e-06, | |
| "loss": 0.9948, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.327752582800518e-06, | |
| "loss": 0.9291, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 7.312377392647694e-06, | |
| "loss": 0.9076, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.296974337378209e-06, | |
| "loss": 0.829, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 7.281543602605333e-06, | |
| "loss": 0.9306, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.266085374275891e-06, | |
| "loss": 0.9055, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.250599838668007e-06, | |
| "loss": 0.8165, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 7.235087182388877e-06, | |
| "loss": 0.939, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.219547592372512e-06, | |
| "loss": 0.9337, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 7.203981255877482e-06, | |
| "loss": 0.8893, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.188388360484667e-06, | |
| "loss": 0.9036, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.17276909409499e-06, | |
| "loss": 0.9001, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.1571236449271575e-06, | |
| "loss": 0.8759, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.141452201515386e-06, | |
| "loss": 0.9875, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 7.125754952707143e-06, | |
| "loss": 0.8603, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 7.110032087660846e-06, | |
| "loss": 0.8247, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 7.094283795843616e-06, | |
| "loss": 0.8525, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 7.0785102670289644e-06, | |
| "loss": 0.8645, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 7.062711691294525e-06, | |
| "loss": 0.8738, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.046888259019757e-06, | |
| "loss": 1.0695, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.031040160883651e-06, | |
| "loss": 0.8732, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.015167587862431e-06, | |
| "loss": 0.892, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.999270731227252e-06, | |
| "loss": 0.8494, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.983349782541901e-06, | |
| "loss": 0.9128, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.96740493366048e-06, | |
| "loss": 0.8998, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.9514363767251024e-06, | |
| "loss": 0.9055, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.9354443041635675e-06, | |
| "loss": 0.8089, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.919428908687057e-06, | |
| "loss": 0.891, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.903390383287795e-06, | |
| "loss": 1.0194, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.887328921236733e-06, | |
| "loss": 0.9762, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.871244716081223e-06, | |
| "loss": 0.943, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.855137961642676e-06, | |
| "loss": 0.7557, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.839008852014233e-06, | |
| "loss": 0.8784, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.822857581558423e-06, | |
| "loss": 0.8028, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.806684344904828e-06, | |
| "loss": 0.7966, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.790489336947723e-06, | |
| "loss": 0.9224, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.774272752843745e-06, | |
| "loss": 0.7707, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.758034788009523e-06, | |
| "loss": 0.8538, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.741775638119345e-06, | |
| "loss": 0.7929, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.7254954991027765e-06, | |
| "loss": 0.8583, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.709194567142316e-06, | |
| "loss": 0.9552, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.6928730386710285e-06, | |
| "loss": 0.8742, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.676531110370168e-06, | |
| "loss": 0.8376, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.66016897916682e-06, | |
| "loss": 0.8969, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.643786842231523e-06, | |
| "loss": 0.8108, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.627384896975896e-06, | |
| "loss": 0.8374, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.6109633410502475e-06, | |
| "loss": 0.8357, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.59452237234121e-06, | |
| "loss": 0.9496, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.57806218896935e-06, | |
| "loss": 1.0033, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.5615829892867715e-06, | |
| "loss": 0.8199, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.545084971874738e-06, | |
| "loss": 0.9517, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 6.528568335541274e-06, | |
| "loss": 0.9681, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 6.512033279318768e-06, | |
| "loss": 0.8775, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 6.495480002461577e-06, | |
| "loss": 0.8583, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.478908704443629e-06, | |
| "loss": 0.8624, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.462319584956e-06, | |
| "loss": 0.7446, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.445712843904542e-06, | |
| "loss": 0.7195, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.4290886814074405e-06, | |
| "loss": 0.9782, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.412447297792818e-06, | |
| "loss": 0.8978, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.3957888935963265e-06, | |
| "loss": 0.881, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.379113669558713e-06, | |
| "loss": 0.9469, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.3624218266234176e-06, | |
| "loss": 0.8274, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.345713565934142e-06, | |
| "loss": 0.8372, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.328989088832431e-06, | |
| "loss": 1.0378, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 6.312248596855241e-06, | |
| "loss": 0.8135, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 6.295492291732519e-06, | |
| "loss": 0.8383, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 6.278720375384759e-06, | |
| "loss": 0.8324, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 6.261933049920587e-06, | |
| "loss": 0.8466, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 6.245130517634307e-06, | |
| "loss": 0.8675, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 6.228312981003476e-06, | |
| "loss": 0.9011, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 6.211480642686458e-06, | |
| "loss": 0.8772, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 6.194633705519983e-06, | |
| "loss": 1.0046, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 6.177772372516706e-06, | |
| "loss": 0.7755, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 6.160896846862754e-06, | |
| "loss": 0.8375, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 6.1440073319152856e-06, | |
| "loss": 0.8476, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 6.127104031200032e-06, | |
| "loss": 0.9187, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 6.110187148408852e-06, | |
| "loss": 0.9529, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 6.093256887397272e-06, | |
| "loss": 0.9052, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 6.076313452182033e-06, | |
| "loss": 0.8845, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 6.05935704693863e-06, | |
| "loss": 0.806, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 6.042387875998852e-06, | |
| "loss": 0.9245, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 6.02540614384832e-06, | |
| "loss": 0.9109, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 6.008412055124024e-06, | |
| "loss": 0.8871, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.991405814611855e-06, | |
| "loss": 0.9606, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.974387627244137e-06, | |
| "loss": 0.9259, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.957357698097163e-06, | |
| "loss": 0.9331, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.940316232388711e-06, | |
| "loss": 0.8568, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.923263435475589e-06, | |
| "loss": 0.9379, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.9061995128511455e-06, | |
| "loss": 0.8497, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.889124670142797e-06, | |
| "loss": 0.8845, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.872039113109557e-06, | |
| "loss": 0.9423, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.854943047639547e-06, | |
| "loss": 0.9536, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.8378366797475184e-06, | |
| "loss": 0.8411, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.820720215572375e-06, | |
| "loss": 0.8476, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.803593861374687e-06, | |
| "loss": 0.922, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.786457823534193e-06, | |
| "loss": 0.9568, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.769312308547334e-06, | |
| "loss": 0.7422, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.752157523024753e-06, | |
| "loss": 0.8765, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 5.734993673688801e-06, | |
| "loss": 0.9319, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 5.71782096737106e-06, | |
| "loss": 0.7123, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.7006396110098306e-06, | |
| "loss": 0.8822, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.683449811647664e-06, | |
| "loss": 0.8585, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 5.666251776428844e-06, | |
| "loss": 0.8678, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 5.6490457125969035e-06, | |
| "loss": 0.7982, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 5.631831827492121e-06, | |
| "loss": 0.8457, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 5.614610328549029e-06, | |
| "loss": 0.732, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 5.5973814232939e-06, | |
| "loss": 0.8866, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 5.5801453193422715e-06, | |
| "loss": 0.8772, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 5.562902224396416e-06, | |
| "loss": 1.0722, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 5.54565234624285e-06, | |
| "loss": 0.8542, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 5.52839589274984e-06, | |
| "loss": 0.7729, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.511133071864881e-06, | |
| "loss": 0.9159, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.493864091612197e-06, | |
| "loss": 0.9416, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 5.476589160090238e-06, | |
| "loss": 0.9376, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 5.459308485469171e-06, | |
| "loss": 0.9561, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 5.442022275988365e-06, | |
| "loss": 0.881, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 5.4247307399538876e-06, | |
| "loss": 0.9091, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 5.407434085735997e-06, | |
| "loss": 0.847, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 5.390132521766626e-06, | |
| "loss": 0.9168, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 5.372826256536867e-06, | |
| "loss": 0.8412, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 5.355515498594472e-06, | |
| "loss": 0.8875, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 5.338200456541329e-06, | |
| "loss": 0.8385, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 5.32088133903095e-06, | |
| "loss": 0.8067, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 5.30355835476596e-06, | |
| "loss": 0.8311, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 5.286231712495578e-06, | |
| "loss": 0.9589, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 5.2689016210131065e-06, | |
| "loss": 0.8565, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.251568289153407e-06, | |
| "loss": 0.862, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.234231925790396e-06, | |
| "loss": 0.7935, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 5.216892739834519e-06, | |
| "loss": 0.9223, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 5.199550940230228e-06, | |
| "loss": 0.8139, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 5.182206735953479e-06, | |
| "loss": 0.926, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 5.1648603360092034e-06, | |
| "loss": 0.9354, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 5.147511949428787e-06, | |
| "loss": 0.9393, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 5.13016178526756e-06, | |
| "loss": 0.9063, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 5.112810052602274e-06, | |
| "loss": 0.8157, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 5.095456960528574e-06, | |
| "loss": 0.7404, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 5.078102718158497e-06, | |
| "loss": 0.8, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 5.060747534617937e-06, | |
| "loss": 0.8425, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 5.043391619044122e-06, | |
| "loss": 0.9117, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.026035180583116e-06, | |
| "loss": 0.9421, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.008678428387273e-06, | |
| "loss": 0.8304, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.99132157161273e-06, | |
| "loss": 0.9645, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.973964819416885e-06, | |
| "loss": 0.7235, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.956608380955877e-06, | |
| "loss": 0.9902, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.939252465382066e-06, | |
| "loss": 0.892, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.9218972818415035e-06, | |
| "loss": 0.7766, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.904543039471427e-06, | |
| "loss": 1.0028, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.887189947397728e-06, | |
| "loss": 0.8625, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.869838214732441e-06, | |
| "loss": 1.0353, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.852488050571214e-06, | |
| "loss": 0.8195, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.8351396639908e-06, | |
| "loss": 0.7516, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.8177932640465216e-06, | |
| "loss": 0.9172, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.800449059769774e-06, | |
| "loss": 0.9456, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.783107260165483e-06, | |
| "loss": 1.0406, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.7657680742096044e-06, | |
| "loss": 1.0074, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.748431710846594e-06, | |
| "loss": 0.6638, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.731098378986897e-06, | |
| "loss": 0.9923, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.7137682875044246e-06, | |
| "loss": 0.8311, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.696441645234042e-06, | |
| "loss": 0.757, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.679118660969051e-06, | |
| "loss": 0.8805, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.6617995434586714e-06, | |
| "loss": 0.9096, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.64448450140553e-06, | |
| "loss": 0.9217, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.627173743463134e-06, | |
| "loss": 0.8868, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.609867478233377e-06, | |
| "loss": 0.8931, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.592565914264004e-06, | |
| "loss": 0.8445, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.575269260046112e-06, | |
| "loss": 0.8162, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.557977724011636e-06, | |
| "loss": 0.8142, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.540691514530831e-06, | |
| "loss": 0.8531, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.523410839909764e-06, | |
| "loss": 0.9631, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.506135908387805e-06, | |
| "loss": 0.7765, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.488866928135121e-06, | |
| "loss": 0.9362, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.4716041072501604e-06, | |
| "loss": 0.8497, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.45434765375715e-06, | |
| "loss": 0.8659, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.437097775603587e-06, | |
| "loss": 0.9575, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.41985468065773e-06, | |
| "loss": 0.8198, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.402618576706101e-06, | |
| "loss": 0.8925, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.385389671450974e-06, | |
| "loss": 0.8702, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.36816817250788e-06, | |
| "loss": 0.7818, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.350954287403099e-06, | |
| "loss": 0.7461, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.333748223571158e-06, | |
| "loss": 0.8872, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.316550188352337e-06, | |
| "loss": 0.9121, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.299360388990171e-06, | |
| "loss": 0.7544, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.282179032628943e-06, | |
| "loss": 0.7822, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.265006326311199e-06, | |
| "loss": 0.9713, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.247842476975249e-06, | |
| "loss": 0.8752, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.230687691452667e-06, | |
| "loss": 0.97, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.213542176465809e-06, | |
| "loss": 0.8821, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.196406138625315e-06, | |
| "loss": 0.8736, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.179279784427625e-06, | |
| "loss": 0.7504, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.1621633202524815e-06, | |
| "loss": 0.7806, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.145056952360456e-06, | |
| "loss": 0.7688, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.127960886890445e-06, | |
| "loss": 0.8505, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.110875329857205e-06, | |
| "loss": 0.8255, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.093800487148857e-06, | |
| "loss": 0.8624, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.076736564524412e-06, | |
| "loss": 0.9774, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.0596837676112905e-06, | |
| "loss": 0.734, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.04264230190284e-06, | |
| "loss": 0.8445, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.025612372755865e-06, | |
| "loss": 0.8627, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.008594185388146e-06, | |
| "loss": 0.9884, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.991587944875977e-06, | |
| "loss": 0.8007, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.97459385615168e-06, | |
| "loss": 0.8553, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.957612124001151e-06, | |
| "loss": 0.8161, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.940642953061371e-06, | |
| "loss": 0.9705, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 3.9236865478179685e-06, | |
| "loss": 0.8085, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 3.906743112602729e-06, | |
| "loss": 0.8702, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 3.889812851591149e-06, | |
| "loss": 0.8137, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 3.872895968799969e-06, | |
| "loss": 0.7632, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 3.855992668084716e-06, | |
| "loss": 0.8425, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 3.839103153137247e-06, | |
| "loss": 0.842, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 3.822227627483295e-06, | |
| "loss": 0.8439, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 3.8053662944800177e-06, | |
| "loss": 0.8059, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 3.7885193573135437e-06, | |
| "loss": 0.8929, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 3.771687018996525e-06, | |
| "loss": 0.8088, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 3.7548694823656945e-06, | |
| "loss": 0.8736, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.7380669500794153e-06, | |
| "loss": 0.886, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 3.721279624615243e-06, | |
| "loss": 0.8559, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 3.704507708267483e-06, | |
| "loss": 0.9221, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 3.6877514031447597e-06, | |
| "loss": 0.8069, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 3.671010911167572e-06, | |
| "loss": 0.9423, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 3.6542864340658602e-06, | |
| "loss": 0.9176, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 3.637578173376584e-06, | |
| "loss": 0.8804, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 3.620886330441289e-06, | |
| "loss": 0.8707, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 3.6042111064036756e-06, | |
| "loss": 0.7642, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 3.5875527022071808e-06, | |
| "loss": 1.0431, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 3.5709113185925615e-06, | |
| "loss": 0.9163, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.554287156095459e-06, | |
| "loss": 0.8734, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.5376804150440002e-06, | |
| "loss": 0.8863, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 3.5210912955563735e-06, | |
| "loss": 0.7968, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 3.5045199975384225e-06, | |
| "loss": 0.876, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.4879667206812316e-06, | |
| "loss": 0.8863, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.471431664458729e-06, | |
| "loss": 0.8485, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 3.4549150281252635e-06, | |
| "loss": 0.8416, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 3.43841701071323e-06, | |
| "loss": 0.8251, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 3.4219378110306523e-06, | |
| "loss": 0.9206, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.4054776276587897e-06, | |
| "loss": 0.8879, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.3890366589497538e-06, | |
| "loss": 1.0349, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 3.3726151030241074e-06, | |
| "loss": 0.8465, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 3.3562131577684778e-06, | |
| "loss": 0.8563, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 3.3398310208331806e-06, | |
| "loss": 0.842, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 3.323468889629834e-06, | |
| "loss": 0.8472, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.3071269613289735e-06, | |
| "loss": 0.8492, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.2908054328576854e-06, | |
| "loss": 0.8731, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.2745045008972255e-06, | |
| "loss": 0.7911, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.2582243618806574e-06, | |
| "loss": 1.068, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 3.2419652119904777e-06, | |
| "loss": 0.8996, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.2257272471562574e-06, | |
| "loss": 0.9333, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.2095106630522778e-06, | |
| "loss": 0.8469, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.193315655095175e-06, | |
| "loss": 0.809, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.177142418441578e-06, | |
| "loss": 0.7341, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 3.160991147985769e-06, | |
| "loss": 0.7984, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 3.144862038357326e-06, | |
| "loss": 0.8389, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.1287552839187784e-06, | |
| "loss": 0.9403, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.1126710787632662e-06, | |
| "loss": 0.8135, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 3.096609616712207e-06, | |
| "loss": 0.7741, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 3.080571091312945e-06, | |
| "loss": 0.835, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 3.064555695836433e-06, | |
| "loss": 0.9042, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.0485636232748992e-06, | |
| "loss": 0.8505, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.032595066339521e-06, | |
| "loss": 0.8271, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.0166502174581012e-06, | |
| "loss": 0.7945, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.00072926877275e-06, | |
| "loss": 0.8788, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 2.9848324121375705e-06, | |
| "loss": 0.8791, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 2.9689598391163497e-06, | |
| "loss": 0.788, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 2.9531117409802432e-06, | |
| "loss": 0.8701, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 2.937288308705475e-06, | |
| "loss": 0.8232, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 2.921489732971038e-06, | |
| "loss": 0.8906, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 2.9057162041563867e-06, | |
| "loss": 0.8842, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 2.8899679123391545e-06, | |
| "loss": 0.8328, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 2.8742450472928595e-06, | |
| "loss": 0.773, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 2.858547798484613e-06, | |
| "loss": 0.8197, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 2.8428763550728446e-06, | |
| "loss": 0.8593, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 2.8272309059050107e-06, | |
| "loss": 0.7967, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 2.8116116395153343e-06, | |
| "loss": 0.9606, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 2.7960187441225185e-06, | |
| "loss": 0.9348, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 2.7804524076274898e-06, | |
| "loss": 0.8325, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 2.764912817611124e-06, | |
| "loss": 0.8465, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 2.7494001613319932e-06, | |
| "loss": 0.8464, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 2.733914625724112e-06, | |
| "loss": 0.969, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 2.7184563973946687e-06, | |
| "loss": 0.9087, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 2.7030256626217932e-06, | |
| "loss": 0.9559, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 2.687622607352307e-06, | |
| "loss": 0.934, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 2.6722474171994826e-06, | |
| "loss": 0.9529, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 2.656900277440807e-06, | |
| "loss": 0.9131, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.6415813730157408e-06, | |
| "loss": 0.9768, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.6262908885235046e-06, | |
| "loss": 0.9021, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 2.611029008220842e-06, | |
| "loss": 0.7654, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 2.5957959160198115e-06, | |
| "loss": 0.8602, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 2.580591795485555e-06, | |
| "loss": 0.7933, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 2.565416829834101e-06, | |
| "loss": 0.9777, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 2.550271201930136e-06, | |
| "loss": 0.7869, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 2.535155094284832e-06, | |
| "loss": 0.738, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 2.5200686890536177e-06, | |
| "loss": 0.9764, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 2.505012168033999e-06, | |
| "loss": 0.8893, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 2.489985712663364e-06, | |
| "loss": 0.9448, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 2.474989504016798e-06, | |
| "loss": 0.8546, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 2.460023722804899e-06, | |
| "loss": 0.9707, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 2.4450885493716026e-06, | |
| "loss": 0.7581, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 2.430184163692008e-06, | |
| "loss": 0.8602, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 2.4153107453702075e-06, | |
| "loss": 0.8751, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 2.4004684736371276e-06, | |
| "loss": 0.8511, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 2.385657527348364e-06, | |
| "loss": 0.9102, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 2.3708780849820218e-06, | |
| "loss": 0.7458, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 2.356130324636578e-06, | |
| "loss": 0.8914, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 2.341414424028723e-06, | |
| "loss": 0.8849, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 2.32673056049123e-06, | |
| "loss": 0.8613, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 2.312078910970808e-06, | |
| "loss": 0.8999, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 2.297459652025972e-06, | |
| "loss": 0.7714, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 2.2828729598249165e-06, | |
| "loss": 0.7473, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 2.2683190101434015e-06, | |
| "loss": 0.9195, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 2.253797978362617e-06, | |
| "loss": 0.8738, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 2.239310039467082e-06, | |
| "loss": 0.8235, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 2.2248553680425308e-06, | |
| "loss": 0.8386, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 2.2104341382738127e-06, | |
| "loss": 1.005, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 2.19604652394279e-06, | |
| "loss": 0.9029, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 2.1816926984262454e-06, | |
| "loss": 0.8638, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 2.1673728346937904e-06, | |
| "loss": 0.8138, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 2.1530871053057843e-06, | |
| "loss": 0.8106, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 2.138835682411257e-06, | |
| "loss": 0.9432, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 2.1246187377458227e-06, | |
| "loss": 0.8596, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 2.1104364426296237e-06, | |
| "loss": 0.8068, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 2.0962889679652576e-06, | |
| "loss": 0.7555, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 2.082176484235728e-06, | |
| "loss": 0.8644, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 2.0680991615023765e-06, | |
| "loss": 0.8524, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 2.05405716940284e-06, | |
| "loss": 0.7569, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 2.040050677149008e-06, | |
| "loss": 0.7284, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 2.026079853524981e-06, | |
| "loss": 0.9232, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 2.0121448668850424e-06, | |
| "loss": 0.8451, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 1.998245885151619e-06, | |
| "loss": 0.9504, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 1.984383075813268e-06, | |
| "loss": 0.9145, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 1.970556605922645e-06, | |
| "loss": 0.8912, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 1.956766642094513e-06, | |
| "loss": 0.8934, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 1.9430133505037133e-06, | |
| "loss": 0.9124, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 1.9292968968831734e-06, | |
| "loss": 0.8914, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 1.9156174465219073e-06, | |
| "loss": 0.8909, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 1.9019751642630252e-06, | |
| "loss": 0.8434, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 1.8883702145017452e-06, | |
| "loss": 0.9324, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 1.8748027611834135e-06, | |
| "loss": 0.9103, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 1.861272967801529e-06, | |
| "loss": 0.7558, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 1.84778099739577e-06, | |
| "loss": 0.9089, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 1.8343270125500379e-06, | |
| "loss": 0.8667, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 1.820911175390488e-06, | |
| "loss": 0.8894, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 1.807533647583577e-06, | |
| "loss": 0.8737, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 1.794194590334123e-06, | |
| "loss": 0.9017, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 1.780894164383355e-06, | |
| "loss": 0.8644, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 1.7676325300069824e-06, | |
| "loss": 0.8413, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 1.754409847013257e-06, | |
| "loss": 0.8635, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 1.741226274741048e-06, | |
| "loss": 1.03, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 1.728081972057925e-06, | |
| "loss": 0.8222, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 1.714977097358248e-06, | |
| "loss": 0.9032, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 1.7019118085612474e-06, | |
| "loss": 0.9348, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 1.6888862631091284e-06, | |
| "loss": 0.8573, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 1.6759006179651726e-06, | |
| "loss": 0.8666, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 1.6629550296118462e-06, | |
| "loss": 0.897, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 1.6500496540489142e-06, | |
| "loss": 1.02, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 1.6371846467915603e-06, | |
| "loss": 0.8099, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 1.624360162868514e-06, | |
| "loss": 0.7988, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 1.6115763568201786e-06, | |
| "loss": 0.8619, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 1.59883338269678e-06, | |
| "loss": 0.8467, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 1.5861313940564915e-06, | |
| "loss": 0.9343, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 1.5734705439636017e-06, | |
| "loss": 0.8878, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 1.5608509849866594e-06, | |
| "loss": 0.7844, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 1.5482728691966377e-06, | |
| "loss": 0.8583, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 1.535736348165106e-06, | |
| "loss": 0.8883, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 1.5232415729623973e-06, | |
| "loss": 0.9003, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 1.5107886941557853e-06, | |
| "loss": 0.7417, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 1.4983778618076783e-06, | |
| "loss": 0.842, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.4860092254738111e-06, | |
| "loss": 0.9448, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.4736829342014336e-06, | |
| "loss": 0.8548, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 1.461399136527522e-06, | |
| "loss": 0.6608, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 1.4491579804769817e-06, | |
| "loss": 0.8563, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 1.4369596135608789e-06, | |
| "loss": 0.6762, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 1.4248041827746445e-06, | |
| "loss": 0.8582, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 1.4126918345963136e-06, | |
| "loss": 0.9294, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 1.4006227149847584e-06, | |
| "loss": 0.8519, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 1.3885969693779277e-06, | |
| "loss": 0.7379, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 1.3766147426910959e-06, | |
| "loss": 0.796, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 1.364676179315117e-06, | |
| "loss": 0.7911, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 1.3527814231146813e-06, | |
| "loss": 0.8022, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 1.3409306174265873e-06, | |
| "loss": 0.7736, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 1.3291239050580085e-06, | |
| "loss": 0.8208, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 1.3173614282847807e-06, | |
| "loss": 0.9625, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 1.3056433288496739e-06, | |
| "loss": 0.8483, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 1.2939697479606993e-06, | |
| "loss": 0.8063, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 1.2823408262893971e-06, | |
| "loss": 0.9138, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 1.2707567039691505e-06, | |
| "loss": 0.8316, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 1.2592175205934865e-06, | |
| "loss": 0.7809, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 1.2477234152143996e-06, | |
| "loss": 0.8831, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 1.236274526340675e-06, | |
| "loss": 0.7254, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 1.2248709919362262e-06, | |
| "loss": 0.8232, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 1.213512949418419e-06, | |
| "loss": 0.7897, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 1.2022005356564264e-06, | |
| "loss": 0.8565, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 1.1909338869695747e-06, | |
| "loss": 0.9481, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 1.1797131391257027e-06, | |
| "loss": 0.8168, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 1.168538427339524e-06, | |
| "loss": 0.8379, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 1.1574098862709993e-06, | |
| "loss": 0.9077, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 1.1463276500237113e-06, | |
| "loss": 0.8559, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 1.1352918521432515e-06, | |
| "loss": 0.7869, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 1.1243026256156103e-06, | |
| "loss": 0.7304, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 1.1133601028655717e-06, | |
| "loss": 0.9448, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 1.1024644157551206e-06, | |
| "loss": 0.8451, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 1.091615695581853e-06, | |
| "loss": 0.8134, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 1.080814073077393e-06, | |
| "loss": 0.8453, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 1.0700596784058205e-06, | |
| "loss": 0.9098, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 1.059352641162098e-06, | |
| "loss": 0.8295, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 1.0486930903705095e-06, | |
| "loss": 0.959, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 1.0380811544831087e-06, | |
| "loss": 0.9064, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 1.0275169613781732e-06, | |
| "loss": 0.9544, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 1.0170006383586561e-06, | |
| "loss": 0.8453, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 1.006532312150658e-06, | |
| "loss": 0.8531, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 9.961121089018933e-07, | |
| "loss": 0.8159, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 9.857401541801814e-07, | |
| "loss": 0.8243, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 9.754165729719222e-07, | |
| "loss": 0.851, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 9.651414896805966e-07, | |
| "loss": 0.778, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 9.549150281252633e-07, | |
| "loss": 0.9498, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 9.447373115390702e-07, | |
| "loss": 0.8119, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 9.346084625677676e-07, | |
| "loss": 0.9721, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 9.245286032682299e-07, | |
| "loss": 0.8062, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 9.144978551069867e-07, | |
| "loss": 0.8626, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 9.045163389587569e-07, | |
| "loss": 0.9832, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 8.945841751049916e-07, | |
| "loss": 0.8116, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 8.847014832324313e-07, | |
| "loss": 0.956, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 8.748683824316518e-07, | |
| "loss": 0.7777, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 8.650849911956388e-07, | |
| "loss": 0.8544, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 8.553514274183561e-07, | |
| "loss": 0.7898, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 8.45667808393329e-07, | |
| "loss": 0.8439, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 8.360342508122238e-07, | |
| "loss": 0.8586, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 8.264508707634472e-07, | |
| "loss": 0.9113, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 8.169177837307418e-07, | |
| "loss": 0.7554, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 8.074351045918055e-07, | |
| "loss": 1.0145, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 7.980029476168943e-07, | |
| "loss": 0.8817, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 7.886214264674525e-07, | |
| "loss": 0.827, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 7.792906541947421e-07, | |
| "loss": 0.8639, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 7.700107432384785e-07, | |
| "loss": 0.8534, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 7.60781805425479e-07, | |
| "loss": 0.88, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 7.516039519683105e-07, | |
| "loss": 0.8416, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 7.424772934639552e-07, | |
| "loss": 0.8425, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 7.334019398924714e-07, | |
| "loss": 0.7865, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 7.243780006156737e-07, | |
| "loss": 0.937, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 7.154055843758118e-07, | |
| "loss": 0.884, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 7.064847992942614e-07, | |
| "loss": 0.884, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 6.976157528702221e-07, | |
| "loss": 0.9562, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 6.887985519794188e-07, | |
| "loss": 0.7733, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 6.80033302872819e-07, | |
| "loss": 0.7185, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 6.713201111753487e-07, | |
| "loss": 0.884, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 6.626590818846163e-07, | |
| "loss": 0.7337, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 6.540503193696551e-07, | |
| "loss": 0.8191, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 6.454939273696631e-07, | |
| "loss": 0.7203, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 6.369900089927488e-07, | |
| "loss": 0.8857, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 6.285386667146937e-07, | |
| "loss": 0.7981, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 6.201400023777105e-07, | |
| "loss": 0.8905, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 6.117941171892272e-07, | |
| "loss": 0.8076, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 6.035011117206574e-07, | |
| "loss": 0.8366, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 5.952610859061925e-07, | |
| "loss": 0.8544, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 5.870741390415963e-07, | |
| "loss": 0.9062, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 5.789403697830104e-07, | |
| "loss": 0.878, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 5.708598761457623e-07, | |
| "loss": 0.8995, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 5.628327555031865e-07, | |
| "loss": 0.7554, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 5.548591045854518e-07, | |
| "loss": 0.8262, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 5.46939019478393e-07, | |
| "loss": 0.8744, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 5.390725956223531e-07, | |
| "loss": 0.8344, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 5.312599278110403e-07, | |
| "loss": 1.0051, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 5.235011101903725e-07, | |
| "loss": 0.9111, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 5.157962362573543e-07, | |
| "loss": 0.8366, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 5.081453988589447e-07, | |
| "loss": 0.7421, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 5.005486901909429e-07, | |
| "loss": 0.924, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 4.930062017968706e-07, | |
| "loss": 0.7788, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 4.855180245668755e-07, | |
| "loss": 0.7902, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 4.780842487366283e-07, | |
| "loss": 0.7286, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 4.7070496388624544e-07, | |
| "loss": 0.9672, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 4.6338025893920167e-07, | |
| "loss": 0.7559, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 4.5611022216126124e-07, | |
| "loss": 0.8186, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 4.488949411594135e-07, | |
| "loss": 0.7844, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 4.4173450288081844e-07, | |
| "loss": 0.7213, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 4.346289936117587e-07, | |
| "loss": 0.9005, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 4.275784989765985e-07, | |
| "loss": 0.9292, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 4.205831039367525e-07, | |
| "loss": 0.8719, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 4.13642892789663e-07, | |
| "loss": 0.8004, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 4.067579491677831e-07, | |
| "loss": 0.9129, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 3.999283560375683e-07, | |
| "loss": 0.8006, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 3.93154195698478e-07, | |
| "loss": 0.9297, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 3.8643554978198385e-07, | |
| "loss": 0.974, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 3.7977249925058303e-07, | |
| "loss": 0.9562, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 3.7316512439682926e-07, | |
| "loss": 0.9122, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 3.6661350484235857e-07, | |
| "loss": 0.8722, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 3.6011771953693044e-07, | |
| "loss": 0.9109, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 3.5367784675747975e-07, | |
| "loss": 0.9652, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 3.472939641071743e-07, | |
| "loss": 0.8378, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 3.409661485144744e-07, | |
| "loss": 0.8497, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 3.346944762322102e-07, | |
| "loss": 0.9019, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 3.284790228366602e-07, | |
| "loss": 0.7724, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 3.2231986322664386e-07, | |
| "loss": 0.7944, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 3.162170716226148e-07, | |
| "loss": 0.9376, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 3.1017072156576957e-07, | |
| "loss": 0.8481, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 3.0418088591716076e-07, | |
| "loss": 0.87, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 2.982476368568177e-07, | |
| "loss": 0.9193, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 2.923710458828777e-07, | |
| "loss": 0.9044, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 2.86551183810726e-07, | |
| "loss": 0.8131, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 2.80788120772138e-07, | |
| "loss": 0.7668, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 2.7508192621443994e-07, | |
| "loss": 0.8382, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 2.6943266889966624e-07, | |
| "loss": 0.8403, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 2.638404169037373e-07, | |
| "loss": 0.923, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 2.583052376156314e-07, | |
| "loss": 0.8127, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 2.528271977365787e-07, | |
| "loss": 0.8425, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 2.474063632792556e-07, | |
| "loss": 0.8664, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 2.4204279956698994e-07, | |
| "loss": 0.9305, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 2.3673657123297166e-07, | |
| "loss": 0.779, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 2.3148774221947667e-07, | |
| "loss": 0.8542, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 2.2629637577709252e-07, | |
| "loss": 0.8933, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 2.2116253446396175e-07, | |
| "loss": 0.8068, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 2.1608628014502364e-07, | |
| "loss": 0.75, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 2.1106767399126982e-07, | |
| "loss": 0.9549, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 2.061067764790087e-07, | |
| "loss": 0.9055, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 2.0120364738913212e-07, | |
| "loss": 0.8934, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 1.9635834580640223e-07, | |
| "loss": 0.8489, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 1.915709301187335e-07, | |
| "loss": 0.7928, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 1.8684145801649067e-07, | |
| "loss": 0.9472, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 1.821699864917953e-07, | |
| "loss": 0.9573, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 1.7755657183783638e-07, | |
| "loss": 0.8264, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 1.7300126964819363e-07, | |
| "loss": 0.8467, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 1.6850413481616868e-07, | |
| "loss": 0.9277, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 1.6406522153411985e-07, | |
| "loss": 0.9041, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 1.596845832928129e-07, | |
| "loss": 0.8498, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 1.5536227288077466e-07, | |
| "loss": 0.8126, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 1.510983423836576e-07, | |
| "loss": 0.9155, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 1.468928431836092e-07, | |
| "loss": 0.8811, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 1.427458259586584e-07, | |
| "loss": 0.7399, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 1.3865734068210124e-07, | |
| "loss": 0.8392, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 1.3462743662189802e-07, | |
| "loss": 0.8258, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 1.3065616234008204e-07, | |
| "loss": 0.9331, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 1.2674356569217282e-07, | |
| "loss": 0.839, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 1.228896938265983e-07, | |
| "loss": 0.9558, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 1.1909459318413086e-07, | |
| "loss": 0.8337, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 1.1535830949732318e-07, | |
| "loss": 0.8119, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 1.1168088778996e-07, | |
| "loss": 0.9339, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 1.080623723765134e-07, | |
| "loss": 0.878, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 1.0450280686161163e-07, | |
| "loss": 0.8696, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 1.010022341395095e-07, | |
| "loss": 0.798, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 9.756069639357602e-08, | |
| "loss": 0.7726, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 9.417823509578316e-08, | |
| "loss": 0.9084, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 9.085489100620737e-08, | |
| "loss": 0.8672, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 8.759070417253768e-08, | |
| "loss": 0.9089, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 8.438571392959338e-08, | |
| "loss": 0.8619, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 8.123995889884995e-08, | |
| "loss": 0.919, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 7.815347698797327e-08, | |
| "loss": 0.9736, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 7.512630539036502e-08, | |
| "loss": 0.8875, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 7.215848058471086e-08, | |
| "loss": 0.8318, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 6.925003833454402e-08, | |
| "loss": 0.8584, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 6.640101368781126e-08, | |
| "loss": 0.9158, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 6.361144097645489e-08, | |
| "loss": 0.7842, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 6.088135381599414e-08, | |
| "loss": 0.6236, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 5.821078510512446e-08, | |
| "loss": 0.8197, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 5.559976702531722e-08, | |
| "loss": 0.8129, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 5.304833104043505e-08, | |
| "loss": 0.9605, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 5.0556507896350473e-08, | |
| "loss": 0.8278, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 4.8124327620576726e-08, | |
| "loss": 0.8557, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.575181952190533e-08, | |
| "loss": 0.8224, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.3439012190053534e-08, | |
| "loss": 0.5978, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 4.1185933495317965e-08, | |
| "loss": 0.8474, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 3.899261058824266e-08, | |
| "loss": 0.7566, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 3.685906989928656e-08, | |
| "loss": 0.8055, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 3.478533713850929e-08, | |
| "loss": 0.8079, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 3.277143729526033e-08, | |
| "loss": 0.8823, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 3.081739463787592e-08, | |
| "loss": 0.8067, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 2.892323271339037e-08, | |
| "loss": 0.8689, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 2.7088974347246888e-08, | |
| "loss": 0.9433, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 2.531464164302888e-08, | |
| "loss": 0.8466, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 2.3600255982187958e-08, | |
| "loss": 0.9036, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 2.19458380237908e-08, | |
| "loss": 0.9569, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 2.0351407704267134e-08, | |
| "loss": 0.9518, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 1.8816984237169378e-08, | |
| "loss": 0.8609, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 1.734258611294448e-08, | |
| "loss": 0.9466, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 1.5928231098706893e-08, | |
| "loss": 0.8711, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 1.4573936238028163e-08, | |
| "loss": 0.94, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 1.3279717850728236e-08, | |
| "loss": 0.8015, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 1.2045591532681145e-08, | |
| "loss": 0.7746, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 1.0871572155626841e-08, | |
| "loss": 0.9479, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 9.757673866990225e-09, | |
| "loss": 0.878, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 8.703910089712387e-09, | |
| "loss": 0.9255, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 7.710293522088518e-09, | |
| "loss": 0.9318, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 6.7768361376152616e-09, | |
| "loss": 0.8398, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 5.903549184844703e-09, | |
| "loss": 0.7752, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 5.090443187251159e-09, | |
| "loss": 0.8166, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 4.337527943101827e-09, | |
| "loss": 0.713, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 3.6448125253413236e-09, | |
| "loss": 0.9034, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 3.0123052814812203e-09, | |
| "loss": 0.867, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 2.440013833498456e-09, | |
| "loss": 0.9285, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.9279450777442976e-09, | |
| "loss": 0.8003, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.4761051848627417e-09, | |
| "loss": 0.8849, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 1.0844995997139064e-09, | |
| "loss": 0.8478, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 7.53133041307974e-10, | |
| "loss": 0.8967, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 4.820095027524563e-10, | |
| "loss": 0.9835, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 2.7113225119834717e-10, | |
| "loss": 0.8379, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 1.205038278051518e-10, | |
| "loss": 0.9426, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 3.0126047709244654e-11, | |
| "loss": 0.8376, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 0.0, | |
| "loss": 0.9024, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "step": 915, | |
| "total_flos": 8.22738094695383e+17, | |
| "train_loss": 0.9721869942920456, | |
| "train_runtime": 16666.1068, | |
| "train_samples_per_second": 0.441, | |
| "train_steps_per_second": 0.055 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 915, | |
| "num_train_epochs": 5, | |
| "save_steps": 1000, | |
| "total_flos": 8.22738094695383e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |