| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.913265306122449, | |
| "eval_steps": 800, | |
| "global_step": 4500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0, | |
| "loss": 2.7231, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 0, | |
| "loss": 2.8688, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0, | |
| "loss": 3.0604, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0, | |
| "loss": 3.3619, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0, | |
| "loss": 3.1263, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 0.0, | |
| "loss": 3.1305, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.5188198451414044e-06, | |
| "loss": 1.2543, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.0735116692035353e-06, | |
| "loss": 1.0489, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.420532010460384e-06, | |
| "loss": 0.8238, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.6736916157045096e-06, | |
| "loss": 0.6357, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.873103126046782e-06, | |
| "loss": 0.9375, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.0376396902828088e-06, | |
| "loss": 1.0042, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.177703030066258e-06, | |
| "loss": 0.7668, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 3.299639602133279e-06, | |
| "loss": 0.9168, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.407608173344298e-06, | |
| "loss": 0.8997, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.5044824703694353e-06, | |
| "loss": 1.0025, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.5923315143449394e-06, | |
| "loss": 0.6651, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.6726945828900284e-06, | |
| "loss": 0.699, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.746747919084026e-06, | |
| "loss": 0.886, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.8154106182475455e-06, | |
| "loss": 0.8937, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.8794145195304064e-06, | |
| "loss": 0.6582, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.9393518556017876e-06, | |
| "loss": 0.7974, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.995708648306083e-06, | |
| "loss": 0.9472, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.0488886794862905e-06, | |
| "loss": 0.8867, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.099231058976525e-06, | |
| "loss": 1.061, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.1470233384070705e-06, | |
| "loss": 0.7292, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.192511460845915e-06, | |
| "loss": 0.9262, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.235907420412398e-06, | |
| "loss": 0.7092, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.27739523661862e-06, | |
| "loss": 0.7174, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.317135669700268e-06, | |
| "loss": 0.8711, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.355269982485126e-06, | |
| "loss": 0.9286, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.391922971188186e-06, | |
| "loss": 0.9947, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.427205429264097e-06, | |
| "loss": 0.8179, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.461216167003915e-06, | |
| "loss": 0.6159, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.494043679663919e-06, | |
| "loss": 0.8981, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.525767535063022e-06, | |
| "loss": 0.9206, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.556459535424214e-06, | |
| "loss": 0.7768, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.5861846961499975e-06, | |
| "loss": 0.8515, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.6150020750907925e-06, | |
| "loss": 0.8131, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.642965478900328e-06, | |
| "loss": 1.1253, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.670124067711698e-06, | |
| "loss": 0.9236, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.689992082159791e-06, | |
| "loss": 0.8775, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.715848379822425e-06, | |
| "loss": 0.8848, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.74101509336297e-06, | |
| "loss": 0.829, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.76552805154028e-06, | |
| "loss": 0.9367, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.789420361336724e-06, | |
| "loss": 0.6566, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.812722676847563e-06, | |
| "loss": 0.8104, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.835463435763974e-06, | |
| "loss": 0.74, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.857669068026358e-06, | |
| "loss": 0.6171, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.879364180487766e-06, | |
| "loss": 1.0545, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.900571720823068e-06, | |
| "loss": 0.9191, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.921313123421507e-06, | |
| "loss": 0.6995, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.941608439588058e-06, | |
| "loss": 0.8479, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.9614764540361516e-06, | |
| "loss": 0.8976, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.980934789368156e-06, | |
| "loss": 0.851, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7368, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.997812135355893e-06, | |
| "loss": 1.0125, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.994894982497083e-06, | |
| "loss": 0.9607, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.991977829638274e-06, | |
| "loss": 0.6321, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.989060676779464e-06, | |
| "loss": 0.7108, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.986143523920654e-06, | |
| "loss": 0.7881, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.983226371061844e-06, | |
| "loss": 0.7147, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.980309218203034e-06, | |
| "loss": 0.9225, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.977392065344224e-06, | |
| "loss": 0.8849, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.974474912485414e-06, | |
| "loss": 0.942, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.971557759626604e-06, | |
| "loss": 1.1629, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.968640606767795e-06, | |
| "loss": 0.8053, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.965723453908986e-06, | |
| "loss": 0.6558, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.962806301050176e-06, | |
| "loss": 0.9606, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.959889148191366e-06, | |
| "loss": 0.9533, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.956971995332556e-06, | |
| "loss": 0.8758, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.954054842473746e-06, | |
| "loss": 0.5825, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.951137689614936e-06, | |
| "loss": 0.8221, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.948220536756126e-06, | |
| "loss": 0.8954, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.945303383897317e-06, | |
| "loss": 0.9549, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.942386231038507e-06, | |
| "loss": 0.8821, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.939469078179697e-06, | |
| "loss": 0.9861, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.936551925320887e-06, | |
| "loss": 0.9681, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.933634772462078e-06, | |
| "loss": 0.647, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.930717619603268e-06, | |
| "loss": 0.8151, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.9278004667444575e-06, | |
| "loss": 1.0373, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.9248833138856475e-06, | |
| "loss": 0.8858, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.921966161026838e-06, | |
| "loss": 0.8392, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.919049008168029e-06, | |
| "loss": 1.0194, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.916131855309218e-06, | |
| "loss": 0.7719, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.913214702450409e-06, | |
| "loss": 0.6672, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.910297549591599e-06, | |
| "loss": 0.8939, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.90738039673279e-06, | |
| "loss": 0.6781, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.90446324387398e-06, | |
| "loss": 0.7391, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.9015460910151695e-06, | |
| "loss": 0.9476, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.8986289381563595e-06, | |
| "loss": 0.6729, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.89571178529755e-06, | |
| "loss": 0.8046, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.89279463243874e-06, | |
| "loss": 0.8521, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.88987747957993e-06, | |
| "loss": 0.8688, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.886960326721121e-06, | |
| "loss": 0.875, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.884043173862311e-06, | |
| "loss": 0.6733, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.881126021003501e-06, | |
| "loss": 0.7755, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.878208868144691e-06, | |
| "loss": 0.7047, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.8752917152858815e-06, | |
| "loss": 0.6979, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.8723745624270714e-06, | |
| "loss": 0.8033, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.869457409568261e-06, | |
| "loss": 0.8039, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.866540256709452e-06, | |
| "loss": 0.9366, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.863623103850642e-06, | |
| "loss": 0.609, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.860705950991833e-06, | |
| "loss": 0.7258, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.857788798133022e-06, | |
| "loss": 1.018, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.854871645274213e-06, | |
| "loss": 0.7338, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.851954492415403e-06, | |
| "loss": 0.6595, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.8490373395565935e-06, | |
| "loss": 0.7456, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.8461201866977834e-06, | |
| "loss": 0.637, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.843203033838973e-06, | |
| "loss": 0.8026, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.840285880980164e-06, | |
| "loss": 0.8439, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.837368728121354e-06, | |
| "loss": 0.646, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.834451575262544e-06, | |
| "loss": 0.8644, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.831534422403734e-06, | |
| "loss": 0.8595, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.828617269544925e-06, | |
| "loss": 1.023, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.825700116686115e-06, | |
| "loss": 0.7745, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.822782963827305e-06, | |
| "loss": 0.9293, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.8198658109684954e-06, | |
| "loss": 0.7542, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.816948658109685e-06, | |
| "loss": 1.0059, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.814031505250875e-06, | |
| "loss": 0.8302, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.811114352392065e-06, | |
| "loss": 0.809, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.808197199533256e-06, | |
| "loss": 0.7628, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.805280046674446e-06, | |
| "loss": 0.6753, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.802362893815637e-06, | |
| "loss": 0.8719, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.799445740956827e-06, | |
| "loss": 0.7059, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.796528588098017e-06, | |
| "loss": 0.6902, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.793611435239207e-06, | |
| "loss": 1.104, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.790694282380397e-06, | |
| "loss": 0.837, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.787777129521587e-06, | |
| "loss": 0.794, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.784859976662777e-06, | |
| "loss": 0.7997, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.781942823803968e-06, | |
| "loss": 0.6738, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.779025670945158e-06, | |
| "loss": 0.963, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.776108518086348e-06, | |
| "loss": 0.6323, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.773191365227539e-06, | |
| "loss": 0.7243, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.770274212368729e-06, | |
| "loss": 0.8873, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.7673570595099186e-06, | |
| "loss": 0.7435, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.7644399066511085e-06, | |
| "loss": 0.8006, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.761522753792299e-06, | |
| "loss": 0.9557, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.758605600933489e-06, | |
| "loss": 0.7131, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.755688448074679e-06, | |
| "loss": 0.8007, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.752771295215869e-06, | |
| "loss": 0.8634, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.74985414235706e-06, | |
| "loss": 0.7848, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.746936989498251e-06, | |
| "loss": 0.6159, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.744019836639441e-06, | |
| "loss": 0.7327, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.7411026837806305e-06, | |
| "loss": 0.733, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.7381855309218205e-06, | |
| "loss": 0.8081, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.735268378063011e-06, | |
| "loss": 0.9809, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.732351225204201e-06, | |
| "loss": 0.7101, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.729434072345391e-06, | |
| "loss": 0.6195, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.726516919486582e-06, | |
| "loss": 0.6483, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.723599766627772e-06, | |
| "loss": 0.819, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.720682613768962e-06, | |
| "loss": 0.7032, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.717765460910152e-06, | |
| "loss": 0.7933, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.7148483080513425e-06, | |
| "loss": 0.9605, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.7119311551925325e-06, | |
| "loss": 0.7783, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.709014002333722e-06, | |
| "loss": 0.7616, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.706096849474912e-06, | |
| "loss": 0.6611, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.703179696616103e-06, | |
| "loss": 0.656, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.700262543757294e-06, | |
| "loss": 0.731, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.697345390898483e-06, | |
| "loss": 0.6204, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.694428238039674e-06, | |
| "loss": 0.695, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.691511085180864e-06, | |
| "loss": 0.9691, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.6885939323220545e-06, | |
| "loss": 0.5099, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.6856767794632445e-06, | |
| "loss": 0.6427, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.682759626604434e-06, | |
| "loss": 0.9209, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.679842473745624e-06, | |
| "loss": 0.891, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.676925320886815e-06, | |
| "loss": 0.7068, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.674008168028005e-06, | |
| "loss": 0.4388, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.671091015169195e-06, | |
| "loss": 0.7769, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.668173862310386e-06, | |
| "loss": 1.03, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.665256709451576e-06, | |
| "loss": 0.9345, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.662339556592766e-06, | |
| "loss": 0.8261, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.659422403733956e-06, | |
| "loss": 0.5926, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.656505250875146e-06, | |
| "loss": 0.618, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.653588098016336e-06, | |
| "loss": 0.8133, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.650670945157526e-06, | |
| "loss": 0.7822, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.647753792298717e-06, | |
| "loss": 0.6367, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.644836639439907e-06, | |
| "loss": 0.7457, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.641919486581098e-06, | |
| "loss": 0.5508, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.639002333722287e-06, | |
| "loss": 0.8247, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.636085180863478e-06, | |
| "loss": 0.8024, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.633168028004668e-06, | |
| "loss": 0.5489, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.630250875145858e-06, | |
| "loss": 0.5789, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.627333722287048e-06, | |
| "loss": 0.7541, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.624416569428238e-06, | |
| "loss": 0.6148, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.621499416569429e-06, | |
| "loss": 0.98, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.618582263710619e-06, | |
| "loss": 0.635, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.615665110851809e-06, | |
| "loss": 0.9664, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.612747957992999e-06, | |
| "loss": 0.8736, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.60983080513419e-06, | |
| "loss": 0.6281, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.60691365227538e-06, | |
| "loss": 0.9843, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.6039964994165695e-06, | |
| "loss": 0.7795, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.60107934655776e-06, | |
| "loss": 0.8437, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.59816219369895e-06, | |
| "loss": 0.6626, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.59524504084014e-06, | |
| "loss": 0.8164, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.59232788798133e-06, | |
| "loss": 0.7359, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.589410735122521e-06, | |
| "loss": 0.7709, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.586493582263711e-06, | |
| "loss": 0.8953, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.583576429404902e-06, | |
| "loss": 0.57, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.5806592765460916e-06, | |
| "loss": 0.9151, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.5777421236872815e-06, | |
| "loss": 0.7838, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.574824970828472e-06, | |
| "loss": 0.8183, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.571907817969662e-06, | |
| "loss": 0.9169, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.568990665110852e-06, | |
| "loss": 0.6786, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.566073512252042e-06, | |
| "loss": 0.7783, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.563156359393233e-06, | |
| "loss": 0.811, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.560239206534423e-06, | |
| "loss": 0.8965, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.557322053675613e-06, | |
| "loss": 0.7526, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5544049008168036e-06, | |
| "loss": 0.6549, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5514877479579935e-06, | |
| "loss": 0.77, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5485705950991834e-06, | |
| "loss": 0.5399, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.545653442240373e-06, | |
| "loss": 0.7214, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.542736289381564e-06, | |
| "loss": 0.7454, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.539819136522754e-06, | |
| "loss": 0.6774, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.536901983663944e-06, | |
| "loss": 0.855, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.533984830805134e-06, | |
| "loss": 0.8038, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.531067677946325e-06, | |
| "loss": 0.6897, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.5281505250875156e-06, | |
| "loss": 0.774, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.525233372228705e-06, | |
| "loss": 0.9166, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.5223162193698954e-06, | |
| "loss": 0.5643, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.519399066511085e-06, | |
| "loss": 0.5427, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.516481913652276e-06, | |
| "loss": 0.6834, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.513564760793466e-06, | |
| "loss": 0.8979, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.510647607934656e-06, | |
| "loss": 0.8804, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.507730455075847e-06, | |
| "loss": 0.6715, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.504813302217037e-06, | |
| "loss": 0.8331, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.501896149358227e-06, | |
| "loss": 0.9623, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.498978996499417e-06, | |
| "loss": 0.7293, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.496061843640607e-06, | |
| "loss": 0.6704, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.493144690781797e-06, | |
| "loss": 0.6452, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.490227537922987e-06, | |
| "loss": 0.7556, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.487310385064177e-06, | |
| "loss": 0.6235, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.484393232205368e-06, | |
| "loss": 0.789, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.481476079346558e-06, | |
| "loss": 0.7959, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.478558926487748e-06, | |
| "loss": 0.8286, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.475641773628939e-06, | |
| "loss": 0.8604, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.472724620770129e-06, | |
| "loss": 0.6135, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.469807467911319e-06, | |
| "loss": 0.4793, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.4668903150525085e-06, | |
| "loss": 0.7225, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.463973162193699e-06, | |
| "loss": 0.5255, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.461056009334889e-06, | |
| "loss": 0.4695, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.45813885647608e-06, | |
| "loss": 0.8741, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.45522170361727e-06, | |
| "loss": 0.6311, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.45230455075846e-06, | |
| "loss": 0.6423, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.449387397899651e-06, | |
| "loss": 0.7879, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.446470245040841e-06, | |
| "loss": 0.9635, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.4435530921820306e-06, | |
| "loss": 0.6497, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.4406359393232205e-06, | |
| "loss": 0.673, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.437718786464411e-06, | |
| "loss": 0.7035, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.434801633605601e-06, | |
| "loss": 0.7664, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.431884480746791e-06, | |
| "loss": 0.6817, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.428967327887982e-06, | |
| "loss": 0.7443, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.426050175029172e-06, | |
| "loss": 0.7205, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.423133022170362e-06, | |
| "loss": 0.6672, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.420215869311552e-06, | |
| "loss": 0.7427, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.4172987164527425e-06, | |
| "loss": 0.659, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.4143815635939325e-06, | |
| "loss": 0.832, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.411464410735123e-06, | |
| "loss": 0.7823, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.408547257876313e-06, | |
| "loss": 0.712, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.405630105017503e-06, | |
| "loss": 0.8165, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.402712952158694e-06, | |
| "loss": 0.6221, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.399795799299884e-06, | |
| "loss": 0.684, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.396878646441074e-06, | |
| "loss": 0.6845, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.393961493582264e-06, | |
| "loss": 0.8462, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.3910443407234545e-06, | |
| "loss": 0.6242, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.3881271878646445e-06, | |
| "loss": 0.5025, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.385210035005834e-06, | |
| "loss": 0.6651, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.382292882147025e-06, | |
| "loss": 0.7653, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.379375729288215e-06, | |
| "loss": 0.9375, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.376458576429405e-06, | |
| "loss": 0.6936, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.373541423570595e-06, | |
| "loss": 0.6709, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.370624270711786e-06, | |
| "loss": 0.7321, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.367707117852976e-06, | |
| "loss": 1.0432, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.364789964994166e-06, | |
| "loss": 0.6687, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.3618728121353565e-06, | |
| "loss": 0.9193, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.358955659276546e-06, | |
| "loss": 0.5975, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.356038506417737e-06, | |
| "loss": 0.7527, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.353121353558927e-06, | |
| "loss": 0.9754, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.350204200700117e-06, | |
| "loss": 0.8299, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.347287047841307e-06, | |
| "loss": 0.8524, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.344369894982498e-06, | |
| "loss": 0.8139, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.341452742123688e-06, | |
| "loss": 0.6144, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.338535589264878e-06, | |
| "loss": 0.8328, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.3356184364060684e-06, | |
| "loss": 0.5855, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.332701283547258e-06, | |
| "loss": 0.8227, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.329784130688448e-06, | |
| "loss": 0.6442, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.326866977829638e-06, | |
| "loss": 0.5148, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.323949824970829e-06, | |
| "loss": 0.8127, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.321032672112019e-06, | |
| "loss": 0.8909, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.318115519253209e-06, | |
| "loss": 0.6379, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.315198366394399e-06, | |
| "loss": 0.5912, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.31228121353559e-06, | |
| "loss": 0.6885, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.3093640606767804e-06, | |
| "loss": 0.8584, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.3064469078179695e-06, | |
| "loss": 0.6441, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.30352975495916e-06, | |
| "loss": 0.6933, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.30061260210035e-06, | |
| "loss": 0.6647, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.297695449241541e-06, | |
| "loss": 0.7737, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.294778296382731e-06, | |
| "loss": 0.8209, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.291861143523921e-06, | |
| "loss": 0.9695, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.288943990665112e-06, | |
| "loss": 0.855, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.286026837806302e-06, | |
| "loss": 0.7051, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.283109684947492e-06, | |
| "loss": 0.4676, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.2801925320886815e-06, | |
| "loss": 0.6563, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.277275379229872e-06, | |
| "loss": 0.6791, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.274358226371062e-06, | |
| "loss": 0.6531, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.271441073512252e-06, | |
| "loss": 0.7638, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.268523920653442e-06, | |
| "loss": 0.8035, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.265606767794633e-06, | |
| "loss": 0.6947, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.262689614935823e-06, | |
| "loss": 0.4111, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.259772462077013e-06, | |
| "loss": 0.787, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.2568553092182036e-06, | |
| "loss": 0.7891, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.2539381563593935e-06, | |
| "loss": 0.6001, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.251021003500584e-06, | |
| "loss": 0.5617, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.248103850641773e-06, | |
| "loss": 0.9523, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.245186697782964e-06, | |
| "loss": 0.8455, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.242269544924154e-06, | |
| "loss": 0.7674, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.239352392065345e-06, | |
| "loss": 0.5396, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.236435239206535e-06, | |
| "loss": 0.5736, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.233518086347725e-06, | |
| "loss": 0.5877, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.2306009334889156e-06, | |
| "loss": 0.6466, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.2276837806301055e-06, | |
| "loss": 0.4551, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.2247666277712954e-06, | |
| "loss": 0.5959, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.221849474912485e-06, | |
| "loss": 0.6903, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.218932322053676e-06, | |
| "loss": 0.7372, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.216015169194866e-06, | |
| "loss": 0.7741, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.213098016336056e-06, | |
| "loss": 0.6861, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.210180863477247e-06, | |
| "loss": 0.7325, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.207263710618437e-06, | |
| "loss": 0.6446, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.204346557759627e-06, | |
| "loss": 0.61, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.201429404900817e-06, | |
| "loss": 0.6286, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.1985122520420074e-06, | |
| "loss": 0.7191, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.195595099183197e-06, | |
| "loss": 0.8621, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.192677946324388e-06, | |
| "loss": 0.7634, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.189760793465578e-06, | |
| "loss": 0.7067, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.186843640606768e-06, | |
| "loss": 0.6049, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.183926487747959e-06, | |
| "loss": 0.3731, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.181009334889149e-06, | |
| "loss": 0.5955, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.178092182030339e-06, | |
| "loss": 0.5576, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.175175029171529e-06, | |
| "loss": 0.6731, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.172257876312719e-06, | |
| "loss": 0.8847, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.169340723453909e-06, | |
| "loss": 0.6158, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.166423570595099e-06, | |
| "loss": 0.7721, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.16350641773629e-06, | |
| "loss": 0.6517, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.16058926487748e-06, | |
| "loss": 0.6014, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.15767211201867e-06, | |
| "loss": 0.8049, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.15475495915986e-06, | |
| "loss": 0.7245, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.151837806301051e-06, | |
| "loss": 0.8411, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.148920653442241e-06, | |
| "loss": 0.6463, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.1460035005834306e-06, | |
| "loss": 0.5053, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.143086347724621e-06, | |
| "loss": 0.4919, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.140169194865811e-06, | |
| "loss": 0.7352, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.137252042007002e-06, | |
| "loss": 0.5577, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.134334889148192e-06, | |
| "loss": 0.7224, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.131417736289382e-06, | |
| "loss": 0.9195, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.128500583430572e-06, | |
| "loss": 0.7652, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.125583430571763e-06, | |
| "loss": 0.756, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.122666277712953e-06, | |
| "loss": 0.7378, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.1197491248541425e-06, | |
| "loss": 0.7264, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.116831971995333e-06, | |
| "loss": 0.7489, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.113914819136523e-06, | |
| "loss": 0.661, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.110997666277713e-06, | |
| "loss": 0.6761, | |
| "step": 1444 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.108080513418903e-06, | |
| "loss": 0.62, | |
| "step": 1448 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.105163360560094e-06, | |
| "loss": 0.5477, | |
| "step": 1452 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.102246207701284e-06, | |
| "loss": 0.7107, | |
| "step": 1456 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.099329054842474e-06, | |
| "loss": 0.9508, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.096411901983664e-06, | |
| "loss": 0.7668, | |
| "step": 1464 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.0934947491248545e-06, | |
| "loss": 0.7971, | |
| "step": 1468 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.0905775962660445e-06, | |
| "loss": 0.7532, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.087660443407234e-06, | |
| "loss": 0.6721, | |
| "step": 1476 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.084743290548425e-06, | |
| "loss": 0.702, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.081826137689615e-06, | |
| "loss": 0.5931, | |
| "step": 1484 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.078908984830806e-06, | |
| "loss": 0.8941, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.075991831971996e-06, | |
| "loss": 0.4728, | |
| "step": 1492 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.073074679113186e-06, | |
| "loss": 0.5366, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.070157526254377e-06, | |
| "loss": 0.6567, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.0672403733955665e-06, | |
| "loss": 0.6594, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.0643232205367565e-06, | |
| "loss": 0.5096, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.061406067677946e-06, | |
| "loss": 0.6852, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.058488914819137e-06, | |
| "loss": 0.7667, | |
| "step": 1516 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.055571761960327e-06, | |
| "loss": 0.662, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.052654609101517e-06, | |
| "loss": 0.6275, | |
| "step": 1524 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.049737456242707e-06, | |
| "loss": 0.715, | |
| "step": 1528 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.046820303383898e-06, | |
| "loss": 0.4791, | |
| "step": 1532 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.043903150525088e-06, | |
| "loss": 0.6893, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.040985997666278e-06, | |
| "loss": 0.4941, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.0380688448074685e-06, | |
| "loss": 0.5576, | |
| "step": 1544 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.035151691948658e-06, | |
| "loss": 0.7089, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.032234539089848e-06, | |
| "loss": 0.4944, | |
| "step": 1552 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.029317386231038e-06, | |
| "loss": 0.4799, | |
| "step": 1556 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.026400233372229e-06, | |
| "loss": 0.6483, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.023483080513419e-06, | |
| "loss": 0.7882, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.02056592765461e-06, | |
| "loss": 0.6979, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.0176487747958e-06, | |
| "loss": 0.6079, | |
| "step": 1572 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.01473162193699e-06, | |
| "loss": 0.7046, | |
| "step": 1576 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.0118144690781804e-06, | |
| "loss": 0.5994, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.00889731621937e-06, | |
| "loss": 0.628, | |
| "step": 1584 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.00598016336056e-06, | |
| "loss": 0.4581, | |
| "step": 1588 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.00306301050175e-06, | |
| "loss": 0.6283, | |
| "step": 1592 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.000145857642941e-06, | |
| "loss": 0.6736, | |
| "step": 1596 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.997228704784131e-06, | |
| "loss": 0.6118, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.994311551925321e-06, | |
| "loss": 0.8692, | |
| "step": 1604 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 3.991394399066512e-06, | |
| "loss": 0.7431, | |
| "step": 1608 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.988477246207702e-06, | |
| "loss": 0.6712, | |
| "step": 1612 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.985560093348892e-06, | |
| "loss": 0.5141, | |
| "step": 1616 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.9826429404900815e-06, | |
| "loss": 0.4037, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.979725787631272e-06, | |
| "loss": 0.5442, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.976808634772462e-06, | |
| "loss": 0.7256, | |
| "step": 1628 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.973891481913652e-06, | |
| "loss": 0.7292, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.970974329054843e-06, | |
| "loss": 0.7739, | |
| "step": 1636 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.968057176196033e-06, | |
| "loss": 0.6673, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.965140023337224e-06, | |
| "loss": 0.8099, | |
| "step": 1644 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.962222870478414e-06, | |
| "loss": 0.5271, | |
| "step": 1648 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.9593057176196036e-06, | |
| "loss": 0.7032, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 3.9563885647607935e-06, | |
| "loss": 0.6766, | |
| "step": 1656 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.953471411901984e-06, | |
| "loss": 0.6286, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.950554259043174e-06, | |
| "loss": 0.6236, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.947637106184364e-06, | |
| "loss": 0.7438, | |
| "step": 1668 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.944719953325555e-06, | |
| "loss": 0.6499, | |
| "step": 1672 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.941802800466745e-06, | |
| "loss": 0.7669, | |
| "step": 1676 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.938885647607935e-06, | |
| "loss": 0.4944, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.935968494749125e-06, | |
| "loss": 0.4765, | |
| "step": 1684 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.9330513418903156e-06, | |
| "loss": 0.7785, | |
| "step": 1688 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.9301341890315055e-06, | |
| "loss": 0.6521, | |
| "step": 1692 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.9272170361726954e-06, | |
| "loss": 0.6744, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.924299883313886e-06, | |
| "loss": 0.6819, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.921382730455076e-06, | |
| "loss": 0.6566, | |
| "step": 1704 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.918465577596267e-06, | |
| "loss": 0.7201, | |
| "step": 1708 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.915548424737456e-06, | |
| "loss": 0.373, | |
| "step": 1712 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.912631271878647e-06, | |
| "loss": 0.8481, | |
| "step": 1716 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.909714119019837e-06, | |
| "loss": 0.6479, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.9067969661610276e-06, | |
| "loss": 0.5682, | |
| "step": 1724 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.9038798133022175e-06, | |
| "loss": 0.651, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.9009626604434074e-06, | |
| "loss": 0.637, | |
| "step": 1732 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.898045507584598e-06, | |
| "loss": 0.5119, | |
| "step": 1736 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.895128354725788e-06, | |
| "loss": 0.5434, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.892211201866978e-06, | |
| "loss": 0.7295, | |
| "step": 1744 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.889294049008168e-06, | |
| "loss": 0.7214, | |
| "step": 1748 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.886376896149359e-06, | |
| "loss": 0.5946, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.883459743290549e-06, | |
| "loss": 0.6668, | |
| "step": 1756 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.880542590431739e-06, | |
| "loss": 0.6054, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.877625437572929e-06, | |
| "loss": 0.6904, | |
| "step": 1764 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.874708284714119e-06, | |
| "loss": 0.7803, | |
| "step": 1768 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.871791131855309e-06, | |
| "loss": 0.6766, | |
| "step": 1772 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.868873978996499e-06, | |
| "loss": 0.6957, | |
| "step": 1776 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.86595682613769e-06, | |
| "loss": 0.8129, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.86303967327888e-06, | |
| "loss": 0.7652, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.860122520420071e-06, | |
| "loss": 0.4505, | |
| "step": 1788 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.85720536756126e-06, | |
| "loss": 0.6318, | |
| "step": 1792 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 3.854288214702451e-06, | |
| "loss": 0.7018, | |
| "step": 1796 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.8513710618436415e-06, | |
| "loss": 0.8348, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.848453908984831e-06, | |
| "loss": 0.5523, | |
| "step": 1804 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.845536756126021e-06, | |
| "loss": 0.5858, | |
| "step": 1808 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.842619603267211e-06, | |
| "loss": 0.3905, | |
| "step": 1812 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.839702450408402e-06, | |
| "loss": 0.7015, | |
| "step": 1816 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.836785297549592e-06, | |
| "loss": 0.4228, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.833868144690782e-06, | |
| "loss": 0.7709, | |
| "step": 1824 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.830950991831972e-06, | |
| "loss": 0.5833, | |
| "step": 1828 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.828033838973163e-06, | |
| "loss": 0.6342, | |
| "step": 1832 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.825116686114353e-06, | |
| "loss": 0.843, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.8221995332555425e-06, | |
| "loss": 0.5876, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 3.819282380396733e-06, | |
| "loss": 0.5432, | |
| "step": 1844 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.816365227537923e-06, | |
| "loss": 0.7204, | |
| "step": 1848 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.8134480746791136e-06, | |
| "loss": 0.4466, | |
| "step": 1852 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.8105309218203036e-06, | |
| "loss": 0.623, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.807613768961494e-06, | |
| "loss": 0.3623, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.804696616102684e-06, | |
| "loss": 0.6491, | |
| "step": 1864 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 3.8017794632438742e-06, | |
| "loss": 0.6469, | |
| "step": 1868 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.7988623103850646e-06, | |
| "loss": 0.5412, | |
| "step": 1872 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.7959451575262545e-06, | |
| "loss": 0.995, | |
| "step": 1876 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.793028004667445e-06, | |
| "loss": 0.6152, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.790110851808635e-06, | |
| "loss": 0.7216, | |
| "step": 1884 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.787193698949825e-06, | |
| "loss": 0.5137, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.784276546091015e-06, | |
| "loss": 0.6703, | |
| "step": 1892 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.7813593932322055e-06, | |
| "loss": 0.5394, | |
| "step": 1896 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.7784422403733963e-06, | |
| "loss": 0.6228, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.7755250875145862e-06, | |
| "loss": 0.6231, | |
| "step": 1904 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.7726079346557766e-06, | |
| "loss": 0.6658, | |
| "step": 1908 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.7696907817969665e-06, | |
| "loss": 0.5279, | |
| "step": 1912 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.766773628938157e-06, | |
| "loss": 0.4709, | |
| "step": 1916 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.763856476079347e-06, | |
| "loss": 0.6442, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.760939323220537e-06, | |
| "loss": 0.6787, | |
| "step": 1924 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.758022170361727e-06, | |
| "loss": 0.6453, | |
| "step": 1928 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.7551050175029175e-06, | |
| "loss": 0.7487, | |
| "step": 1932 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.752187864644108e-06, | |
| "loss": 0.6078, | |
| "step": 1936 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.749270711785298e-06, | |
| "loss": 0.5252, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.746353558926488e-06, | |
| "loss": 0.4936, | |
| "step": 1944 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.743436406067678e-06, | |
| "loss": 0.4545, | |
| "step": 1948 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.7405192532088685e-06, | |
| "loss": 0.7937, | |
| "step": 1952 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.7376021003500584e-06, | |
| "loss": 0.8007, | |
| "step": 1956 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.7346849474912488e-06, | |
| "loss": 0.4401, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.731767794632439e-06, | |
| "loss": 0.8051, | |
| "step": 1964 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.728850641773629e-06, | |
| "loss": 0.7178, | |
| "step": 1968 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.72593348891482e-06, | |
| "loss": 0.5673, | |
| "step": 1972 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.7230163360560094e-06, | |
| "loss": 0.8238, | |
| "step": 1976 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.7200991831972e-06, | |
| "loss": 0.646, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.7171820303383897e-06, | |
| "loss": 0.5166, | |
| "step": 1984 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.7142648774795804e-06, | |
| "loss": 0.5725, | |
| "step": 1988 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.7113477246207704e-06, | |
| "loss": 0.8298, | |
| "step": 1992 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.7084305717619607e-06, | |
| "loss": 0.5858, | |
| "step": 1996 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.705513418903151e-06, | |
| "loss": 0.5432, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.702596266044341e-06, | |
| "loss": 0.5454, | |
| "step": 2004 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.6996791131855314e-06, | |
| "loss": 0.6, | |
| "step": 2008 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.6967619603267213e-06, | |
| "loss": 0.5889, | |
| "step": 2012 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.6938448074679117e-06, | |
| "loss": 0.7542, | |
| "step": 2016 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.6909276546091016e-06, | |
| "loss": 0.7573, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.688010501750292e-06, | |
| "loss": 0.7429, | |
| "step": 2024 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.685093348891482e-06, | |
| "loss": 0.6015, | |
| "step": 2028 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.6821761960326723e-06, | |
| "loss": 0.6009, | |
| "step": 2032 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.6792590431738627e-06, | |
| "loss": 0.4447, | |
| "step": 2036 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.6763418903150526e-06, | |
| "loss": 0.667, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.673424737456243e-06, | |
| "loss": 0.594, | |
| "step": 2044 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.670507584597433e-06, | |
| "loss": 0.6368, | |
| "step": 2048 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.6675904317386237e-06, | |
| "loss": 0.5883, | |
| "step": 2052 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.6646732788798132e-06, | |
| "loss": 0.7004, | |
| "step": 2056 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.661756126021004e-06, | |
| "loss": 0.6169, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.6588389731621944e-06, | |
| "loss": 0.6757, | |
| "step": 2064 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.6559218203033843e-06, | |
| "loss": 0.6134, | |
| "step": 2068 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.6530046674445747e-06, | |
| "loss": 0.6381, | |
| "step": 2072 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.6500875145857646e-06, | |
| "loss": 0.715, | |
| "step": 2076 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.647170361726955e-06, | |
| "loss": 0.6319, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.644253208868145e-06, | |
| "loss": 0.4187, | |
| "step": 2084 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.6413360560093353e-06, | |
| "loss": 0.5733, | |
| "step": 2088 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.638418903150525e-06, | |
| "loss": 0.6943, | |
| "step": 2092 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.6355017502917156e-06, | |
| "loss": 0.4808, | |
| "step": 2096 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.632584597432906e-06, | |
| "loss": 0.4813, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.629667444574096e-06, | |
| "loss": 0.5777, | |
| "step": 2104 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.6267502917152862e-06, | |
| "loss": 0.5967, | |
| "step": 2108 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.623833138856476e-06, | |
| "loss": 0.5644, | |
| "step": 2112 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.6209159859976665e-06, | |
| "loss": 0.6584, | |
| "step": 2116 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.6179988331388565e-06, | |
| "loss": 0.5162, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.615081680280047e-06, | |
| "loss": 0.6585, | |
| "step": 2124 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.6121645274212368e-06, | |
| "loss": 0.7716, | |
| "step": 2128 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.6092473745624276e-06, | |
| "loss": 0.5249, | |
| "step": 2132 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.606330221703618e-06, | |
| "loss": 0.722, | |
| "step": 2136 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.603413068844808e-06, | |
| "loss": 0.6634, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.6004959159859982e-06, | |
| "loss": 0.6547, | |
| "step": 2144 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.597578763127188e-06, | |
| "loss": 0.7804, | |
| "step": 2148 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.5946616102683785e-06, | |
| "loss": 0.5373, | |
| "step": 2152 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.5917444574095685e-06, | |
| "loss": 0.6122, | |
| "step": 2156 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.588827304550759e-06, | |
| "loss": 0.5015, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.585910151691949e-06, | |
| "loss": 0.6872, | |
| "step": 2164 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.582992998833139e-06, | |
| "loss": 0.3367, | |
| "step": 2168 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.5800758459743295e-06, | |
| "loss": 0.4729, | |
| "step": 2172 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.5771586931155194e-06, | |
| "loss": 0.7766, | |
| "step": 2176 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.5742415402567098e-06, | |
| "loss": 0.7324, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.5713243873978997e-06, | |
| "loss": 0.5244, | |
| "step": 2184 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.56840723453909e-06, | |
| "loss": 0.6321, | |
| "step": 2188 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.56549008168028e-06, | |
| "loss": 0.6109, | |
| "step": 2192 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.5625729288214704e-06, | |
| "loss": 0.5534, | |
| "step": 2196 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.559655775962661e-06, | |
| "loss": 0.5453, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5567386231038507e-06, | |
| "loss": 0.4379, | |
| "step": 2204 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5538214702450415e-06, | |
| "loss": 0.5268, | |
| "step": 2208 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5509043173862314e-06, | |
| "loss": 0.7081, | |
| "step": 2212 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5479871645274218e-06, | |
| "loss": 0.5149, | |
| "step": 2216 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5450700116686117e-06, | |
| "loss": 0.4048, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.542152858809802e-06, | |
| "loss": 0.7552, | |
| "step": 2224 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.539235705950992e-06, | |
| "loss": 0.4624, | |
| "step": 2228 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.5363185530921824e-06, | |
| "loss": 0.5872, | |
| "step": 2232 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.5334014002333727e-06, | |
| "loss": 0.5265, | |
| "step": 2236 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.5304842473745627e-06, | |
| "loss": 0.4802, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.527567094515753e-06, | |
| "loss": 0.6172, | |
| "step": 2244 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.524649941656943e-06, | |
| "loss": 0.5788, | |
| "step": 2248 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.5217327887981333e-06, | |
| "loss": 0.5426, | |
| "step": 2252 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.5188156359393233e-06, | |
| "loss": 0.6427, | |
| "step": 2256 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.5158984830805136e-06, | |
| "loss": 0.6474, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.512981330221704e-06, | |
| "loss": 0.5518, | |
| "step": 2264 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.510064177362894e-06, | |
| "loss": 0.6028, | |
| "step": 2268 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.5071470245040843e-06, | |
| "loss": 0.4178, | |
| "step": 2272 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.5042298716452742e-06, | |
| "loss": 0.761, | |
| "step": 2276 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.501312718786465e-06, | |
| "loss": 0.7076, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.4983955659276545e-06, | |
| "loss": 0.5466, | |
| "step": 2284 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.4954784130688453e-06, | |
| "loss": 0.4539, | |
| "step": 2288 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.4925612602100353e-06, | |
| "loss": 0.6436, | |
| "step": 2292 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.4896441073512256e-06, | |
| "loss": 0.471, | |
| "step": 2296 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.486726954492416e-06, | |
| "loss": 0.6848, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.483809801633606e-06, | |
| "loss": 0.6043, | |
| "step": 2304 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.4808926487747963e-06, | |
| "loss": 0.6458, | |
| "step": 2308 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.4779754959159862e-06, | |
| "loss": 0.6336, | |
| "step": 2312 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.4750583430571766e-06, | |
| "loss": 0.5843, | |
| "step": 2316 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.4721411901983665e-06, | |
| "loss": 0.4364, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.469224037339557e-06, | |
| "loss": 0.5796, | |
| "step": 2324 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.466306884480747e-06, | |
| "loss": 0.5754, | |
| "step": 2328 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.463389731621937e-06, | |
| "loss": 0.6848, | |
| "step": 2332 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.4604725787631276e-06, | |
| "loss": 0.6489, | |
| "step": 2336 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.4575554259043175e-06, | |
| "loss": 0.6255, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.454638273045508e-06, | |
| "loss": 0.4827, | |
| "step": 2344 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.451721120186698e-06, | |
| "loss": 0.5178, | |
| "step": 2348 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.448803967327888e-06, | |
| "loss": 0.5294, | |
| "step": 2352 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.445886814469078e-06, | |
| "loss": 0.4968, | |
| "step": 2356 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.442969661610269e-06, | |
| "loss": 0.7301, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.4400525087514592e-06, | |
| "loss": 0.5503, | |
| "step": 2364 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.437135355892649e-06, | |
| "loss": 0.4724, | |
| "step": 2368 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.4342182030338395e-06, | |
| "loss": 0.5471, | |
| "step": 2372 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.4313010501750295e-06, | |
| "loss": 0.5414, | |
| "step": 2376 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.42838389731622e-06, | |
| "loss": 0.4228, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.4254667444574098e-06, | |
| "loss": 0.6098, | |
| "step": 2384 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.4225495915986e-06, | |
| "loss": 0.5935, | |
| "step": 2388 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.41963243873979e-06, | |
| "loss": 0.5567, | |
| "step": 2392 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.4167152858809804e-06, | |
| "loss": 0.4576, | |
| "step": 2396 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.413798133022171e-06, | |
| "loss": 0.5125, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.4108809801633607e-06, | |
| "loss": 0.4188, | |
| "step": 2404 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.407963827304551e-06, | |
| "loss": 0.6133, | |
| "step": 2408 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.405046674445741e-06, | |
| "loss": 0.5203, | |
| "step": 2412 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.4021295215869314e-06, | |
| "loss": 0.5936, | |
| "step": 2416 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.3992123687281213e-06, | |
| "loss": 0.5783, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.3962952158693117e-06, | |
| "loss": 0.5501, | |
| "step": 2424 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.3933780630105017e-06, | |
| "loss": 0.6859, | |
| "step": 2428 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.390460910151692e-06, | |
| "loss": 0.5361, | |
| "step": 2432 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.387543757292883e-06, | |
| "loss": 0.3481, | |
| "step": 2436 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.3846266044340727e-06, | |
| "loss": 0.5475, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.381709451575263e-06, | |
| "loss": 0.4606, | |
| "step": 2444 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.378792298716453e-06, | |
| "loss": 0.3753, | |
| "step": 2448 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.3758751458576434e-06, | |
| "loss": 0.5286, | |
| "step": 2452 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.3729579929988333e-06, | |
| "loss": 0.5214, | |
| "step": 2456 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.3700408401400237e-06, | |
| "loss": 0.4971, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.367123687281214e-06, | |
| "loss": 0.5731, | |
| "step": 2464 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.364206534422404e-06, | |
| "loss": 0.6563, | |
| "step": 2468 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.3612893815635944e-06, | |
| "loss": 0.3885, | |
| "step": 2472 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.3583722287047843e-06, | |
| "loss": 0.7304, | |
| "step": 2476 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.3554550758459747e-06, | |
| "loss": 0.6601, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.3525379229871646e-06, | |
| "loss": 0.4629, | |
| "step": 2484 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.349620770128355e-06, | |
| "loss": 0.5769, | |
| "step": 2488 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.346703617269545e-06, | |
| "loss": 0.5616, | |
| "step": 2492 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.3437864644107353e-06, | |
| "loss": 0.4976, | |
| "step": 2496 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.340869311551926e-06, | |
| "loss": 0.6104, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.3379521586931156e-06, | |
| "loss": 0.4875, | |
| "step": 2504 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.3350350058343064e-06, | |
| "loss": 0.3093, | |
| "step": 2508 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.332117852975496e-06, | |
| "loss": 0.3887, | |
| "step": 2512 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.3292007001166867e-06, | |
| "loss": 0.6688, | |
| "step": 2516 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.3262835472578766e-06, | |
| "loss": 0.479, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.323366394399067e-06, | |
| "loss": 0.616, | |
| "step": 2524 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.320449241540257e-06, | |
| "loss": 0.8513, | |
| "step": 2528 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.3175320886814473e-06, | |
| "loss": 0.4224, | |
| "step": 2532 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.3146149358226376e-06, | |
| "loss": 0.4577, | |
| "step": 2536 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.3116977829638276e-06, | |
| "loss": 0.4286, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.308780630105018e-06, | |
| "loss": 0.5298, | |
| "step": 2544 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.305863477246208e-06, | |
| "loss": 0.4531, | |
| "step": 2548 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.3029463243873982e-06, | |
| "loss": 0.561, | |
| "step": 2552 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.300029171528588e-06, | |
| "loss": 0.4931, | |
| "step": 2556 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.2971120186697785e-06, | |
| "loss": 0.6341, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.294194865810969e-06, | |
| "loss": 0.5096, | |
| "step": 2564 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.291277712952159e-06, | |
| "loss": 0.3863, | |
| "step": 2568 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.288360560093349e-06, | |
| "loss": 0.6633, | |
| "step": 2572 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.285443407234539e-06, | |
| "loss": 0.7305, | |
| "step": 2576 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.2825262543757295e-06, | |
| "loss": 0.627, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.2796091015169194e-06, | |
| "loss": 0.4865, | |
| "step": 2584 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.27669194865811e-06, | |
| "loss": 0.507, | |
| "step": 2588 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.2737747957992997e-06, | |
| "loss": 0.5226, | |
| "step": 2592 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.2708576429404905e-06, | |
| "loss": 0.5523, | |
| "step": 2596 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.267940490081681e-06, | |
| "loss": 0.5195, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.265023337222871e-06, | |
| "loss": 0.3034, | |
| "step": 2604 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.262106184364061e-06, | |
| "loss": 0.7056, | |
| "step": 2608 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.259189031505251e-06, | |
| "loss": 0.6256, | |
| "step": 2612 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.2562718786464415e-06, | |
| "loss": 0.4434, | |
| "step": 2616 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.2533547257876314e-06, | |
| "loss": 0.3871, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.2504375729288218e-06, | |
| "loss": 0.6902, | |
| "step": 2624 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.2475204200700117e-06, | |
| "loss": 0.3734, | |
| "step": 2628 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.244603267211202e-06, | |
| "loss": 0.4771, | |
| "step": 2632 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.2416861143523924e-06, | |
| "loss": 0.5457, | |
| "step": 2636 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.2387689614935824e-06, | |
| "loss": 0.5787, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.2358518086347727e-06, | |
| "loss": 0.7326, | |
| "step": 2644 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.2329346557759627e-06, | |
| "loss": 0.4606, | |
| "step": 2648 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.230017502917153e-06, | |
| "loss": 0.5832, | |
| "step": 2652 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.227100350058343e-06, | |
| "loss": 0.4674, | |
| "step": 2656 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.2241831971995333e-06, | |
| "loss": 0.3263, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.221266044340724e-06, | |
| "loss": 0.433, | |
| "step": 2664 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.218348891481914e-06, | |
| "loss": 0.562, | |
| "step": 2668 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.2154317386231044e-06, | |
| "loss": 0.5465, | |
| "step": 2672 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.2125145857642944e-06, | |
| "loss": 0.5806, | |
| "step": 2676 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.2095974329054847e-06, | |
| "loss": 0.5941, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.2066802800466747e-06, | |
| "loss": 0.5672, | |
| "step": 2684 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.203763127187865e-06, | |
| "loss": 0.4044, | |
| "step": 2688 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.200845974329055e-06, | |
| "loss": 0.6372, | |
| "step": 2692 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.1979288214702453e-06, | |
| "loss": 0.5306, | |
| "step": 2696 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.1950116686114357e-06, | |
| "loss": 0.6751, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.1920945157526256e-06, | |
| "loss": 0.3999, | |
| "step": 2704 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.189177362893816e-06, | |
| "loss": 0.6431, | |
| "step": 2708 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.186260210035006e-06, | |
| "loss": 0.7637, | |
| "step": 2712 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.1833430571761963e-06, | |
| "loss": 0.4698, | |
| "step": 2716 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.1804259043173862e-06, | |
| "loss": 0.5558, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.1775087514585766e-06, | |
| "loss": 0.5202, | |
| "step": 2724 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.1745915985997665e-06, | |
| "loss": 0.4455, | |
| "step": 2728 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.171674445740957e-06, | |
| "loss": 0.6568, | |
| "step": 2732 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.1687572928821477e-06, | |
| "loss": 0.6879, | |
| "step": 2736 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.165840140023337e-06, | |
| "loss": 0.474, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.162922987164528e-06, | |
| "loss": 0.5385, | |
| "step": 2744 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.160005834305718e-06, | |
| "loss": 0.6955, | |
| "step": 2748 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.1570886814469083e-06, | |
| "loss": 0.501, | |
| "step": 2752 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.1541715285880982e-06, | |
| "loss": 0.5093, | |
| "step": 2756 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.1512543757292886e-06, | |
| "loss": 0.3698, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.148337222870479e-06, | |
| "loss": 0.4271, | |
| "step": 2764 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.145420070011669e-06, | |
| "loss": 0.5814, | |
| "step": 2768 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.1425029171528592e-06, | |
| "loss": 0.3312, | |
| "step": 2772 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.139585764294049e-06, | |
| "loss": 0.5727, | |
| "step": 2776 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.1366686114352395e-06, | |
| "loss": 0.6723, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.1337514585764295e-06, | |
| "loss": 0.5772, | |
| "step": 2784 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.13083430571762e-06, | |
| "loss": 0.2121, | |
| "step": 2788 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.1279171528588098e-06, | |
| "loss": 0.5148, | |
| "step": 2792 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.125e-06, | |
| "loss": 0.402, | |
| "step": 2796 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.1220828471411905e-06, | |
| "loss": 0.4757, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.1191656942823805e-06, | |
| "loss": 0.5157, | |
| "step": 2804 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.1162485414235712e-06, | |
| "loss": 0.5417, | |
| "step": 2808 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.1133313885647608e-06, | |
| "loss": 0.3401, | |
| "step": 2812 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.1104142357059515e-06, | |
| "loss": 0.6862, | |
| "step": 2816 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.107497082847141e-06, | |
| "loss": 0.516, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.104579929988332e-06, | |
| "loss": 0.5201, | |
| "step": 2824 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.1016627771295218e-06, | |
| "loss": 0.4436, | |
| "step": 2828 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.098745624270712e-06, | |
| "loss": 0.5983, | |
| "step": 2832 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0958284714119025e-06, | |
| "loss": 0.5076, | |
| "step": 2836 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0929113185530924e-06, | |
| "loss": 0.4357, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.089994165694283e-06, | |
| "loss": 0.5015, | |
| "step": 2844 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.0870770128354727e-06, | |
| "loss": 0.5186, | |
| "step": 2848 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.084159859976663e-06, | |
| "loss": 0.5653, | |
| "step": 2852 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.081242707117853e-06, | |
| "loss": 0.6339, | |
| "step": 2856 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.0783255542590434e-06, | |
| "loss": 0.5534, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.0754084014002338e-06, | |
| "loss": 0.5144, | |
| "step": 2864 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.0724912485414237e-06, | |
| "loss": 0.393, | |
| "step": 2868 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.069574095682614e-06, | |
| "loss": 0.4453, | |
| "step": 2872 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.066656942823804e-06, | |
| "loss": 0.5278, | |
| "step": 2876 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.0637397899649944e-06, | |
| "loss": 0.7833, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0608226371061843e-06, | |
| "loss": 0.4581, | |
| "step": 2884 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0579054842473747e-06, | |
| "loss": 0.3942, | |
| "step": 2888 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0549883313885646e-06, | |
| "loss": 0.6298, | |
| "step": 2892 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0520711785297554e-06, | |
| "loss": 0.6276, | |
| "step": 2896 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0491540256709458e-06, | |
| "loss": 0.737, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.0462368728121357e-06, | |
| "loss": 0.5208, | |
| "step": 2904 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.043319719953326e-06, | |
| "loss": 0.5953, | |
| "step": 2908 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.040402567094516e-06, | |
| "loss": 0.5373, | |
| "step": 2912 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.0374854142357064e-06, | |
| "loss": 0.3865, | |
| "step": 2916 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.0345682613768963e-06, | |
| "loss": 0.5451, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.0316511085180867e-06, | |
| "loss": 0.4434, | |
| "step": 2924 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.0287339556592766e-06, | |
| "loss": 0.6256, | |
| "step": 2928 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.025816802800467e-06, | |
| "loss": 0.4372, | |
| "step": 2932 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.0228996499416573e-06, | |
| "loss": 0.4225, | |
| "step": 2936 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.0199824970828473e-06, | |
| "loss": 0.6176, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.0170653442240376e-06, | |
| "loss": 0.5814, | |
| "step": 2944 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.0141481913652276e-06, | |
| "loss": 0.6453, | |
| "step": 2948 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.011231038506418e-06, | |
| "loss": 0.599, | |
| "step": 2952 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.008313885647608e-06, | |
| "loss": 0.4661, | |
| "step": 2956 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.0053967327887982e-06, | |
| "loss": 0.3218, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.003208868144691e-06, | |
| "loss": 0.6456, | |
| "step": 2964 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.0002917152858813e-06, | |
| "loss": 0.2891, | |
| "step": 2968 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.9973745624270716e-06, | |
| "loss": 0.4821, | |
| "step": 2972 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9944574095682616e-06, | |
| "loss": 0.733, | |
| "step": 2976 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.991540256709452e-06, | |
| "loss": 0.2193, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.988623103850642e-06, | |
| "loss": 0.4148, | |
| "step": 2984 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9857059509918322e-06, | |
| "loss": 0.5474, | |
| "step": 2988 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.982788798133022e-06, | |
| "loss": 0.4005, | |
| "step": 2992 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.9798716452742125e-06, | |
| "loss": 0.5799, | |
| "step": 2996 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.9769544924154025e-06, | |
| "loss": 0.4967, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.974037339556593e-06, | |
| "loss": 0.4672, | |
| "step": 3004 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.9711201866977836e-06, | |
| "loss": 0.4892, | |
| "step": 3008 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.9682030338389735e-06, | |
| "loss": 0.554, | |
| "step": 3012 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.965285880980164e-06, | |
| "loss": 0.3553, | |
| "step": 3016 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 2.962368728121354e-06, | |
| "loss": 0.5259, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.959451575262544e-06, | |
| "loss": 0.4461, | |
| "step": 3024 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.956534422403734e-06, | |
| "loss": 0.4768, | |
| "step": 3028 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.9536172695449245e-06, | |
| "loss": 0.4537, | |
| "step": 3032 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.9507001166861144e-06, | |
| "loss": 0.5611, | |
| "step": 3036 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.947782963827305e-06, | |
| "loss": 0.5684, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.944865810968495e-06, | |
| "loss": 0.2357, | |
| "step": 3044 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.941948658109685e-06, | |
| "loss": 0.4909, | |
| "step": 3048 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.9390315052508755e-06, | |
| "loss": 0.5225, | |
| "step": 3052 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.9361143523920654e-06, | |
| "loss": 0.391, | |
| "step": 3056 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.9331971995332558e-06, | |
| "loss": 0.4058, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.9302800466744457e-06, | |
| "loss": 0.446, | |
| "step": 3064 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 2.927362893815636e-06, | |
| "loss": 0.3136, | |
| "step": 3068 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.924445740956827e-06, | |
| "loss": 0.4259, | |
| "step": 3072 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.9215285880980164e-06, | |
| "loss": 0.4293, | |
| "step": 3076 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.918611435239207e-06, | |
| "loss": 0.6366, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.9156942823803967e-06, | |
| "loss": 0.5103, | |
| "step": 3084 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.9127771295215875e-06, | |
| "loss": 0.4994, | |
| "step": 3088 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.9098599766627774e-06, | |
| "loss": 0.5369, | |
| "step": 3092 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.9069428238039678e-06, | |
| "loss": 0.4632, | |
| "step": 3096 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.9040256709451577e-06, | |
| "loss": 0.3315, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.901108518086348e-06, | |
| "loss": 0.5776, | |
| "step": 3104 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.8981913652275384e-06, | |
| "loss": 0.4545, | |
| "step": 3108 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.8952742123687284e-06, | |
| "loss": 0.6122, | |
| "step": 3112 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 2.8923570595099187e-06, | |
| "loss": 0.2894, | |
| "step": 3116 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.8894399066511087e-06, | |
| "loss": 0.5543, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.886522753792299e-06, | |
| "loss": 0.524, | |
| "step": 3124 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.883605600933489e-06, | |
| "loss": 0.5279, | |
| "step": 3128 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.8806884480746793e-06, | |
| "loss": 0.3323, | |
| "step": 3132 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 2.8777712952158693e-06, | |
| "loss": 0.5169, | |
| "step": 3136 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.8748541423570596e-06, | |
| "loss": 0.4117, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.87193698949825e-06, | |
| "loss": 0.2958, | |
| "step": 3144 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.86901983663944e-06, | |
| "loss": 0.5434, | |
| "step": 3148 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.8661026837806307e-06, | |
| "loss": 0.3813, | |
| "step": 3152 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.8631855309218202e-06, | |
| "loss": 0.4985, | |
| "step": 3156 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 2.860268378063011e-06, | |
| "loss": 0.5145, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.8573512252042005e-06, | |
| "loss": 0.4176, | |
| "step": 3164 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.8544340723453913e-06, | |
| "loss": 0.4034, | |
| "step": 3168 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.8515169194865817e-06, | |
| "loss": 0.514, | |
| "step": 3172 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.8485997666277716e-06, | |
| "loss": 0.4951, | |
| "step": 3176 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.845682613768962e-06, | |
| "loss": 0.3946, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 2.842765460910152e-06, | |
| "loss": 0.4, | |
| "step": 3184 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8398483080513423e-06, | |
| "loss": 0.5394, | |
| "step": 3188 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8369311551925322e-06, | |
| "loss": 0.6328, | |
| "step": 3192 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8340140023337226e-06, | |
| "loss": 0.4449, | |
| "step": 3196 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8310968494749125e-06, | |
| "loss": 0.4787, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.828179696616103e-06, | |
| "loss": 0.3408, | |
| "step": 3204 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8252625437572932e-06, | |
| "loss": 0.4688, | |
| "step": 3208 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.822345390898483e-06, | |
| "loss": 0.557, | |
| "step": 3212 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.8194282380396735e-06, | |
| "loss": 0.5299, | |
| "step": 3216 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.8165110851808635e-06, | |
| "loss": 0.468, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.813593932322054e-06, | |
| "loss": 0.3655, | |
| "step": 3224 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.8106767794632438e-06, | |
| "loss": 0.2575, | |
| "step": 3228 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 2.807759626604434e-06, | |
| "loss": 0.5705, | |
| "step": 3232 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.804842473745624e-06, | |
| "loss": 0.4812, | |
| "step": 3236 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.801925320886815e-06, | |
| "loss": 0.5761, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7990081680280052e-06, | |
| "loss": 0.6039, | |
| "step": 3244 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.796091015169195e-06, | |
| "loss": 0.3454, | |
| "step": 3248 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7931738623103855e-06, | |
| "loss": 0.5554, | |
| "step": 3252 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.7902567094515755e-06, | |
| "loss": 0.4531, | |
| "step": 3256 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.787339556592766e-06, | |
| "loss": 0.3332, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7844224037339558e-06, | |
| "loss": 0.3159, | |
| "step": 3264 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.781505250875146e-06, | |
| "loss": 0.498, | |
| "step": 3268 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7785880980163365e-06, | |
| "loss": 0.4386, | |
| "step": 3272 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.7756709451575264e-06, | |
| "loss": 0.4108, | |
| "step": 3276 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 2.772753792298717e-06, | |
| "loss": 0.3739, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.7698366394399067e-06, | |
| "loss": 0.3242, | |
| "step": 3284 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.766919486581097e-06, | |
| "loss": 0.5226, | |
| "step": 3288 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.764002333722287e-06, | |
| "loss": 0.655, | |
| "step": 3292 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.7610851808634774e-06, | |
| "loss": 0.3379, | |
| "step": 3296 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.7581680280046673e-06, | |
| "loss": 0.5461, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 2.7552508751458577e-06, | |
| "loss": 0.4614, | |
| "step": 3304 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.7523337222870485e-06, | |
| "loss": 0.6258, | |
| "step": 3308 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.749416569428238e-06, | |
| "loss": 0.3237, | |
| "step": 3312 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.7464994165694288e-06, | |
| "loss": 0.4256, | |
| "step": 3316 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.7435822637106187e-06, | |
| "loss": 0.4114, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.740665110851809e-06, | |
| "loss": 0.7195, | |
| "step": 3324 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 2.737747957992999e-06, | |
| "loss": 0.5899, | |
| "step": 3328 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.7348308051341894e-06, | |
| "loss": 0.4853, | |
| "step": 3332 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.7319136522753793e-06, | |
| "loss": 0.4043, | |
| "step": 3336 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.7289964994165697e-06, | |
| "loss": 0.5738, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.72607934655776e-06, | |
| "loss": 0.5704, | |
| "step": 3344 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 2.72316219369895e-06, | |
| "loss": 0.6312, | |
| "step": 3348 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.7202450408401404e-06, | |
| "loss": 0.3233, | |
| "step": 3352 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.7173278879813303e-06, | |
| "loss": 0.5598, | |
| "step": 3356 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.7144107351225207e-06, | |
| "loss": 0.3501, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.7114935822637106e-06, | |
| "loss": 0.3945, | |
| "step": 3364 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.708576429404901e-06, | |
| "loss": 0.4801, | |
| "step": 3368 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 2.7056592765460913e-06, | |
| "loss": 0.5775, | |
| "step": 3372 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.7027421236872813e-06, | |
| "loss": 0.3939, | |
| "step": 3376 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.699824970828472e-06, | |
| "loss": 0.4818, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.6969078179696616e-06, | |
| "loss": 0.4698, | |
| "step": 3384 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.6939906651108523e-06, | |
| "loss": 0.3852, | |
| "step": 3388 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.691073512252042e-06, | |
| "loss": 0.4516, | |
| "step": 3392 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.6881563593932326e-06, | |
| "loss": 0.3766, | |
| "step": 3396 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6852392065344226e-06, | |
| "loss": 0.315, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.682322053675613e-06, | |
| "loss": 0.4293, | |
| "step": 3404 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6794049008168033e-06, | |
| "loss": 0.5251, | |
| "step": 3408 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6764877479579932e-06, | |
| "loss": 0.4517, | |
| "step": 3412 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6735705950991836e-06, | |
| "loss": 0.568, | |
| "step": 3416 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 2.6706534422403735e-06, | |
| "loss": 0.5349, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.667736289381564e-06, | |
| "loss": 0.4316, | |
| "step": 3424 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.664819136522754e-06, | |
| "loss": 0.372, | |
| "step": 3428 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.661901983663944e-06, | |
| "loss": 0.4546, | |
| "step": 3432 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.658984830805134e-06, | |
| "loss": 0.3791, | |
| "step": 3436 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.6560676779463245e-06, | |
| "loss": 0.4281, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 2.653150525087515e-06, | |
| "loss": 0.3564, | |
| "step": 3444 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.650233372228705e-06, | |
| "loss": 0.443, | |
| "step": 3448 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.647316219369895e-06, | |
| "loss": 0.6713, | |
| "step": 3452 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.644399066511085e-06, | |
| "loss": 0.2832, | |
| "step": 3456 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.641481913652276e-06, | |
| "loss": 0.5267, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.6385647607934654e-06, | |
| "loss": 0.688, | |
| "step": 3464 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.635647607934656e-06, | |
| "loss": 0.6487, | |
| "step": 3468 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.6327304550758466e-06, | |
| "loss": 0.6653, | |
| "step": 3472 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.6298133022170365e-06, | |
| "loss": 0.4218, | |
| "step": 3476 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.626896149358227e-06, | |
| "loss": 0.5679, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.623978996499417e-06, | |
| "loss": 0.3429, | |
| "step": 3484 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.621061843640607e-06, | |
| "loss": 0.4428, | |
| "step": 3488 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.618144690781797e-06, | |
| "loss": 0.4073, | |
| "step": 3492 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.6152275379229875e-06, | |
| "loss": 0.476, | |
| "step": 3496 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.6123103850641774e-06, | |
| "loss": 0.3964, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.6093932322053678e-06, | |
| "loss": 0.4971, | |
| "step": 3504 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.606476079346558e-06, | |
| "loss": 0.5309, | |
| "step": 3508 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.603558926487748e-06, | |
| "loss": 0.7094, | |
| "step": 3512 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.6006417736289384e-06, | |
| "loss": 0.4345, | |
| "step": 3516 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.5977246207701284e-06, | |
| "loss": 0.5559, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.5948074679113187e-06, | |
| "loss": 0.519, | |
| "step": 3524 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.5918903150525087e-06, | |
| "loss": 0.4054, | |
| "step": 3528 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.588973162193699e-06, | |
| "loss": 0.4334, | |
| "step": 3532 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.586056009334889e-06, | |
| "loss": 0.3535, | |
| "step": 3536 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5831388564760793e-06, | |
| "loss": 0.6168, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.58022170361727e-06, | |
| "loss": 0.3484, | |
| "step": 3544 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.57730455075846e-06, | |
| "loss": 0.3373, | |
| "step": 3548 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5743873978996504e-06, | |
| "loss": 0.5002, | |
| "step": 3552 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5714702450408404e-06, | |
| "loss": 0.4713, | |
| "step": 3556 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5685530921820307e-06, | |
| "loss": 0.4236, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.5656359393232207e-06, | |
| "loss": 0.49, | |
| "step": 3564 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.562718786464411e-06, | |
| "loss": 0.3107, | |
| "step": 3568 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.5598016336056014e-06, | |
| "loss": 0.6111, | |
| "step": 3572 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.5568844807467913e-06, | |
| "loss": 0.4386, | |
| "step": 3576 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.5539673278879817e-06, | |
| "loss": 0.434, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.5510501750291716e-06, | |
| "loss": 0.3085, | |
| "step": 3584 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.548133022170362e-06, | |
| "loss": 0.6368, | |
| "step": 3588 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.545215869311552e-06, | |
| "loss": 0.3797, | |
| "step": 3592 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.5422987164527423e-06, | |
| "loss": 0.5329, | |
| "step": 3596 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.5393815635939322e-06, | |
| "loss": 0.5015, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.5364644107351226e-06, | |
| "loss": 0.333, | |
| "step": 3604 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.5335472578763134e-06, | |
| "loss": 0.349, | |
| "step": 3608 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.530630105017503e-06, | |
| "loss": 0.5365, | |
| "step": 3612 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.5277129521586937e-06, | |
| "loss": 0.6777, | |
| "step": 3616 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.524795799299883e-06, | |
| "loss": 0.5432, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.521878646441074e-06, | |
| "loss": 0.2763, | |
| "step": 3624 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.518961493582264e-06, | |
| "loss": 0.5183, | |
| "step": 3628 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.5160443407234543e-06, | |
| "loss": 0.4486, | |
| "step": 3632 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.513127187864644e-06, | |
| "loss": 0.3521, | |
| "step": 3636 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.5102100350058346e-06, | |
| "loss": 0.5112, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.507292882147025e-06, | |
| "loss": 0.378, | |
| "step": 3644 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.504375729288215e-06, | |
| "loss": 0.4282, | |
| "step": 3648 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.5014585764294052e-06, | |
| "loss": 0.5283, | |
| "step": 3652 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.4985414235705956e-06, | |
| "loss": 0.5269, | |
| "step": 3656 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.4956242707117855e-06, | |
| "loss": 0.3568, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.492707117852976e-06, | |
| "loss": 0.3911, | |
| "step": 3664 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.489789964994166e-06, | |
| "loss": 0.554, | |
| "step": 3668 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.486872812135356e-06, | |
| "loss": 0.6132, | |
| "step": 3672 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.483955659276546e-06, | |
| "loss": 0.5473, | |
| "step": 3676 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.4810385064177365e-06, | |
| "loss": 0.472, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4781213535589264e-06, | |
| "loss": 0.4616, | |
| "step": 3684 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4752042007001172e-06, | |
| "loss": 0.5357, | |
| "step": 3688 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.472287047841307e-06, | |
| "loss": 0.5629, | |
| "step": 3692 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4693698949824975e-06, | |
| "loss": 0.5284, | |
| "step": 3696 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.4664527421236875e-06, | |
| "loss": 0.3988, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.463535589264878e-06, | |
| "loss": 0.7007, | |
| "step": 3704 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.4606184364060678e-06, | |
| "loss": 0.2642, | |
| "step": 3708 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.457701283547258e-06, | |
| "loss": 0.6179, | |
| "step": 3712 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.454784130688448e-06, | |
| "loss": 0.5595, | |
| "step": 3716 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.4518669778296384e-06, | |
| "loss": 0.4074, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.448949824970829e-06, | |
| "loss": 0.3635, | |
| "step": 3724 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.446032672112019e-06, | |
| "loss": 0.3888, | |
| "step": 3728 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.443115519253209e-06, | |
| "loss": 0.2755, | |
| "step": 3732 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.4401983663943995e-06, | |
| "loss": 0.4511, | |
| "step": 3736 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.4372812135355894e-06, | |
| "loss": 0.5516, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.4343640606767798e-06, | |
| "loss": 0.4041, | |
| "step": 3744 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.4314469078179697e-06, | |
| "loss": 0.1592, | |
| "step": 3748 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.42852975495916e-06, | |
| "loss": 0.5583, | |
| "step": 3752 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.4256126021003504e-06, | |
| "loss": 0.445, | |
| "step": 3756 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.4226954492415404e-06, | |
| "loss": 0.3201, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.4197782963827307e-06, | |
| "loss": 0.521, | |
| "step": 3764 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.416861143523921e-06, | |
| "loss": 0.4229, | |
| "step": 3768 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.413943990665111e-06, | |
| "loss": 0.4319, | |
| "step": 3772 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.4110268378063014e-06, | |
| "loss": 0.4708, | |
| "step": 3776 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.4081096849474913e-06, | |
| "loss": 0.5419, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.4051925320886817e-06, | |
| "loss": 0.4454, | |
| "step": 3784 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.402275379229872e-06, | |
| "loss": 0.4804, | |
| "step": 3788 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.399358226371062e-06, | |
| "loss": 0.5686, | |
| "step": 3792 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.3964410735122523e-06, | |
| "loss": 0.5354, | |
| "step": 3796 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3935239206534423e-06, | |
| "loss": 0.5101, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3906067677946326e-06, | |
| "loss": 0.3946, | |
| "step": 3804 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.387689614935823e-06, | |
| "loss": 0.4951, | |
| "step": 3808 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.384772462077013e-06, | |
| "loss": 0.4952, | |
| "step": 3812 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3818553092182033e-06, | |
| "loss": 0.5854, | |
| "step": 3816 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.3789381563593932e-06, | |
| "loss": 0.4074, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3760210035005836e-06, | |
| "loss": 0.4058, | |
| "step": 3824 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.373103850641774e-06, | |
| "loss": 0.4952, | |
| "step": 3828 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.370186697782964e-06, | |
| "loss": 0.5502, | |
| "step": 3832 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3672695449241543e-06, | |
| "loss": 0.4379, | |
| "step": 3836 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.364352392065344e-06, | |
| "loss": 0.4695, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.3614352392065346e-06, | |
| "loss": 0.5127, | |
| "step": 3844 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.3585180863477245e-06, | |
| "loss": 0.4037, | |
| "step": 3848 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.355600933488915e-06, | |
| "loss": 0.3387, | |
| "step": 3852 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.3526837806301052e-06, | |
| "loss": 0.5302, | |
| "step": 3856 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.3497666277712956e-06, | |
| "loss": 0.4549, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.3468494749124855e-06, | |
| "loss": 0.4871, | |
| "step": 3864 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.343932322053676e-06, | |
| "loss": 0.4167, | |
| "step": 3868 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.341015169194866e-06, | |
| "loss": 0.4498, | |
| "step": 3872 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.338098016336056e-06, | |
| "loss": 0.3591, | |
| "step": 3876 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.335180863477246e-06, | |
| "loss": 0.3389, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.3322637106184365e-06, | |
| "loss": 0.4688, | |
| "step": 3884 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.329346557759627e-06, | |
| "loss": 0.4151, | |
| "step": 3888 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.3264294049008172e-06, | |
| "loss": 0.5132, | |
| "step": 3892 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.323512252042007e-06, | |
| "loss": 0.3682, | |
| "step": 3896 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.3205950991831975e-06, | |
| "loss": 0.3668, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.3176779463243875e-06, | |
| "loss": 0.2637, | |
| "step": 3904 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.314760793465578e-06, | |
| "loss": 0.5291, | |
| "step": 3908 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.3118436406067678e-06, | |
| "loss": 0.5459, | |
| "step": 3912 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.308926487747958e-06, | |
| "loss": 0.5774, | |
| "step": 3916 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.306009334889148e-06, | |
| "loss": 0.5955, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.303092182030339e-06, | |
| "loss": 0.2941, | |
| "step": 3924 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.300175029171529e-06, | |
| "loss": 0.4735, | |
| "step": 3928 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.297257876312719e-06, | |
| "loss": 0.27, | |
| "step": 3932 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.294340723453909e-06, | |
| "loss": 0.496, | |
| "step": 3936 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2914235705950995e-06, | |
| "loss": 0.279, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2885064177362894e-06, | |
| "loss": 0.3848, | |
| "step": 3944 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2855892648774798e-06, | |
| "loss": 0.4868, | |
| "step": 3948 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2826721120186697e-06, | |
| "loss": 0.5328, | |
| "step": 3952 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2797549591598605e-06, | |
| "loss": 0.3838, | |
| "step": 3956 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.2768378063010504e-06, | |
| "loss": 0.4603, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2739206534422408e-06, | |
| "loss": 0.515, | |
| "step": 3964 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2710035005834307e-06, | |
| "loss": 0.4371, | |
| "step": 3968 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.268086347724621e-06, | |
| "loss": 0.5561, | |
| "step": 3972 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.265169194865811e-06, | |
| "loss": 0.3882, | |
| "step": 3976 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2622520420070014e-06, | |
| "loss": 0.4774, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.2593348891481913e-06, | |
| "loss": 0.6546, | |
| "step": 3984 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.2564177362893817e-06, | |
| "loss": 0.4335, | |
| "step": 3988 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.253500583430572e-06, | |
| "loss": 0.3352, | |
| "step": 3992 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.2505834305717624e-06, | |
| "loss": 0.2946, | |
| "step": 3996 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.2476662777129523e-06, | |
| "loss": 0.2657, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.2447491248541427e-06, | |
| "loss": 0.4859, | |
| "step": 4004 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.2418319719953326e-06, | |
| "loss": 0.3047, | |
| "step": 4008 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.238914819136523e-06, | |
| "loss": 0.5828, | |
| "step": 4012 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.235997666277713e-06, | |
| "loss": 0.4141, | |
| "step": 4016 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.2330805134189033e-06, | |
| "loss": 0.4643, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.2301633605600937e-06, | |
| "loss": 0.5517, | |
| "step": 4024 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.2272462077012836e-06, | |
| "loss": 0.551, | |
| "step": 4028 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.224329054842474e-06, | |
| "loss": 0.4237, | |
| "step": 4032 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.2214119019836643e-06, | |
| "loss": 0.4345, | |
| "step": 4036 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.2184947491248543e-06, | |
| "loss": 0.428, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.2155775962660446e-06, | |
| "loss": 0.3503, | |
| "step": 4044 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.2126604434072346e-06, | |
| "loss": 0.4281, | |
| "step": 4048 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.209743290548425e-06, | |
| "loss": 0.5451, | |
| "step": 4052 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.2068261376896153e-06, | |
| "loss": 0.6199, | |
| "step": 4056 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.2039089848308052e-06, | |
| "loss": 0.6851, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.2009918319719956e-06, | |
| "loss": 0.3541, | |
| "step": 4064 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.1980746791131855e-06, | |
| "loss": 0.4333, | |
| "step": 4068 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.195157526254376e-06, | |
| "loss": 0.413, | |
| "step": 4072 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.1922403733955663e-06, | |
| "loss": 0.4385, | |
| "step": 4076 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.189323220536756e-06, | |
| "loss": 0.3163, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.1864060676779466e-06, | |
| "loss": 0.4678, | |
| "step": 4084 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.183488914819137e-06, | |
| "loss": 0.2889, | |
| "step": 4088 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.180571761960327e-06, | |
| "loss": 0.3339, | |
| "step": 4092 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.1776546091015172e-06, | |
| "loss": 0.4381, | |
| "step": 4096 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.174737456242707e-06, | |
| "loss": 0.4926, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.1718203033838975e-06, | |
| "loss": 0.4147, | |
| "step": 4104 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.1689031505250875e-06, | |
| "loss": 0.525, | |
| "step": 4108 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.165985997666278e-06, | |
| "loss": 0.6573, | |
| "step": 4112 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.163068844807468e-06, | |
| "loss": 0.4188, | |
| "step": 4116 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.1601516919486586e-06, | |
| "loss": 0.2622, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.1572345390898485e-06, | |
| "loss": 0.4533, | |
| "step": 4124 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.154317386231039e-06, | |
| "loss": 0.5007, | |
| "step": 4128 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.151400233372229e-06, | |
| "loss": 0.5307, | |
| "step": 4132 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.148483080513419e-06, | |
| "loss": 0.4071, | |
| "step": 4136 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.145565927654609e-06, | |
| "loss": 0.5252, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.1426487747957995e-06, | |
| "loss": 0.3672, | |
| "step": 4144 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.1397316219369894e-06, | |
| "loss": 0.3608, | |
| "step": 4148 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.1368144690781798e-06, | |
| "loss": 0.4581, | |
| "step": 4152 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.13389731621937e-06, | |
| "loss": 0.4953, | |
| "step": 4156 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.1309801633605605e-06, | |
| "loss": 0.515, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.1280630105017504e-06, | |
| "loss": 0.4272, | |
| "step": 4164 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.1251458576429408e-06, | |
| "loss": 0.5713, | |
| "step": 4168 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.1222287047841307e-06, | |
| "loss": 0.3837, | |
| "step": 4172 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.119311551925321e-06, | |
| "loss": 0.4367, | |
| "step": 4176 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.116394399066511e-06, | |
| "loss": 0.3561, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.1134772462077014e-06, | |
| "loss": 0.2825, | |
| "step": 4184 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.1105600933488917e-06, | |
| "loss": 0.2891, | |
| "step": 4188 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.107642940490082e-06, | |
| "loss": 0.4617, | |
| "step": 4192 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.104725787631272e-06, | |
| "loss": 0.4423, | |
| "step": 4196 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.1018086347724624e-06, | |
| "loss": 0.2344, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0988914819136523e-06, | |
| "loss": 0.5355, | |
| "step": 4204 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0959743290548427e-06, | |
| "loss": 0.427, | |
| "step": 4208 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0930571761960326e-06, | |
| "loss": 0.3997, | |
| "step": 4212 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.090140023337223e-06, | |
| "loss": 0.3945, | |
| "step": 4216 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.0872228704784134e-06, | |
| "loss": 0.3998, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0843057176196037e-06, | |
| "loss": 0.4695, | |
| "step": 4224 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0813885647607937e-06, | |
| "loss": 0.3051, | |
| "step": 4228 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.078471411901984e-06, | |
| "loss": 0.4195, | |
| "step": 4232 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.075554259043174e-06, | |
| "loss": 0.4064, | |
| "step": 4236 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0726371061843643e-06, | |
| "loss": 0.4709, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.0697199533255543e-06, | |
| "loss": 0.6638, | |
| "step": 4244 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0668028004667446e-06, | |
| "loss": 0.4436, | |
| "step": 4248 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0638856476079346e-06, | |
| "loss": 0.3434, | |
| "step": 4252 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.060968494749125e-06, | |
| "loss": 0.2871, | |
| "step": 4256 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0580513418903153e-06, | |
| "loss": 0.2696, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0551341890315057e-06, | |
| "loss": 0.373, | |
| "step": 4264 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.0522170361726956e-06, | |
| "loss": 0.4656, | |
| "step": 4268 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.049299883313886e-06, | |
| "loss": 0.3027, | |
| "step": 4272 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.046382730455076e-06, | |
| "loss": 0.7287, | |
| "step": 4276 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.0434655775962663e-06, | |
| "loss": 0.3299, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.040548424737456e-06, | |
| "loss": 0.3556, | |
| "step": 4284 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.0376312718786466e-06, | |
| "loss": 0.3275, | |
| "step": 4288 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.034714119019837e-06, | |
| "loss": 0.334, | |
| "step": 4292 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.031796966161027e-06, | |
| "loss": 0.4437, | |
| "step": 4296 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.0288798133022172e-06, | |
| "loss": 0.4776, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.0259626604434076e-06, | |
| "loss": 0.497, | |
| "step": 4304 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.0230455075845975e-06, | |
| "loss": 0.6054, | |
| "step": 4308 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.020128354725788e-06, | |
| "loss": 0.3877, | |
| "step": 4312 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.017211201866978e-06, | |
| "loss": 0.3442, | |
| "step": 4316 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.014294049008168e-06, | |
| "loss": 0.5557, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.0113768961493586e-06, | |
| "loss": 0.5137, | |
| "step": 4324 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.0084597432905485e-06, | |
| "loss": 0.6162, | |
| "step": 4328 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.005542590431739e-06, | |
| "loss": 0.4181, | |
| "step": 4332 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.002625437572929e-06, | |
| "loss": 0.3354, | |
| "step": 4336 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.999708284714119e-06, | |
| "loss": 0.3924, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.9967911318553095e-06, | |
| "loss": 0.3781, | |
| "step": 4344 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.9938739789964995e-06, | |
| "loss": 0.4388, | |
| "step": 4348 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.99095682613769e-06, | |
| "loss": 0.4512, | |
| "step": 4352 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.98803967327888e-06, | |
| "loss": 0.4171, | |
| "step": 4356 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.98512252042007e-06, | |
| "loss": 0.3768, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9822053675612605e-06, | |
| "loss": 0.3517, | |
| "step": 4364 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9792882147024504e-06, | |
| "loss": 0.412, | |
| "step": 4368 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9763710618436408e-06, | |
| "loss": 0.2887, | |
| "step": 4372 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9734539089848307e-06, | |
| "loss": 0.3032, | |
| "step": 4376 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.970536756126021e-06, | |
| "loss": 0.3046, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 1.9676196032672114e-06, | |
| "loss": 0.3144, | |
| "step": 4384 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.964702450408402e-06, | |
| "loss": 0.533, | |
| "step": 4388 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.9617852975495917e-06, | |
| "loss": 0.3993, | |
| "step": 4392 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.958868144690782e-06, | |
| "loss": 0.3253, | |
| "step": 4396 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.955950991831972e-06, | |
| "loss": 0.4888, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.9530338389731624e-06, | |
| "loss": 0.5357, | |
| "step": 4404 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 1.9501166861143524e-06, | |
| "loss": 0.3057, | |
| "step": 4408 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9471995332555427e-06, | |
| "loss": 0.3553, | |
| "step": 4412 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9442823803967327e-06, | |
| "loss": 0.3878, | |
| "step": 4416 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9413652275379234e-06, | |
| "loss": 0.315, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9384480746791134e-06, | |
| "loss": 0.3791, | |
| "step": 4424 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9355309218203037e-06, | |
| "loss": 0.4092, | |
| "step": 4428 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 1.9326137689614937e-06, | |
| "loss": 0.3911, | |
| "step": 4432 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.929696616102684e-06, | |
| "loss": 0.3135, | |
| "step": 4436 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.926779463243874e-06, | |
| "loss": 0.3574, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.9238623103850643e-06, | |
| "loss": 0.3147, | |
| "step": 4444 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.9209451575262543e-06, | |
| "loss": 0.5802, | |
| "step": 4448 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.9180280046674446e-06, | |
| "loss": 0.5169, | |
| "step": 4452 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 1.915110851808635e-06, | |
| "loss": 0.4599, | |
| "step": 4456 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.9121936989498254e-06, | |
| "loss": 0.3763, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.9092765460910153e-06, | |
| "loss": 0.4633, | |
| "step": 4464 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.9063593932322055e-06, | |
| "loss": 0.462, | |
| "step": 4468 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.9034422403733956e-06, | |
| "loss": 0.4908, | |
| "step": 4472 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.900525087514586e-06, | |
| "loss": 0.3367, | |
| "step": 4476 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 1.8976079346557761e-06, | |
| "loss": 0.4497, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.8946907817969663e-06, | |
| "loss": 0.3253, | |
| "step": 4484 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.8917736289381566e-06, | |
| "loss": 0.5464, | |
| "step": 4488 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.8888564760793468e-06, | |
| "loss": 0.3004, | |
| "step": 4492 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.885939323220537e-06, | |
| "loss": 0.3876, | |
| "step": 4496 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 1.883022170361727e-06, | |
| "loss": 0.4443, | |
| "step": 4500 | |
| } | |
| ], | |
| "logging_steps": 4, | |
| "max_steps": 7056, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "total_flos": 34569317253120.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |