| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 20.0, | |
| "eval_steps": 500, | |
| "global_step": 900, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 6.3594, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.4814814814814815e-05, | |
| "loss": 6.3594, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 2.2222222222222223e-05, | |
| "loss": 6.3008, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 2.962962962962963e-05, | |
| "loss": 6.0938, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 3.7037037037037037e-05, | |
| "loss": 5.7578, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 5.4258, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 5.185185185185185e-05, | |
| "loss": 5.5195, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 5.925925925925926e-05, | |
| "loss": 5.457, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 5.4727, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 7.407407407407407e-05, | |
| "loss": 5.4062, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 8.148148148148148e-05, | |
| "loss": 5.3789, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 8.888888888888889e-05, | |
| "loss": 5.3086, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.62962962962963e-05, | |
| "loss": 5.375, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0001037037037037037, | |
| "loss": 5.1719, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.00011111111111111112, | |
| "loss": 5.1719, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 0.00011851851851851852, | |
| "loss": 5.1836, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.00012592592592592592, | |
| "loss": 5.2617, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 5.0781, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 0.00014074074074074076, | |
| "loss": 5.0859, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 0.00014814814814814815, | |
| "loss": 5.1914, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 0.00015555555555555556, | |
| "loss": 5.2227, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.00016296296296296295, | |
| "loss": 5.1094, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.00017037037037037037, | |
| "loss": 5.0156, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 0.00017777777777777779, | |
| "loss": 5.1289, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 0.0001851851851851852, | |
| "loss": 5.0703, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 0.0001925925925925926, | |
| "loss": 4.9531, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.0002, | |
| "loss": 4.8281, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 0.00019999935249847796, | |
| "loss": 4.7852, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 0.00019999741000229694, | |
| "loss": 4.7969, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 0.00019999417253661235, | |
| "loss": 4.8047, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 0.00019998964014334946, | |
| "loss": 4.6992, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 0.00019998381288120295, | |
| "loss": 4.7617, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 0.00019997669082563597, | |
| "loss": 4.8281, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 0.0001999682740688794, | |
| "loss": 4.625, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 0.0001999585627199305, | |
| "loss": 4.5703, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 0.00019994755690455152, | |
| "loss": 4.6289, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 0.00019993525676526805, | |
| "loss": 4.582, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 0.00019992166246136738, | |
| "loss": 4.5508, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 0.00019990677416889608, | |
| "loss": 4.5039, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 0.000199890592080658, | |
| "loss": 4.4609, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 0.00019987311640621173, | |
| "loss": 4.4102, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 0.0001998543473718677, | |
| "loss": 4.3477, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 0.00019983428522068556, | |
| "loss": 4.2051, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 0.0001998129302124707, | |
| "loss": 4.25, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 0.00019979028262377118, | |
| "loss": 4.0957, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 0.00019976634274787393, | |
| "loss": 4.0957, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 0.00019974111089480112, | |
| "loss": 4.084, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 0.00019971458739130598, | |
| "loss": 4.0977, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 0.00019968677258086866, | |
| "loss": 4.0059, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 0.00019965766682369186, | |
| "loss": 4.0938, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 0.000199627270496696, | |
| "loss": 4.0527, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 0.00019959558399351444, | |
| "loss": 4.1016, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 0.00019956260772448833, | |
| "loss": 4.0449, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 0.0001995283421166614, | |
| "loss": 3.9648, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 0.0001994927876137743, | |
| "loss": 3.9746, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 0.00019945594467625895, | |
| "loss": 3.8984, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 0.00019941781378123244, | |
| "loss": 3.873, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 0.00019937839542249108, | |
| "loss": 3.7598, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 0.0001993376901105038, | |
| "loss": 3.8691, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 0.00019929569837240564, | |
| "loss": 3.8906, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 0.0001992524207519909, | |
| "loss": 3.7227, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 0.00019920785780970604, | |
| "loss": 3.7578, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 0.00019916201012264254, | |
| "loss": 3.7109, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 0.00019911487828452932, | |
| "loss": 3.6836, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 0.00019906646290572514, | |
| "loss": 3.6445, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 0.00019901676461321068, | |
| "loss": 3.7051, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 0.00019896578405058027, | |
| "loss": 3.6348, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 0.00019891352187803376, | |
| "loss": 3.582, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 0.00019885997877236788, | |
| "loss": 3.668, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 0.0001988051554269675, | |
| "loss": 3.6387, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 0.0001987490525517966, | |
| "loss": 3.4805, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 0.00019869167087338907, | |
| "loss": 3.4434, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 0.00019863301113483948, | |
| "loss": 3.4824, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 0.00019857307409579318, | |
| "loss": 3.4902, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 0.00019851186053243666, | |
| "loss": 3.3926, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 0.0001984493712374874, | |
| "loss": 3.4238, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 0.00019838560702018374, | |
| "loss": 3.375, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 0.00019832056870627417, | |
| "loss": 3.5664, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 0.00019825425713800688, | |
| "loss": 3.3262, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 0.00019818667317411865, | |
| "loss": 3.5371, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 0.0001981178176898239, | |
| "loss": 3.5586, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 0.00019804769157680328, | |
| "loss": 3.4258, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 0.000197976295743192, | |
| "loss": 3.334, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 0.00019790363111356837, | |
| "loss": 3.3652, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 0.0001978296986289415, | |
| "loss": 3.5098, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 0.00019775449924673934, | |
| "loss": 3.4902, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 0.00019767803394079615, | |
| "loss": 3.3887, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 0.00019760030370134004, | |
| "loss": 3.4141, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 0.0001975213095349799, | |
| "loss": 3.2051, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 0.00019744105246469263, | |
| "loss": 2.9727, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 0.0001973595335298097, | |
| "loss": 3.0605, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 0.0001972767537860038, | |
| "loss": 3.1074, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 0.0001971927143052752, | |
| "loss": 2.9453, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 0.00019710741617593767, | |
| "loss": 2.9258, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 0.00019702086050260456, | |
| "loss": 3.0449, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 0.00019693304840617457, | |
| "loss": 2.9863, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 0.00019684398102381694, | |
| "loss": 2.8984, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 0.000196753659508957, | |
| "loss": 2.8652, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 0.00019666208503126112, | |
| "loss": 3.0098, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 0.0001965692587766216, | |
| "loss": 2.7969, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 0.00019647518194714124, | |
| "loss": 2.873, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 0.00019637985576111778, | |
| "loss": 3.0078, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 0.00019628328145302837, | |
| "loss": 2.8848, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 0.00019618546027351316, | |
| "loss": 2.8301, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 0.0001960863934893594, | |
| "loss": 3.1016, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 0.00019598608238348493, | |
| "loss": 2.7852, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 0.0001958845282549217, | |
| "loss": 3.0176, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 0.00019578173241879872, | |
| "loss": 2.8535, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 0.0001956776962063252, | |
| "loss": 2.9648, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 0.00019557242096477327, | |
| "loss": 2.9141, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 0.00019546590805746052, | |
| "loss": 2.9336, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 0.00019535815886373233, | |
| "loss": 2.9141, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 0.00019524917477894404, | |
| "loss": 2.8945, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 0.00019513895721444286, | |
| "loss": 2.7754, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 0.00019502750759754962, | |
| "loss": 2.793, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 0.0001949148273715402, | |
| "loss": 2.8418, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 0.00019480091799562704, | |
| "loss": 2.707, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 0.00019468578094493998, | |
| "loss": 2.7285, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 0.00019456941771050734, | |
| "loss": 2.9492, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 0.00019445182979923654, | |
| "loss": 2.7891, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 0.00019433301873389463, | |
| "loss": 2.793, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 0.00019421298605308847, | |
| "loss": 2.7969, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 0.000194091733311245, | |
| "loss": 2.873, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 0.00019396926207859084, | |
| "loss": 2.8613, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 0.00019384557394113228, | |
| "loss": 2.8047, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 0.00019372067050063438, | |
| "loss": 2.9355, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 0.00019359455337460054, | |
| "loss": 2.8828, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 0.00019346722419625136, | |
| "loss": 2.8652, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 0.0001933386846145036, | |
| "loss": 2.9199, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 0.00019320893629394873, | |
| "loss": 2.9102, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 0.00019307798091483145, | |
| "loss": 2.7422, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 0.00019294582017302797, | |
| "loss": 2.8496, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 0.0001928124557800238, | |
| "loss": 2.8379, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 0.000192677889462892, | |
| "loss": 2.7012, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 0.00019254212296427044, | |
| "loss": 2.6133, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 0.00019240515804233937, | |
| "loss": 2.498, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 0.00019226699647079874, | |
| "loss": 2.4883, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 0.0001921276400388451, | |
| "loss": 2.2422, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 0.0001919870905511485, | |
| "loss": 2.4473, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 0.00019184534982782904, | |
| "loss": 2.541, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 0.00019170241970443343, | |
| "loss": 2.416, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 0.00019155830203191113, | |
| "loss": 2.4648, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 0.00019141299867659036, | |
| "loss": 2.5273, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 0.00019126651152015403, | |
| "loss": 2.4902, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 0.00019111884245961522, | |
| "loss": 2.4023, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 0.00019096999340729282, | |
| "loss": 2.4238, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 0.00019081996629078657, | |
| "loss": 2.4004, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 0.00019066876305295216, | |
| "loss": 2.3242, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 0.00019051638565187616, | |
| "loss": 2.4141, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 0.00019036283606085053, | |
| "loss": 2.5547, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 0.0001902081162683472, | |
| "loss": 2.4004, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 0.00019005222827799212, | |
| "loss": 2.3369, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 0.00018989517410853955, | |
| "loss": 2.4609, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 0.00018973695579384576, | |
| "loss": 2.332, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 0.00018957757538284273, | |
| "loss": 2.2852, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 0.00018941703493951164, | |
| "loss": 2.3906, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 0.00018925533654285615, | |
| "loss": 2.3008, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 0.00018909248228687536, | |
| "loss": 2.3008, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 0.00018892847428053693, | |
| "loss": 2.3271, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 0.00018876331464774945, | |
| "loss": 2.3242, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 0.00018859700552733516, | |
| "loss": 2.3652, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 0.00018842954907300236, | |
| "loss": 2.3555, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 0.00018826094745331715, | |
| "loss": 2.3613, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 0.00018809120285167565, | |
| "loss": 2.3691, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 0.00018792031746627563, | |
| "loss": 2.3066, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 0.000187748293510088, | |
| "loss": 2.3613, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 0.0001875751332108283, | |
| "loss": 2.3457, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 0.0001874008388109276, | |
| "loss": 2.3652, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 0.0001872254125675037, | |
| "loss": 2.3506, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 0.0001870488567523318, | |
| "loss": 2.3672, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 0.00018687117365181512, | |
| "loss": 2.2666, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 0.00018669236556695515, | |
| "loss": 2.5156, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 0.00018651243481332213, | |
| "loss": 2.3926, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 0.00018633138372102468, | |
| "loss": 2.4102, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 0.00018614921463468002, | |
| "loss": 2.4082, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 0.00018596592991338334, | |
| "loss": 2.2695, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 0.00018578153193067745, | |
| "loss": 2.418, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 0.0001855960230745217, | |
| "loss": 2.3223, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 0.0001854094057472615, | |
| "loss": 2.3398, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 0.00018522168236559695, | |
| "loss": 2.2422, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 0.00018503285536055147, | |
| "loss": 2.0654, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 0.0001848429271774406, | |
| "loss": 2.0186, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 0.00018465190027584005, | |
| "loss": 2.0049, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 0.00018445977712955403, | |
| "loss": 2.0869, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 0.0001842665602265831, | |
| "loss": 2.0547, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 0.00018407225206909208, | |
| "loss": 2.0264, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 0.0001838768551733775, | |
| "loss": 1.8994, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 0.00018368037206983515, | |
| "loss": 2.0342, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 0.00018348280530292713, | |
| "loss": 2.0576, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 0.00018328415743114912, | |
| "loss": 2.0615, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 0.00018308443102699713, | |
| "loss": 2.0518, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 0.00018288362867693414, | |
| "loss": 2.0537, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 0.0001826817529813567, | |
| "loss": 1.9531, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 0.00018247880655456125, | |
| "loss": 2.0049, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 0.00018227479202471015, | |
| "loss": 2.0693, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 0.00018206971203379773, | |
| "loss": 2.0088, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 0.00018186356923761626, | |
| "loss": 1.9189, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 0.0001816563663057211, | |
| "loss": 2.0371, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 0.00018144810592139656, | |
| "loss": 2.1172, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 0.00018123879078162097, | |
| "loss": 1.9873, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 0.00018102842359703176, | |
| "loss": 2.0508, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 0.0001808170070918904, | |
| "loss": 2.0654, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 0.00018060454400404695, | |
| "loss": 2.0928, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 0.000180391037084905, | |
| "loss": 2.0166, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 0.0001801764890993856, | |
| "loss": 2.0518, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 0.0001799609028258917, | |
| "loss": 1.9688, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 0.00017974428105627208, | |
| "loss": 1.9512, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 0.0001795266265957853, | |
| "loss": 2.0537, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 0.0001793079422630632, | |
| "loss": 2.0195, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 0.00017908823089007457, | |
| "loss": 2.0605, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 0.00017886749532208837, | |
| "loss": 2.1709, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 0.00017864573841763694, | |
| "loss": 1.9629, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 0.00017842296304847893, | |
| "loss": 2.126, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 0.00017819917209956215, | |
| "loss": 2.0469, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 0.00017797436846898619, | |
| "loss": 2.0107, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 0.00017774855506796496, | |
| "loss": 2.0693, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 0.0001775217348207888, | |
| "loss": 2.0762, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 0.00017729391066478688, | |
| "loss": 2.0322, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 0.00017706508555028893, | |
| "loss": 2.0098, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 0.00017683526244058716, | |
| "loss": 2.0449, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 0.0001766044443118978, | |
| "loss": 2.0615, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 0.0001763726341533227, | |
| "loss": 1.9961, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 0.00017613983496681044, | |
| "loss": 2.0596, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 0.00017590604976711754, | |
| "loss": 2.0498, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 0.00017567128158176953, | |
| "loss": 2.0449, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 0.00017543553345102152, | |
| "loss": 1.7109, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 0.00017519880842781894, | |
| "loss": 1.6914, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 5.07, | |
| "learning_rate": 0.0001749611095777581, | |
| "loss": 1.8711, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 0.00017472243997904626, | |
| "loss": 1.6826, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 0.00017448280272246212, | |
| "loss": 1.8135, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 5.13, | |
| "learning_rate": 0.00017424220091131535, | |
| "loss": 1.7578, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 0.00017400063766140678, | |
| "loss": 1.7686, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 0.0001737581161009878, | |
| "loss": 1.7383, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 0.00017351463937072004, | |
| "loss": 1.6729, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 5.22, | |
| "learning_rate": 0.00017327021062363458, | |
| "loss": 1.6133, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 0.00017302483302509108, | |
| "loss": 1.7119, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 5.27, | |
| "learning_rate": 0.00017277850975273696, | |
| "loss": 1.6914, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 0.00017253124399646606, | |
| "loss": 1.7627, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 0.00017228303895837748, | |
| "loss": 1.7412, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "learning_rate": 0.000172033897852734, | |
| "loss": 1.7412, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 0.00017178382390592057, | |
| "loss": 1.8037, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 0.0001715328203564023, | |
| "loss": 1.6709, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 0.00017128089045468294, | |
| "loss": 1.6562, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 5.42, | |
| "learning_rate": 0.00017102803746326227, | |
| "loss": 1.6533, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 5.44, | |
| "learning_rate": 0.00017077426465659433, | |
| "loss": 1.752, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 0.0001705195753210446, | |
| "loss": 1.7041, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 0.00017026397275484773, | |
| "loss": 1.7969, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 0.0001700074602680648, | |
| "loss": 1.7539, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 0.0001697500411825403, | |
| "loss": 1.79, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 5.56, | |
| "learning_rate": 0.00016949171883185918, | |
| "loss": 1.8535, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 5.58, | |
| "learning_rate": 0.0001692324965613038, | |
| "loss": 1.8213, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 0.00016897237772781044, | |
| "loss": 1.8232, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 5.62, | |
| "learning_rate": 0.00016871136569992587, | |
| "loss": 1.8584, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 0.00016844946385776384, | |
| "loss": 1.8145, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 0.0001681866755929612, | |
| "loss": 1.7588, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 5.69, | |
| "learning_rate": 0.00016792300430863396, | |
| "loss": 1.8535, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 0.0001676584534193332, | |
| "loss": 1.8047, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 5.73, | |
| "learning_rate": 0.00016739302635100108, | |
| "loss": 1.7617, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 5.76, | |
| "learning_rate": 0.00016712672654092622, | |
| "loss": 1.8027, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 5.78, | |
| "learning_rate": 0.0001668595574376992, | |
| "loss": 1.8447, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 0.00016659152250116812, | |
| "loss": 1.8174, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 0.0001663226252023935, | |
| "loss": 1.707, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 5.84, | |
| "learning_rate": 0.00016605286902360357, | |
| "loss": 1.8584, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 0.00016578225745814907, | |
| "loss": 1.7969, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 5.89, | |
| "learning_rate": 0.000165510794010458, | |
| "loss": 1.7988, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 5.91, | |
| "learning_rate": 0.00016523848219599023, | |
| "loss": 1.793, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 0.00016496532554119214, | |
| "loss": 1.8994, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 0.0001646913275834506, | |
| "loss": 1.7725, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 5.98, | |
| "learning_rate": 0.00016441649187104763, | |
| "loss": 1.8682, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 0.000164140821963114, | |
| "loss": 1.7598, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 6.02, | |
| "learning_rate": 0.00016386432142958342, | |
| "loss": 1.4678, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 6.04, | |
| "learning_rate": 0.00016358699385114625, | |
| "loss": 1.4043, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 0.000163308842819203, | |
| "loss": 1.5, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 0.00016302987193581806, | |
| "loss": 1.5537, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 0.00016275008481367287, | |
| "loss": 1.4404, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 0.00016246948507601914, | |
| "loss": 1.5049, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 6.16, | |
| "learning_rate": 0.00016218807635663202, | |
| "loss": 1.4834, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 6.18, | |
| "learning_rate": 0.00016190586229976304, | |
| "loss": 1.5586, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 6.2, | |
| "learning_rate": 0.00016162284656009274, | |
| "loss": 1.501, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 6.22, | |
| "learning_rate": 0.00016133903280268362, | |
| "loss": 1.4824, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 6.24, | |
| "learning_rate": 0.0001610544247029325, | |
| "loss": 1.5479, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 6.27, | |
| "learning_rate": 0.0001607690259465229, | |
| "loss": 1.4863, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 0.0001604828402293774, | |
| "loss": 1.3975, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 0.00016019587125760978, | |
| "loss": 1.4707, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 6.33, | |
| "learning_rate": 0.00015990812274747692, | |
| "loss": 1.5244, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 6.36, | |
| "learning_rate": 0.00015961959842533083, | |
| "loss": 1.4512, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 6.38, | |
| "learning_rate": 0.00015933030202757022, | |
| "loss": 1.4209, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 0.00015904023730059228, | |
| "loss": 1.4961, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 0.00015874940800074402, | |
| "loss": 1.4639, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 6.44, | |
| "learning_rate": 0.00015845781789427377, | |
| "loss": 1.5713, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 6.47, | |
| "learning_rate": 0.00015816547075728226, | |
| "loss": 1.4365, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 0.00015787237037567385, | |
| "loss": 1.5361, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 0.0001575785205451073, | |
| "loss": 1.4453, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 0.000157283925070947, | |
| "loss": 1.5576, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "learning_rate": 0.0001569885877682132, | |
| "loss": 1.4775, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 6.58, | |
| "learning_rate": 0.000156692512461533, | |
| "loss": 1.542, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 0.00015639570298509064, | |
| "loss": 1.5146, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 0.00015609816318257788, | |
| "loss": 1.582, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 6.64, | |
| "learning_rate": 0.00015579989690714423, | |
| "loss": 1.4658, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 0.000155500908021347, | |
| "loss": 1.6279, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 6.69, | |
| "learning_rate": 0.00015520120039710139, | |
| "loss": 1.5908, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 0.0001549007779156302, | |
| "loss": 1.4941, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "learning_rate": 0.00015459964446741382, | |
| "loss": 1.5684, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 6.76, | |
| "learning_rate": 0.0001542978039521395, | |
| "loss": 1.6045, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 6.78, | |
| "learning_rate": 0.00015399526027865107, | |
| "loss": 1.5527, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 6.8, | |
| "learning_rate": 0.0001536920173648984, | |
| "loss": 1.5781, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 0.00015338807913788636, | |
| "loss": 1.5352, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 6.84, | |
| "learning_rate": 0.0001530834495336243, | |
| "loss": 1.6094, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 0.00015277813249707487, | |
| "loss": 1.6562, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 0.000152472131982103, | |
| "loss": 1.6064, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 6.91, | |
| "learning_rate": 0.0001521654519514246, | |
| "loss": 1.5303, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 6.93, | |
| "learning_rate": 0.0001518580963765555, | |
| "loss": 1.5273, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 6.96, | |
| "learning_rate": 0.00015155006923775965, | |
| "loss": 1.6074, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 0.00015124137452399795, | |
| "loss": 1.6553, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 0.00015093201623287631, | |
| "loss": 1.5859, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 7.02, | |
| "learning_rate": 0.00015062199837059405, | |
| "loss": 1.3838, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 0.00015031132495189187, | |
| "loss": 1.2764, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 7.07, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 1.2939, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 0.00014968802754658614, | |
| "loss": 1.4043, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 0.0001493754116317029, | |
| "loss": 1.3164, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 7.13, | |
| "learning_rate": 0.00014906215630373606, | |
| "loss": 1.3623, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 7.16, | |
| "learning_rate": 0.00014874826561935155, | |
| "loss": 1.2451, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 7.18, | |
| "learning_rate": 0.00014843374364344333, | |
| "loss": 1.3223, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 0.00014811859444908052, | |
| "loss": 1.3125, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 7.22, | |
| "learning_rate": 0.0001478028221174548, | |
| "loss": 1.3672, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 7.24, | |
| "learning_rate": 0.00014748643073782752, | |
| "loss": 1.2632, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 7.27, | |
| "learning_rate": 0.00014716942440747664, | |
| "loss": 1.2617, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 7.29, | |
| "learning_rate": 0.00014685180723164376, | |
| "loss": 1.3564, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 0.000146533583323481, | |
| "loss": 1.3525, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 7.33, | |
| "learning_rate": 0.0001462147568039977, | |
| "loss": 1.3721, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 0.00014589533180200693, | |
| "loss": 1.2017, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 7.38, | |
| "learning_rate": 0.00014557531245407225, | |
| "loss": 1.3311, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 0.00014525470290445392, | |
| "loss": 1.3828, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 7.42, | |
| "learning_rate": 0.00014493350730505533, | |
| "loss": 1.2324, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 7.44, | |
| "learning_rate": 0.0001446117298153693, | |
| "loss": 1.3467, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 7.47, | |
| "learning_rate": 0.00014428937460242417, | |
| "loss": 1.3291, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 7.49, | |
| "learning_rate": 0.00014396644584072972, | |
| "loss": 1.29, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "learning_rate": 0.00014364294771222324, | |
| "loss": 1.2803, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 0.00014331888440621533, | |
| "loss": 1.4258, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 7.56, | |
| "learning_rate": 0.00014299426011933568, | |
| "loss": 1.2871, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 7.58, | |
| "learning_rate": 0.0001426690790554787, | |
| "loss": 1.4482, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 0.00014234334542574906, | |
| "loss": 1.3438, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 0.00014201706344840712, | |
| "loss": 1.3965, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 7.64, | |
| "learning_rate": 0.00014169023734881452, | |
| "loss": 1.3115, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 0.00014136287135937915, | |
| "loss": 1.3428, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 7.69, | |
| "learning_rate": 0.00014103496971950053, | |
| "loss": 1.291, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 7.71, | |
| "learning_rate": 0.0001407065366755149, | |
| "loss": 1.4287, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 7.73, | |
| "learning_rate": 0.00014037757648064018, | |
| "loss": 1.3296, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 7.76, | |
| "learning_rate": 0.00014004809339492088, | |
| "loss": 1.3203, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 0.00013971809168517298, | |
| "loss": 1.3623, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 0.00013938757562492873, | |
| "loss": 1.3008, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "learning_rate": 0.0001390565494943811, | |
| "loss": 1.4365, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 0.00013872501758032863, | |
| "loss": 1.2852, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 0.00013839298417611963, | |
| "loss": 1.4785, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 0.00013806045358159683, | |
| "loss": 1.2998, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 7.91, | |
| "learning_rate": 0.00013772743010304154, | |
| "loss": 1.373, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 7.93, | |
| "learning_rate": 0.00013739391805311793, | |
| "loss": 1.416, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 0.00013705992175081728, | |
| "loss": 1.5264, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 7.98, | |
| "learning_rate": 0.00013672544552140176, | |
| "loss": 1.4072, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 0.00013639049369634876, | |
| "loss": 1.2705, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 0.00013605507061329464, | |
| "loss": 1.1445, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "learning_rate": 0.0001357191806159785, | |
| "loss": 1.1514, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 8.07, | |
| "learning_rate": 0.0001353828280541861, | |
| "loss": 1.0825, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 8.09, | |
| "learning_rate": 0.00013504601728369327, | |
| "loss": 1.1479, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 8.11, | |
| "learning_rate": 0.00013470875266620978, | |
| "loss": 1.0366, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "learning_rate": 0.00013437103856932264, | |
| "loss": 1.1152, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 0.00013403287936643977, | |
| "loss": 1.1719, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 8.18, | |
| "learning_rate": 0.00013369427943673312, | |
| "loss": 1.1543, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 0.00013335524316508208, | |
| "loss": 1.0898, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 8.22, | |
| "learning_rate": 0.00013301577494201664, | |
| "loss": 1.2358, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 8.24, | |
| "learning_rate": 0.0001326758791636607, | |
| "loss": 1.0977, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 8.27, | |
| "learning_rate": 0.00013233556023167485, | |
| "loss": 1.1689, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 8.29, | |
| "learning_rate": 0.0001319948225531997, | |
| "loss": 1.1787, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 8.31, | |
| "learning_rate": 0.0001316536705407985, | |
| "loss": 1.2651, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 8.33, | |
| "learning_rate": 0.00013131210861240026, | |
| "loss": 1.1973, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 0.0001309701411912423, | |
| "loss": 1.1792, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 0.00013062777270581312, | |
| "loss": 1.208, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 0.00013028500758979506, | |
| "loss": 1.1719, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 8.42, | |
| "learning_rate": 0.00012994185028200684, | |
| "loss": 1.1865, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 8.44, | |
| "learning_rate": 0.00012959830522634596, | |
| "loss": 1.1963, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 8.47, | |
| "learning_rate": 0.00012925437687173142, | |
| "loss": 1.1533, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 8.49, | |
| "learning_rate": 0.00012891006967204584, | |
| "loss": 1.1924, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 8.51, | |
| "learning_rate": 0.00012856538808607795, | |
| "loss": 1.208, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 8.53, | |
| "learning_rate": 0.00012822033657746478, | |
| "loss": 1.1924, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 8.56, | |
| "learning_rate": 0.0001278749196146339, | |
| "loss": 1.1641, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 8.58, | |
| "learning_rate": 0.0001275291416707454, | |
| "loss": 1.228, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 0.0001271830072236343, | |
| "loss": 1.207, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 0.00012683652075575218, | |
| "loss": 1.1543, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 8.64, | |
| "learning_rate": 0.00012648968675410928, | |
| "loss": 1.23, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 8.67, | |
| "learning_rate": 0.00012614250971021657, | |
| "loss": 1.2354, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 8.69, | |
| "learning_rate": 0.0001257949941200273, | |
| "loss": 1.207, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 8.71, | |
| "learning_rate": 0.00012544714448387893, | |
| "loss": 1.2148, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 0.00012509896530643488, | |
| "loss": 1.2241, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 8.76, | |
| "learning_rate": 0.00012475046109662605, | |
| "loss": 1.1982, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 8.78, | |
| "learning_rate": 0.0001244016363675926, | |
| "loss": 1.1904, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 8.8, | |
| "learning_rate": 0.00012405249563662537, | |
| "loss": 1.2168, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 8.82, | |
| "learning_rate": 0.00012370304342510747, | |
| "loss": 1.1934, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 8.84, | |
| "learning_rate": 0.00012335328425845565, | |
| "loss": 1.2246, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 8.87, | |
| "learning_rate": 0.00012300322266606178, | |
| "loss": 1.2129, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 0.00012265286318123415, | |
| "loss": 1.3145, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 8.91, | |
| "learning_rate": 0.00012230221034113874, | |
| "loss": 1.2236, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 8.93, | |
| "learning_rate": 0.00012195126868674051, | |
| "loss": 1.1782, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 0.00012160004276274453, | |
| "loss": 1.1338, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 8.98, | |
| "learning_rate": 0.00012124853711753727, | |
| "loss": 1.2559, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "learning_rate": 0.00012089675630312754, | |
| "loss": 1.1895, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 0.00012054470487508751, | |
| "loss": 0.9648, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 0.00012019238739249397, | |
| "loss": 1.082, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 9.07, | |
| "learning_rate": 0.000119839808417869, | |
| "loss": 1.0996, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 0.00011948697251712109, | |
| "loss": 1.1094, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 9.11, | |
| "learning_rate": 0.00011913388425948584, | |
| "loss": 1.0088, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 9.13, | |
| "learning_rate": 0.00011878054821746703, | |
| "loss": 1.0249, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 9.16, | |
| "learning_rate": 0.00011842696896677708, | |
| "loss": 0.9722, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 0.00011807315108627806, | |
| "loss": 0.9707, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 9.2, | |
| "learning_rate": 0.0001177190991579223, | |
| "loss": 1.0815, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 9.22, | |
| "learning_rate": 0.00011736481776669306, | |
| "loss": 1.0107, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 9.24, | |
| "learning_rate": 0.0001170103115005451, | |
| "loss": 1.0142, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 9.27, | |
| "learning_rate": 0.00011665558495034546, | |
| "loss": 1.0698, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 9.29, | |
| "learning_rate": 0.00011630064270981367, | |
| "loss": 0.9937, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 0.00011594548937546257, | |
| "loss": 0.9829, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 9.33, | |
| "learning_rate": 0.00011559012954653865, | |
| "loss": 1.0083, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 9.36, | |
| "learning_rate": 0.00011523456782496253, | |
| "loss": 1.0469, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 9.38, | |
| "learning_rate": 0.00011487880881526934, | |
| "loss": 1.1035, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 0.00011452285712454904, | |
| "loss": 1.0723, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 9.42, | |
| "learning_rate": 0.00011416671736238689, | |
| "loss": 1.0435, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 9.44, | |
| "learning_rate": 0.00011381039414080365, | |
| "loss": 1.1118, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 9.47, | |
| "learning_rate": 0.00011345389207419588, | |
| "loss": 1.0879, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 9.49, | |
| "learning_rate": 0.00011309721577927619, | |
| "loss": 1.0337, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 0.00011274036987501348, | |
| "loss": 0.9663, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 9.53, | |
| "learning_rate": 0.00011238335898257304, | |
| "loss": 0.979, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 9.56, | |
| "learning_rate": 0.0001120261877252568, | |
| "loss": 1.0454, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 9.58, | |
| "learning_rate": 0.00011166886072844342, | |
| "loss": 1.0103, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 9.6, | |
| "learning_rate": 0.00011131138261952845, | |
| "loss": 1.1553, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 9.62, | |
| "learning_rate": 0.00011095375802786419, | |
| "loss": 1.0342, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 9.64, | |
| "learning_rate": 0.00011059599158470002, | |
| "loss": 1.0591, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 9.67, | |
| "learning_rate": 0.00011023808792312227, | |
| "loss": 1.0347, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 0.00010988005167799427, | |
| "loss": 1.0762, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 9.71, | |
| "learning_rate": 0.00010952188748589625, | |
| "loss": 1.1216, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 9.73, | |
| "learning_rate": 0.0001091635999850655, | |
| "loss": 1.1182, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 9.76, | |
| "learning_rate": 0.00010880519381533592, | |
| "loss": 1.0396, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 9.78, | |
| "learning_rate": 0.00010844667361807842, | |
| "loss": 0.9829, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 9.8, | |
| "learning_rate": 0.00010808804403614043, | |
| "loss": 1.0903, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 9.82, | |
| "learning_rate": 0.00010772930971378596, | |
| "loss": 1.0249, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 9.84, | |
| "learning_rate": 0.00010737047529663545, | |
| "loss": 1.0493, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 0.00010701154543160541, | |
| "loss": 1.1426, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 0.00010665252476684864, | |
| "loss": 1.0591, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 0.0001062934179516936, | |
| "loss": 1.0083, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 9.93, | |
| "learning_rate": 0.00010593422963658452, | |
| "loss": 1.0356, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 9.96, | |
| "learning_rate": 0.00010557496447302102, | |
| "loss": 1.1387, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 0.00010521562711349788, | |
| "loss": 1.0908, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 0.00010485622221144484, | |
| "loss": 1.1177, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 10.02, | |
| "learning_rate": 0.00010449675442116634, | |
| "loss": 0.9346, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 10.04, | |
| "learning_rate": 0.00010413722839778117, | |
| "loss": 0.96, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 10.07, | |
| "learning_rate": 0.00010377764879716234, | |
| "loss": 0.9424, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 0.00010341802027587659, | |
| "loss": 0.9385, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 0.00010305834749112421, | |
| "loss": 0.8706, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 10.13, | |
| "learning_rate": 0.00010269863510067872, | |
| "loss": 0.8843, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "learning_rate": 0.00010233888776282649, | |
| "loss": 0.8813, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 10.18, | |
| "learning_rate": 0.00010197911013630659, | |
| "loss": 0.9258, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 10.2, | |
| "learning_rate": 0.00010161930688025017, | |
| "loss": 0.9229, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 10.22, | |
| "learning_rate": 0.00010125948265412033, | |
| "loss": 0.8975, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 10.24, | |
| "learning_rate": 0.0001008996421176518, | |
| "loss": 0.8848, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 10.27, | |
| "learning_rate": 0.00010053978993079045, | |
| "loss": 0.9873, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 10.29, | |
| "learning_rate": 0.00010017993075363305, | |
| "loss": 0.9463, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 9.982006924636697e-05, | |
| "loss": 0.9058, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 9.946021006920959e-05, | |
| "loss": 1.0679, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 10.36, | |
| "learning_rate": 9.910035788234822e-05, | |
| "loss": 0.9067, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 10.38, | |
| "learning_rate": 9.874051734587968e-05, | |
| "loss": 0.9453, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 9.838069311974986e-05, | |
| "loss": 0.9468, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 10.42, | |
| "learning_rate": 9.802088986369342e-05, | |
| "loss": 0.9111, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 10.44, | |
| "learning_rate": 9.766111223717352e-05, | |
| "loss": 1.0132, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 10.47, | |
| "learning_rate": 9.730136489932133e-05, | |
| "loss": 0.9922, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 9.694165250887584e-05, | |
| "loss": 0.9736, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 10.51, | |
| "learning_rate": 9.658197972412345e-05, | |
| "loss": 0.9551, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 10.53, | |
| "learning_rate": 9.622235120283769e-05, | |
| "loss": 0.9575, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 10.56, | |
| "learning_rate": 9.586277160221884e-05, | |
| "loss": 0.9448, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 10.58, | |
| "learning_rate": 9.550324557883373e-05, | |
| "loss": 0.9775, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 10.6, | |
| "learning_rate": 9.514377778855521e-05, | |
| "loss": 1.0044, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 10.62, | |
| "learning_rate": 9.478437288650213e-05, | |
| "loss": 0.9067, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 10.64, | |
| "learning_rate": 9.442503552697899e-05, | |
| "loss": 0.9453, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "learning_rate": 9.406577036341548e-05, | |
| "loss": 1.0103, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 10.69, | |
| "learning_rate": 9.37065820483064e-05, | |
| "loss": 1.0801, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 10.71, | |
| "learning_rate": 9.334747523315137e-05, | |
| "loss": 0.957, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 10.73, | |
| "learning_rate": 9.298845456839459e-05, | |
| "loss": 0.9092, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 10.76, | |
| "learning_rate": 9.262952470336458e-05, | |
| "loss": 0.9854, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 10.78, | |
| "learning_rate": 9.227069028621406e-05, | |
| "loss": 0.9238, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 9.19119559638596e-05, | |
| "loss": 0.9814, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 10.82, | |
| "learning_rate": 9.15533263819216e-05, | |
| "loss": 1.0024, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 10.84, | |
| "learning_rate": 9.119480618466409e-05, | |
| "loss": 0.9043, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 10.87, | |
| "learning_rate": 9.083640001493454e-05, | |
| "loss": 0.98, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 9.047811251410376e-05, | |
| "loss": 1.0547, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 9.011994832200577e-05, | |
| "loss": 0.9736, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 10.93, | |
| "learning_rate": 8.976191207687775e-05, | |
| "loss": 1.0371, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 8.94040084153e-05, | |
| "loss": 0.9771, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "learning_rate": 8.904624197213585e-05, | |
| "loss": 0.9429, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 8.868861738047158e-05, | |
| "loss": 0.9692, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 11.02, | |
| "learning_rate": 8.83311392715566e-05, | |
| "loss": 0.8188, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 11.04, | |
| "learning_rate": 8.797381227474324e-05, | |
| "loss": 0.8862, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 11.07, | |
| "learning_rate": 8.7616641017427e-05, | |
| "loss": 0.8545, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 11.09, | |
| "learning_rate": 8.725963012498657e-05, | |
| "loss": 0.8179, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 8.690278422072384e-05, | |
| "loss": 0.9062, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 11.13, | |
| "learning_rate": 8.654610792580415e-05, | |
| "loss": 0.811, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 11.16, | |
| "learning_rate": 8.61896058591964e-05, | |
| "loss": 0.7476, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 11.18, | |
| "learning_rate": 8.583328263761316e-05, | |
| "loss": 0.835, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 11.2, | |
| "learning_rate": 8.5477142875451e-05, | |
| "loss": 0.8276, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 11.22, | |
| "learning_rate": 8.512119118473067e-05, | |
| "loss": 0.8247, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 11.24, | |
| "learning_rate": 8.476543217503748e-05, | |
| "loss": 0.8823, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 11.27, | |
| "learning_rate": 8.440987045346134e-05, | |
| "loss": 0.8638, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 11.29, | |
| "learning_rate": 8.405451062453744e-05, | |
| "loss": 0.8462, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 11.31, | |
| "learning_rate": 8.369935729018634e-05, | |
| "loss": 0.8252, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 11.33, | |
| "learning_rate": 8.334441504965455e-05, | |
| "loss": 0.7397, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 11.36, | |
| "learning_rate": 8.29896884994549e-05, | |
| "loss": 0.8184, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 11.38, | |
| "learning_rate": 8.263518223330697e-05, | |
| "loss": 0.9321, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 8.228090084207774e-05, | |
| "loss": 0.8481, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 11.42, | |
| "learning_rate": 8.192684891372198e-05, | |
| "loss": 0.8745, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 11.44, | |
| "learning_rate": 8.157303103322296e-05, | |
| "loss": 0.8618, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 11.47, | |
| "learning_rate": 8.1219451782533e-05, | |
| "loss": 0.9106, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 11.49, | |
| "learning_rate": 8.086611574051417e-05, | |
| "loss": 0.8784, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 11.51, | |
| "learning_rate": 8.051302748287895e-05, | |
| "loss": 0.8667, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 11.53, | |
| "learning_rate": 8.016019158213101e-05, | |
| "loss": 0.8765, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 11.56, | |
| "learning_rate": 7.980761260750607e-05, | |
| "loss": 0.9463, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 11.58, | |
| "learning_rate": 7.945529512491251e-05, | |
| "loss": 0.9443, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 7.91032436968725e-05, | |
| "loss": 0.9102, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 11.62, | |
| "learning_rate": 7.875146288246275e-05, | |
| "loss": 0.9639, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 11.64, | |
| "learning_rate": 7.839995723725548e-05, | |
| "loss": 0.8325, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 11.67, | |
| "learning_rate": 7.804873131325954e-05, | |
| "loss": 0.9619, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 11.69, | |
| "learning_rate": 7.76977896588613e-05, | |
| "loss": 0.8569, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 11.71, | |
| "learning_rate": 7.734713681876589e-05, | |
| "loss": 0.8643, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 11.73, | |
| "learning_rate": 7.699677733393826e-05, | |
| "loss": 0.8789, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 11.76, | |
| "learning_rate": 7.66467157415444e-05, | |
| "loss": 0.8901, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 11.78, | |
| "learning_rate": 7.629695657489257e-05, | |
| "loss": 0.9199, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 11.8, | |
| "learning_rate": 7.594750436337467e-05, | |
| "loss": 0.9204, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 11.82, | |
| "learning_rate": 7.55983636324074e-05, | |
| "loss": 0.9624, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 11.84, | |
| "learning_rate": 7.524953890337395e-05, | |
| "loss": 0.9517, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 11.87, | |
| "learning_rate": 7.490103469356513e-05, | |
| "loss": 0.8242, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 7.455285551612105e-05, | |
| "loss": 0.979, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 11.91, | |
| "learning_rate": 7.42050058799727e-05, | |
| "loss": 0.9175, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 11.93, | |
| "learning_rate": 7.385749028978346e-05, | |
| "loss": 0.936, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 11.96, | |
| "learning_rate": 7.351031324589074e-05, | |
| "loss": 0.833, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 11.98, | |
| "learning_rate": 7.316347924424787e-05, | |
| "loss": 0.9663, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 7.281699277636572e-05, | |
| "loss": 0.8643, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 12.02, | |
| "learning_rate": 7.24708583292546e-05, | |
| "loss": 0.7593, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 12.04, | |
| "learning_rate": 7.212508038536613e-05, | |
| "loss": 0.7949, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 12.07, | |
| "learning_rate": 7.177966342253524e-05, | |
| "loss": 0.7622, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 12.09, | |
| "learning_rate": 7.143461191392206e-05, | |
| "loss": 0.7656, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 12.11, | |
| "learning_rate": 7.108993032795418e-05, | |
| "loss": 0.749, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 12.13, | |
| "learning_rate": 7.07456231282686e-05, | |
| "loss": 0.7344, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 12.16, | |
| "learning_rate": 7.040169477365403e-05, | |
| "loss": 0.7153, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 12.18, | |
| "learning_rate": 7.005814971799318e-05, | |
| "loss": 0.7266, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 12.2, | |
| "learning_rate": 6.971499241020495e-05, | |
| "loss": 0.7949, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 12.22, | |
| "learning_rate": 6.93722272941869e-05, | |
| "loss": 0.7397, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 12.24, | |
| "learning_rate": 6.902985880875773e-05, | |
| "loss": 0.7422, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 12.27, | |
| "learning_rate": 6.868789138759976e-05, | |
| "loss": 0.7798, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 12.29, | |
| "learning_rate": 6.83463294592015e-05, | |
| "loss": 0.8408, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 12.31, | |
| "learning_rate": 6.800517744680032e-05, | |
| "loss": 0.8447, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 12.33, | |
| "learning_rate": 6.766443976832517e-05, | |
| "loss": 0.8174, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 12.36, | |
| "learning_rate": 6.732412083633936e-05, | |
| "loss": 0.7407, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 12.38, | |
| "learning_rate": 6.698422505798338e-05, | |
| "loss": 0.7998, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 6.664475683491796e-05, | |
| "loss": 0.8213, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 12.42, | |
| "learning_rate": 6.630572056326687e-05, | |
| "loss": 0.7622, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 12.44, | |
| "learning_rate": 6.59671206335602e-05, | |
| "loss": 0.8169, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 12.47, | |
| "learning_rate": 6.562896143067734e-05, | |
| "loss": 0.7935, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 12.49, | |
| "learning_rate": 6.529124733379024e-05, | |
| "loss": 0.7744, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 12.51, | |
| "learning_rate": 6.495398271630675e-05, | |
| "loss": 0.835, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 12.53, | |
| "learning_rate": 6.461717194581393e-05, | |
| "loss": 0.8018, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 12.56, | |
| "learning_rate": 6.428081938402149e-05, | |
| "loss": 0.7261, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 12.58, | |
| "learning_rate": 6.394492938670538e-05, | |
| "loss": 0.8813, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 12.6, | |
| "learning_rate": 6.360950630365126e-05, | |
| "loss": 0.8125, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 6.327455447859827e-05, | |
| "loss": 0.8042, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 12.64, | |
| "learning_rate": 6.294007824918276e-05, | |
| "loss": 0.7339, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 12.67, | |
| "learning_rate": 6.260608194688206e-05, | |
| "loss": 0.8389, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 12.69, | |
| "learning_rate": 6.227256989695848e-05, | |
| "loss": 0.8213, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 12.71, | |
| "learning_rate": 6.193954641840318e-05, | |
| "loss": 0.7446, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 12.73, | |
| "learning_rate": 6.160701582388038e-05, | |
| "loss": 0.8755, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 12.76, | |
| "learning_rate": 6.12749824196714e-05, | |
| "loss": 0.7798, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 12.78, | |
| "learning_rate": 6.0943450505618917e-05, | |
| "loss": 0.8774, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 12.8, | |
| "learning_rate": 6.061242437507131e-05, | |
| "loss": 0.7964, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 12.82, | |
| "learning_rate": 6.028190831482703e-05, | |
| "loss": 0.793, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 12.84, | |
| "learning_rate": 5.995190660507915e-05, | |
| "loss": 0.8257, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 12.87, | |
| "learning_rate": 5.962242351935985e-05, | |
| "loss": 0.8042, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 5.929346332448511e-05, | |
| "loss": 0.832, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 5.89650302804995e-05, | |
| "loss": 0.8218, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 12.93, | |
| "learning_rate": 5.863712864062089e-05, | |
| "loss": 0.8047, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 12.96, | |
| "learning_rate": 5.8309762651185484e-05, | |
| "loss": 0.8052, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 12.98, | |
| "learning_rate": 5.7982936551592906e-05, | |
| "loss": 0.791, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "learning_rate": 5.765665457425102e-05, | |
| "loss": 0.8032, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 13.02, | |
| "learning_rate": 5.733092094452135e-05, | |
| "loss": 0.7715, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 13.04, | |
| "learning_rate": 5.700573988066433e-05, | |
| "loss": 0.7065, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 13.07, | |
| "learning_rate": 5.668111559378471e-05, | |
| "loss": 0.7202, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 13.09, | |
| "learning_rate": 5.6357052287776765e-05, | |
| "loss": 0.7153, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 5.6033554159270294e-05, | |
| "loss": 0.7456, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 13.13, | |
| "learning_rate": 5.571062539757581e-05, | |
| "loss": 0.7344, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 13.16, | |
| "learning_rate": 5.538827018463069e-05, | |
| "loss": 0.6982, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 13.18, | |
| "learning_rate": 5.50664926949447e-05, | |
| "loss": 0.7637, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 13.2, | |
| "learning_rate": 5.474529709554612e-05, | |
| "loss": 0.7227, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 13.22, | |
| "learning_rate": 5.4424687545927776e-05, | |
| "loss": 0.6914, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 13.24, | |
| "learning_rate": 5.410466819799306e-05, | |
| "loss": 0.7603, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 13.27, | |
| "learning_rate": 5.378524319600231e-05, | |
| "loss": 0.707, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 13.29, | |
| "learning_rate": 5.346641667651897e-05, | |
| "loss": 0.7505, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 13.31, | |
| "learning_rate": 5.314819276835625e-05, | |
| "loss": 0.7998, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 13.33, | |
| "learning_rate": 5.283057559252341e-05, | |
| "loss": 0.7847, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 13.36, | |
| "learning_rate": 5.25135692621725e-05, | |
| "loss": 0.7686, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 13.38, | |
| "learning_rate": 5.219717788254521e-05, | |
| "loss": 0.687, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 13.4, | |
| "learning_rate": 5.1881405550919493e-05, | |
| "loss": 0.7051, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 13.42, | |
| "learning_rate": 5.156625635655672e-05, | |
| "loss": 0.7412, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 13.44, | |
| "learning_rate": 5.12517343806485e-05, | |
| "loss": 0.729, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 13.47, | |
| "learning_rate": 5.0937843696263966e-05, | |
| "loss": 0.7505, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 13.49, | |
| "learning_rate": 5.062458836829711e-05, | |
| "loss": 0.7002, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 13.51, | |
| "learning_rate": 5.03119724534139e-05, | |
| "loss": 0.7573, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 13.53, | |
| "learning_rate": 5.000000000000002e-05, | |
| "loss": 0.7241, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 13.56, | |
| "learning_rate": 4.96886750481082e-05, | |
| "loss": 0.6753, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 13.58, | |
| "learning_rate": 4.9378001629406e-05, | |
| "loss": 0.7456, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 13.6, | |
| "learning_rate": 4.9067983767123736e-05, | |
| "loss": 0.7471, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 13.62, | |
| "learning_rate": 4.875862547600207e-05, | |
| "loss": 0.6768, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 13.64, | |
| "learning_rate": 4.8449930762240355e-05, | |
| "loss": 0.7339, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 13.67, | |
| "learning_rate": 4.814190362344454e-05, | |
| "loss": 0.7017, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 13.69, | |
| "learning_rate": 4.783454804857539e-05, | |
| "loss": 0.8296, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "learning_rate": 4.752786801789703e-05, | |
| "loss": 0.7427, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 13.73, | |
| "learning_rate": 4.722186750292511e-05, | |
| "loss": 0.7837, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 13.76, | |
| "learning_rate": 4.6916550466375684e-05, | |
| "loss": 0.7305, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 13.78, | |
| "learning_rate": 4.661192086211366e-05, | |
| "loss": 0.7734, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 13.8, | |
| "learning_rate": 4.630798263510162e-05, | |
| "loss": 0.7188, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 13.82, | |
| "learning_rate": 4.600473972134894e-05, | |
| "loss": 0.7505, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 13.84, | |
| "learning_rate": 4.570219604786051e-05, | |
| "loss": 0.6426, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 13.87, | |
| "learning_rate": 4.540035553258619e-05, | |
| "loss": 0.7739, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 13.89, | |
| "learning_rate": 4.5099222084369805e-05, | |
| "loss": 0.7842, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 13.91, | |
| "learning_rate": 4.479879960289863e-05, | |
| "loss": 0.7319, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 13.93, | |
| "learning_rate": 4.449909197865303e-05, | |
| "loss": 0.7788, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 4.420010309285577e-05, | |
| "loss": 0.7427, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 13.98, | |
| "learning_rate": 4.3901836817422124e-05, | |
| "loss": 0.7358, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "learning_rate": 4.360429701490934e-05, | |
| "loss": 0.7964, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 14.02, | |
| "learning_rate": 4.3307487538467006e-05, | |
| "loss": 0.7236, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 14.04, | |
| "learning_rate": 4.301141223178684e-05, | |
| "loss": 0.689, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 14.07, | |
| "learning_rate": 4.271607492905303e-05, | |
| "loss": 0.6389, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 14.09, | |
| "learning_rate": 4.242147945489272e-05, | |
| "loss": 0.6812, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 14.11, | |
| "learning_rate": 4.212762962432619e-05, | |
| "loss": 0.6782, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 14.13, | |
| "learning_rate": 4.183452924271776e-05, | |
| "loss": 0.6899, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 14.16, | |
| "learning_rate": 4.154218210572627e-05, | |
| "loss": 0.7407, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 14.18, | |
| "learning_rate": 4.125059199925599e-05, | |
| "loss": 0.6567, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 4.0959762699407766e-05, | |
| "loss": 0.6587, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 14.22, | |
| "learning_rate": 4.06696979724298e-05, | |
| "loss": 0.7441, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 14.24, | |
| "learning_rate": 4.038040157466918e-05, | |
| "loss": 0.707, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 14.27, | |
| "learning_rate": 4.009187725252309e-05, | |
| "loss": 0.7041, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 14.29, | |
| "learning_rate": 3.980412874239021e-05, | |
| "loss": 0.6216, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 14.31, | |
| "learning_rate": 3.95171597706226e-05, | |
| "loss": 0.6748, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 14.33, | |
| "learning_rate": 3.9230974053477086e-05, | |
| "loss": 0.6362, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 14.36, | |
| "learning_rate": 3.8945575297067506e-05, | |
| "loss": 0.6045, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 14.38, | |
| "learning_rate": 3.866096719731639e-05, | |
| "loss": 0.7173, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 14.4, | |
| "learning_rate": 3.8377153439907266e-05, | |
| "loss": 0.6299, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 14.42, | |
| "learning_rate": 3.809413770023701e-05, | |
| "loss": 0.6499, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 14.44, | |
| "learning_rate": 3.7811923643367974e-05, | |
| "loss": 0.6914, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 14.47, | |
| "learning_rate": 3.7530514923980884e-05, | |
| "loss": 0.7236, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 14.49, | |
| "learning_rate": 3.724991518632717e-05, | |
| "loss": 0.7598, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 14.51, | |
| "learning_rate": 3.697012806418194e-05, | |
| "loss": 0.6279, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 14.53, | |
| "learning_rate": 3.669115718079702e-05, | |
| "loss": 0.7622, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 14.56, | |
| "learning_rate": 3.641300614885378e-05, | |
| "loss": 0.6772, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 14.58, | |
| "learning_rate": 3.61356785704166e-05, | |
| "loss": 0.7568, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 3.585917803688603e-05, | |
| "loss": 0.7065, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 14.62, | |
| "learning_rate": 3.558350812895238e-05, | |
| "loss": 0.6743, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 14.64, | |
| "learning_rate": 3.530867241654942e-05, | |
| "loss": 0.6709, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 14.67, | |
| "learning_rate": 3.503467445880789e-05, | |
| "loss": 0.71, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 14.69, | |
| "learning_rate": 3.476151780400979e-05, | |
| "loss": 0.6836, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 14.71, | |
| "learning_rate": 3.448920598954203e-05, | |
| "loss": 0.6973, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 14.73, | |
| "learning_rate": 3.421774254185096e-05, | |
| "loss": 0.7256, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 14.76, | |
| "learning_rate": 3.394713097639647e-05, | |
| "loss": 0.7334, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 14.78, | |
| "learning_rate": 3.367737479760652e-05, | |
| "loss": 0.6675, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 14.8, | |
| "learning_rate": 3.340847749883191e-05, | |
| "loss": 0.7363, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 14.82, | |
| "learning_rate": 3.31404425623008e-05, | |
| "loss": 0.7139, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 14.84, | |
| "learning_rate": 3.287327345907381e-05, | |
| "loss": 0.6792, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 14.87, | |
| "learning_rate": 3.2606973648998915e-05, | |
| "loss": 0.6704, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 14.89, | |
| "learning_rate": 3.2341546580666796e-05, | |
| "loss": 0.7329, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 14.91, | |
| "learning_rate": 3.207699569136608e-05, | |
| "loss": 0.709, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 14.93, | |
| "learning_rate": 3.1813324407038825e-05, | |
| "loss": 0.6294, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 14.96, | |
| "learning_rate": 3.1550536142236145e-05, | |
| "loss": 0.7065, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "learning_rate": 3.128863430007414e-05, | |
| "loss": 0.7358, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "learning_rate": 3.102762227218957e-05, | |
| "loss": 0.7622, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 15.02, | |
| "learning_rate": 3.0767503438696213e-05, | |
| "loss": 0.6484, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 15.04, | |
| "learning_rate": 3.0508281168140806e-05, | |
| "loss": 0.6318, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 15.07, | |
| "learning_rate": 3.0249958817459722e-05, | |
| "loss": 0.6851, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 15.09, | |
| "learning_rate": 2.999253973193522e-05, | |
| "loss": 0.6123, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 15.11, | |
| "learning_rate": 2.9736027245152275e-05, | |
| "loss": 0.603, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 15.13, | |
| "learning_rate": 2.9480424678955443e-05, | |
| "loss": 0.6152, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 15.16, | |
| "learning_rate": 2.9225735343405693e-05, | |
| "loss": 0.6743, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 15.18, | |
| "learning_rate": 2.897196253673773e-05, | |
| "loss": 0.6562, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 15.2, | |
| "learning_rate": 2.8719109545317103e-05, | |
| "loss": 0.6621, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 15.22, | |
| "learning_rate": 2.8467179643597697e-05, | |
| "loss": 0.6233, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 15.24, | |
| "learning_rate": 2.8216176094079482e-05, | |
| "loss": 0.6812, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 15.27, | |
| "learning_rate": 2.7966102147265994e-05, | |
| "loss": 0.687, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 15.29, | |
| "learning_rate": 2.7716961041622534e-05, | |
| "loss": 0.6611, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 15.31, | |
| "learning_rate": 2.746875600353398e-05, | |
| "loss": 0.6675, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 15.33, | |
| "learning_rate": 2.722149024726307e-05, | |
| "loss": 0.6821, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 15.36, | |
| "learning_rate": 2.697516697490896e-05, | |
| "loss": 0.6733, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 15.38, | |
| "learning_rate": 2.6729789376365456e-05, | |
| "loss": 0.6606, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 15.4, | |
| "learning_rate": 2.6485360629279987e-05, | |
| "loss": 0.6973, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 15.42, | |
| "learning_rate": 2.624188389901221e-05, | |
| "loss": 0.6748, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 15.44, | |
| "learning_rate": 2.599936233859326e-05, | |
| "loss": 0.7324, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 15.47, | |
| "learning_rate": 2.5757799088684654e-05, | |
| "loss": 0.6538, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 15.49, | |
| "learning_rate": 2.5517197277537886e-05, | |
| "loss": 0.6245, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 15.51, | |
| "learning_rate": 2.527756002095373e-05, | |
| "loss": 0.5972, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 15.53, | |
| "learning_rate": 2.5038890422241958e-05, | |
| "loss": 0.6587, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 15.56, | |
| "learning_rate": 2.480119157218108e-05, | |
| "loss": 0.6523, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 15.58, | |
| "learning_rate": 2.4564466548978525e-05, | |
| "loss": 0.5957, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 15.6, | |
| "learning_rate": 2.432871841823047e-05, | |
| "loss": 0.625, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 15.62, | |
| "learning_rate": 2.4093950232882456e-05, | |
| "loss": 0.6392, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 15.64, | |
| "learning_rate": 2.3860165033189587e-05, | |
| "loss": 0.6982, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 15.67, | |
| "learning_rate": 2.3627365846677306e-05, | |
| "loss": 0.6118, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 15.69, | |
| "learning_rate": 2.339555568810221e-05, | |
| "loss": 0.6895, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 15.71, | |
| "learning_rate": 2.3164737559412854e-05, | |
| "loss": 0.6025, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 15.73, | |
| "learning_rate": 2.2934914449711087e-05, | |
| "loss": 0.6487, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 15.76, | |
| "learning_rate": 2.2706089335213122e-05, | |
| "loss": 0.6846, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 15.78, | |
| "learning_rate": 2.247826517921121e-05, | |
| "loss": 0.6377, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 15.8, | |
| "learning_rate": 2.2251444932035094e-05, | |
| "loss": 0.6797, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 15.82, | |
| "learning_rate": 2.2025631531013824e-05, | |
| "loss": 0.7083, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 15.84, | |
| "learning_rate": 2.1800827900437894e-05, | |
| "loss": 0.7168, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 15.87, | |
| "learning_rate": 2.157703695152109e-05, | |
| "loss": 0.6914, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 15.89, | |
| "learning_rate": 2.135426158236309e-05, | |
| "loss": 0.6123, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 15.91, | |
| "learning_rate": 2.1132504677911658e-05, | |
| "loss": 0.6733, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 15.93, | |
| "learning_rate": 2.091176910992545e-05, | |
| "loss": 0.645, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 15.96, | |
| "learning_rate": 2.069205773693683e-05, | |
| "loss": 0.7017, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 15.98, | |
| "learning_rate": 2.0473373404214723e-05, | |
| "loss": 0.6704, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 2.025571894372794e-05, | |
| "loss": 0.6509, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 16.02, | |
| "learning_rate": 2.003909717410831e-05, | |
| "loss": 0.6914, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 16.04, | |
| "learning_rate": 1.9823510900614417e-05, | |
| "loss": 0.6296, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 16.07, | |
| "learning_rate": 1.9608962915094996e-05, | |
| "loss": 0.6333, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 16.09, | |
| "learning_rate": 1.9395455995953036e-05, | |
| "loss": 0.6514, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 16.11, | |
| "learning_rate": 1.9182992908109644e-05, | |
| "loss": 0.6104, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 16.13, | |
| "learning_rate": 1.897157640296825e-05, | |
| "loss": 0.6318, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 16.16, | |
| "learning_rate": 1.8761209218379016e-05, | |
| "loss": 0.606, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 16.18, | |
| "learning_rate": 1.855189407860344e-05, | |
| "loss": 0.6172, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 16.2, | |
| "learning_rate": 1.8343633694278895e-05, | |
| "loss": 0.6108, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 16.22, | |
| "learning_rate": 1.813643076238375e-05, | |
| "loss": 0.6021, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 16.24, | |
| "learning_rate": 1.7930287966202265e-05, | |
| "loss": 0.5974, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 16.27, | |
| "learning_rate": 1.772520797528988e-05, | |
| "loss": 0.6309, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 16.29, | |
| "learning_rate": 1.752119344543879e-05, | |
| "loss": 0.623, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 16.31, | |
| "learning_rate": 1.731824701864331e-05, | |
| "loss": 0.668, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 16.33, | |
| "learning_rate": 1.7116371323065883e-05, | |
| "loss": 0.5984, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 16.36, | |
| "learning_rate": 1.6915568973002905e-05, | |
| "loss": 0.6904, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 16.38, | |
| "learning_rate": 1.6715842568850893e-05, | |
| "loss": 0.627, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 1.65171946970729e-05, | |
| "loss": 0.5972, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 16.42, | |
| "learning_rate": 1.631962793016487e-05, | |
| "loss": 0.623, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 16.44, | |
| "learning_rate": 1.6123144826622504e-05, | |
| "loss": 0.6499, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 16.47, | |
| "learning_rate": 1.592774793090792e-05, | |
| "loss": 0.6597, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 16.49, | |
| "learning_rate": 1.5733439773416915e-05, | |
| "loss": 0.6045, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 16.51, | |
| "learning_rate": 1.554022287044602e-05, | |
| "loss": 0.6943, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 16.53, | |
| "learning_rate": 1.534809972415998e-05, | |
| "loss": 0.5886, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 16.56, | |
| "learning_rate": 1.5157072822559437e-05, | |
| "loss": 0.6047, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 16.58, | |
| "learning_rate": 1.4967144639448538e-05, | |
| "loss": 0.6401, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 16.6, | |
| "learning_rate": 1.4778317634403083e-05, | |
| "loss": 0.6108, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 16.62, | |
| "learning_rate": 1.4590594252738522e-05, | |
| "loss": 0.6514, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 16.64, | |
| "learning_rate": 1.4403976925478312e-05, | |
| "loss": 0.6289, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 16.67, | |
| "learning_rate": 1.4218468069322578e-05, | |
| "loss": 0.6323, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 16.69, | |
| "learning_rate": 1.4034070086616647e-05, | |
| "loss": 0.6553, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 16.71, | |
| "learning_rate": 1.3850785365319984e-05, | |
| "loss": 0.6082, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 16.73, | |
| "learning_rate": 1.3668616278975343e-05, | |
| "loss": 0.6484, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "learning_rate": 1.3487565186677897e-05, | |
| "loss": 0.6587, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 16.78, | |
| "learning_rate": 1.3307634433044846e-05, | |
| "loss": 0.6204, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 16.8, | |
| "learning_rate": 1.3128826348184887e-05, | |
| "loss": 0.6255, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 16.82, | |
| "learning_rate": 1.2951143247668197e-05, | |
| "loss": 0.5898, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 16.84, | |
| "learning_rate": 1.2774587432496321e-05, | |
| "loss": 0.624, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 16.87, | |
| "learning_rate": 1.2599161189072427e-05, | |
| "loss": 0.6282, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 16.89, | |
| "learning_rate": 1.2424866789171729e-05, | |
| "loss": 0.6191, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 16.91, | |
| "learning_rate": 1.2251706489911984e-05, | |
| "loss": 0.6267, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 16.93, | |
| "learning_rate": 1.2079682533724379e-05, | |
| "loss": 0.6519, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 16.96, | |
| "learning_rate": 1.1908797148324358e-05, | |
| "loss": 0.6577, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 16.98, | |
| "learning_rate": 1.173905254668285e-05, | |
| "loss": 0.6426, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 1.1570450926997655e-05, | |
| "loss": 0.6863, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 17.02, | |
| "learning_rate": 1.140299447266483e-05, | |
| "loss": 0.6089, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 17.04, | |
| "learning_rate": 1.1236685352250597e-05, | |
| "loss": 0.6086, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 17.07, | |
| "learning_rate": 1.1071525719463095e-05, | |
| "loss": 0.6118, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 17.09, | |
| "learning_rate": 1.0907517713124638e-05, | |
| "loss": 0.5825, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 17.11, | |
| "learning_rate": 1.0744663457143878e-05, | |
| "loss": 0.6562, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 17.13, | |
| "learning_rate": 1.0582965060488359e-05, | |
| "loss": 0.5364, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 17.16, | |
| "learning_rate": 1.042242461715729e-05, | |
| "loss": 0.5894, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 17.18, | |
| "learning_rate": 1.026304420615426e-05, | |
| "loss": 0.5815, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 17.2, | |
| "learning_rate": 1.010482589146048e-05, | |
| "loss": 0.606, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 17.22, | |
| "learning_rate": 9.947771722007915e-06, | |
| "loss": 0.5908, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 17.24, | |
| "learning_rate": 9.791883731652828e-06, | |
| "loss": 0.5938, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 17.27, | |
| "learning_rate": 9.637163939149485e-06, | |
| "loss": 0.594, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 17.29, | |
| "learning_rate": 9.48361434812386e-06, | |
| "loss": 0.6211, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 17.31, | |
| "learning_rate": 9.33123694704784e-06, | |
| "loss": 0.6226, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 17.33, | |
| "learning_rate": 9.180033709213454e-06, | |
| "loss": 0.573, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 17.36, | |
| "learning_rate": 9.030006592707174e-06, | |
| "loss": 0.5986, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 17.38, | |
| "learning_rate": 8.881157540384777e-06, | |
| "loss": 0.6294, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 17.4, | |
| "learning_rate": 8.733488479845997e-06, | |
| "loss": 0.627, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 17.42, | |
| "learning_rate": 8.587001323409638e-06, | |
| "loss": 0.5654, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 17.44, | |
| "learning_rate": 8.441697968088891e-06, | |
| "loss": 0.5955, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 17.47, | |
| "learning_rate": 8.297580295566575e-06, | |
| "loss": 0.6597, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 17.49, | |
| "learning_rate": 8.154650172170975e-06, | |
| "loss": 0.6575, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 17.51, | |
| "learning_rate": 8.012909448851514e-06, | |
| "loss": 0.6304, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 17.53, | |
| "learning_rate": 7.872359961154906e-06, | |
| "loss": 0.6387, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 17.56, | |
| "learning_rate": 7.733003529201278e-06, | |
| "loss": 0.6089, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 17.58, | |
| "learning_rate": 7.594841957660637e-06, | |
| "loss": 0.6738, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 17.6, | |
| "learning_rate": 7.457877035729588e-06, | |
| "loss": 0.6523, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 17.62, | |
| "learning_rate": 7.322110537108007e-06, | |
| "loss": 0.627, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 17.64, | |
| "learning_rate": 7.187544219976205e-06, | |
| "loss": 0.6855, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 17.67, | |
| "learning_rate": 7.054179826972074e-06, | |
| "loss": 0.5837, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 17.69, | |
| "learning_rate": 6.9220190851685516e-06, | |
| "loss": 0.6162, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 17.71, | |
| "learning_rate": 6.7910637060512924e-06, | |
| "loss": 0.6091, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 17.73, | |
| "learning_rate": 6.661315385496425e-06, | |
| "loss": 0.5933, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 17.76, | |
| "learning_rate": 6.5327758037486585e-06, | |
| "loss": 0.5962, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "learning_rate": 6.405446625399481e-06, | |
| "loss": 0.6714, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 17.8, | |
| "learning_rate": 6.2793294993656494e-06, | |
| "loss": 0.5962, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 17.82, | |
| "learning_rate": 6.1544260588677575e-06, | |
| "loss": 0.6211, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 17.84, | |
| "learning_rate": 6.030737921409169e-06, | |
| "loss": 0.6191, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 17.87, | |
| "learning_rate": 5.908266688755049e-06, | |
| "loss": 0.6064, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 17.89, | |
| "learning_rate": 5.787013946911546e-06, | |
| "loss": 0.6123, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 17.91, | |
| "learning_rate": 5.666981266105398e-06, | |
| "loss": 0.6287, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 17.93, | |
| "learning_rate": 5.54817020076347e-06, | |
| "loss": 0.6648, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 17.96, | |
| "learning_rate": 5.430582289492659e-06, | |
| "loss": 0.6086, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 17.98, | |
| "learning_rate": 5.314219055060022e-06, | |
| "loss": 0.6147, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "learning_rate": 5.199082004372957e-06, | |
| "loss": 0.6768, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 18.02, | |
| "learning_rate": 5.085172628459778e-06, | |
| "loss": 0.6069, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 18.04, | |
| "learning_rate": 4.972492402450402e-06, | |
| "loss": 0.6294, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 18.07, | |
| "learning_rate": 4.861042785557146e-06, | |
| "loss": 0.5703, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 18.09, | |
| "learning_rate": 4.750825221055965e-06, | |
| "loss": 0.5938, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 18.11, | |
| "learning_rate": 4.641841136267666e-06, | |
| "loss": 0.6108, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 18.13, | |
| "learning_rate": 4.534091942539475e-06, | |
| "loss": 0.605, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 18.16, | |
| "learning_rate": 4.427579035226725e-06, | |
| "loss": 0.5735, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 18.18, | |
| "learning_rate": 4.322303793674798e-06, | |
| "loss": 0.603, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 18.2, | |
| "learning_rate": 4.2182675812012965e-06, | |
| "loss": 0.572, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 18.22, | |
| "learning_rate": 4.115471745078314e-06, | |
| "loss": 0.6404, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 18.24, | |
| "learning_rate": 4.0139176165150835e-06, | |
| "loss": 0.5867, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 18.27, | |
| "learning_rate": 3.913606510640644e-06, | |
| "loss": 0.5813, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 18.29, | |
| "learning_rate": 3.8145397264868656e-06, | |
| "loss": 0.645, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 18.31, | |
| "learning_rate": 3.7167185469716426e-06, | |
| "loss": 0.6445, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 18.33, | |
| "learning_rate": 3.620144238882206e-06, | |
| "loss": 0.5747, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 18.36, | |
| "learning_rate": 3.5248180528588024e-06, | |
| "loss": 0.5996, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 18.38, | |
| "learning_rate": 3.4307412233784308e-06, | |
| "loss": 0.5754, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 18.4, | |
| "learning_rate": 3.3379149687388867e-06, | |
| "loss": 0.6387, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 18.42, | |
| "learning_rate": 3.2463404910430206e-06, | |
| "loss": 0.5884, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 18.44, | |
| "learning_rate": 3.1560189761830728e-06, | |
| "loss": 0.6265, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 18.47, | |
| "learning_rate": 3.06695159382544e-06, | |
| "loss": 0.572, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 18.49, | |
| "learning_rate": 2.9791394973954225e-06, | |
| "loss": 0.583, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 18.51, | |
| "learning_rate": 2.892583824062334e-06, | |
| "loss": 0.5796, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 18.53, | |
| "learning_rate": 2.8072856947248037e-06, | |
| "loss": 0.623, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 18.56, | |
| "learning_rate": 2.723246213996178e-06, | |
| "loss": 0.5913, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 18.58, | |
| "learning_rate": 2.640466470190317e-06, | |
| "loss": 0.5737, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 18.6, | |
| "learning_rate": 2.5589475353073988e-06, | |
| "loss": 0.6025, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 18.62, | |
| "learning_rate": 2.478690465020117e-06, | |
| "loss": 0.6367, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 18.64, | |
| "learning_rate": 2.3996962986599814e-06, | |
| "loss": 0.5872, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 18.67, | |
| "learning_rate": 2.3219660592038285e-06, | |
| "loss": 0.5713, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 18.69, | |
| "learning_rate": 2.2455007532606655e-06, | |
| "loss": 0.6096, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 18.71, | |
| "learning_rate": 2.170301371058503e-06, | |
| "loss": 0.6313, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 18.73, | |
| "learning_rate": 2.0963688864316323e-06, | |
| "loss": 0.6387, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 18.76, | |
| "learning_rate": 2.0237042568080012e-06, | |
| "loss": 0.6155, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 18.78, | |
| "learning_rate": 1.9523084231967358e-06, | |
| "loss": 0.6047, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 18.8, | |
| "learning_rate": 1.882182310176095e-06, | |
| "loss": 0.5659, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 18.82, | |
| "learning_rate": 1.8133268258813563e-06, | |
| "loss": 0.6431, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 18.84, | |
| "learning_rate": 1.745742861993138e-06, | |
| "loss": 0.5503, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 18.87, | |
| "learning_rate": 1.6794312937258417e-06, | |
| "loss": 0.647, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 18.89, | |
| "learning_rate": 1.6143929798162704e-06, | |
| "loss": 0.5608, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 18.91, | |
| "learning_rate": 1.550628762512596e-06, | |
| "loss": 0.5947, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 18.93, | |
| "learning_rate": 1.488139467563354e-06, | |
| "loss": 0.624, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 18.96, | |
| "learning_rate": 1.4269259042068327e-06, | |
| "loss": 0.5898, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 18.98, | |
| "learning_rate": 1.3669888651605345e-06, | |
| "loss": 0.6182, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "learning_rate": 1.30832912661093e-06, | |
| "loss": 0.5952, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 19.02, | |
| "learning_rate": 1.2509474482034433e-06, | |
| "loss": 0.6436, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 19.04, | |
| "learning_rate": 1.1948445730325163e-06, | |
| "loss": 0.6177, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 19.07, | |
| "learning_rate": 1.1400212276321376e-06, | |
| "loss": 0.5146, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 19.09, | |
| "learning_rate": 1.0864781219662611e-06, | |
| "loss": 0.6079, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 19.11, | |
| "learning_rate": 1.034215949419748e-06, | |
| "loss": 0.625, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 19.13, | |
| "learning_rate": 9.832353867893386e-07, | |
| "loss": 0.5764, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 19.16, | |
| "learning_rate": 9.335370942748389e-07, | |
| "loss": 0.6313, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 19.18, | |
| "learning_rate": 8.851217154706826e-07, | |
| "loss": 0.5654, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 19.2, | |
| "learning_rate": 8.379898773574924e-07, | |
| "loss": 0.6265, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 19.22, | |
| "learning_rate": 7.921421902939874e-07, | |
| "loss": 0.6167, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 19.24, | |
| "learning_rate": 7.475792480091226e-07, | |
| "loss": 0.5459, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 19.27, | |
| "learning_rate": 7.043016275943615e-07, | |
| "loss": 0.5786, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 19.29, | |
| "learning_rate": 6.623098894962044e-07, | |
| "loss": 0.5356, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 19.31, | |
| "learning_rate": 6.216045775089275e-07, | |
| "loss": 0.5835, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 19.33, | |
| "learning_rate": 5.821862187675775e-07, | |
| "loss": 0.6123, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 19.36, | |
| "learning_rate": 5.440553237410772e-07, | |
| "loss": 0.6187, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 19.38, | |
| "learning_rate": 5.072123862256972e-07, | |
| "loss": 0.5815, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 19.4, | |
| "learning_rate": 4.7165788333860536e-07, | |
| "loss": 0.6494, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 19.42, | |
| "learning_rate": 4.373922755116722e-07, | |
| "loss": 0.5793, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 19.44, | |
| "learning_rate": 4.044160064855751e-07, | |
| "loss": 0.5752, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 19.47, | |
| "learning_rate": 3.727295033040035e-07, | |
| "loss": 0.6265, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 19.49, | |
| "learning_rate": 3.4233317630814053e-07, | |
| "loss": 0.6616, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 19.51, | |
| "learning_rate": 3.132274191313345e-07, | |
| "loss": 0.5957, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 19.53, | |
| "learning_rate": 2.854126086940356e-07, | |
| "loss": 0.624, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 19.56, | |
| "learning_rate": 2.588891051988895e-07, | |
| "loss": 0.5566, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 19.58, | |
| "learning_rate": 2.3365725212607381e-07, | |
| "loss": 0.5947, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 19.6, | |
| "learning_rate": 2.0971737622883515e-07, | |
| "loss": 0.5977, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 19.62, | |
| "learning_rate": 1.8706978752931482e-07, | |
| "loss": 0.6406, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 19.64, | |
| "learning_rate": 1.6571477931446312e-07, | |
| "loss": 0.624, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 19.67, | |
| "learning_rate": 1.4565262813230894e-07, | |
| "loss": 0.5942, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 19.69, | |
| "learning_rate": 1.2688359378829618e-07, | |
| "loss": 0.5977, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 19.71, | |
| "learning_rate": 1.0940791934198613e-07, | |
| "loss": 0.6108, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 19.73, | |
| "learning_rate": 9.32258311039269e-08, | |
| "loss": 0.5933, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 19.76, | |
| "learning_rate": 7.833753863263349e-08, | |
| "loss": 0.6123, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 19.78, | |
| "learning_rate": 6.474323473194543e-08, | |
| "loss": 0.5894, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 19.8, | |
| "learning_rate": 5.2443095448506674e-08, | |
| "loss": 0.5991, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 19.82, | |
| "learning_rate": 4.143728006951175e-08, | |
| "loss": 0.6077, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 19.84, | |
| "learning_rate": 3.1725931120596405e-08, | |
| "loss": 0.6455, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 19.87, | |
| "learning_rate": 2.3309174364027907e-08, | |
| "loss": 0.5962, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 19.89, | |
| "learning_rate": 1.6187118797061917e-08, | |
| "loss": 0.6064, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 19.91, | |
| "learning_rate": 1.0359856650532519e-08, | |
| "loss": 0.5718, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 19.93, | |
| "learning_rate": 5.827463387653165e-09, | |
| "loss": 0.5679, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 19.96, | |
| "learning_rate": 2.589997703072999e-09, | |
| "loss": 0.5688, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 19.98, | |
| "learning_rate": 6.475015220552827e-10, | |
| "loss": 0.6479, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.6077, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "step": 900, | |
| "total_flos": 1.5961750176771604e+18, | |
| "train_loss": 1.5027020941840277, | |
| "train_runtime": 6381.0295, | |
| "train_samples_per_second": 71.835, | |
| "train_steps_per_second": 0.141 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 900, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 20, | |
| "save_steps": 500, | |
| "total_flos": 1.5961750176771604e+18, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |