| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 17.142857142857142, | |
| "eval_steps": 500, | |
| "global_step": 600, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.000999404761904762, | |
| "loss": 3.0944, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.0009988095238095238, | |
| "loss": 2.7802, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.0009982142857142857, | |
| "loss": 2.7798, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0009976190476190477, | |
| "loss": 2.6729, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0009970238095238096, | |
| "loss": 2.7544, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.0009964285714285715, | |
| "loss": 2.7115, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.0009958333333333334, | |
| "loss": 2.7491, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.0009952380952380953, | |
| "loss": 2.681, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0009946428571428571, | |
| "loss": 2.7396, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 0.000994047619047619, | |
| "loss": 2.6911, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.0009934523809523809, | |
| "loss": 2.6725, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.000992857142857143, | |
| "loss": 2.6951, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 0.0009922619047619049, | |
| "loss": 2.6741, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 0.0009916666666666667, | |
| "loss": 2.6623, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 0.0009910714285714286, | |
| "loss": 2.6268, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 0.0009904761904761905, | |
| "loss": 2.6727, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 0.0009898809523809524, | |
| "loss": 2.6174, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 0.0009892857142857142, | |
| "loss": 2.6544, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 0.0009886904761904763, | |
| "loss": 2.5404, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 0.0009880952380952382, | |
| "loss": 2.6286, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 0.0009875, | |
| "loss": 2.6824, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 0.000986904761904762, | |
| "loss": 2.6012, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 0.0009863095238095239, | |
| "loss": 2.6446, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 0.0009857142857142857, | |
| "loss": 2.6437, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 0.0009851190476190476, | |
| "loss": 2.6596, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 0.0009845238095238097, | |
| "loss": 2.6554, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 0.0009839285714285714, | |
| "loss": 2.6064, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 0.0009833333333333332, | |
| "loss": 2.5994, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 0.0009827380952380951, | |
| "loss": 2.545, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 0.0009821428571428572, | |
| "loss": 2.5704, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 0.000981547619047619, | |
| "loss": 2.6461, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 0.000980952380952381, | |
| "loss": 2.631, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 0.0009803571428571428, | |
| "loss": 2.6678, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 0.0009797619047619047, | |
| "loss": 2.5964, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 0.0009791666666666666, | |
| "loss": 2.6566, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 0.0009785714285714285, | |
| "loss": 2.4962, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 0.0009779761904761906, | |
| "loss": 2.4815, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 0.0009773809523809524, | |
| "loss": 2.4172, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 0.0009767857142857143, | |
| "loss": 2.4641, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 0.0009761904761904762, | |
| "loss": 2.4875, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 0.0009755952380952381, | |
| "loss": 2.4486, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 0.000975, | |
| "loss": 2.4463, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 0.0009744047619047619, | |
| "loss": 2.5063, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 0.0009738095238095238, | |
| "loss": 2.4506, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 0.0009732142857142857, | |
| "loss": 2.3862, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 0.0009726190476190476, | |
| "loss": 2.4925, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 0.0009720238095238096, | |
| "loss": 2.5, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 0.0009714285714285714, | |
| "loss": 2.4658, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 0.0009708333333333333, | |
| "loss": 2.3983, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 0.0009702380952380953, | |
| "loss": 2.369, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 0.0009696428571428572, | |
| "loss": 2.4849, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 0.0009690476190476191, | |
| "loss": 2.4106, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 0.0009684523809523809, | |
| "loss": 2.4363, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 0.0009678571428571429, | |
| "loss": 2.4935, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 0.0009672619047619048, | |
| "loss": 2.479, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 0.0009666666666666667, | |
| "loss": 2.4299, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 0.0009660714285714285, | |
| "loss": 2.3964, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 0.0009654761904761905, | |
| "loss": 2.4865, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 0.0009648809523809524, | |
| "loss": 2.3831, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 0.0009642857142857143, | |
| "loss": 2.4304, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 0.0009636904761904763, | |
| "loss": 2.4273, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 0.0009630952380952382, | |
| "loss": 2.4427, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 0.0009625, | |
| "loss": 2.4191, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 0.0009619047619047619, | |
| "loss": 2.3902, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 0.0009613095238095239, | |
| "loss": 2.4699, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 0.0009607142857142858, | |
| "loss": 2.3978, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 0.0009601190476190476, | |
| "loss": 2.4128, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 0.0009595238095238095, | |
| "loss": 2.4585, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 0.0009589285714285715, | |
| "loss": 2.4476, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 0.0009583333333333334, | |
| "loss": 2.4231, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 0.0009577380952380953, | |
| "loss": 2.2655, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 0.0009571428571428573, | |
| "loss": 2.208, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 0.0009565476190476191, | |
| "loss": 2.1758, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 0.000955952380952381, | |
| "loss": 2.3113, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 0.0009553571428571429, | |
| "loss": 2.1739, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 0.0009547619047619049, | |
| "loss": 2.2234, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 0.0009541666666666667, | |
| "loss": 2.21, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 0.0009535714285714286, | |
| "loss": 2.2316, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 0.0009529761904761904, | |
| "loss": 2.2044, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 0.0009523809523809524, | |
| "loss": 2.1784, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 0.0009517857142857143, | |
| "loss": 2.2489, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 0.0009511904761904761, | |
| "loss": 2.2003, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 0.0009505952380952381, | |
| "loss": 2.169, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 0.00095, | |
| "loss": 2.2303, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 0.0009494047619047619, | |
| "loss": 2.1744, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 0.0009488095238095238, | |
| "loss": 2.1904, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 0.0009482142857142857, | |
| "loss": 2.222, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 0.0009476190476190476, | |
| "loss": 2.2467, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 0.0009470238095238095, | |
| "loss": 2.2241, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 0.0009464285714285714, | |
| "loss": 2.2339, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 0.0009458333333333334, | |
| "loss": 2.2339, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 0.0009452380952380952, | |
| "loss": 2.1653, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 0.0009446428571428571, | |
| "loss": 2.2954, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 0.0009440476190476191, | |
| "loss": 2.2859, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 0.000943452380952381, | |
| "loss": 2.2107, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 0.0009428571428571429, | |
| "loss": 2.2142, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 0.0009422619047619047, | |
| "loss": 2.2433, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 0.0009416666666666667, | |
| "loss": 2.2417, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 0.0009410714285714286, | |
| "loss": 2.2045, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 0.0009404761904761905, | |
| "loss": 2.293, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 0.0009398809523809523, | |
| "loss": 2.2051, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 0.0009392857142857143, | |
| "loss": 2.2889, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 0.0009386904761904762, | |
| "loss": 2.2, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 0.0009380952380952381, | |
| "loss": 2.2298, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 0.0009375, | |
| "loss": 2.2722, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 0.000936904761904762, | |
| "loss": 2.0032, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 0.0009363095238095238, | |
| "loss": 1.9269, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 0.0009357142857142857, | |
| "loss": 1.916, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 0.0009351190476190477, | |
| "loss": 1.9165, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 0.0009345238095238096, | |
| "loss": 1.9296, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 0.0009339285714285714, | |
| "loss": 1.954, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 0.0009333333333333333, | |
| "loss": 1.9457, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 0.0009327380952380953, | |
| "loss": 1.9135, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 0.0009321428571428572, | |
| "loss": 1.9383, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 0.0009315476190476191, | |
| "loss": 1.9057, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 0.0009309523809523809, | |
| "loss": 1.9541, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 0.0009303571428571429, | |
| "loss": 1.8827, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 0.0009297619047619048, | |
| "loss": 1.899, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 0.0009291666666666667, | |
| "loss": 1.9095, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 0.0009285714285714287, | |
| "loss": 1.936, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 0.0009279761904761905, | |
| "loss": 1.9224, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 0.0009273809523809524, | |
| "loss": 1.9547, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 0.0009267857142857143, | |
| "loss": 1.9323, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 0.0009261904761904763, | |
| "loss": 1.9938, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 0.0009255952380952382, | |
| "loss": 1.8674, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 0.000925, | |
| "loss": 1.95, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 0.0009244047619047619, | |
| "loss": 1.9374, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 0.0009238095238095239, | |
| "loss": 1.9456, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 0.0009232142857142858, | |
| "loss": 1.9791, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 0.0009226190476190477, | |
| "loss": 2.0047, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 0.0009220238095238096, | |
| "loss": 1.9971, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 0.0009214285714285714, | |
| "loss": 2.0267, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 0.0009208333333333333, | |
| "loss": 1.9374, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 0.0009202380952380952, | |
| "loss": 1.9793, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 0.0009196428571428572, | |
| "loss": 2.0483, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 0.000919047619047619, | |
| "loss": 2.0003, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 0.0009184523809523809, | |
| "loss": 2.0185, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 0.0009178571428571428, | |
| "loss": 2.0517, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 0.0009172619047619048, | |
| "loss": 1.9824, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 0.0009166666666666666, | |
| "loss": 2.0383, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 0.0009160714285714285, | |
| "loss": 1.6818, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 0.0009154761904761905, | |
| "loss": 1.6208, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 0.0009148809523809524, | |
| "loss": 1.6843, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 0.0009142857142857143, | |
| "loss": 1.5885, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 0.0009136904761904761, | |
| "loss": 1.5799, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 0.0009130952380952381, | |
| "loss": 1.6334, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 0.0009125, | |
| "loss": 1.6297, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 0.0009119047619047619, | |
| "loss": 1.5929, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 0.0009113095238095238, | |
| "loss": 1.6621, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 0.0009107142857142857, | |
| "loss": 1.626, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 0.0009101190476190476, | |
| "loss": 1.6138, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 0.0009095238095238095, | |
| "loss": 1.6465, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 0.0009089285714285715, | |
| "loss": 1.6622, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 0.0009083333333333334, | |
| "loss": 1.6662, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 0.0009077380952380952, | |
| "loss": 1.6348, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 0.0009071428571428571, | |
| "loss": 1.6196, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 0.0009065476190476191, | |
| "loss": 1.6766, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 0.000905952380952381, | |
| "loss": 1.7069, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 0.0009053571428571429, | |
| "loss": 1.6848, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 0.0009047619047619047, | |
| "loss": 1.6884, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 0.0009041666666666667, | |
| "loss": 1.6721, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 0.0009035714285714286, | |
| "loss": 1.7116, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 0.0009029761904761905, | |
| "loss": 1.693, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 0.0009023809523809525, | |
| "loss": 1.6826, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 0.0009017857142857143, | |
| "loss": 1.7061, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 0.0009011904761904762, | |
| "loss": 1.6964, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 0.0009005952380952381, | |
| "loss": 1.7903, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 0.0009000000000000001, | |
| "loss": 1.6829, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 0.000899404761904762, | |
| "loss": 1.7047, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 0.0008988095238095238, | |
| "loss": 1.7671, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 0.0008982142857142857, | |
| "loss": 1.7184, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 0.0008976190476190477, | |
| "loss": 1.8213, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 0.0008970238095238096, | |
| "loss": 1.7688, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 0.0008964285714285715, | |
| "loss": 1.7522, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 0.0008958333333333334, | |
| "loss": 1.7862, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 5.03, | |
| "learning_rate": 0.0008952380952380953, | |
| "loss": 1.3871, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 5.06, | |
| "learning_rate": 0.0008946428571428572, | |
| "loss": 1.3491, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 0.0008940476190476191, | |
| "loss": 1.3399, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 0.0008934523809523811, | |
| "loss": 1.3569, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 5.14, | |
| "learning_rate": 0.0008928571428571429, | |
| "loss": 1.3734, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 5.17, | |
| "learning_rate": 0.0008922619047619048, | |
| "loss": 1.3151, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 0.0008916666666666667, | |
| "loss": 1.3243, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 0.0008910714285714287, | |
| "loss": 1.342, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 0.0008904761904761904, | |
| "loss": 1.3664, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 0.0008898809523809523, | |
| "loss": 1.3493, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 0.0008892857142857142, | |
| "loss": 1.32, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 5.34, | |
| "learning_rate": 0.0008886904761904762, | |
| "loss": 1.3978, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 0.0008880952380952381, | |
| "loss": 1.3762, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 0.0008874999999999999, | |
| "loss": 1.4172, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 0.0008869047619047619, | |
| "loss": 1.3817, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 5.46, | |
| "learning_rate": 0.0008863095238095238, | |
| "loss": 1.3779, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 0.0008857142857142857, | |
| "loss": 1.378, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 0.0008851190476190476, | |
| "loss": 1.4245, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 5.54, | |
| "learning_rate": 0.0008845238095238095, | |
| "loss": 1.4425, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 0.0008839285714285714, | |
| "loss": 1.4324, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 0.0008833333333333333, | |
| "loss": 1.4264, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 0.0008827380952380952, | |
| "loss": 1.4395, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 5.66, | |
| "learning_rate": 0.0008821428571428572, | |
| "loss": 1.4549, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 5.69, | |
| "learning_rate": 0.000881547619047619, | |
| "loss": 1.501, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 0.0008809523809523809, | |
| "loss": 1.432, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 0.0008803571428571429, | |
| "loss": 1.4922, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 0.0008797619047619048, | |
| "loss": 1.4622, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 0.0008791666666666667, | |
| "loss": 1.4794, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 0.0008785714285714285, | |
| "loss": 1.4938, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 5.86, | |
| "learning_rate": 0.0008779761904761905, | |
| "loss": 1.4792, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 5.89, | |
| "learning_rate": 0.0008773809523809524, | |
| "loss": 1.5192, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 5.91, | |
| "learning_rate": 0.0008767857142857143, | |
| "loss": 1.5055, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 5.94, | |
| "learning_rate": 0.0008761904761904762, | |
| "loss": 1.5484, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 0.0008755952380952381, | |
| "loss": 1.5096, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 0.000875, | |
| "loss": 1.5298, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 6.03, | |
| "learning_rate": 0.0008744047619047619, | |
| "loss": 1.1704, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 6.06, | |
| "learning_rate": 0.0008738095238095239, | |
| "loss": 1.1261, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 0.0008732142857142858, | |
| "loss": 1.1144, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 0.0008726190476190476, | |
| "loss": 1.0984, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 6.14, | |
| "learning_rate": 0.0008720238095238095, | |
| "loss": 1.0704, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 6.17, | |
| "learning_rate": 0.0008714285714285715, | |
| "loss": 1.0655, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 6.2, | |
| "learning_rate": 0.0008708333333333334, | |
| "loss": 1.09, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 0.0008702380952380953, | |
| "loss": 1.0619, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "learning_rate": 0.0008696428571428571, | |
| "loss": 1.1633, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 0.0008690476190476191, | |
| "loss": 1.1022, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 0.000868452380952381, | |
| "loss": 1.1057, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 0.0008678571428571429, | |
| "loss": 1.1279, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 0.0008672619047619049, | |
| "loss": 1.0915, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 0.0008666666666666667, | |
| "loss": 1.1731, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 0.0008660714285714286, | |
| "loss": 1.1352, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 6.46, | |
| "learning_rate": 0.0008654761904761905, | |
| "loss": 1.1632, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 0.0008648809523809525, | |
| "loss": 1.1691, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 0.0008642857142857144, | |
| "loss": 1.181, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 6.54, | |
| "learning_rate": 0.0008636904761904762, | |
| "loss": 1.1635, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 0.0008630952380952381, | |
| "loss": 1.1802, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 0.0008625000000000001, | |
| "loss": 1.2111, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 6.63, | |
| "learning_rate": 0.000861904761904762, | |
| "loss": 1.2503, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 6.66, | |
| "learning_rate": 0.0008613095238095238, | |
| "loss": 1.2305, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 6.69, | |
| "learning_rate": 0.0008607142857142858, | |
| "loss": 1.2446, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 0.0008601190476190477, | |
| "loss": 1.263, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 6.74, | |
| "learning_rate": 0.0008595238095238096, | |
| "loss": 1.2407, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 6.77, | |
| "learning_rate": 0.0008589285714285714, | |
| "loss": 1.303, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 6.8, | |
| "learning_rate": 0.0008583333333333333, | |
| "loss": 1.2309, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 6.83, | |
| "learning_rate": 0.0008577380952380952, | |
| "loss": 1.2669, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 6.86, | |
| "learning_rate": 0.0008571428571428571, | |
| "loss": 1.226, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 0.000856547619047619, | |
| "loss": 1.2862, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 6.91, | |
| "learning_rate": 0.000855952380952381, | |
| "loss": 1.2472, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 6.94, | |
| "learning_rate": 0.0008553571428571428, | |
| "loss": 1.2928, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 6.97, | |
| "learning_rate": 0.0008547619047619047, | |
| "loss": 1.2427, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 0.0008541666666666666, | |
| "loss": 1.3195, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 7.03, | |
| "learning_rate": 0.0008535714285714286, | |
| "loss": 0.8949, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 0.0008529761904761905, | |
| "loss": 0.8907, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 0.0008523809523809523, | |
| "loss": 0.8813, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 0.0008517857142857143, | |
| "loss": 0.8702, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 0.0008511904761904762, | |
| "loss": 0.9105, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 0.0008505952380952381, | |
| "loss": 0.9096, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 0.00085, | |
| "loss": 0.9121, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "learning_rate": 0.0008494047619047619, | |
| "loss": 0.9063, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 0.0008488095238095238, | |
| "loss": 0.8976, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 7.29, | |
| "learning_rate": 0.0008482142857142857, | |
| "loss": 0.9283, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 0.0008476190476190476, | |
| "loss": 0.9409, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 0.0008470238095238096, | |
| "loss": 0.9311, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "learning_rate": 0.0008464285714285714, | |
| "loss": 0.926, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 0.0008458333333333333, | |
| "loss": 0.9704, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 0.0008452380952380953, | |
| "loss": 0.9515, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 7.46, | |
| "learning_rate": 0.0008446428571428572, | |
| "loss": 0.9069, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 7.49, | |
| "learning_rate": 0.000844047619047619, | |
| "loss": 0.9359, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "learning_rate": 0.0008434523809523809, | |
| "loss": 0.9482, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 7.54, | |
| "learning_rate": 0.0008428571428571429, | |
| "loss": 0.9717, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "learning_rate": 0.0008422619047619048, | |
| "loss": 0.9869, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 0.0008416666666666667, | |
| "loss": 0.9728, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 7.63, | |
| "learning_rate": 0.0008410714285714285, | |
| "loss": 0.9516, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 7.66, | |
| "learning_rate": 0.0008404761904761905, | |
| "loss": 0.9838, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 7.69, | |
| "learning_rate": 0.0008398809523809524, | |
| "loss": 1.0044, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 7.71, | |
| "learning_rate": 0.0008392857142857143, | |
| "loss": 1.0153, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 7.74, | |
| "learning_rate": 0.0008386904761904763, | |
| "loss": 1.0382, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 7.77, | |
| "learning_rate": 0.0008380952380952382, | |
| "loss": 1.0109, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 0.0008375, | |
| "loss": 0.9989, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 7.83, | |
| "learning_rate": 0.0008369047619047619, | |
| "loss": 1.0631, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 7.86, | |
| "learning_rate": 0.0008363095238095239, | |
| "loss": 1.0546, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 0.0008357142857142858, | |
| "loss": 1.0827, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 7.91, | |
| "learning_rate": 0.0008351190476190476, | |
| "loss": 1.087, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 0.0008345238095238095, | |
| "loss": 1.041, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 7.97, | |
| "learning_rate": 0.0008339285714285715, | |
| "loss": 1.0633, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 0.0008333333333333334, | |
| "loss": 1.0709, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 8.03, | |
| "learning_rate": 0.0008327380952380953, | |
| "loss": 0.7273, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 0.0008321428571428573, | |
| "loss": 0.726, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 8.09, | |
| "learning_rate": 0.0008315476190476191, | |
| "loss": 0.6943, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 8.11, | |
| "learning_rate": 0.000830952380952381, | |
| "loss": 0.7127, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 8.14, | |
| "learning_rate": 0.0008303571428571429, | |
| "loss": 0.6915, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 8.17, | |
| "learning_rate": 0.0008297619047619049, | |
| "loss": 0.7138, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 0.0008291666666666667, | |
| "loss": 0.7356, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 8.23, | |
| "learning_rate": 0.0008285714285714286, | |
| "loss": 0.678, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 0.0008279761904761904, | |
| "loss": 0.7375, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 8.29, | |
| "learning_rate": 0.0008273809523809524, | |
| "loss": 0.7284, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 8.31, | |
| "learning_rate": 0.0008267857142857143, | |
| "loss": 0.7304, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 8.34, | |
| "learning_rate": 0.0008261904761904761, | |
| "loss": 0.7633, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 8.37, | |
| "learning_rate": 0.0008255952380952381, | |
| "loss": 0.7416, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 0.000825, | |
| "loss": 0.7895, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 8.43, | |
| "learning_rate": 0.0008244047619047619, | |
| "loss": 0.8037, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 8.46, | |
| "learning_rate": 0.0008238095238095238, | |
| "loss": 0.7736, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 8.49, | |
| "learning_rate": 0.0008232142857142857, | |
| "loss": 0.778, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 8.51, | |
| "learning_rate": 0.0008226190476190476, | |
| "loss": 0.7644, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 8.54, | |
| "learning_rate": 0.0008220238095238095, | |
| "loss": 0.7942, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 8.57, | |
| "learning_rate": 0.0008214285714285714, | |
| "loss": 0.7715, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 0.0008208333333333334, | |
| "loss": 0.8288, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 8.63, | |
| "learning_rate": 0.0008202380952380952, | |
| "loss": 0.8263, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 8.66, | |
| "learning_rate": 0.0008196428571428571, | |
| "loss": 0.7923, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 8.69, | |
| "learning_rate": 0.0008190476190476191, | |
| "loss": 0.8063, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 8.71, | |
| "learning_rate": 0.000818452380952381, | |
| "loss": 0.8016, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 8.74, | |
| "learning_rate": 0.0008178571428571428, | |
| "loss": 0.8467, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 8.77, | |
| "learning_rate": 0.0008172619047619047, | |
| "loss": 0.8353, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 8.8, | |
| "learning_rate": 0.0008166666666666667, | |
| "loss": 0.8272, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 8.83, | |
| "learning_rate": 0.0008160714285714286, | |
| "loss": 0.8852, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 8.86, | |
| "learning_rate": 0.0008154761904761905, | |
| "loss": 0.8541, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 0.0008148809523809523, | |
| "loss": 0.8236, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 8.91, | |
| "learning_rate": 0.0008142857142857143, | |
| "loss": 0.8609, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 8.94, | |
| "learning_rate": 0.0008136904761904762, | |
| "loss": 0.8802, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 8.97, | |
| "learning_rate": 0.0008130952380952381, | |
| "loss": 0.8615, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "learning_rate": 0.0008125000000000001, | |
| "loss": 0.8514, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 9.03, | |
| "learning_rate": 0.000811904761904762, | |
| "loss": 0.5529, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 9.06, | |
| "learning_rate": 0.0008113095238095238, | |
| "loss": 0.5736, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 0.0008107142857142857, | |
| "loss": 0.5647, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 9.11, | |
| "learning_rate": 0.0008101190476190477, | |
| "loss": 0.5677, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "learning_rate": 0.0008095238095238096, | |
| "loss": 0.5991, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 9.17, | |
| "learning_rate": 0.0008089285714285714, | |
| "loss": 0.5666, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 9.2, | |
| "learning_rate": 0.0008083333333333333, | |
| "loss": 0.5902, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 9.23, | |
| "learning_rate": 0.0008077380952380953, | |
| "loss": 0.5961, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 0.0008071428571428572, | |
| "loss": 0.5684, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 9.29, | |
| "learning_rate": 0.0008065476190476191, | |
| "loss": 0.5976, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 0.0008059523809523809, | |
| "loss": 0.6033, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 0.0008053571428571429, | |
| "loss": 0.5877, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 9.37, | |
| "learning_rate": 0.0008047619047619048, | |
| "loss": 0.5943, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 0.0008041666666666667, | |
| "loss": 0.6176, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 0.0008035714285714287, | |
| "loss": 0.6143, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 9.46, | |
| "learning_rate": 0.0008029761904761905, | |
| "loss": 0.597, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 9.49, | |
| "learning_rate": 0.0008023809523809524, | |
| "loss": 0.604, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 0.0008017857142857143, | |
| "loss": 0.6036, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 9.54, | |
| "learning_rate": 0.0008011904761904763, | |
| "loss": 0.6243, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 0.0008005952380952382, | |
| "loss": 0.6301, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 9.6, | |
| "learning_rate": 0.0008, | |
| "loss": 0.6271, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 0.0007994047619047619, | |
| "loss": 0.6246, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 9.66, | |
| "learning_rate": 0.0007988095238095239, | |
| "loss": 0.6597, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 0.0007982142857142858, | |
| "loss": 0.6517, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 9.71, | |
| "learning_rate": 0.0007976190476190477, | |
| "loss": 0.6645, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 9.74, | |
| "learning_rate": 0.0007970238095238096, | |
| "loss": 0.6542, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 9.77, | |
| "learning_rate": 0.0007964285714285714, | |
| "loss": 0.6496, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 9.8, | |
| "learning_rate": 0.0007958333333333333, | |
| "loss": 0.6309, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 0.0007952380952380952, | |
| "loss": 0.6668, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 9.86, | |
| "learning_rate": 0.0007946428571428572, | |
| "loss": 0.6841, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 0.000794047619047619, | |
| "loss": 0.6958, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 0.0007934523809523809, | |
| "loss": 0.6592, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 0.0007928571428571428, | |
| "loss": 0.6968, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "learning_rate": 0.0007922619047619048, | |
| "loss": 0.6916, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 0.0007916666666666666, | |
| "loss": 0.7155, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 10.03, | |
| "learning_rate": 0.0007910714285714285, | |
| "loss": 0.4288, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 10.06, | |
| "learning_rate": 0.0007904761904761905, | |
| "loss": 0.4493, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 0.0007898809523809524, | |
| "loss": 0.4152, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 0.0007892857142857143, | |
| "loss": 0.4324, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 10.14, | |
| "learning_rate": 0.0007886904761904761, | |
| "loss": 0.4334, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 10.17, | |
| "learning_rate": 0.0007880952380952381, | |
| "loss": 0.4479, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 10.2, | |
| "learning_rate": 0.0007875, | |
| "loss": 0.4391, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 0.0007869047619047619, | |
| "loss": 0.4534, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 10.26, | |
| "learning_rate": 0.0007863095238095238, | |
| "loss": 0.4494, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 10.29, | |
| "learning_rate": 0.0007857142857142857, | |
| "loss": 0.4519, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 0.0007851190476190476, | |
| "loss": 0.4673, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 10.34, | |
| "learning_rate": 0.0007845238095238095, | |
| "loss": 0.4628, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 10.37, | |
| "learning_rate": 0.0007839285714285715, | |
| "loss": 0.4608, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 0.0007833333333333334, | |
| "loss": 0.4755, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 10.43, | |
| "learning_rate": 0.0007827380952380952, | |
| "loss": 0.4771, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 10.46, | |
| "learning_rate": 0.0007821428571428571, | |
| "loss": 0.4679, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 0.0007815476190476191, | |
| "loss": 0.4985, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 10.51, | |
| "learning_rate": 0.000780952380952381, | |
| "loss": 0.5242, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 10.54, | |
| "learning_rate": 0.0007803571428571429, | |
| "loss": 0.478, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 10.57, | |
| "learning_rate": 0.0007797619047619047, | |
| "loss": 0.5072, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 10.6, | |
| "learning_rate": 0.0007791666666666667, | |
| "loss": 0.5001, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 10.63, | |
| "learning_rate": 0.0007785714285714286, | |
| "loss": 0.5119, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 10.66, | |
| "learning_rate": 0.0007779761904761905, | |
| "loss": 0.5212, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 10.69, | |
| "learning_rate": 0.0007773809523809525, | |
| "loss": 0.5073, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 10.71, | |
| "learning_rate": 0.0007767857142857143, | |
| "loss": 0.5089, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 0.0007761904761904762, | |
| "loss": 0.5161, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 0.0007755952380952381, | |
| "loss": 0.4861, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 0.0007750000000000001, | |
| "loss": 0.531, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 10.83, | |
| "learning_rate": 0.000774404761904762, | |
| "loss": 0.5244, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 10.86, | |
| "learning_rate": 0.0007738095238095238, | |
| "loss": 0.5446, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 0.0007732142857142857, | |
| "loss": 0.5515, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 0.0007726190476190477, | |
| "loss": 0.5345, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 10.94, | |
| "learning_rate": 0.0007720238095238096, | |
| "loss": 0.537, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 10.97, | |
| "learning_rate": 0.0007714285714285715, | |
| "loss": 0.5589, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 0.0007708333333333334, | |
| "loss": 0.5459, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 11.03, | |
| "learning_rate": 0.0007702380952380953, | |
| "loss": 0.3344, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 11.06, | |
| "learning_rate": 0.0007696428571428572, | |
| "loss": 0.3352, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 11.09, | |
| "learning_rate": 0.0007690476190476191, | |
| "loss": 0.3263, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 0.0007684523809523811, | |
| "loss": 0.3501, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 11.14, | |
| "learning_rate": 0.0007678571428571429, | |
| "loss": 0.3523, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "learning_rate": 0.0007672619047619048, | |
| "loss": 0.3379, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 11.2, | |
| "learning_rate": 0.0007666666666666667, | |
| "loss": 0.3456, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 11.23, | |
| "learning_rate": 0.0007660714285714287, | |
| "loss": 0.347, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 11.26, | |
| "learning_rate": 0.0007654761904761904, | |
| "loss": 0.3622, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 11.29, | |
| "learning_rate": 0.0007648809523809523, | |
| "loss": 0.3612, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 11.31, | |
| "learning_rate": 0.0007642857142857142, | |
| "loss": 0.3789, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 11.34, | |
| "learning_rate": 0.0007636904761904762, | |
| "loss": 0.3491, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 11.37, | |
| "learning_rate": 0.0007630952380952381, | |
| "loss": 0.3578, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 0.0007624999999999999, | |
| "loss": 0.3524, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 11.43, | |
| "learning_rate": 0.0007619047619047619, | |
| "loss": 0.3671, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 11.46, | |
| "learning_rate": 0.0007613095238095238, | |
| "loss": 0.374, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 11.49, | |
| "learning_rate": 0.0007607142857142857, | |
| "loss": 0.3872, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 11.51, | |
| "learning_rate": 0.0007601190476190476, | |
| "loss": 0.3801, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 11.54, | |
| "learning_rate": 0.0007595238095238095, | |
| "loss": 0.3618, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 11.57, | |
| "learning_rate": 0.0007589285714285714, | |
| "loss": 0.3929, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 0.0007583333333333333, | |
| "loss": 0.4099, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 11.63, | |
| "learning_rate": 0.0007577380952380952, | |
| "loss": 0.3778, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 11.66, | |
| "learning_rate": 0.0007571428571428572, | |
| "loss": 0.3965, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 11.69, | |
| "learning_rate": 0.000756547619047619, | |
| "loss": 0.4013, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 11.71, | |
| "learning_rate": 0.0007559523809523809, | |
| "loss": 0.4031, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 11.74, | |
| "learning_rate": 0.0007553571428571429, | |
| "loss": 0.398, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 11.77, | |
| "learning_rate": 0.0007547619047619048, | |
| "loss": 0.4106, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 11.8, | |
| "learning_rate": 0.0007541666666666667, | |
| "loss": 0.4031, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 11.83, | |
| "learning_rate": 0.0007535714285714285, | |
| "loss": 0.4199, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 11.86, | |
| "learning_rate": 0.0007529761904761905, | |
| "loss": 0.4012, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 0.0007523809523809524, | |
| "loss": 0.4096, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 11.91, | |
| "learning_rate": 0.0007517857142857143, | |
| "loss": 0.4237, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "learning_rate": 0.0007511904761904762, | |
| "loss": 0.4115, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 11.97, | |
| "learning_rate": 0.0007505952380952381, | |
| "loss": 0.4459, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 0.00075, | |
| "loss": 0.4406, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 12.03, | |
| "learning_rate": 0.0007494047619047619, | |
| "loss": 0.2561, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "learning_rate": 0.0007488095238095239, | |
| "loss": 0.2662, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 12.09, | |
| "learning_rate": 0.0007482142857142858, | |
| "loss": 0.2538, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 12.11, | |
| "learning_rate": 0.0007476190476190476, | |
| "loss": 0.2571, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 12.14, | |
| "learning_rate": 0.0007470238095238095, | |
| "loss": 0.2527, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 12.17, | |
| "learning_rate": 0.0007464285714285715, | |
| "loss": 0.2739, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 12.2, | |
| "learning_rate": 0.0007458333333333334, | |
| "loss": 0.2642, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 12.23, | |
| "learning_rate": 0.0007452380952380953, | |
| "loss": 0.2736, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 12.26, | |
| "learning_rate": 0.0007446428571428571, | |
| "loss": 0.2774, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 12.29, | |
| "learning_rate": 0.0007440476190476191, | |
| "loss": 0.2684, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 12.31, | |
| "learning_rate": 0.000743452380952381, | |
| "loss": 0.2727, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 12.34, | |
| "learning_rate": 0.0007428571428571429, | |
| "loss": 0.2801, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 12.37, | |
| "learning_rate": 0.0007422619047619049, | |
| "loss": 0.2685, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 0.0007416666666666667, | |
| "loss": 0.2958, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 12.43, | |
| "learning_rate": 0.0007410714285714286, | |
| "loss": 0.2868, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 12.46, | |
| "learning_rate": 0.0007404761904761905, | |
| "loss": 0.2838, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 12.49, | |
| "learning_rate": 0.0007398809523809525, | |
| "loss": 0.2976, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 12.51, | |
| "learning_rate": 0.0007392857142857144, | |
| "loss": 0.2963, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 12.54, | |
| "learning_rate": 0.0007386904761904762, | |
| "loss": 0.2891, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 12.57, | |
| "learning_rate": 0.0007380952380952381, | |
| "loss": 0.2897, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 12.6, | |
| "learning_rate": 0.0007375000000000001, | |
| "loss": 0.3013, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 12.63, | |
| "learning_rate": 0.000736904761904762, | |
| "loss": 0.2977, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 12.66, | |
| "learning_rate": 0.0007363095238095238, | |
| "loss": 0.2985, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 12.69, | |
| "learning_rate": 0.0007357142857142858, | |
| "loss": 0.3097, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 12.71, | |
| "learning_rate": 0.0007351190476190477, | |
| "loss": 0.3183, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 12.74, | |
| "learning_rate": 0.0007345238095238096, | |
| "loss": 0.3192, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 12.77, | |
| "learning_rate": 0.0007339285714285714, | |
| "loss": 0.3265, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 12.8, | |
| "learning_rate": 0.0007333333333333333, | |
| "loss": 0.3033, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 12.83, | |
| "learning_rate": 0.0007327380952380952, | |
| "loss": 0.313, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 12.86, | |
| "learning_rate": 0.0007321428571428571, | |
| "loss": 0.3186, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 0.000731547619047619, | |
| "loss": 0.3275, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 0.000730952380952381, | |
| "loss": 0.3272, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 12.94, | |
| "learning_rate": 0.0007303571428571428, | |
| "loss": 0.3336, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 12.97, | |
| "learning_rate": 0.0007297619047619047, | |
| "loss": 0.3244, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "learning_rate": 0.0007291666666666666, | |
| "loss": 0.3403, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 13.03, | |
| "learning_rate": 0.0007285714285714286, | |
| "loss": 0.1931, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 13.06, | |
| "learning_rate": 0.0007279761904761905, | |
| "loss": 0.2104, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 13.09, | |
| "learning_rate": 0.0007273809523809523, | |
| "loss": 0.2025, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 0.0007267857142857143, | |
| "loss": 0.1933, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 13.14, | |
| "learning_rate": 0.0007261904761904762, | |
| "loss": 0.2002, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 13.17, | |
| "learning_rate": 0.0007255952380952381, | |
| "loss": 0.2097, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 13.2, | |
| "learning_rate": 0.000725, | |
| "loss": 0.2097, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 13.23, | |
| "learning_rate": 0.0007244047619047619, | |
| "loss": 0.2061, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 13.26, | |
| "learning_rate": 0.0007238095238095238, | |
| "loss": 0.2, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 13.29, | |
| "learning_rate": 0.0007232142857142857, | |
| "loss": 0.2082, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 13.31, | |
| "learning_rate": 0.0007226190476190476, | |
| "loss": 0.209, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 13.34, | |
| "learning_rate": 0.0007220238095238096, | |
| "loss": 0.2083, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 13.37, | |
| "learning_rate": 0.0007214285714285714, | |
| "loss": 0.215, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 13.4, | |
| "learning_rate": 0.0007208333333333333, | |
| "loss": 0.2113, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 13.43, | |
| "learning_rate": 0.0007202380952380953, | |
| "loss": 0.2186, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 13.46, | |
| "learning_rate": 0.0007196428571428572, | |
| "loss": 0.2164, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 13.49, | |
| "learning_rate": 0.000719047619047619, | |
| "loss": 0.2197, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 13.51, | |
| "learning_rate": 0.0007184523809523809, | |
| "loss": 0.2256, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 13.54, | |
| "learning_rate": 0.0007178571428571429, | |
| "loss": 0.2165, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 13.57, | |
| "learning_rate": 0.0007172619047619048, | |
| "loss": 0.2227, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 13.6, | |
| "learning_rate": 0.0007166666666666667, | |
| "loss": 0.2245, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 13.63, | |
| "learning_rate": 0.0007160714285714285, | |
| "loss": 0.2239, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 13.66, | |
| "learning_rate": 0.0007154761904761905, | |
| "loss": 0.2251, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 13.69, | |
| "learning_rate": 0.0007148809523809524, | |
| "loss": 0.2305, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "learning_rate": 0.0007142857142857143, | |
| "loss": 0.2334, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 13.74, | |
| "learning_rate": 0.0007136904761904763, | |
| "loss": 0.229, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 13.77, | |
| "learning_rate": 0.0007130952380952381, | |
| "loss": 0.2363, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 13.8, | |
| "learning_rate": 0.0007125, | |
| "loss": 0.2382, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 13.83, | |
| "learning_rate": 0.0007119047619047619, | |
| "loss": 0.2374, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 13.86, | |
| "learning_rate": 0.0007113095238095239, | |
| "loss": 0.2433, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 13.89, | |
| "learning_rate": 0.0007107142857142858, | |
| "loss": 0.2349, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 13.91, | |
| "learning_rate": 0.0007101190476190476, | |
| "loss": 0.2549, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 13.94, | |
| "learning_rate": 0.0007095238095238095, | |
| "loss": 0.2481, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 13.97, | |
| "learning_rate": 0.0007089285714285715, | |
| "loss": 0.2504, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "learning_rate": 0.0007083333333333334, | |
| "loss": 0.2521, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 14.03, | |
| "learning_rate": 0.0007077380952380953, | |
| "loss": 0.1489, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 14.06, | |
| "learning_rate": 0.0007071428571428572, | |
| "loss": 0.15, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 14.09, | |
| "learning_rate": 0.0007065476190476191, | |
| "loss": 0.1499, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 14.11, | |
| "learning_rate": 0.000705952380952381, | |
| "loss": 0.1515, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 14.14, | |
| "learning_rate": 0.0007053571428571429, | |
| "loss": 0.1542, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 14.17, | |
| "learning_rate": 0.0007047619047619049, | |
| "loss": 0.1523, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 0.0007041666666666667, | |
| "loss": 0.1605, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 14.23, | |
| "learning_rate": 0.0007035714285714286, | |
| "loss": 0.1572, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 14.26, | |
| "learning_rate": 0.0007029761904761904, | |
| "loss": 0.1634, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 14.29, | |
| "learning_rate": 0.0007023809523809524, | |
| "loss": 0.1548, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 14.31, | |
| "learning_rate": 0.0007017857142857143, | |
| "loss": 0.1521, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 14.34, | |
| "learning_rate": 0.0007011904761904761, | |
| "loss": 0.1644, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 14.37, | |
| "learning_rate": 0.0007005952380952381, | |
| "loss": 0.155, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 14.4, | |
| "learning_rate": 0.0007, | |
| "loss": 0.1674, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 14.43, | |
| "learning_rate": 0.0006994047619047619, | |
| "loss": 0.1619, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 14.46, | |
| "learning_rate": 0.0006988095238095237, | |
| "loss": 0.1644, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 14.49, | |
| "learning_rate": 0.0006982142857142857, | |
| "loss": 0.1723, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 14.51, | |
| "learning_rate": 0.0006976190476190476, | |
| "loss": 0.1621, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 14.54, | |
| "learning_rate": 0.0006970238095238095, | |
| "loss": 0.1647, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 14.57, | |
| "learning_rate": 0.0006964285714285714, | |
| "loss": 0.1741, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 0.0006958333333333334, | |
| "loss": 0.1673, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 14.63, | |
| "learning_rate": 0.0006952380952380952, | |
| "loss": 0.1752, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 14.66, | |
| "learning_rate": 0.0006946428571428571, | |
| "loss": 0.167, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 14.69, | |
| "learning_rate": 0.0006940476190476191, | |
| "loss": 0.1718, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 14.71, | |
| "learning_rate": 0.000693452380952381, | |
| "loss": 0.1787, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 14.74, | |
| "learning_rate": 0.0006928571428571428, | |
| "loss": 0.1747, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 14.77, | |
| "learning_rate": 0.0006922619047619047, | |
| "loss": 0.1766, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 14.8, | |
| "learning_rate": 0.0006916666666666667, | |
| "loss": 0.1782, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 14.83, | |
| "learning_rate": 0.0006910714285714286, | |
| "loss": 0.1799, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 14.86, | |
| "learning_rate": 0.0006904761904761905, | |
| "loss": 0.169, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 14.89, | |
| "learning_rate": 0.0006898809523809523, | |
| "loss": 0.1802, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 14.91, | |
| "learning_rate": 0.0006892857142857143, | |
| "loss": 0.18, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 14.94, | |
| "learning_rate": 0.0006886904761904762, | |
| "loss": 0.1823, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 14.97, | |
| "learning_rate": 0.0006880952380952381, | |
| "loss": 0.1829, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "learning_rate": 0.0006875, | |
| "loss": 0.1885, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 15.03, | |
| "learning_rate": 0.000686904761904762, | |
| "loss": 0.1171, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 15.06, | |
| "learning_rate": 0.0006863095238095238, | |
| "loss": 0.1126, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 15.09, | |
| "learning_rate": 0.0006857142857142857, | |
| "loss": 0.1186, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 15.11, | |
| "learning_rate": 0.0006851190476190477, | |
| "loss": 0.1164, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 15.14, | |
| "learning_rate": 0.0006845238095238096, | |
| "loss": 0.1163, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 15.17, | |
| "learning_rate": 0.0006839285714285714, | |
| "loss": 0.1204, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 15.2, | |
| "learning_rate": 0.0006833333333333333, | |
| "loss": 0.1212, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 15.23, | |
| "learning_rate": 0.0006827380952380953, | |
| "loss": 0.1188, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 15.26, | |
| "learning_rate": 0.0006821428571428572, | |
| "loss": 0.1154, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 15.29, | |
| "learning_rate": 0.0006815476190476191, | |
| "loss": 0.1244, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 15.31, | |
| "learning_rate": 0.0006809523809523809, | |
| "loss": 0.1214, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 15.34, | |
| "learning_rate": 0.0006803571428571429, | |
| "loss": 0.1294, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 15.37, | |
| "learning_rate": 0.0006797619047619048, | |
| "loss": 0.1232, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 15.4, | |
| "learning_rate": 0.0006791666666666667, | |
| "loss": 0.1262, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 15.43, | |
| "learning_rate": 0.0006785714285714287, | |
| "loss": 0.1195, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 15.46, | |
| "learning_rate": 0.0006779761904761905, | |
| "loss": 0.123, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 15.49, | |
| "learning_rate": 0.0006773809523809524, | |
| "loss": 0.1266, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 15.51, | |
| "learning_rate": 0.0006767857142857143, | |
| "loss": 0.1345, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 15.54, | |
| "learning_rate": 0.0006761904761904763, | |
| "loss": 0.1174, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 15.57, | |
| "learning_rate": 0.0006755952380952382, | |
| "loss": 0.1293, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 15.6, | |
| "learning_rate": 0.000675, | |
| "loss": 0.1293, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 15.63, | |
| "learning_rate": 0.0006744047619047619, | |
| "loss": 0.1269, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 15.66, | |
| "learning_rate": 0.0006738095238095239, | |
| "loss": 0.1321, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 15.69, | |
| "learning_rate": 0.0006732142857142858, | |
| "loss": 0.1318, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 15.71, | |
| "learning_rate": 0.0006726190476190477, | |
| "loss": 0.1283, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 15.74, | |
| "learning_rate": 0.0006720238095238096, | |
| "loss": 0.128, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 15.77, | |
| "learning_rate": 0.0006714285714285714, | |
| "loss": 0.1295, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 15.8, | |
| "learning_rate": 0.0006708333333333333, | |
| "loss": 0.1323, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 15.83, | |
| "learning_rate": 0.0006702380952380952, | |
| "loss": 0.1348, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 15.86, | |
| "learning_rate": 0.0006696428571428571, | |
| "loss": 0.1276, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 15.89, | |
| "learning_rate": 0.000669047619047619, | |
| "loss": 0.1356, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 15.91, | |
| "learning_rate": 0.0006684523809523809, | |
| "loss": 0.1404, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 15.94, | |
| "learning_rate": 0.0006678571428571428, | |
| "loss": 0.1311, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 15.97, | |
| "learning_rate": 0.0006672619047619048, | |
| "loss": 0.1401, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 0.0006666666666666666, | |
| "loss": 0.1411, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 16.03, | |
| "learning_rate": 0.0006660714285714285, | |
| "loss": 0.0862, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 16.06, | |
| "learning_rate": 0.0006654761904761905, | |
| "loss": 0.0902, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 16.09, | |
| "learning_rate": 0.0006648809523809524, | |
| "loss": 0.0871, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 16.11, | |
| "learning_rate": 0.0006642857142857143, | |
| "loss": 0.0906, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 16.14, | |
| "learning_rate": 0.0006636904761904761, | |
| "loss": 0.0891, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 16.17, | |
| "learning_rate": 0.0006630952380952381, | |
| "loss": 0.0917, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 16.2, | |
| "learning_rate": 0.0006625, | |
| "loss": 0.0906, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 16.23, | |
| "learning_rate": 0.0006619047619047619, | |
| "loss": 0.0927, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 16.26, | |
| "learning_rate": 0.0006613095238095238, | |
| "loss": 0.0927, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 16.29, | |
| "learning_rate": 0.0006607142857142857, | |
| "loss": 0.0934, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 16.31, | |
| "learning_rate": 0.0006601190476190476, | |
| "loss": 0.0956, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 16.34, | |
| "learning_rate": 0.0006595238095238095, | |
| "loss": 0.0933, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 16.37, | |
| "learning_rate": 0.0006589285714285715, | |
| "loss": 0.0993, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 0.0006583333333333334, | |
| "loss": 0.095, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 16.43, | |
| "learning_rate": 0.0006577380952380952, | |
| "loss": 0.0963, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 16.46, | |
| "learning_rate": 0.0006571428571428571, | |
| "loss": 0.0948, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 16.49, | |
| "learning_rate": 0.0006565476190476191, | |
| "loss": 0.0952, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 16.51, | |
| "learning_rate": 0.000655952380952381, | |
| "loss": 0.1001, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 16.54, | |
| "learning_rate": 0.0006553571428571429, | |
| "loss": 0.0924, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 16.57, | |
| "learning_rate": 0.0006547619047619047, | |
| "loss": 0.0962, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 16.6, | |
| "learning_rate": 0.0006541666666666667, | |
| "loss": 0.0949, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 16.63, | |
| "learning_rate": 0.0006535714285714286, | |
| "loss": 0.1, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 16.66, | |
| "learning_rate": 0.0006529761904761905, | |
| "loss": 0.1009, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 16.69, | |
| "learning_rate": 0.0006523809523809525, | |
| "loss": 0.1023, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 16.71, | |
| "learning_rate": 0.0006517857142857143, | |
| "loss": 0.0995, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 16.74, | |
| "learning_rate": 0.0006511904761904762, | |
| "loss": 0.1015, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 16.77, | |
| "learning_rate": 0.0006505952380952381, | |
| "loss": 0.0966, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 16.8, | |
| "learning_rate": 0.0006500000000000001, | |
| "loss": 0.1019, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 16.83, | |
| "learning_rate": 0.000649404761904762, | |
| "loss": 0.0996, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 16.86, | |
| "learning_rate": 0.0006488095238095238, | |
| "loss": 0.103, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 16.89, | |
| "learning_rate": 0.0006482142857142857, | |
| "loss": 0.1042, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 16.91, | |
| "learning_rate": 0.0006476190476190477, | |
| "loss": 0.1039, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 16.94, | |
| "learning_rate": 0.0006470238095238096, | |
| "loss": 0.1058, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 16.97, | |
| "learning_rate": 0.0006464285714285715, | |
| "loss": 0.0994, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 0.0006458333333333334, | |
| "loss": 0.1062, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 17.03, | |
| "learning_rate": 0.0006452380952380953, | |
| "loss": 0.0709, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 17.06, | |
| "learning_rate": 0.0006446428571428572, | |
| "loss": 0.0733, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 17.09, | |
| "learning_rate": 0.0006440476190476191, | |
| "loss": 0.0724, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 17.11, | |
| "learning_rate": 0.0006434523809523811, | |
| "loss": 0.0733, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 17.14, | |
| "learning_rate": 0.0006428571428571429, | |
| "loss": 0.0741, | |
| "step": 600 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1680, | |
| "num_train_epochs": 48, | |
| "save_steps": 100, | |
| "total_flos": 3.463615849187021e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |