{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 0, "global_step": 1594, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012547051442910915, "grad_norm": 0.498046875, "learning_rate": 9.993726474278545e-06, "loss": 1.942, "step": 1 }, { "epoch": 0.002509410288582183, "grad_norm": 0.53125, "learning_rate": 9.98745294855709e-06, "loss": 1.9052, "step": 2 }, { "epoch": 0.0037641154328732747, "grad_norm": 0.474609375, "learning_rate": 9.981179422835635e-06, "loss": 1.84, "step": 3 }, { "epoch": 0.005018820577164366, "grad_norm": 0.4765625, "learning_rate": 9.974905897114179e-06, "loss": 1.8541, "step": 4 }, { "epoch": 0.006273525721455458, "grad_norm": 0.474609375, "learning_rate": 9.968632371392723e-06, "loss": 1.9384, "step": 5 }, { "epoch": 0.0075282308657465494, "grad_norm": 0.43359375, "learning_rate": 9.962358845671269e-06, "loss": 1.8467, "step": 6 }, { "epoch": 0.00878293601003764, "grad_norm": 0.443359375, "learning_rate": 9.956085319949813e-06, "loss": 1.8768, "step": 7 }, { "epoch": 0.010037641154328732, "grad_norm": 0.423828125, "learning_rate": 9.949811794228357e-06, "loss": 1.9057, "step": 8 }, { "epoch": 0.011292346298619825, "grad_norm": 0.400390625, "learning_rate": 9.943538268506901e-06, "loss": 1.8385, "step": 9 }, { "epoch": 0.012547051442910916, "grad_norm": 0.431640625, "learning_rate": 9.937264742785447e-06, "loss": 1.7961, "step": 10 }, { "epoch": 0.013801756587202008, "grad_norm": 0.34765625, "learning_rate": 9.930991217063991e-06, "loss": 1.6866, "step": 11 }, { "epoch": 0.015056461731493099, "grad_norm": 0.375, "learning_rate": 9.924717691342535e-06, "loss": 1.7956, "step": 12 }, { "epoch": 0.01631116687578419, "grad_norm": 0.375, "learning_rate": 9.918444165621079e-06, "loss": 1.7926, "step": 13 }, { "epoch": 0.01756587202007528, "grad_norm": 0.337890625, "learning_rate": 9.912170639899625e-06, "loss": 1.7063, "step": 14 }, { "epoch": 0.018820577164366373, "grad_norm": 0.341796875, "learning_rate": 9.905897114178169e-06, "loss": 1.6948, "step": 15 }, { "epoch": 0.020075282308657464, "grad_norm": 0.333984375, "learning_rate": 9.899623588456713e-06, "loss": 1.675, "step": 16 }, { "epoch": 0.02132998745294856, "grad_norm": 0.326171875, "learning_rate": 9.893350062735259e-06, "loss": 1.6552, "step": 17 }, { "epoch": 0.02258469259723965, "grad_norm": 0.34375, "learning_rate": 9.887076537013803e-06, "loss": 1.7244, "step": 18 }, { "epoch": 0.02383939774153074, "grad_norm": 0.33203125, "learning_rate": 9.880803011292347e-06, "loss": 1.682, "step": 19 }, { "epoch": 0.025094102885821833, "grad_norm": 0.306640625, "learning_rate": 9.874529485570891e-06, "loss": 1.6312, "step": 20 }, { "epoch": 0.026348808030112924, "grad_norm": 0.310546875, "learning_rate": 9.868255959849437e-06, "loss": 1.568, "step": 21 }, { "epoch": 0.027603513174404015, "grad_norm": 0.279296875, "learning_rate": 9.861982434127981e-06, "loss": 1.5441, "step": 22 }, { "epoch": 0.028858218318695106, "grad_norm": 0.2734375, "learning_rate": 9.855708908406525e-06, "loss": 1.6236, "step": 23 }, { "epoch": 0.030112923462986198, "grad_norm": 0.259765625, "learning_rate": 9.849435382685069e-06, "loss": 1.5282, "step": 24 }, { "epoch": 0.03136762860727729, "grad_norm": 0.2431640625, "learning_rate": 9.843161856963615e-06, "loss": 1.5391, "step": 25 }, { "epoch": 0.03262233375156838, "grad_norm": 0.263671875, "learning_rate": 9.836888331242159e-06, "loss": 1.511, "step": 26 }, { "epoch": 0.033877038895859475, "grad_norm": 0.2421875, "learning_rate": 9.830614805520703e-06, "loss": 1.4423, "step": 27 }, { "epoch": 0.03513174404015056, "grad_norm": 0.236328125, "learning_rate": 9.824341279799247e-06, "loss": 1.4665, "step": 28 }, { "epoch": 0.03638644918444166, "grad_norm": 0.248046875, "learning_rate": 9.818067754077793e-06, "loss": 1.5649, "step": 29 }, { "epoch": 0.037641154328732745, "grad_norm": 0.2373046875, "learning_rate": 9.811794228356337e-06, "loss": 1.4689, "step": 30 }, { "epoch": 0.03889585947302384, "grad_norm": 0.228515625, "learning_rate": 9.805520702634881e-06, "loss": 1.4904, "step": 31 }, { "epoch": 0.04015056461731493, "grad_norm": 0.2578125, "learning_rate": 9.799247176913425e-06, "loss": 1.4185, "step": 32 }, { "epoch": 0.04140526976160602, "grad_norm": 0.232421875, "learning_rate": 9.792973651191971e-06, "loss": 1.4938, "step": 33 }, { "epoch": 0.04265997490589712, "grad_norm": 0.212890625, "learning_rate": 9.786700125470515e-06, "loss": 1.4119, "step": 34 }, { "epoch": 0.043914680050188205, "grad_norm": 0.2119140625, "learning_rate": 9.78042659974906e-06, "loss": 1.4329, "step": 35 }, { "epoch": 0.0451693851944793, "grad_norm": 0.197265625, "learning_rate": 9.774153074027605e-06, "loss": 1.3964, "step": 36 }, { "epoch": 0.04642409033877039, "grad_norm": 0.2890625, "learning_rate": 9.767879548306149e-06, "loss": 1.4299, "step": 37 }, { "epoch": 0.04767879548306148, "grad_norm": 0.19921875, "learning_rate": 9.761606022584693e-06, "loss": 1.4048, "step": 38 }, { "epoch": 0.04893350062735257, "grad_norm": 0.2080078125, "learning_rate": 9.755332496863237e-06, "loss": 1.4785, "step": 39 }, { "epoch": 0.050188205771643665, "grad_norm": 0.19921875, "learning_rate": 9.749058971141783e-06, "loss": 1.406, "step": 40 }, { "epoch": 0.05144291091593475, "grad_norm": 0.392578125, "learning_rate": 9.742785445420327e-06, "loss": 1.4591, "step": 41 }, { "epoch": 0.05269761606022585, "grad_norm": 0.208984375, "learning_rate": 9.736511919698871e-06, "loss": 1.3938, "step": 42 }, { "epoch": 0.053952321204516936, "grad_norm": 0.1748046875, "learning_rate": 9.730238393977415e-06, "loss": 1.3415, "step": 43 }, { "epoch": 0.05520702634880803, "grad_norm": 0.1943359375, "learning_rate": 9.723964868255961e-06, "loss": 1.3972, "step": 44 }, { "epoch": 0.056461731493099125, "grad_norm": 0.1796875, "learning_rate": 9.717691342534505e-06, "loss": 1.3859, "step": 45 }, { "epoch": 0.05771643663739021, "grad_norm": 0.1748046875, "learning_rate": 9.711417816813051e-06, "loss": 1.3514, "step": 46 }, { "epoch": 0.05897114178168131, "grad_norm": 0.1767578125, "learning_rate": 9.705144291091593e-06, "loss": 1.3994, "step": 47 }, { "epoch": 0.060225846925972396, "grad_norm": 0.177734375, "learning_rate": 9.69887076537014e-06, "loss": 1.393, "step": 48 }, { "epoch": 0.06148055207026349, "grad_norm": 0.1689453125, "learning_rate": 9.692597239648683e-06, "loss": 1.3489, "step": 49 }, { "epoch": 0.06273525721455459, "grad_norm": 0.1728515625, "learning_rate": 9.686323713927227e-06, "loss": 1.2711, "step": 50 }, { "epoch": 0.06398996235884567, "grad_norm": 0.169921875, "learning_rate": 9.680050188205773e-06, "loss": 1.337, "step": 51 }, { "epoch": 0.06524466750313676, "grad_norm": 0.2021484375, "learning_rate": 9.673776662484317e-06, "loss": 1.3172, "step": 52 }, { "epoch": 0.06649937264742785, "grad_norm": 0.2158203125, "learning_rate": 9.667503136762861e-06, "loss": 1.2469, "step": 53 }, { "epoch": 0.06775407779171895, "grad_norm": 0.171875, "learning_rate": 9.661229611041405e-06, "loss": 1.3057, "step": 54 }, { "epoch": 0.06900878293601004, "grad_norm": 0.1552734375, "learning_rate": 9.654956085319951e-06, "loss": 1.2991, "step": 55 }, { "epoch": 0.07026348808030113, "grad_norm": 0.18359375, "learning_rate": 9.648682559598495e-06, "loss": 1.2801, "step": 56 }, { "epoch": 0.07151819322459223, "grad_norm": 0.185546875, "learning_rate": 9.64240903387704e-06, "loss": 1.3001, "step": 57 }, { "epoch": 0.07277289836888332, "grad_norm": 0.17578125, "learning_rate": 9.636135508155583e-06, "loss": 1.3372, "step": 58 }, { "epoch": 0.0740276035131744, "grad_norm": 0.2060546875, "learning_rate": 9.62986198243413e-06, "loss": 1.3672, "step": 59 }, { "epoch": 0.07528230865746549, "grad_norm": 0.1767578125, "learning_rate": 9.623588456712673e-06, "loss": 1.2997, "step": 60 }, { "epoch": 0.07653701380175659, "grad_norm": 0.16015625, "learning_rate": 9.617314930991219e-06, "loss": 1.2769, "step": 61 }, { "epoch": 0.07779171894604768, "grad_norm": 0.162109375, "learning_rate": 9.611041405269761e-06, "loss": 1.3084, "step": 62 }, { "epoch": 0.07904642409033877, "grad_norm": 0.197265625, "learning_rate": 9.604767879548307e-06, "loss": 1.3025, "step": 63 }, { "epoch": 0.08030112923462986, "grad_norm": 0.169921875, "learning_rate": 9.598494353826851e-06, "loss": 1.3005, "step": 64 }, { "epoch": 0.08155583437892096, "grad_norm": 0.2001953125, "learning_rate": 9.592220828105397e-06, "loss": 1.298, "step": 65 }, { "epoch": 0.08281053952321205, "grad_norm": 0.154296875, "learning_rate": 9.585947302383941e-06, "loss": 1.2837, "step": 66 }, { "epoch": 0.08406524466750313, "grad_norm": 0.1533203125, "learning_rate": 9.579673776662485e-06, "loss": 1.2571, "step": 67 }, { "epoch": 0.08531994981179424, "grad_norm": 0.251953125, "learning_rate": 9.57340025094103e-06, "loss": 1.3062, "step": 68 }, { "epoch": 0.08657465495608532, "grad_norm": 0.158203125, "learning_rate": 9.567126725219574e-06, "loss": 1.2688, "step": 69 }, { "epoch": 0.08782936010037641, "grad_norm": 0.1611328125, "learning_rate": 9.56085319949812e-06, "loss": 1.2268, "step": 70 }, { "epoch": 0.0890840652446675, "grad_norm": 0.1494140625, "learning_rate": 9.554579673776663e-06, "loss": 1.2604, "step": 71 }, { "epoch": 0.0903387703889586, "grad_norm": 0.171875, "learning_rate": 9.548306148055207e-06, "loss": 1.3662, "step": 72 }, { "epoch": 0.09159347553324969, "grad_norm": 0.1533203125, "learning_rate": 9.542032622333752e-06, "loss": 1.2705, "step": 73 }, { "epoch": 0.09284818067754078, "grad_norm": 0.158203125, "learning_rate": 9.535759096612297e-06, "loss": 1.2313, "step": 74 }, { "epoch": 0.09410288582183186, "grad_norm": 0.1650390625, "learning_rate": 9.529485570890841e-06, "loss": 1.325, "step": 75 }, { "epoch": 0.09535759096612297, "grad_norm": 0.15625, "learning_rate": 9.523212045169386e-06, "loss": 1.2819, "step": 76 }, { "epoch": 0.09661229611041405, "grad_norm": 0.1533203125, "learning_rate": 9.51693851944793e-06, "loss": 1.2636, "step": 77 }, { "epoch": 0.09786700125470514, "grad_norm": 0.1591796875, "learning_rate": 9.510664993726475e-06, "loss": 1.2181, "step": 78 }, { "epoch": 0.09912170639899624, "grad_norm": 0.15234375, "learning_rate": 9.50439146800502e-06, "loss": 1.1886, "step": 79 }, { "epoch": 0.10037641154328733, "grad_norm": 0.158203125, "learning_rate": 9.498117942283565e-06, "loss": 1.2199, "step": 80 }, { "epoch": 0.10163111668757842, "grad_norm": 0.177734375, "learning_rate": 9.491844416562108e-06, "loss": 1.2421, "step": 81 }, { "epoch": 0.1028858218318695, "grad_norm": 0.1689453125, "learning_rate": 9.485570890840653e-06, "loss": 1.2717, "step": 82 }, { "epoch": 0.10414052697616061, "grad_norm": 0.1484375, "learning_rate": 9.479297365119198e-06, "loss": 1.2014, "step": 83 }, { "epoch": 0.1053952321204517, "grad_norm": 0.19921875, "learning_rate": 9.473023839397743e-06, "loss": 1.242, "step": 84 }, { "epoch": 0.10664993726474278, "grad_norm": 0.15234375, "learning_rate": 9.466750313676287e-06, "loss": 1.2134, "step": 85 }, { "epoch": 0.10790464240903387, "grad_norm": 0.1455078125, "learning_rate": 9.460476787954832e-06, "loss": 1.2176, "step": 86 }, { "epoch": 0.10915934755332497, "grad_norm": 0.16015625, "learning_rate": 9.454203262233376e-06, "loss": 1.2264, "step": 87 }, { "epoch": 0.11041405269761606, "grad_norm": 0.173828125, "learning_rate": 9.44792973651192e-06, "loss": 1.1823, "step": 88 }, { "epoch": 0.11166875784190715, "grad_norm": 0.189453125, "learning_rate": 9.441656210790465e-06, "loss": 1.1788, "step": 89 }, { "epoch": 0.11292346298619825, "grad_norm": 0.15234375, "learning_rate": 9.43538268506901e-06, "loss": 1.2218, "step": 90 }, { "epoch": 0.11417816813048934, "grad_norm": 0.173828125, "learning_rate": 9.429109159347554e-06, "loss": 1.2314, "step": 91 }, { "epoch": 0.11543287327478043, "grad_norm": 0.2001953125, "learning_rate": 9.422835633626098e-06, "loss": 1.2038, "step": 92 }, { "epoch": 0.11668757841907151, "grad_norm": 0.27734375, "learning_rate": 9.416562107904644e-06, "loss": 1.2453, "step": 93 }, { "epoch": 0.11794228356336262, "grad_norm": 0.2099609375, "learning_rate": 9.410288582183188e-06, "loss": 1.2143, "step": 94 }, { "epoch": 0.1191969887076537, "grad_norm": 0.1767578125, "learning_rate": 9.404015056461733e-06, "loss": 1.252, "step": 95 }, { "epoch": 0.12045169385194479, "grad_norm": 0.1650390625, "learning_rate": 9.397741530740276e-06, "loss": 1.2321, "step": 96 }, { "epoch": 0.12170639899623588, "grad_norm": 0.171875, "learning_rate": 9.391468005018822e-06, "loss": 1.1837, "step": 97 }, { "epoch": 0.12296110414052698, "grad_norm": 0.16015625, "learning_rate": 9.385194479297366e-06, "loss": 1.2344, "step": 98 }, { "epoch": 0.12421580928481807, "grad_norm": 0.1630859375, "learning_rate": 9.378920953575911e-06, "loss": 1.226, "step": 99 }, { "epoch": 0.12547051442910917, "grad_norm": 0.1845703125, "learning_rate": 9.372647427854456e-06, "loss": 1.256, "step": 100 }, { "epoch": 0.12672521957340024, "grad_norm": 0.20703125, "learning_rate": 9.366373902133e-06, "loss": 1.2032, "step": 101 }, { "epoch": 0.12797992471769135, "grad_norm": 0.177734375, "learning_rate": 9.360100376411544e-06, "loss": 1.2402, "step": 102 }, { "epoch": 0.12923462986198245, "grad_norm": 0.1650390625, "learning_rate": 9.35382685069009e-06, "loss": 1.1702, "step": 103 }, { "epoch": 0.13048933500627352, "grad_norm": 0.1689453125, "learning_rate": 9.347553324968634e-06, "loss": 1.1564, "step": 104 }, { "epoch": 0.13174404015056462, "grad_norm": 0.166015625, "learning_rate": 9.341279799247178e-06, "loss": 1.1618, "step": 105 }, { "epoch": 0.1329987452948557, "grad_norm": 0.1708984375, "learning_rate": 9.335006273525722e-06, "loss": 1.1585, "step": 106 }, { "epoch": 0.1342534504391468, "grad_norm": 0.6171875, "learning_rate": 9.328732747804266e-06, "loss": 1.2589, "step": 107 }, { "epoch": 0.1355081555834379, "grad_norm": 0.275390625, "learning_rate": 9.322459222082812e-06, "loss": 1.1961, "step": 108 }, { "epoch": 0.13676286072772897, "grad_norm": 0.1630859375, "learning_rate": 9.316185696361356e-06, "loss": 1.1551, "step": 109 }, { "epoch": 0.13801756587202008, "grad_norm": 0.1787109375, "learning_rate": 9.309912170639902e-06, "loss": 1.1134, "step": 110 }, { "epoch": 0.13927227101631118, "grad_norm": 0.158203125, "learning_rate": 9.303638644918444e-06, "loss": 1.1523, "step": 111 }, { "epoch": 0.14052697616060225, "grad_norm": 0.2099609375, "learning_rate": 9.29736511919699e-06, "loss": 1.1655, "step": 112 }, { "epoch": 0.14178168130489335, "grad_norm": 0.1748046875, "learning_rate": 9.291091593475534e-06, "loss": 1.1678, "step": 113 }, { "epoch": 0.14303638644918445, "grad_norm": 0.166015625, "learning_rate": 9.28481806775408e-06, "loss": 1.1475, "step": 114 }, { "epoch": 0.14429109159347553, "grad_norm": 0.1630859375, "learning_rate": 9.278544542032622e-06, "loss": 1.1367, "step": 115 }, { "epoch": 0.14554579673776663, "grad_norm": 0.158203125, "learning_rate": 9.272271016311168e-06, "loss": 1.2162, "step": 116 }, { "epoch": 0.1468005018820577, "grad_norm": 0.1748046875, "learning_rate": 9.265997490589712e-06, "loss": 1.1494, "step": 117 }, { "epoch": 0.1480552070263488, "grad_norm": 0.169921875, "learning_rate": 9.259723964868258e-06, "loss": 1.1306, "step": 118 }, { "epoch": 0.1493099121706399, "grad_norm": 0.1611328125, "learning_rate": 9.253450439146802e-06, "loss": 1.173, "step": 119 }, { "epoch": 0.15056461731493098, "grad_norm": 0.2392578125, "learning_rate": 9.247176913425346e-06, "loss": 1.1578, "step": 120 }, { "epoch": 0.15181932245922208, "grad_norm": 0.2060546875, "learning_rate": 9.24090338770389e-06, "loss": 1.1069, "step": 121 }, { "epoch": 0.15307402760351319, "grad_norm": 0.1611328125, "learning_rate": 9.234629861982434e-06, "loss": 1.1782, "step": 122 }, { "epoch": 0.15432873274780426, "grad_norm": 0.169921875, "learning_rate": 9.22835633626098e-06, "loss": 1.2029, "step": 123 }, { "epoch": 0.15558343789209536, "grad_norm": 0.17578125, "learning_rate": 9.222082810539524e-06, "loss": 1.1524, "step": 124 }, { "epoch": 0.15683814303638646, "grad_norm": 0.1982421875, "learning_rate": 9.215809284818068e-06, "loss": 1.1127, "step": 125 }, { "epoch": 0.15809284818067754, "grad_norm": 0.18359375, "learning_rate": 9.209535759096612e-06, "loss": 1.1433, "step": 126 }, { "epoch": 0.15934755332496864, "grad_norm": 0.169921875, "learning_rate": 9.203262233375158e-06, "loss": 1.1837, "step": 127 }, { "epoch": 0.1606022584692597, "grad_norm": 0.189453125, "learning_rate": 9.196988707653702e-06, "loss": 1.1743, "step": 128 }, { "epoch": 0.1618569636135508, "grad_norm": 0.1806640625, "learning_rate": 9.190715181932248e-06, "loss": 1.1444, "step": 129 }, { "epoch": 0.16311166875784192, "grad_norm": 0.181640625, "learning_rate": 9.18444165621079e-06, "loss": 1.1661, "step": 130 }, { "epoch": 0.164366373902133, "grad_norm": 0.171875, "learning_rate": 9.178168130489336e-06, "loss": 1.1499, "step": 131 }, { "epoch": 0.1656210790464241, "grad_norm": 0.1943359375, "learning_rate": 9.17189460476788e-06, "loss": 1.1517, "step": 132 }, { "epoch": 0.1668757841907152, "grad_norm": 0.2080078125, "learning_rate": 9.165621079046426e-06, "loss": 1.154, "step": 133 }, { "epoch": 0.16813048933500627, "grad_norm": 0.1689453125, "learning_rate": 9.15934755332497e-06, "loss": 1.1418, "step": 134 }, { "epoch": 0.16938519447929737, "grad_norm": 0.1689453125, "learning_rate": 9.153074027603514e-06, "loss": 1.1471, "step": 135 }, { "epoch": 0.17063989962358847, "grad_norm": 0.171875, "learning_rate": 9.146800501882058e-06, "loss": 1.1473, "step": 136 }, { "epoch": 0.17189460476787954, "grad_norm": 0.1865234375, "learning_rate": 9.140526976160604e-06, "loss": 1.1212, "step": 137 }, { "epoch": 0.17314930991217065, "grad_norm": 0.16796875, "learning_rate": 9.134253450439148e-06, "loss": 1.1714, "step": 138 }, { "epoch": 0.17440401505646172, "grad_norm": 0.1787109375, "learning_rate": 9.127979924717692e-06, "loss": 1.2051, "step": 139 }, { "epoch": 0.17565872020075282, "grad_norm": 0.197265625, "learning_rate": 9.121706398996236e-06, "loss": 1.1596, "step": 140 }, { "epoch": 0.17691342534504392, "grad_norm": 0.18359375, "learning_rate": 9.11543287327478e-06, "loss": 1.1252, "step": 141 }, { "epoch": 0.178168130489335, "grad_norm": 0.1611328125, "learning_rate": 9.109159347553326e-06, "loss": 1.1467, "step": 142 }, { "epoch": 0.1794228356336261, "grad_norm": 0.1904296875, "learning_rate": 9.10288582183187e-06, "loss": 1.1456, "step": 143 }, { "epoch": 0.1806775407779172, "grad_norm": 0.171875, "learning_rate": 9.096612296110416e-06, "loss": 1.132, "step": 144 }, { "epoch": 0.18193224592220827, "grad_norm": 0.169921875, "learning_rate": 9.090338770388958e-06, "loss": 1.1319, "step": 145 }, { "epoch": 0.18318695106649938, "grad_norm": 0.1650390625, "learning_rate": 9.084065244667504e-06, "loss": 1.1503, "step": 146 }, { "epoch": 0.18444165621079048, "grad_norm": 0.185546875, "learning_rate": 9.077791718946048e-06, "loss": 1.1487, "step": 147 }, { "epoch": 0.18569636135508155, "grad_norm": 0.1689453125, "learning_rate": 9.071518193224594e-06, "loss": 1.134, "step": 148 }, { "epoch": 0.18695106649937265, "grad_norm": 0.1640625, "learning_rate": 9.065244667503138e-06, "loss": 1.1319, "step": 149 }, { "epoch": 0.18820577164366373, "grad_norm": 0.1962890625, "learning_rate": 9.058971141781682e-06, "loss": 1.1466, "step": 150 }, { "epoch": 0.18946047678795483, "grad_norm": 0.1943359375, "learning_rate": 9.052697616060226e-06, "loss": 1.1513, "step": 151 }, { "epoch": 0.19071518193224593, "grad_norm": 0.1640625, "learning_rate": 9.046424090338772e-06, "loss": 1.1264, "step": 152 }, { "epoch": 0.191969887076537, "grad_norm": 0.1748046875, "learning_rate": 9.040150564617316e-06, "loss": 1.117, "step": 153 }, { "epoch": 0.1932245922208281, "grad_norm": 0.1806640625, "learning_rate": 9.03387703889586e-06, "loss": 1.1816, "step": 154 }, { "epoch": 0.1944792973651192, "grad_norm": 0.1796875, "learning_rate": 9.027603513174404e-06, "loss": 1.1734, "step": 155 }, { "epoch": 0.19573400250941028, "grad_norm": 0.1630859375, "learning_rate": 9.02132998745295e-06, "loss": 1.1002, "step": 156 }, { "epoch": 0.19698870765370138, "grad_norm": 0.1845703125, "learning_rate": 9.015056461731494e-06, "loss": 1.1671, "step": 157 }, { "epoch": 0.19824341279799249, "grad_norm": 0.1767578125, "learning_rate": 9.008782936010038e-06, "loss": 1.0889, "step": 158 }, { "epoch": 0.19949811794228356, "grad_norm": 0.1689453125, "learning_rate": 9.002509410288582e-06, "loss": 1.1309, "step": 159 }, { "epoch": 0.20075282308657466, "grad_norm": 0.1728515625, "learning_rate": 8.996235884567126e-06, "loss": 1.1929, "step": 160 }, { "epoch": 0.20200752823086573, "grad_norm": 0.1669921875, "learning_rate": 8.989962358845672e-06, "loss": 1.1352, "step": 161 }, { "epoch": 0.20326223337515684, "grad_norm": 0.1787109375, "learning_rate": 8.983688833124216e-06, "loss": 1.0916, "step": 162 }, { "epoch": 0.20451693851944794, "grad_norm": 0.1923828125, "learning_rate": 8.977415307402762e-06, "loss": 1.1583, "step": 163 }, { "epoch": 0.205771643663739, "grad_norm": 0.1728515625, "learning_rate": 8.971141781681304e-06, "loss": 1.0858, "step": 164 }, { "epoch": 0.20702634880803011, "grad_norm": 0.1689453125, "learning_rate": 8.96486825595985e-06, "loss": 1.1407, "step": 165 }, { "epoch": 0.20828105395232122, "grad_norm": 0.1923828125, "learning_rate": 8.958594730238394e-06, "loss": 1.1098, "step": 166 }, { "epoch": 0.2095357590966123, "grad_norm": 0.1630859375, "learning_rate": 8.95232120451694e-06, "loss": 1.1061, "step": 167 }, { "epoch": 0.2107904642409034, "grad_norm": 0.1728515625, "learning_rate": 8.946047678795484e-06, "loss": 1.1356, "step": 168 }, { "epoch": 0.2120451693851945, "grad_norm": 0.287109375, "learning_rate": 8.939774153074028e-06, "loss": 1.1593, "step": 169 }, { "epoch": 0.21329987452948557, "grad_norm": 0.32421875, "learning_rate": 8.933500627352572e-06, "loss": 1.1497, "step": 170 }, { "epoch": 0.21455457967377667, "grad_norm": 0.1875, "learning_rate": 8.927227101631118e-06, "loss": 1.0603, "step": 171 }, { "epoch": 0.21580928481806774, "grad_norm": 0.171875, "learning_rate": 8.920953575909662e-06, "loss": 1.0541, "step": 172 }, { "epoch": 0.21706398996235884, "grad_norm": 0.1884765625, "learning_rate": 8.914680050188206e-06, "loss": 1.1198, "step": 173 }, { "epoch": 0.21831869510664995, "grad_norm": 0.185546875, "learning_rate": 8.90840652446675e-06, "loss": 1.1629, "step": 174 }, { "epoch": 0.21957340025094102, "grad_norm": 0.173828125, "learning_rate": 8.902132998745296e-06, "loss": 1.1691, "step": 175 }, { "epoch": 0.22082810539523212, "grad_norm": 0.1796875, "learning_rate": 8.89585947302384e-06, "loss": 1.1136, "step": 176 }, { "epoch": 0.22208281053952322, "grad_norm": 0.17578125, "learning_rate": 8.889585947302384e-06, "loss": 1.1333, "step": 177 }, { "epoch": 0.2233375156838143, "grad_norm": 0.1650390625, "learning_rate": 8.88331242158093e-06, "loss": 1.1083, "step": 178 }, { "epoch": 0.2245922208281054, "grad_norm": 0.181640625, "learning_rate": 8.877038895859473e-06, "loss": 1.1126, "step": 179 }, { "epoch": 0.2258469259723965, "grad_norm": 0.17578125, "learning_rate": 8.870765370138018e-06, "loss": 1.1364, "step": 180 }, { "epoch": 0.22710163111668757, "grad_norm": 0.1865234375, "learning_rate": 8.864491844416562e-06, "loss": 1.1168, "step": 181 }, { "epoch": 0.22835633626097868, "grad_norm": 0.1708984375, "learning_rate": 8.858218318695108e-06, "loss": 1.1352, "step": 182 }, { "epoch": 0.22961104140526975, "grad_norm": 0.1767578125, "learning_rate": 8.851944792973652e-06, "loss": 1.1343, "step": 183 }, { "epoch": 0.23086574654956085, "grad_norm": 0.1904296875, "learning_rate": 8.845671267252196e-06, "loss": 1.0849, "step": 184 }, { "epoch": 0.23212045169385195, "grad_norm": 0.177734375, "learning_rate": 8.83939774153074e-06, "loss": 1.1027, "step": 185 }, { "epoch": 0.23337515683814303, "grad_norm": 0.23046875, "learning_rate": 8.833124215809286e-06, "loss": 1.0837, "step": 186 }, { "epoch": 0.23462986198243413, "grad_norm": 0.177734375, "learning_rate": 8.82685069008783e-06, "loss": 1.1269, "step": 187 }, { "epoch": 0.23588456712672523, "grad_norm": 0.1796875, "learning_rate": 8.820577164366374e-06, "loss": 1.082, "step": 188 }, { "epoch": 0.2371392722710163, "grad_norm": 0.1943359375, "learning_rate": 8.814303638644919e-06, "loss": 1.0682, "step": 189 }, { "epoch": 0.2383939774153074, "grad_norm": 0.2001953125, "learning_rate": 8.808030112923464e-06, "loss": 1.0743, "step": 190 }, { "epoch": 0.2396486825595985, "grad_norm": 0.1767578125, "learning_rate": 8.801756587202008e-06, "loss": 1.1165, "step": 191 }, { "epoch": 0.24090338770388958, "grad_norm": 0.169921875, "learning_rate": 8.795483061480552e-06, "loss": 1.1129, "step": 192 }, { "epoch": 0.24215809284818068, "grad_norm": 0.1982421875, "learning_rate": 8.789209535759098e-06, "loss": 1.1087, "step": 193 }, { "epoch": 0.24341279799247176, "grad_norm": 0.2890625, "learning_rate": 8.782936010037642e-06, "loss": 1.1832, "step": 194 }, { "epoch": 0.24466750313676286, "grad_norm": 0.2353515625, "learning_rate": 8.776662484316186e-06, "loss": 1.1368, "step": 195 }, { "epoch": 0.24592220828105396, "grad_norm": 0.2021484375, "learning_rate": 8.77038895859473e-06, "loss": 1.0946, "step": 196 }, { "epoch": 0.24717691342534504, "grad_norm": 0.1796875, "learning_rate": 8.764115432873276e-06, "loss": 1.0906, "step": 197 }, { "epoch": 0.24843161856963614, "grad_norm": 0.1826171875, "learning_rate": 8.75784190715182e-06, "loss": 1.076, "step": 198 }, { "epoch": 0.24968632371392724, "grad_norm": 0.185546875, "learning_rate": 8.751568381430364e-06, "loss": 1.0663, "step": 199 }, { "epoch": 0.25094102885821834, "grad_norm": 0.173828125, "learning_rate": 8.745294855708909e-06, "loss": 1.12, "step": 200 }, { "epoch": 0.2521957340025094, "grad_norm": 0.177734375, "learning_rate": 8.739021329987454e-06, "loss": 1.0858, "step": 201 }, { "epoch": 0.2534504391468005, "grad_norm": 0.1943359375, "learning_rate": 8.732747804265998e-06, "loss": 1.0915, "step": 202 }, { "epoch": 0.2547051442910916, "grad_norm": 0.1806640625, "learning_rate": 8.726474278544543e-06, "loss": 1.0664, "step": 203 }, { "epoch": 0.2559598494353827, "grad_norm": 0.1845703125, "learning_rate": 8.720200752823087e-06, "loss": 1.0621, "step": 204 }, { "epoch": 0.2572145545796738, "grad_norm": 0.185546875, "learning_rate": 8.713927227101632e-06, "loss": 1.0917, "step": 205 }, { "epoch": 0.2584692597239649, "grad_norm": 0.193359375, "learning_rate": 8.707653701380176e-06, "loss": 1.1201, "step": 206 }, { "epoch": 0.25972396486825594, "grad_norm": 0.23046875, "learning_rate": 8.70138017565872e-06, "loss": 1.0627, "step": 207 }, { "epoch": 0.26097867001254704, "grad_norm": 0.2177734375, "learning_rate": 8.695106649937265e-06, "loss": 1.045, "step": 208 }, { "epoch": 0.26223337515683814, "grad_norm": 0.18359375, "learning_rate": 8.68883312421581e-06, "loss": 1.0962, "step": 209 }, { "epoch": 0.26348808030112925, "grad_norm": 0.1943359375, "learning_rate": 8.682559598494355e-06, "loss": 1.103, "step": 210 }, { "epoch": 0.26474278544542035, "grad_norm": 0.2373046875, "learning_rate": 8.676286072772899e-06, "loss": 1.0934, "step": 211 }, { "epoch": 0.2659974905897114, "grad_norm": 0.1806640625, "learning_rate": 8.670012547051444e-06, "loss": 1.1168, "step": 212 }, { "epoch": 0.2672521957340025, "grad_norm": 0.1845703125, "learning_rate": 8.663739021329989e-06, "loss": 1.0541, "step": 213 }, { "epoch": 0.2685069008782936, "grad_norm": 0.181640625, "learning_rate": 8.657465495608533e-06, "loss": 1.1042, "step": 214 }, { "epoch": 0.2697616060225847, "grad_norm": 0.1767578125, "learning_rate": 8.651191969887077e-06, "loss": 1.0692, "step": 215 }, { "epoch": 0.2710163111668758, "grad_norm": 0.1875, "learning_rate": 8.644918444165622e-06, "loss": 1.0942, "step": 216 }, { "epoch": 0.2722710163111669, "grad_norm": 0.2021484375, "learning_rate": 8.638644918444167e-06, "loss": 1.036, "step": 217 }, { "epoch": 0.27352572145545795, "grad_norm": 0.19921875, "learning_rate": 8.63237139272271e-06, "loss": 1.1155, "step": 218 }, { "epoch": 0.27478042659974905, "grad_norm": 0.197265625, "learning_rate": 8.626097867001255e-06, "loss": 1.094, "step": 219 }, { "epoch": 0.27603513174404015, "grad_norm": 0.228515625, "learning_rate": 8.6198243412798e-06, "loss": 1.1424, "step": 220 }, { "epoch": 0.27728983688833125, "grad_norm": 0.205078125, "learning_rate": 8.613550815558345e-06, "loss": 1.059, "step": 221 }, { "epoch": 0.27854454203262236, "grad_norm": 0.1767578125, "learning_rate": 8.607277289836889e-06, "loss": 1.0716, "step": 222 }, { "epoch": 0.2797992471769134, "grad_norm": 0.181640625, "learning_rate": 8.601003764115433e-06, "loss": 1.0815, "step": 223 }, { "epoch": 0.2810539523212045, "grad_norm": 0.1865234375, "learning_rate": 8.594730238393979e-06, "loss": 1.0617, "step": 224 }, { "epoch": 0.2823086574654956, "grad_norm": 0.1796875, "learning_rate": 8.588456712672523e-06, "loss": 1.0549, "step": 225 }, { "epoch": 0.2835633626097867, "grad_norm": 0.1962890625, "learning_rate": 8.582183186951067e-06, "loss": 1.0264, "step": 226 }, { "epoch": 0.2848180677540778, "grad_norm": 0.189453125, "learning_rate": 8.575909661229613e-06, "loss": 1.0628, "step": 227 }, { "epoch": 0.2860727728983689, "grad_norm": 0.197265625, "learning_rate": 8.569636135508157e-06, "loss": 1.0491, "step": 228 }, { "epoch": 0.28732747804265996, "grad_norm": 0.232421875, "learning_rate": 8.5633626097867e-06, "loss": 1.0761, "step": 229 }, { "epoch": 0.28858218318695106, "grad_norm": 0.2314453125, "learning_rate": 8.557089084065245e-06, "loss": 1.087, "step": 230 }, { "epoch": 0.28983688833124216, "grad_norm": 0.185546875, "learning_rate": 8.55081555834379e-06, "loss": 1.0627, "step": 231 }, { "epoch": 0.29109159347553326, "grad_norm": 0.1845703125, "learning_rate": 8.544542032622335e-06, "loss": 1.1257, "step": 232 }, { "epoch": 0.29234629861982436, "grad_norm": 0.2080078125, "learning_rate": 8.538268506900879e-06, "loss": 1.1212, "step": 233 }, { "epoch": 0.2936010037641154, "grad_norm": 0.2041015625, "learning_rate": 8.531994981179423e-06, "loss": 1.0462, "step": 234 }, { "epoch": 0.2948557089084065, "grad_norm": 0.205078125, "learning_rate": 8.525721455457969e-06, "loss": 1.1078, "step": 235 }, { "epoch": 0.2961104140526976, "grad_norm": 0.1845703125, "learning_rate": 8.519447929736513e-06, "loss": 1.0946, "step": 236 }, { "epoch": 0.2973651191969887, "grad_norm": 0.60546875, "learning_rate": 8.513174404015057e-06, "loss": 1.0422, "step": 237 }, { "epoch": 0.2986198243412798, "grad_norm": 0.2001953125, "learning_rate": 8.506900878293601e-06, "loss": 1.0799, "step": 238 }, { "epoch": 0.2998745294855709, "grad_norm": 0.220703125, "learning_rate": 8.500627352572147e-06, "loss": 1.0707, "step": 239 }, { "epoch": 0.30112923462986196, "grad_norm": 0.18359375, "learning_rate": 8.49435382685069e-06, "loss": 1.0565, "step": 240 }, { "epoch": 0.30238393977415307, "grad_norm": 0.1982421875, "learning_rate": 8.488080301129235e-06, "loss": 1.0532, "step": 241 }, { "epoch": 0.30363864491844417, "grad_norm": 0.2021484375, "learning_rate": 8.48180677540778e-06, "loss": 1.1413, "step": 242 }, { "epoch": 0.30489335006273527, "grad_norm": 0.1982421875, "learning_rate": 8.475533249686325e-06, "loss": 1.133, "step": 243 }, { "epoch": 0.30614805520702637, "grad_norm": 0.189453125, "learning_rate": 8.469259723964869e-06, "loss": 1.0553, "step": 244 }, { "epoch": 0.3074027603513174, "grad_norm": 0.2109375, "learning_rate": 8.462986198243413e-06, "loss": 1.0661, "step": 245 }, { "epoch": 0.3086574654956085, "grad_norm": 0.2041015625, "learning_rate": 8.456712672521959e-06, "loss": 1.056, "step": 246 }, { "epoch": 0.3099121706398996, "grad_norm": 0.21875, "learning_rate": 8.450439146800503e-06, "loss": 1.1469, "step": 247 }, { "epoch": 0.3111668757841907, "grad_norm": 0.2412109375, "learning_rate": 8.444165621079047e-06, "loss": 1.0669, "step": 248 }, { "epoch": 0.3124215809284818, "grad_norm": 0.2099609375, "learning_rate": 8.437892095357591e-06, "loss": 1.099, "step": 249 }, { "epoch": 0.3136762860727729, "grad_norm": 0.2001953125, "learning_rate": 8.431618569636137e-06, "loss": 1.1291, "step": 250 }, { "epoch": 0.31493099121706397, "grad_norm": 0.1904296875, "learning_rate": 8.425345043914681e-06, "loss": 1.087, "step": 251 }, { "epoch": 0.3161856963613551, "grad_norm": 0.203125, "learning_rate": 8.419071518193225e-06, "loss": 1.1106, "step": 252 }, { "epoch": 0.3174404015056462, "grad_norm": 0.205078125, "learning_rate": 8.412797992471769e-06, "loss": 1.0771, "step": 253 }, { "epoch": 0.3186951066499373, "grad_norm": 0.19921875, "learning_rate": 8.406524466750315e-06, "loss": 1.064, "step": 254 }, { "epoch": 0.3199498117942284, "grad_norm": 0.2041015625, "learning_rate": 8.400250941028859e-06, "loss": 1.0344, "step": 255 }, { "epoch": 0.3212045169385194, "grad_norm": 0.2021484375, "learning_rate": 8.393977415307403e-06, "loss": 1.0636, "step": 256 }, { "epoch": 0.3224592220828105, "grad_norm": 0.2001953125, "learning_rate": 8.387703889585947e-06, "loss": 1.0692, "step": 257 }, { "epoch": 0.3237139272271016, "grad_norm": 0.19140625, "learning_rate": 8.381430363864493e-06, "loss": 1.0365, "step": 258 }, { "epoch": 0.32496863237139273, "grad_norm": 0.201171875, "learning_rate": 8.375156838143037e-06, "loss": 1.0516, "step": 259 }, { "epoch": 0.32622333751568383, "grad_norm": 0.2001953125, "learning_rate": 8.368883312421581e-06, "loss": 1.0558, "step": 260 }, { "epoch": 0.32747804265997493, "grad_norm": 0.2373046875, "learning_rate": 8.362609786700127e-06, "loss": 1.1051, "step": 261 }, { "epoch": 0.328732747804266, "grad_norm": 0.1884765625, "learning_rate": 8.356336260978671e-06, "loss": 1.028, "step": 262 }, { "epoch": 0.3299874529485571, "grad_norm": 0.2041015625, "learning_rate": 8.350062735257215e-06, "loss": 1.0849, "step": 263 }, { "epoch": 0.3312421580928482, "grad_norm": 0.1962890625, "learning_rate": 8.343789209535759e-06, "loss": 1.0684, "step": 264 }, { "epoch": 0.3324968632371393, "grad_norm": 0.1923828125, "learning_rate": 8.337515683814305e-06, "loss": 1.0105, "step": 265 }, { "epoch": 0.3337515683814304, "grad_norm": 0.2021484375, "learning_rate": 8.331242158092849e-06, "loss": 1.0608, "step": 266 }, { "epoch": 0.33500627352572143, "grad_norm": 0.27734375, "learning_rate": 8.324968632371393e-06, "loss": 1.0286, "step": 267 }, { "epoch": 0.33626097867001253, "grad_norm": 0.189453125, "learning_rate": 8.318695106649937e-06, "loss": 1.0726, "step": 268 }, { "epoch": 0.33751568381430364, "grad_norm": 0.19140625, "learning_rate": 8.312421580928483e-06, "loss": 1.0911, "step": 269 }, { "epoch": 0.33877038895859474, "grad_norm": 0.201171875, "learning_rate": 8.306148055207027e-06, "loss": 1.0299, "step": 270 }, { "epoch": 0.34002509410288584, "grad_norm": 0.1982421875, "learning_rate": 8.299874529485573e-06, "loss": 1.0606, "step": 271 }, { "epoch": 0.34127979924717694, "grad_norm": 0.2294921875, "learning_rate": 8.293601003764115e-06, "loss": 1.0658, "step": 272 }, { "epoch": 0.342534504391468, "grad_norm": 0.2021484375, "learning_rate": 8.287327478042661e-06, "loss": 0.9989, "step": 273 }, { "epoch": 0.3437892095357591, "grad_norm": 0.20703125, "learning_rate": 8.281053952321205e-06, "loss": 1.051, "step": 274 }, { "epoch": 0.3450439146800502, "grad_norm": 0.19921875, "learning_rate": 8.27478042659975e-06, "loss": 1.1025, "step": 275 }, { "epoch": 0.3462986198243413, "grad_norm": 0.1796875, "learning_rate": 8.268506900878295e-06, "loss": 1.0706, "step": 276 }, { "epoch": 0.3475533249686324, "grad_norm": 0.20703125, "learning_rate": 8.262233375156839e-06, "loss": 1.094, "step": 277 }, { "epoch": 0.34880803011292344, "grad_norm": 0.2001953125, "learning_rate": 8.255959849435383e-06, "loss": 1.0697, "step": 278 }, { "epoch": 0.35006273525721454, "grad_norm": 0.1884765625, "learning_rate": 8.249686323713927e-06, "loss": 1.0619, "step": 279 }, { "epoch": 0.35131744040150564, "grad_norm": 0.283203125, "learning_rate": 8.243412797992473e-06, "loss": 1.0414, "step": 280 }, { "epoch": 0.35257214554579674, "grad_norm": 0.244140625, "learning_rate": 8.237139272271017e-06, "loss": 1.1023, "step": 281 }, { "epoch": 0.35382685069008785, "grad_norm": 0.201171875, "learning_rate": 8.230865746549561e-06, "loss": 1.0512, "step": 282 }, { "epoch": 0.35508155583437895, "grad_norm": 0.2265625, "learning_rate": 8.224592220828105e-06, "loss": 1.0255, "step": 283 }, { "epoch": 0.35633626097867, "grad_norm": 0.2275390625, "learning_rate": 8.218318695106651e-06, "loss": 1.0444, "step": 284 }, { "epoch": 0.3575909661229611, "grad_norm": 0.1982421875, "learning_rate": 8.212045169385195e-06, "loss": 1.0198, "step": 285 }, { "epoch": 0.3588456712672522, "grad_norm": 0.240234375, "learning_rate": 8.205771643663741e-06, "loss": 1.0218, "step": 286 }, { "epoch": 0.3601003764115433, "grad_norm": 0.2431640625, "learning_rate": 8.199498117942283e-06, "loss": 1.037, "step": 287 }, { "epoch": 0.3613550815558344, "grad_norm": 0.203125, "learning_rate": 8.193224592220829e-06, "loss": 1.0115, "step": 288 }, { "epoch": 0.36260978670012545, "grad_norm": 0.22265625, "learning_rate": 8.186951066499373e-06, "loss": 1.0836, "step": 289 }, { "epoch": 0.36386449184441655, "grad_norm": 0.1953125, "learning_rate": 8.180677540777919e-06, "loss": 1.0397, "step": 290 }, { "epoch": 0.36511919698870765, "grad_norm": 0.1982421875, "learning_rate": 8.174404015056461e-06, "loss": 1.0344, "step": 291 }, { "epoch": 0.36637390213299875, "grad_norm": 0.2021484375, "learning_rate": 8.168130489335007e-06, "loss": 1.0117, "step": 292 }, { "epoch": 0.36762860727728985, "grad_norm": 0.1953125, "learning_rate": 8.161856963613551e-06, "loss": 1.0735, "step": 293 }, { "epoch": 0.36888331242158096, "grad_norm": 0.1982421875, "learning_rate": 8.155583437892095e-06, "loss": 1.0265, "step": 294 }, { "epoch": 0.370138017565872, "grad_norm": 0.2890625, "learning_rate": 8.149309912170641e-06, "loss": 1.0526, "step": 295 }, { "epoch": 0.3713927227101631, "grad_norm": 0.1884765625, "learning_rate": 8.143036386449185e-06, "loss": 1.0359, "step": 296 }, { "epoch": 0.3726474278544542, "grad_norm": 0.2041015625, "learning_rate": 8.13676286072773e-06, "loss": 1.0763, "step": 297 }, { "epoch": 0.3739021329987453, "grad_norm": 0.1943359375, "learning_rate": 8.130489335006273e-06, "loss": 1.066, "step": 298 }, { "epoch": 0.3751568381430364, "grad_norm": 0.2138671875, "learning_rate": 8.12421580928482e-06, "loss": 1.0578, "step": 299 }, { "epoch": 0.37641154328732745, "grad_norm": 0.2412109375, "learning_rate": 8.117942283563363e-06, "loss": 1.076, "step": 300 }, { "epoch": 0.37766624843161856, "grad_norm": 0.19921875, "learning_rate": 8.111668757841907e-06, "loss": 1.0188, "step": 301 }, { "epoch": 0.37892095357590966, "grad_norm": 0.1982421875, "learning_rate": 8.105395232120451e-06, "loss": 1.009, "step": 302 }, { "epoch": 0.38017565872020076, "grad_norm": 0.21484375, "learning_rate": 8.099121706398997e-06, "loss": 1.0595, "step": 303 }, { "epoch": 0.38143036386449186, "grad_norm": 0.41015625, "learning_rate": 8.092848180677541e-06, "loss": 1.0747, "step": 304 }, { "epoch": 0.38268506900878296, "grad_norm": 0.19921875, "learning_rate": 8.086574654956087e-06, "loss": 1.0314, "step": 305 }, { "epoch": 0.383939774153074, "grad_norm": 0.2177734375, "learning_rate": 8.08030112923463e-06, "loss": 1.0971, "step": 306 }, { "epoch": 0.3851944792973651, "grad_norm": 0.251953125, "learning_rate": 8.074027603513175e-06, "loss": 1.0279, "step": 307 }, { "epoch": 0.3864491844416562, "grad_norm": 0.224609375, "learning_rate": 8.06775407779172e-06, "loss": 1.0518, "step": 308 }, { "epoch": 0.3877038895859473, "grad_norm": 0.2138671875, "learning_rate": 8.061480552070265e-06, "loss": 1.05, "step": 309 }, { "epoch": 0.3889585947302384, "grad_norm": 0.2275390625, "learning_rate": 8.05520702634881e-06, "loss": 1.0721, "step": 310 }, { "epoch": 0.39021329987452946, "grad_norm": 0.2177734375, "learning_rate": 8.048933500627353e-06, "loss": 1.0648, "step": 311 }, { "epoch": 0.39146800501882056, "grad_norm": 0.2041015625, "learning_rate": 8.042659974905897e-06, "loss": 1.0472, "step": 312 }, { "epoch": 0.39272271016311167, "grad_norm": 0.240234375, "learning_rate": 8.036386449184442e-06, "loss": 1.0006, "step": 313 }, { "epoch": 0.39397741530740277, "grad_norm": 0.212890625, "learning_rate": 8.030112923462987e-06, "loss": 1.0319, "step": 314 }, { "epoch": 0.39523212045169387, "grad_norm": 0.22265625, "learning_rate": 8.023839397741531e-06, "loss": 1.0058, "step": 315 }, { "epoch": 0.39648682559598497, "grad_norm": 0.19921875, "learning_rate": 8.017565872020076e-06, "loss": 1.0115, "step": 316 }, { "epoch": 0.397741530740276, "grad_norm": 0.21875, "learning_rate": 8.01129234629862e-06, "loss": 1.0273, "step": 317 }, { "epoch": 0.3989962358845671, "grad_norm": 0.21875, "learning_rate": 8.005018820577165e-06, "loss": 1.0251, "step": 318 }, { "epoch": 0.4002509410288582, "grad_norm": 0.203125, "learning_rate": 7.99874529485571e-06, "loss": 1.0013, "step": 319 }, { "epoch": 0.4015056461731493, "grad_norm": 0.1962890625, "learning_rate": 7.992471769134255e-06, "loss": 0.9909, "step": 320 }, { "epoch": 0.4027603513174404, "grad_norm": 0.1982421875, "learning_rate": 7.986198243412798e-06, "loss": 1.006, "step": 321 }, { "epoch": 0.40401505646173147, "grad_norm": 0.2099609375, "learning_rate": 7.979924717691343e-06, "loss": 1.0249, "step": 322 }, { "epoch": 0.40526976160602257, "grad_norm": 0.22265625, "learning_rate": 7.973651191969888e-06, "loss": 1.058, "step": 323 }, { "epoch": 0.4065244667503137, "grad_norm": 0.216796875, "learning_rate": 7.967377666248433e-06, "loss": 0.9824, "step": 324 }, { "epoch": 0.4077791718946048, "grad_norm": 0.2060546875, "learning_rate": 7.961104140526977e-06, "loss": 0.9865, "step": 325 }, { "epoch": 0.4090338770388959, "grad_norm": 0.2041015625, "learning_rate": 7.954830614805521e-06, "loss": 1.0238, "step": 326 }, { "epoch": 0.410288582183187, "grad_norm": 0.224609375, "learning_rate": 7.948557089084066e-06, "loss": 1.0541, "step": 327 }, { "epoch": 0.411543287327478, "grad_norm": 0.23828125, "learning_rate": 7.942283563362611e-06, "loss": 1.0218, "step": 328 }, { "epoch": 0.4127979924717691, "grad_norm": 0.208984375, "learning_rate": 7.936010037641155e-06, "loss": 1.0333, "step": 329 }, { "epoch": 0.41405269761606023, "grad_norm": 0.205078125, "learning_rate": 7.9297365119197e-06, "loss": 1.0099, "step": 330 }, { "epoch": 0.41530740276035133, "grad_norm": 0.234375, "learning_rate": 7.923462986198244e-06, "loss": 1.0408, "step": 331 }, { "epoch": 0.41656210790464243, "grad_norm": 0.2421875, "learning_rate": 7.917189460476788e-06, "loss": 1.0841, "step": 332 }, { "epoch": 0.4178168130489335, "grad_norm": 0.2099609375, "learning_rate": 7.910915934755334e-06, "loss": 1.0308, "step": 333 }, { "epoch": 0.4190715181932246, "grad_norm": 0.2255859375, "learning_rate": 7.904642409033878e-06, "loss": 1.0417, "step": 334 }, { "epoch": 0.4203262233375157, "grad_norm": 0.259765625, "learning_rate": 7.898368883312422e-06, "loss": 1.0166, "step": 335 }, { "epoch": 0.4215809284818068, "grad_norm": 0.27734375, "learning_rate": 7.892095357590966e-06, "loss": 1.0378, "step": 336 }, { "epoch": 0.4228356336260979, "grad_norm": 0.2138671875, "learning_rate": 7.885821831869512e-06, "loss": 0.9998, "step": 337 }, { "epoch": 0.424090338770389, "grad_norm": 0.2099609375, "learning_rate": 7.879548306148056e-06, "loss": 1.0485, "step": 338 }, { "epoch": 0.42534504391468003, "grad_norm": 0.2255859375, "learning_rate": 7.873274780426601e-06, "loss": 0.9984, "step": 339 }, { "epoch": 0.42659974905897113, "grad_norm": 0.279296875, "learning_rate": 7.867001254705144e-06, "loss": 1.0603, "step": 340 }, { "epoch": 0.42785445420326224, "grad_norm": 0.203125, "learning_rate": 7.86072772898369e-06, "loss": 1.0619, "step": 341 }, { "epoch": 0.42910915934755334, "grad_norm": 0.1982421875, "learning_rate": 7.854454203262234e-06, "loss": 1.0599, "step": 342 }, { "epoch": 0.43036386449184444, "grad_norm": 0.6015625, "learning_rate": 7.84818067754078e-06, "loss": 0.9961, "step": 343 }, { "epoch": 0.4316185696361355, "grad_norm": 0.2265625, "learning_rate": 7.841907151819324e-06, "loss": 0.9853, "step": 344 }, { "epoch": 0.4328732747804266, "grad_norm": 0.201171875, "learning_rate": 7.835633626097868e-06, "loss": 1.0197, "step": 345 }, { "epoch": 0.4341279799247177, "grad_norm": 0.2119140625, "learning_rate": 7.829360100376412e-06, "loss": 1.0452, "step": 346 }, { "epoch": 0.4353826850690088, "grad_norm": 0.22265625, "learning_rate": 7.823086574654956e-06, "loss": 0.9994, "step": 347 }, { "epoch": 0.4366373902132999, "grad_norm": 0.2392578125, "learning_rate": 7.816813048933502e-06, "loss": 1.1041, "step": 348 }, { "epoch": 0.437892095357591, "grad_norm": 0.232421875, "learning_rate": 7.810539523212046e-06, "loss": 1.0084, "step": 349 }, { "epoch": 0.43914680050188204, "grad_norm": 0.359375, "learning_rate": 7.80426599749059e-06, "loss": 1.0615, "step": 350 }, { "epoch": 0.44040150564617314, "grad_norm": 0.2177734375, "learning_rate": 7.797992471769134e-06, "loss": 0.9738, "step": 351 }, { "epoch": 0.44165621079046424, "grad_norm": 0.2138671875, "learning_rate": 7.79171894604768e-06, "loss": 1.0617, "step": 352 }, { "epoch": 0.44291091593475534, "grad_norm": 0.2158203125, "learning_rate": 7.785445420326224e-06, "loss": 1.0699, "step": 353 }, { "epoch": 0.44416562107904645, "grad_norm": 0.201171875, "learning_rate": 7.77917189460477e-06, "loss": 1.0082, "step": 354 }, { "epoch": 0.4454203262233375, "grad_norm": 0.2021484375, "learning_rate": 7.772898368883312e-06, "loss": 1.0524, "step": 355 }, { "epoch": 0.4466750313676286, "grad_norm": 0.220703125, "learning_rate": 7.766624843161858e-06, "loss": 0.9862, "step": 356 }, { "epoch": 0.4479297365119197, "grad_norm": 0.20703125, "learning_rate": 7.760351317440402e-06, "loss": 1.0315, "step": 357 }, { "epoch": 0.4491844416562108, "grad_norm": 0.201171875, "learning_rate": 7.754077791718948e-06, "loss": 1.0394, "step": 358 }, { "epoch": 0.4504391468005019, "grad_norm": 0.2119140625, "learning_rate": 7.747804265997492e-06, "loss": 1.0103, "step": 359 }, { "epoch": 0.451693851944793, "grad_norm": 0.2099609375, "learning_rate": 7.741530740276036e-06, "loss": 1.0186, "step": 360 }, { "epoch": 0.45294855708908405, "grad_norm": 0.2041015625, "learning_rate": 7.73525721455458e-06, "loss": 1.0353, "step": 361 }, { "epoch": 0.45420326223337515, "grad_norm": 0.2177734375, "learning_rate": 7.728983688833126e-06, "loss": 1.0083, "step": 362 }, { "epoch": 0.45545796737766625, "grad_norm": 0.212890625, "learning_rate": 7.72271016311167e-06, "loss": 1.0701, "step": 363 }, { "epoch": 0.45671267252195735, "grad_norm": 0.2001953125, "learning_rate": 7.716436637390214e-06, "loss": 1.0321, "step": 364 }, { "epoch": 0.45796737766624845, "grad_norm": 0.2177734375, "learning_rate": 7.710163111668758e-06, "loss": 1.056, "step": 365 }, { "epoch": 0.4592220828105395, "grad_norm": 0.2021484375, "learning_rate": 7.703889585947302e-06, "loss": 1.0262, "step": 366 }, { "epoch": 0.4604767879548306, "grad_norm": 0.28515625, "learning_rate": 7.697616060225848e-06, "loss": 1.0355, "step": 367 }, { "epoch": 0.4617314930991217, "grad_norm": 0.2177734375, "learning_rate": 7.691342534504392e-06, "loss": 0.9801, "step": 368 }, { "epoch": 0.4629861982434128, "grad_norm": 0.275390625, "learning_rate": 7.685069008782938e-06, "loss": 0.9967, "step": 369 }, { "epoch": 0.4642409033877039, "grad_norm": 0.2216796875, "learning_rate": 7.67879548306148e-06, "loss": 1.0582, "step": 370 }, { "epoch": 0.465495608531995, "grad_norm": 0.287109375, "learning_rate": 7.672521957340026e-06, "loss": 1.0034, "step": 371 }, { "epoch": 0.46675031367628605, "grad_norm": 0.216796875, "learning_rate": 7.66624843161857e-06, "loss": 1.0526, "step": 372 }, { "epoch": 0.46800501882057716, "grad_norm": 0.2158203125, "learning_rate": 7.659974905897116e-06, "loss": 1.0316, "step": 373 }, { "epoch": 0.46925972396486826, "grad_norm": 0.2197265625, "learning_rate": 7.65370138017566e-06, "loss": 1.0175, "step": 374 }, { "epoch": 0.47051442910915936, "grad_norm": 0.21875, "learning_rate": 7.647427854454204e-06, "loss": 1.0513, "step": 375 }, { "epoch": 0.47176913425345046, "grad_norm": 0.2392578125, "learning_rate": 7.641154328732748e-06, "loss": 1.0369, "step": 376 }, { "epoch": 0.4730238393977415, "grad_norm": 0.2236328125, "learning_rate": 7.634880803011294e-06, "loss": 1.0862, "step": 377 }, { "epoch": 0.4742785445420326, "grad_norm": 0.22265625, "learning_rate": 7.628607277289838e-06, "loss": 1.0211, "step": 378 }, { "epoch": 0.4755332496863237, "grad_norm": 0.21484375, "learning_rate": 7.622333751568381e-06, "loss": 1.0367, "step": 379 }, { "epoch": 0.4767879548306148, "grad_norm": 0.294921875, "learning_rate": 7.616060225846926e-06, "loss": 1.0147, "step": 380 }, { "epoch": 0.4780426599749059, "grad_norm": 0.224609375, "learning_rate": 7.609786700125471e-06, "loss": 1.005, "step": 381 }, { "epoch": 0.479297365119197, "grad_norm": 0.208984375, "learning_rate": 7.603513174404016e-06, "loss": 1.0502, "step": 382 }, { "epoch": 0.48055207026348806, "grad_norm": 0.212890625, "learning_rate": 7.597239648682561e-06, "loss": 1.0237, "step": 383 }, { "epoch": 0.48180677540777916, "grad_norm": 0.2080078125, "learning_rate": 7.590966122961104e-06, "loss": 1.06, "step": 384 }, { "epoch": 0.48306148055207027, "grad_norm": 0.203125, "learning_rate": 7.584692597239649e-06, "loss": 0.9772, "step": 385 }, { "epoch": 0.48431618569636137, "grad_norm": 0.21484375, "learning_rate": 7.578419071518194e-06, "loss": 0.983, "step": 386 }, { "epoch": 0.48557089084065247, "grad_norm": 0.2080078125, "learning_rate": 7.572145545796739e-06, "loss": 0.9959, "step": 387 }, { "epoch": 0.4868255959849435, "grad_norm": 0.2060546875, "learning_rate": 7.565872020075283e-06, "loss": 1.002, "step": 388 }, { "epoch": 0.4880803011292346, "grad_norm": 0.2158203125, "learning_rate": 7.559598494353827e-06, "loss": 1.0067, "step": 389 }, { "epoch": 0.4893350062735257, "grad_norm": 0.21484375, "learning_rate": 7.553324968632372e-06, "loss": 1.0152, "step": 390 }, { "epoch": 0.4905897114178168, "grad_norm": 0.2255859375, "learning_rate": 7.547051442910916e-06, "loss": 1.0054, "step": 391 }, { "epoch": 0.4918444165621079, "grad_norm": 0.205078125, "learning_rate": 7.540777917189461e-06, "loss": 0.9975, "step": 392 }, { "epoch": 0.493099121706399, "grad_norm": 0.2353515625, "learning_rate": 7.534504391468006e-06, "loss": 0.9547, "step": 393 }, { "epoch": 0.49435382685069007, "grad_norm": 0.271484375, "learning_rate": 7.52823086574655e-06, "loss": 1.0871, "step": 394 }, { "epoch": 0.49560853199498117, "grad_norm": 0.2294921875, "learning_rate": 7.521957340025094e-06, "loss": 1.0393, "step": 395 }, { "epoch": 0.4968632371392723, "grad_norm": 0.2138671875, "learning_rate": 7.515683814303639e-06, "loss": 1.0192, "step": 396 }, { "epoch": 0.4981179422835634, "grad_norm": 0.2421875, "learning_rate": 7.509410288582184e-06, "loss": 1.0178, "step": 397 }, { "epoch": 0.4993726474278545, "grad_norm": 0.2041015625, "learning_rate": 7.503136762860729e-06, "loss": 1.0357, "step": 398 }, { "epoch": 0.5006273525721455, "grad_norm": 0.2333984375, "learning_rate": 7.496863237139272e-06, "loss": 0.9985, "step": 399 }, { "epoch": 0.5018820577164367, "grad_norm": 0.2392578125, "learning_rate": 7.490589711417817e-06, "loss": 1.0675, "step": 400 }, { "epoch": 0.5031367628607277, "grad_norm": 0.22265625, "learning_rate": 7.484316185696362e-06, "loss": 1.0444, "step": 401 }, { "epoch": 0.5043914680050188, "grad_norm": 0.2080078125, "learning_rate": 7.478042659974907e-06, "loss": 1.0283, "step": 402 }, { "epoch": 0.5056461731493099, "grad_norm": 0.212890625, "learning_rate": 7.471769134253451e-06, "loss": 0.98, "step": 403 }, { "epoch": 0.506900878293601, "grad_norm": 0.2333984375, "learning_rate": 7.465495608531995e-06, "loss": 1.0182, "step": 404 }, { "epoch": 0.5081555834378921, "grad_norm": 0.271484375, "learning_rate": 7.45922208281054e-06, "loss": 0.9183, "step": 405 }, { "epoch": 0.5094102885821832, "grad_norm": 0.2236328125, "learning_rate": 7.452948557089085e-06, "loss": 1.0021, "step": 406 }, { "epoch": 0.5106649937264742, "grad_norm": 0.2109375, "learning_rate": 7.446675031367629e-06, "loss": 1.0338, "step": 407 }, { "epoch": 0.5119196988707654, "grad_norm": 0.2119140625, "learning_rate": 7.440401505646174e-06, "loss": 1.0483, "step": 408 }, { "epoch": 0.5131744040150564, "grad_norm": 0.2265625, "learning_rate": 7.434127979924718e-06, "loss": 0.9954, "step": 409 }, { "epoch": 0.5144291091593476, "grad_norm": 0.240234375, "learning_rate": 7.427854454203262e-06, "loss": 1.0083, "step": 410 }, { "epoch": 0.5156838143036386, "grad_norm": 0.2294921875, "learning_rate": 7.421580928481807e-06, "loss": 1.1175, "step": 411 }, { "epoch": 0.5169385194479298, "grad_norm": 0.2099609375, "learning_rate": 7.415307402760352e-06, "loss": 1.0056, "step": 412 }, { "epoch": 0.5181932245922208, "grad_norm": 0.2236328125, "learning_rate": 7.409033877038897e-06, "loss": 1.0064, "step": 413 }, { "epoch": 0.5194479297365119, "grad_norm": 0.21484375, "learning_rate": 7.40276035131744e-06, "loss": 1.0692, "step": 414 }, { "epoch": 0.520702634880803, "grad_norm": 0.2177734375, "learning_rate": 7.396486825595985e-06, "loss": 1.0873, "step": 415 }, { "epoch": 0.5219573400250941, "grad_norm": 0.21875, "learning_rate": 7.39021329987453e-06, "loss": 1.1036, "step": 416 }, { "epoch": 0.5232120451693852, "grad_norm": 0.216796875, "learning_rate": 7.383939774153075e-06, "loss": 0.9454, "step": 417 }, { "epoch": 0.5244667503136763, "grad_norm": 0.2333984375, "learning_rate": 7.37766624843162e-06, "loss": 0.9912, "step": 418 }, { "epoch": 0.5257214554579673, "grad_norm": 0.2197265625, "learning_rate": 7.371392722710163e-06, "loss": 0.9896, "step": 419 }, { "epoch": 0.5269761606022585, "grad_norm": 0.2138671875, "learning_rate": 7.365119196988708e-06, "loss": 1.0081, "step": 420 }, { "epoch": 0.5282308657465495, "grad_norm": 0.228515625, "learning_rate": 7.358845671267253e-06, "loss": 1.041, "step": 421 }, { "epoch": 0.5294855708908407, "grad_norm": 0.2158203125, "learning_rate": 7.352572145545797e-06, "loss": 1.0657, "step": 422 }, { "epoch": 0.5307402760351317, "grad_norm": 0.22265625, "learning_rate": 7.346298619824341e-06, "loss": 1.0857, "step": 423 }, { "epoch": 0.5319949811794228, "grad_norm": 0.2265625, "learning_rate": 7.340025094102886e-06, "loss": 1.084, "step": 424 }, { "epoch": 0.533249686323714, "grad_norm": 0.22265625, "learning_rate": 7.333751568381431e-06, "loss": 0.9957, "step": 425 }, { "epoch": 0.534504391468005, "grad_norm": 0.2099609375, "learning_rate": 7.327478042659975e-06, "loss": 1.0152, "step": 426 }, { "epoch": 0.5357590966122961, "grad_norm": 0.2158203125, "learning_rate": 7.32120451693852e-06, "loss": 1.04, "step": 427 }, { "epoch": 0.5370138017565872, "grad_norm": 0.2197265625, "learning_rate": 7.314930991217064e-06, "loss": 1.0416, "step": 428 }, { "epoch": 0.5382685069008782, "grad_norm": 0.2060546875, "learning_rate": 7.3086574654956085e-06, "loss": 1.025, "step": 429 }, { "epoch": 0.5395232120451694, "grad_norm": 0.2021484375, "learning_rate": 7.302383939774153e-06, "loss": 0.9931, "step": 430 }, { "epoch": 0.5407779171894604, "grad_norm": 0.259765625, "learning_rate": 7.296110414052698e-06, "loss": 1.023, "step": 431 }, { "epoch": 0.5420326223337516, "grad_norm": 0.2236328125, "learning_rate": 7.289836888331243e-06, "loss": 1.0244, "step": 432 }, { "epoch": 0.5432873274780426, "grad_norm": 0.2158203125, "learning_rate": 7.2835633626097865e-06, "loss": 1.046, "step": 433 }, { "epoch": 0.5445420326223338, "grad_norm": 0.2099609375, "learning_rate": 7.2772898368883315e-06, "loss": 1.0521, "step": 434 }, { "epoch": 0.5457967377666249, "grad_norm": 0.2119140625, "learning_rate": 7.271016311166876e-06, "loss": 0.969, "step": 435 }, { "epoch": 0.5470514429109159, "grad_norm": 0.234375, "learning_rate": 7.264742785445421e-06, "loss": 0.9776, "step": 436 }, { "epoch": 0.548306148055207, "grad_norm": 0.2275390625, "learning_rate": 7.258469259723966e-06, "loss": 1.002, "step": 437 }, { "epoch": 0.5495608531994981, "grad_norm": 0.2265625, "learning_rate": 7.2521957340025095e-06, "loss": 1.0279, "step": 438 }, { "epoch": 0.5508155583437893, "grad_norm": 0.212890625, "learning_rate": 7.2459222082810545e-06, "loss": 0.9688, "step": 439 }, { "epoch": 0.5520702634880803, "grad_norm": 0.21484375, "learning_rate": 7.239648682559599e-06, "loss": 0.9864, "step": 440 }, { "epoch": 0.5533249686323714, "grad_norm": 0.21484375, "learning_rate": 7.2333751568381435e-06, "loss": 1.0034, "step": 441 }, { "epoch": 0.5545796737766625, "grad_norm": 0.2099609375, "learning_rate": 7.2271016311166884e-06, "loss": 0.9908, "step": 442 }, { "epoch": 0.5558343789209536, "grad_norm": 0.21875, "learning_rate": 7.2208281053952325e-06, "loss": 0.9927, "step": 443 }, { "epoch": 0.5570890840652447, "grad_norm": 0.2255859375, "learning_rate": 7.2145545796737775e-06, "loss": 1.0138, "step": 444 }, { "epoch": 0.5583437892095358, "grad_norm": 0.21875, "learning_rate": 7.2082810539523215e-06, "loss": 1.0306, "step": 445 }, { "epoch": 0.5595984943538268, "grad_norm": 0.2177734375, "learning_rate": 7.2020075282308665e-06, "loss": 0.969, "step": 446 }, { "epoch": 0.560853199498118, "grad_norm": 0.21484375, "learning_rate": 7.195734002509411e-06, "loss": 0.9617, "step": 447 }, { "epoch": 0.562107904642409, "grad_norm": 0.2216796875, "learning_rate": 7.189460476787955e-06, "loss": 1.0538, "step": 448 }, { "epoch": 0.5633626097867002, "grad_norm": 0.232421875, "learning_rate": 7.1831869510665e-06, "loss": 1.0007, "step": 449 }, { "epoch": 0.5646173149309912, "grad_norm": 0.251953125, "learning_rate": 7.1769134253450445e-06, "loss": 1.0418, "step": 450 }, { "epoch": 0.5658720200752823, "grad_norm": 0.216796875, "learning_rate": 7.1706398996235895e-06, "loss": 0.9931, "step": 451 }, { "epoch": 0.5671267252195734, "grad_norm": 0.228515625, "learning_rate": 7.164366373902134e-06, "loss": 1.0609, "step": 452 }, { "epoch": 0.5683814303638645, "grad_norm": 0.2119140625, "learning_rate": 7.158092848180678e-06, "loss": 0.9577, "step": 453 }, { "epoch": 0.5696361355081556, "grad_norm": 0.224609375, "learning_rate": 7.151819322459223e-06, "loss": 1.0269, "step": 454 }, { "epoch": 0.5708908406524467, "grad_norm": 0.220703125, "learning_rate": 7.1455457967377675e-06, "loss": 1.0278, "step": 455 }, { "epoch": 0.5721455457967378, "grad_norm": 0.2265625, "learning_rate": 7.1392722710163125e-06, "loss": 0.9647, "step": 456 }, { "epoch": 0.5734002509410289, "grad_norm": 0.27734375, "learning_rate": 7.1329987452948566e-06, "loss": 1.0457, "step": 457 }, { "epoch": 0.5746549560853199, "grad_norm": 0.2099609375, "learning_rate": 7.126725219573401e-06, "loss": 1.0258, "step": 458 }, { "epoch": 0.5759096612296111, "grad_norm": 0.251953125, "learning_rate": 7.120451693851946e-06, "loss": 1.0312, "step": 459 }, { "epoch": 0.5771643663739021, "grad_norm": 0.2392578125, "learning_rate": 7.11417816813049e-06, "loss": 0.9868, "step": 460 }, { "epoch": 0.5784190715181933, "grad_norm": 0.216796875, "learning_rate": 7.107904642409035e-06, "loss": 1.0312, "step": 461 }, { "epoch": 0.5796737766624843, "grad_norm": 0.2392578125, "learning_rate": 7.1016311166875795e-06, "loss": 1.0408, "step": 462 }, { "epoch": 0.5809284818067754, "grad_norm": 0.2314453125, "learning_rate": 7.095357590966123e-06, "loss": 1.0226, "step": 463 }, { "epoch": 0.5821831869510665, "grad_norm": 0.228515625, "learning_rate": 7.089084065244668e-06, "loss": 1.032, "step": 464 }, { "epoch": 0.5834378920953576, "grad_norm": 0.2255859375, "learning_rate": 7.082810539523213e-06, "loss": 1.0121, "step": 465 }, { "epoch": 0.5846925972396487, "grad_norm": 0.353515625, "learning_rate": 7.076537013801758e-06, "loss": 1.0505, "step": 466 }, { "epoch": 0.5859473023839398, "grad_norm": 0.234375, "learning_rate": 7.070263488080301e-06, "loss": 1.0132, "step": 467 }, { "epoch": 0.5872020075282308, "grad_norm": 0.287109375, "learning_rate": 7.063989962358846e-06, "loss": 1.0211, "step": 468 }, { "epoch": 0.588456712672522, "grad_norm": 0.232421875, "learning_rate": 7.057716436637391e-06, "loss": 1.0142, "step": 469 }, { "epoch": 0.589711417816813, "grad_norm": 0.21484375, "learning_rate": 7.051442910915936e-06, "loss": 1.0247, "step": 470 }, { "epoch": 0.5909661229611042, "grad_norm": 0.2314453125, "learning_rate": 7.045169385194481e-06, "loss": 1.0399, "step": 471 }, { "epoch": 0.5922208281053952, "grad_norm": 0.212890625, "learning_rate": 7.038895859473024e-06, "loss": 0.9943, "step": 472 }, { "epoch": 0.5934755332496863, "grad_norm": 0.2255859375, "learning_rate": 7.032622333751569e-06, "loss": 1.0616, "step": 473 }, { "epoch": 0.5947302383939774, "grad_norm": 0.21484375, "learning_rate": 7.026348808030114e-06, "loss": 1.0144, "step": 474 }, { "epoch": 0.5959849435382685, "grad_norm": 0.2197265625, "learning_rate": 7.020075282308658e-06, "loss": 1.0428, "step": 475 }, { "epoch": 0.5972396486825596, "grad_norm": 0.240234375, "learning_rate": 7.013801756587203e-06, "loss": 1.0131, "step": 476 }, { "epoch": 0.5984943538268507, "grad_norm": 0.216796875, "learning_rate": 7.007528230865747e-06, "loss": 1.0477, "step": 477 }, { "epoch": 0.5997490589711418, "grad_norm": 0.208984375, "learning_rate": 7.001254705144292e-06, "loss": 1.0386, "step": 478 }, { "epoch": 0.6010037641154329, "grad_norm": 0.2431640625, "learning_rate": 6.994981179422836e-06, "loss": 0.9963, "step": 479 }, { "epoch": 0.6022584692597239, "grad_norm": 0.224609375, "learning_rate": 6.988707653701381e-06, "loss": 0.9886, "step": 480 }, { "epoch": 0.6035131744040151, "grad_norm": 0.216796875, "learning_rate": 6.982434127979926e-06, "loss": 1.0472, "step": 481 }, { "epoch": 0.6047678795483061, "grad_norm": 0.2109375, "learning_rate": 6.976160602258469e-06, "loss": 1.0632, "step": 482 }, { "epoch": 0.6060225846925973, "grad_norm": 0.21875, "learning_rate": 6.969887076537014e-06, "loss": 1.0271, "step": 483 }, { "epoch": 0.6072772898368883, "grad_norm": 0.2275390625, "learning_rate": 6.963613550815559e-06, "loss": 1.0109, "step": 484 }, { "epoch": 0.6085319949811794, "grad_norm": 0.2314453125, "learning_rate": 6.957340025094104e-06, "loss": 0.9945, "step": 485 }, { "epoch": 0.6097867001254705, "grad_norm": 0.2373046875, "learning_rate": 6.951066499372649e-06, "loss": 1.0228, "step": 486 }, { "epoch": 0.6110414052697616, "grad_norm": 0.2373046875, "learning_rate": 6.944792973651192e-06, "loss": 1.041, "step": 487 }, { "epoch": 0.6122961104140527, "grad_norm": 0.234375, "learning_rate": 6.938519447929737e-06, "loss": 1.0175, "step": 488 }, { "epoch": 0.6135508155583438, "grad_norm": 0.2412109375, "learning_rate": 6.932245922208282e-06, "loss": 1.0233, "step": 489 }, { "epoch": 0.6148055207026348, "grad_norm": 0.2392578125, "learning_rate": 6.925972396486827e-06, "loss": 1.0153, "step": 490 }, { "epoch": 0.616060225846926, "grad_norm": 0.2255859375, "learning_rate": 6.919698870765371e-06, "loss": 0.9937, "step": 491 }, { "epoch": 0.617314930991217, "grad_norm": 0.2275390625, "learning_rate": 6.913425345043915e-06, "loss": 0.9822, "step": 492 }, { "epoch": 0.6185696361355082, "grad_norm": 0.240234375, "learning_rate": 6.90715181932246e-06, "loss": 1.0323, "step": 493 }, { "epoch": 0.6198243412797992, "grad_norm": 0.224609375, "learning_rate": 6.900878293601004e-06, "loss": 1.0243, "step": 494 }, { "epoch": 0.6210790464240903, "grad_norm": 0.2177734375, "learning_rate": 6.894604767879549e-06, "loss": 0.9982, "step": 495 }, { "epoch": 0.6223337515683814, "grad_norm": 0.216796875, "learning_rate": 6.888331242158094e-06, "loss": 1.0098, "step": 496 }, { "epoch": 0.6235884567126725, "grad_norm": 0.2158203125, "learning_rate": 6.882057716436638e-06, "loss": 0.9781, "step": 497 }, { "epoch": 0.6248431618569636, "grad_norm": 0.2431640625, "learning_rate": 6.875784190715182e-06, "loss": 1.0312, "step": 498 }, { "epoch": 0.6260978670012547, "grad_norm": 0.2451171875, "learning_rate": 6.869510664993727e-06, "loss": 1.0932, "step": 499 }, { "epoch": 0.6273525721455459, "grad_norm": 0.2158203125, "learning_rate": 6.863237139272272e-06, "loss": 0.9969, "step": 500 }, { "epoch": 0.6286072772898369, "grad_norm": 0.2275390625, "learning_rate": 6.856963613550817e-06, "loss": 1.0032, "step": 501 }, { "epoch": 0.6298619824341279, "grad_norm": 0.244140625, "learning_rate": 6.85069008782936e-06, "loss": 1.074, "step": 502 }, { "epoch": 0.6311166875784191, "grad_norm": 0.2392578125, "learning_rate": 6.844416562107905e-06, "loss": 1.031, "step": 503 }, { "epoch": 0.6323713927227101, "grad_norm": 0.2236328125, "learning_rate": 6.83814303638645e-06, "loss": 1.0139, "step": 504 }, { "epoch": 0.6336260978670013, "grad_norm": 0.26171875, "learning_rate": 6.831869510664995e-06, "loss": 1.0592, "step": 505 }, { "epoch": 0.6348808030112923, "grad_norm": 0.232421875, "learning_rate": 6.825595984943539e-06, "loss": 0.9784, "step": 506 }, { "epoch": 0.6361355081555834, "grad_norm": 0.2373046875, "learning_rate": 6.819322459222083e-06, "loss": 1.0075, "step": 507 }, { "epoch": 0.6373902132998746, "grad_norm": 0.2412109375, "learning_rate": 6.813048933500628e-06, "loss": 1.008, "step": 508 }, { "epoch": 0.6386449184441656, "grad_norm": 0.21875, "learning_rate": 6.806775407779173e-06, "loss": 1.0333, "step": 509 }, { "epoch": 0.6398996235884568, "grad_norm": 0.240234375, "learning_rate": 6.800501882057717e-06, "loss": 0.973, "step": 510 }, { "epoch": 0.6411543287327478, "grad_norm": 0.34375, "learning_rate": 6.794228356336261e-06, "loss": 1.0299, "step": 511 }, { "epoch": 0.6424090338770388, "grad_norm": 0.2333984375, "learning_rate": 6.787954830614806e-06, "loss": 1.0069, "step": 512 }, { "epoch": 0.64366373902133, "grad_norm": 0.25, "learning_rate": 6.78168130489335e-06, "loss": 1.056, "step": 513 }, { "epoch": 0.644918444165621, "grad_norm": 0.21875, "learning_rate": 6.775407779171895e-06, "loss": 0.9837, "step": 514 }, { "epoch": 0.6461731493099122, "grad_norm": 0.2255859375, "learning_rate": 6.76913425345044e-06, "loss": 1.0136, "step": 515 }, { "epoch": 0.6474278544542033, "grad_norm": 0.240234375, "learning_rate": 6.762860727728984e-06, "loss": 1.0288, "step": 516 }, { "epoch": 0.6486825595984943, "grad_norm": 0.22265625, "learning_rate": 6.756587202007528e-06, "loss": 0.9805, "step": 517 }, { "epoch": 0.6499372647427855, "grad_norm": 0.240234375, "learning_rate": 6.750313676286073e-06, "loss": 1.0803, "step": 518 }, { "epoch": 0.6511919698870765, "grad_norm": 0.251953125, "learning_rate": 6.744040150564618e-06, "loss": 0.9999, "step": 519 }, { "epoch": 0.6524466750313677, "grad_norm": 0.2333984375, "learning_rate": 6.737766624843163e-06, "loss": 1.0007, "step": 520 }, { "epoch": 0.6537013801756587, "grad_norm": 0.23046875, "learning_rate": 6.731493099121706e-06, "loss": 1.0622, "step": 521 }, { "epoch": 0.6549560853199499, "grad_norm": 0.2314453125, "learning_rate": 6.725219573400251e-06, "loss": 1.0222, "step": 522 }, { "epoch": 0.6562107904642409, "grad_norm": 0.2265625, "learning_rate": 6.718946047678796e-06, "loss": 1.0317, "step": 523 }, { "epoch": 0.657465495608532, "grad_norm": 0.2314453125, "learning_rate": 6.712672521957341e-06, "loss": 1.0315, "step": 524 }, { "epoch": 0.6587202007528231, "grad_norm": 0.232421875, "learning_rate": 6.706398996235885e-06, "loss": 0.9917, "step": 525 }, { "epoch": 0.6599749058971142, "grad_norm": 0.2490234375, "learning_rate": 6.700125470514429e-06, "loss": 1.0201, "step": 526 }, { "epoch": 0.6612296110414053, "grad_norm": 0.2314453125, "learning_rate": 6.693851944792974e-06, "loss": 1.0103, "step": 527 }, { "epoch": 0.6624843161856964, "grad_norm": 0.265625, "learning_rate": 6.687578419071519e-06, "loss": 0.9997, "step": 528 }, { "epoch": 0.6637390213299874, "grad_norm": 0.2392578125, "learning_rate": 6.681304893350063e-06, "loss": 1.0081, "step": 529 }, { "epoch": 0.6649937264742786, "grad_norm": 0.2138671875, "learning_rate": 6.675031367628608e-06, "loss": 1.014, "step": 530 }, { "epoch": 0.6662484316185696, "grad_norm": 0.2333984375, "learning_rate": 6.668757841907152e-06, "loss": 1.0614, "step": 531 }, { "epoch": 0.6675031367628608, "grad_norm": 0.228515625, "learning_rate": 6.662484316185696e-06, "loss": 1.0032, "step": 532 }, { "epoch": 0.6687578419071518, "grad_norm": 0.2177734375, "learning_rate": 6.656210790464241e-06, "loss": 1.003, "step": 533 }, { "epoch": 0.6700125470514429, "grad_norm": 0.2216796875, "learning_rate": 6.649937264742786e-06, "loss": 1.0225, "step": 534 }, { "epoch": 0.671267252195734, "grad_norm": 0.2275390625, "learning_rate": 6.643663739021331e-06, "loss": 0.951, "step": 535 }, { "epoch": 0.6725219573400251, "grad_norm": 0.21875, "learning_rate": 6.637390213299874e-06, "loss": 0.9905, "step": 536 }, { "epoch": 0.6737766624843162, "grad_norm": 0.2373046875, "learning_rate": 6.631116687578419e-06, "loss": 0.9888, "step": 537 }, { "epoch": 0.6750313676286073, "grad_norm": 0.271484375, "learning_rate": 6.624843161856964e-06, "loss": 1.0246, "step": 538 }, { "epoch": 0.6762860727728983, "grad_norm": 0.2275390625, "learning_rate": 6.618569636135509e-06, "loss": 0.9412, "step": 539 }, { "epoch": 0.6775407779171895, "grad_norm": 0.265625, "learning_rate": 6.612296110414054e-06, "loss": 1.0455, "step": 540 }, { "epoch": 0.6787954830614805, "grad_norm": 0.2421875, "learning_rate": 6.606022584692597e-06, "loss": 1.0466, "step": 541 }, { "epoch": 0.6800501882057717, "grad_norm": 0.251953125, "learning_rate": 6.599749058971142e-06, "loss": 1.0005, "step": 542 }, { "epoch": 0.6813048933500627, "grad_norm": 0.2265625, "learning_rate": 6.593475533249687e-06, "loss": 1.0148, "step": 543 }, { "epoch": 0.6825595984943539, "grad_norm": 0.248046875, "learning_rate": 6.587202007528231e-06, "loss": 0.985, "step": 544 }, { "epoch": 0.6838143036386449, "grad_norm": 0.23046875, "learning_rate": 6.580928481806776e-06, "loss": 1.0537, "step": 545 }, { "epoch": 0.685069008782936, "grad_norm": 0.25390625, "learning_rate": 6.57465495608532e-06, "loss": 1.0159, "step": 546 }, { "epoch": 0.6863237139272271, "grad_norm": 0.271484375, "learning_rate": 6.568381430363865e-06, "loss": 0.9791, "step": 547 }, { "epoch": 0.6875784190715182, "grad_norm": 0.2177734375, "learning_rate": 6.562107904642409e-06, "loss": 0.9641, "step": 548 }, { "epoch": 0.6888331242158093, "grad_norm": 0.30859375, "learning_rate": 6.555834378920954e-06, "loss": 1.0173, "step": 549 }, { "epoch": 0.6900878293601004, "grad_norm": 0.2314453125, "learning_rate": 6.549560853199499e-06, "loss": 1.0059, "step": 550 }, { "epoch": 0.6913425345043914, "grad_norm": 0.232421875, "learning_rate": 6.5432873274780425e-06, "loss": 1.0071, "step": 551 }, { "epoch": 0.6925972396486826, "grad_norm": 0.2294921875, "learning_rate": 6.5370138017565874e-06, "loss": 1.03, "step": 552 }, { "epoch": 0.6938519447929736, "grad_norm": 0.23828125, "learning_rate": 6.530740276035132e-06, "loss": 1.0033, "step": 553 }, { "epoch": 0.6951066499372648, "grad_norm": 0.2333984375, "learning_rate": 6.524466750313677e-06, "loss": 0.98, "step": 554 }, { "epoch": 0.6963613550815558, "grad_norm": 0.23828125, "learning_rate": 6.5181932245922206e-06, "loss": 1.0204, "step": 555 }, { "epoch": 0.6976160602258469, "grad_norm": 0.236328125, "learning_rate": 6.5119196988707655e-06, "loss": 1.0, "step": 556 }, { "epoch": 0.698870765370138, "grad_norm": 0.251953125, "learning_rate": 6.5056461731493104e-06, "loss": 1.0319, "step": 557 }, { "epoch": 0.7001254705144291, "grad_norm": 0.2265625, "learning_rate": 6.499372647427855e-06, "loss": 0.9701, "step": 558 }, { "epoch": 0.7013801756587202, "grad_norm": 0.244140625, "learning_rate": 6.4930991217064e-06, "loss": 0.9973, "step": 559 }, { "epoch": 0.7026348808030113, "grad_norm": 0.23828125, "learning_rate": 6.4868255959849435e-06, "loss": 0.9552, "step": 560 }, { "epoch": 0.7038895859473023, "grad_norm": 0.2236328125, "learning_rate": 6.4805520702634885e-06, "loss": 0.9786, "step": 561 }, { "epoch": 0.7051442910915935, "grad_norm": 0.22265625, "learning_rate": 6.474278544542033e-06, "loss": 1.0019, "step": 562 }, { "epoch": 0.7063989962358845, "grad_norm": 0.2353515625, "learning_rate": 6.4680050188205775e-06, "loss": 1.0359, "step": 563 }, { "epoch": 0.7076537013801757, "grad_norm": 0.26171875, "learning_rate": 6.4617314930991224e-06, "loss": 0.9592, "step": 564 }, { "epoch": 0.7089084065244667, "grad_norm": 0.287109375, "learning_rate": 6.4554579673776665e-06, "loss": 1.0433, "step": 565 }, { "epoch": 0.7101631116687579, "grad_norm": 0.2392578125, "learning_rate": 6.4491844416562115e-06, "loss": 1.0679, "step": 566 }, { "epoch": 0.7114178168130489, "grad_norm": 0.24609375, "learning_rate": 6.4429109159347556e-06, "loss": 1.028, "step": 567 }, { "epoch": 0.71267252195734, "grad_norm": 0.23046875, "learning_rate": 6.4366373902133005e-06, "loss": 1.0134, "step": 568 }, { "epoch": 0.7139272271016311, "grad_norm": 0.224609375, "learning_rate": 6.4303638644918454e-06, "loss": 0.9715, "step": 569 }, { "epoch": 0.7151819322459222, "grad_norm": 0.234375, "learning_rate": 6.424090338770389e-06, "loss": 0.9781, "step": 570 }, { "epoch": 0.7164366373902133, "grad_norm": 0.2734375, "learning_rate": 6.417816813048934e-06, "loss": 0.9516, "step": 571 }, { "epoch": 0.7176913425345044, "grad_norm": 0.2255859375, "learning_rate": 6.4115432873274786e-06, "loss": 1.0163, "step": 572 }, { "epoch": 0.7189460476787954, "grad_norm": 0.2177734375, "learning_rate": 6.4052697616060235e-06, "loss": 0.9777, "step": 573 }, { "epoch": 0.7202007528230866, "grad_norm": 0.224609375, "learning_rate": 6.3989962358845684e-06, "loss": 0.9474, "step": 574 }, { "epoch": 0.7214554579673776, "grad_norm": 0.23828125, "learning_rate": 6.392722710163112e-06, "loss": 0.9935, "step": 575 }, { "epoch": 0.7227101631116688, "grad_norm": 0.2314453125, "learning_rate": 6.386449184441657e-06, "loss": 1.0086, "step": 576 }, { "epoch": 0.7239648682559598, "grad_norm": 0.244140625, "learning_rate": 6.3801756587202015e-06, "loss": 1.0281, "step": 577 }, { "epoch": 0.7252195734002509, "grad_norm": 0.2265625, "learning_rate": 6.373902132998746e-06, "loss": 0.9463, "step": 578 }, { "epoch": 0.726474278544542, "grad_norm": 0.2412109375, "learning_rate": 6.3676286072772906e-06, "loss": 0.9626, "step": 579 }, { "epoch": 0.7277289836888331, "grad_norm": 0.2451171875, "learning_rate": 6.361355081555835e-06, "loss": 1.0673, "step": 580 }, { "epoch": 0.7289836888331243, "grad_norm": 0.224609375, "learning_rate": 6.35508155583438e-06, "loss": 1.0027, "step": 581 }, { "epoch": 0.7302383939774153, "grad_norm": 0.232421875, "learning_rate": 6.348808030112924e-06, "loss": 0.9524, "step": 582 }, { "epoch": 0.7314930991217063, "grad_norm": 0.224609375, "learning_rate": 6.342534504391469e-06, "loss": 0.9515, "step": 583 }, { "epoch": 0.7327478042659975, "grad_norm": 0.23828125, "learning_rate": 6.3362609786700136e-06, "loss": 0.9853, "step": 584 }, { "epoch": 0.7340025094102886, "grad_norm": 0.2265625, "learning_rate": 6.329987452948557e-06, "loss": 0.9573, "step": 585 }, { "epoch": 0.7352572145545797, "grad_norm": 0.2578125, "learning_rate": 6.323713927227102e-06, "loss": 1.0266, "step": 586 }, { "epoch": 0.7365119196988708, "grad_norm": 0.236328125, "learning_rate": 6.317440401505647e-06, "loss": 0.9818, "step": 587 }, { "epoch": 0.7377666248431619, "grad_norm": 0.234375, "learning_rate": 6.311166875784192e-06, "loss": 1.0315, "step": 588 }, { "epoch": 0.739021329987453, "grad_norm": 0.248046875, "learning_rate": 6.3048933500627365e-06, "loss": 1.0251, "step": 589 }, { "epoch": 0.740276035131744, "grad_norm": 0.2451171875, "learning_rate": 6.29861982434128e-06, "loss": 1.0012, "step": 590 }, { "epoch": 0.7415307402760352, "grad_norm": 0.294921875, "learning_rate": 6.292346298619825e-06, "loss": 0.9821, "step": 591 }, { "epoch": 0.7427854454203262, "grad_norm": 0.2236328125, "learning_rate": 6.28607277289837e-06, "loss": 1.0166, "step": 592 }, { "epoch": 0.7440401505646174, "grad_norm": 0.2275390625, "learning_rate": 6.279799247176915e-06, "loss": 0.9963, "step": 593 }, { "epoch": 0.7452948557089084, "grad_norm": 0.2294921875, "learning_rate": 6.273525721455459e-06, "loss": 0.9921, "step": 594 }, { "epoch": 0.7465495608531995, "grad_norm": 0.23046875, "learning_rate": 6.267252195734003e-06, "loss": 1.0109, "step": 595 }, { "epoch": 0.7478042659974906, "grad_norm": 0.23046875, "learning_rate": 6.260978670012548e-06, "loss": 0.9927, "step": 596 }, { "epoch": 0.7490589711417817, "grad_norm": 0.22265625, "learning_rate": 6.254705144291092e-06, "loss": 0.9984, "step": 597 }, { "epoch": 0.7503136762860728, "grad_norm": 0.2353515625, "learning_rate": 6.248431618569637e-06, "loss": 1.0286, "step": 598 }, { "epoch": 0.7515683814303639, "grad_norm": 0.271484375, "learning_rate": 6.242158092848181e-06, "loss": 0.9725, "step": 599 }, { "epoch": 0.7528230865746549, "grad_norm": 0.24609375, "learning_rate": 6.235884567126726e-06, "loss": 0.9851, "step": 600 }, { "epoch": 0.7540777917189461, "grad_norm": 0.2353515625, "learning_rate": 6.22961104140527e-06, "loss": 0.9679, "step": 601 }, { "epoch": 0.7553324968632371, "grad_norm": 0.2255859375, "learning_rate": 6.223337515683815e-06, "loss": 0.981, "step": 602 }, { "epoch": 0.7565872020075283, "grad_norm": 0.23046875, "learning_rate": 6.21706398996236e-06, "loss": 1.0109, "step": 603 }, { "epoch": 0.7578419071518193, "grad_norm": 0.244140625, "learning_rate": 6.210790464240903e-06, "loss": 0.9815, "step": 604 }, { "epoch": 0.7590966122961104, "grad_norm": 0.2392578125, "learning_rate": 6.204516938519448e-06, "loss": 0.976, "step": 605 }, { "epoch": 0.7603513174404015, "grad_norm": 0.244140625, "learning_rate": 6.198243412797993e-06, "loss": 0.9372, "step": 606 }, { "epoch": 0.7616060225846926, "grad_norm": 0.220703125, "learning_rate": 6.191969887076538e-06, "loss": 0.9484, "step": 607 }, { "epoch": 0.7628607277289837, "grad_norm": 0.234375, "learning_rate": 6.185696361355083e-06, "loss": 0.9565, "step": 608 }, { "epoch": 0.7641154328732748, "grad_norm": 0.2158203125, "learning_rate": 6.179422835633626e-06, "loss": 0.9878, "step": 609 }, { "epoch": 0.7653701380175659, "grad_norm": 0.236328125, "learning_rate": 6.173149309912171e-06, "loss": 1.0041, "step": 610 }, { "epoch": 0.766624843161857, "grad_norm": 0.244140625, "learning_rate": 6.166875784190716e-06, "loss": 0.979, "step": 611 }, { "epoch": 0.767879548306148, "grad_norm": 0.2431640625, "learning_rate": 6.160602258469261e-06, "loss": 0.9777, "step": 612 }, { "epoch": 0.7691342534504392, "grad_norm": 0.2265625, "learning_rate": 6.154328732747805e-06, "loss": 1.0164, "step": 613 }, { "epoch": 0.7703889585947302, "grad_norm": 0.22265625, "learning_rate": 6.148055207026349e-06, "loss": 0.9936, "step": 614 }, { "epoch": 0.7716436637390214, "grad_norm": 0.248046875, "learning_rate": 6.141781681304894e-06, "loss": 0.9012, "step": 615 }, { "epoch": 0.7728983688833124, "grad_norm": 0.2373046875, "learning_rate": 6.135508155583438e-06, "loss": 1.0325, "step": 616 }, { "epoch": 0.7741530740276035, "grad_norm": 0.25390625, "learning_rate": 6.129234629861983e-06, "loss": 0.9771, "step": 617 }, { "epoch": 0.7754077791718946, "grad_norm": 0.24609375, "learning_rate": 6.122961104140528e-06, "loss": 0.9804, "step": 618 }, { "epoch": 0.7766624843161857, "grad_norm": 0.232421875, "learning_rate": 6.116687578419072e-06, "loss": 0.9996, "step": 619 }, { "epoch": 0.7779171894604768, "grad_norm": 0.2353515625, "learning_rate": 6.110414052697616e-06, "loss": 0.9465, "step": 620 }, { "epoch": 0.7791718946047679, "grad_norm": 0.2333984375, "learning_rate": 6.104140526976161e-06, "loss": 1.0249, "step": 621 }, { "epoch": 0.7804265997490589, "grad_norm": 0.2216796875, "learning_rate": 6.097867001254706e-06, "loss": 0.9768, "step": 622 }, { "epoch": 0.7816813048933501, "grad_norm": 0.2333984375, "learning_rate": 6.091593475533251e-06, "loss": 1.008, "step": 623 }, { "epoch": 0.7829360100376411, "grad_norm": 0.2314453125, "learning_rate": 6.085319949811794e-06, "loss": 0.9911, "step": 624 }, { "epoch": 0.7841907151819323, "grad_norm": 0.244140625, "learning_rate": 6.079046424090339e-06, "loss": 0.9873, "step": 625 }, { "epoch": 0.7854454203262233, "grad_norm": 0.2353515625, "learning_rate": 6.072772898368884e-06, "loss": 0.9975, "step": 626 }, { "epoch": 0.7867001254705144, "grad_norm": 0.25, "learning_rate": 6.066499372647429e-06, "loss": 1.0364, "step": 627 }, { "epoch": 0.7879548306148055, "grad_norm": 0.244140625, "learning_rate": 6.060225846925973e-06, "loss": 1.0059, "step": 628 }, { "epoch": 0.7892095357590966, "grad_norm": 0.240234375, "learning_rate": 6.053952321204517e-06, "loss": 0.9842, "step": 629 }, { "epoch": 0.7904642409033877, "grad_norm": 0.263671875, "learning_rate": 6.047678795483062e-06, "loss": 0.9922, "step": 630 }, { "epoch": 0.7917189460476788, "grad_norm": 0.2490234375, "learning_rate": 6.041405269761607e-06, "loss": 1.0593, "step": 631 }, { "epoch": 0.7929736511919699, "grad_norm": 0.2333984375, "learning_rate": 6.035131744040151e-06, "loss": 1.0491, "step": 632 }, { "epoch": 0.794228356336261, "grad_norm": 0.234375, "learning_rate": 6.028858218318696e-06, "loss": 0.9905, "step": 633 }, { "epoch": 0.795483061480552, "grad_norm": 0.248046875, "learning_rate": 6.02258469259724e-06, "loss": 0.9623, "step": 634 }, { "epoch": 0.7967377666248432, "grad_norm": 0.232421875, "learning_rate": 6.016311166875784e-06, "loss": 0.9974, "step": 635 }, { "epoch": 0.7979924717691342, "grad_norm": 0.232421875, "learning_rate": 6.010037641154329e-06, "loss": 0.9685, "step": 636 }, { "epoch": 0.7992471769134254, "grad_norm": 0.248046875, "learning_rate": 6.003764115432874e-06, "loss": 0.9959, "step": 637 }, { "epoch": 0.8005018820577164, "grad_norm": 0.275390625, "learning_rate": 5.997490589711419e-06, "loss": 0.9808, "step": 638 }, { "epoch": 0.8017565872020075, "grad_norm": 0.294921875, "learning_rate": 5.991217063989962e-06, "loss": 0.981, "step": 639 }, { "epoch": 0.8030112923462986, "grad_norm": 0.248046875, "learning_rate": 5.984943538268507e-06, "loss": 1.033, "step": 640 }, { "epoch": 0.8042659974905897, "grad_norm": 0.267578125, "learning_rate": 5.978670012547052e-06, "loss": 1.026, "step": 641 }, { "epoch": 0.8055207026348808, "grad_norm": 0.27734375, "learning_rate": 5.972396486825597e-06, "loss": 1.0095, "step": 642 }, { "epoch": 0.8067754077791719, "grad_norm": 0.2412109375, "learning_rate": 5.96612296110414e-06, "loss": 0.9649, "step": 643 }, { "epoch": 0.8080301129234629, "grad_norm": 0.2373046875, "learning_rate": 5.959849435382685e-06, "loss": 0.9814, "step": 644 }, { "epoch": 0.8092848180677541, "grad_norm": 0.23828125, "learning_rate": 5.95357590966123e-06, "loss": 0.9885, "step": 645 }, { "epoch": 0.8105395232120451, "grad_norm": 0.236328125, "learning_rate": 5.947302383939775e-06, "loss": 0.937, "step": 646 }, { "epoch": 0.8117942283563363, "grad_norm": 0.2412109375, "learning_rate": 5.941028858218319e-06, "loss": 1.0283, "step": 647 }, { "epoch": 0.8130489335006273, "grad_norm": 0.244140625, "learning_rate": 5.934755332496863e-06, "loss": 0.9984, "step": 648 }, { "epoch": 0.8143036386449184, "grad_norm": 0.2314453125, "learning_rate": 5.928481806775408e-06, "loss": 1.0018, "step": 649 }, { "epoch": 0.8155583437892095, "grad_norm": 0.2431640625, "learning_rate": 5.922208281053953e-06, "loss": 1.0605, "step": 650 }, { "epoch": 0.8168130489335006, "grad_norm": 0.2294921875, "learning_rate": 5.915934755332497e-06, "loss": 0.9697, "step": 651 }, { "epoch": 0.8180677540777918, "grad_norm": 0.251953125, "learning_rate": 5.909661229611042e-06, "loss": 1.017, "step": 652 }, { "epoch": 0.8193224592220828, "grad_norm": 0.232421875, "learning_rate": 5.903387703889586e-06, "loss": 0.9963, "step": 653 }, { "epoch": 0.820577164366374, "grad_norm": 0.244140625, "learning_rate": 5.89711417816813e-06, "loss": 1.0128, "step": 654 }, { "epoch": 0.821831869510665, "grad_norm": 0.2333984375, "learning_rate": 5.890840652446675e-06, "loss": 0.9942, "step": 655 }, { "epoch": 0.823086574654956, "grad_norm": 0.23828125, "learning_rate": 5.88456712672522e-06, "loss": 1.0162, "step": 656 }, { "epoch": 0.8243412797992472, "grad_norm": 0.244140625, "learning_rate": 5.878293601003765e-06, "loss": 1.0365, "step": 657 }, { "epoch": 0.8255959849435383, "grad_norm": 0.244140625, "learning_rate": 5.872020075282308e-06, "loss": 1.0091, "step": 658 }, { "epoch": 0.8268506900878294, "grad_norm": 0.2451171875, "learning_rate": 5.865746549560853e-06, "loss": 1.0597, "step": 659 }, { "epoch": 0.8281053952321205, "grad_norm": 0.3046875, "learning_rate": 5.859473023839398e-06, "loss": 0.9988, "step": 660 }, { "epoch": 0.8293601003764115, "grad_norm": 0.23828125, "learning_rate": 5.853199498117943e-06, "loss": 1.0082, "step": 661 }, { "epoch": 0.8306148055207027, "grad_norm": 0.25390625, "learning_rate": 5.846925972396488e-06, "loss": 1.0225, "step": 662 }, { "epoch": 0.8318695106649937, "grad_norm": 0.291015625, "learning_rate": 5.840652446675031e-06, "loss": 1.0269, "step": 663 }, { "epoch": 0.8331242158092849, "grad_norm": 0.2470703125, "learning_rate": 5.834378920953576e-06, "loss": 0.9613, "step": 664 }, { "epoch": 0.8343789209535759, "grad_norm": 0.2353515625, "learning_rate": 5.828105395232121e-06, "loss": 1.0025, "step": 665 }, { "epoch": 0.835633626097867, "grad_norm": 0.267578125, "learning_rate": 5.821831869510665e-06, "loss": 1.0103, "step": 666 }, { "epoch": 0.8368883312421581, "grad_norm": 0.306640625, "learning_rate": 5.81555834378921e-06, "loss": 0.9712, "step": 667 }, { "epoch": 0.8381430363864492, "grad_norm": 0.23828125, "learning_rate": 5.809284818067754e-06, "loss": 0.9557, "step": 668 }, { "epoch": 0.8393977415307403, "grad_norm": 0.28125, "learning_rate": 5.803011292346299e-06, "loss": 0.9834, "step": 669 }, { "epoch": 0.8406524466750314, "grad_norm": 0.2470703125, "learning_rate": 5.796737766624843e-06, "loss": 1.0025, "step": 670 }, { "epoch": 0.8419071518193224, "grad_norm": 0.23828125, "learning_rate": 5.790464240903388e-06, "loss": 0.9785, "step": 671 }, { "epoch": 0.8431618569636136, "grad_norm": 0.2197265625, "learning_rate": 5.784190715181933e-06, "loss": 0.9546, "step": 672 }, { "epoch": 0.8444165621079046, "grad_norm": 0.23828125, "learning_rate": 5.7779171894604765e-06, "loss": 1.0014, "step": 673 }, { "epoch": 0.8456712672521958, "grad_norm": 0.2373046875, "learning_rate": 5.7716436637390215e-06, "loss": 0.954, "step": 674 }, { "epoch": 0.8469259723964868, "grad_norm": 0.236328125, "learning_rate": 5.765370138017566e-06, "loss": 0.994, "step": 675 }, { "epoch": 0.848180677540778, "grad_norm": 0.25390625, "learning_rate": 5.759096612296111e-06, "loss": 0.9832, "step": 676 }, { "epoch": 0.849435382685069, "grad_norm": 0.2314453125, "learning_rate": 5.752823086574656e-06, "loss": 0.9771, "step": 677 }, { "epoch": 0.8506900878293601, "grad_norm": 0.248046875, "learning_rate": 5.7465495608531995e-06, "loss": 1.0424, "step": 678 }, { "epoch": 0.8519447929736512, "grad_norm": 0.2255859375, "learning_rate": 5.7402760351317444e-06, "loss": 0.9387, "step": 679 }, { "epoch": 0.8531994981179423, "grad_norm": 0.318359375, "learning_rate": 5.734002509410289e-06, "loss": 0.986, "step": 680 }, { "epoch": 0.8544542032622334, "grad_norm": 0.2412109375, "learning_rate": 5.727728983688834e-06, "loss": 1.0346, "step": 681 }, { "epoch": 0.8557089084065245, "grad_norm": 0.2236328125, "learning_rate": 5.721455457967378e-06, "loss": 0.9562, "step": 682 }, { "epoch": 0.8569636135508155, "grad_norm": 0.2265625, "learning_rate": 5.7151819322459225e-06, "loss": 0.9859, "step": 683 }, { "epoch": 0.8582183186951067, "grad_norm": 0.2421875, "learning_rate": 5.7089084065244674e-06, "loss": 0.9454, "step": 684 }, { "epoch": 0.8594730238393977, "grad_norm": 0.2412109375, "learning_rate": 5.7026348808030115e-06, "loss": 0.9751, "step": 685 }, { "epoch": 0.8607277289836889, "grad_norm": 0.2392578125, "learning_rate": 5.6963613550815565e-06, "loss": 0.9906, "step": 686 }, { "epoch": 0.8619824341279799, "grad_norm": 0.2314453125, "learning_rate": 5.6900878293601006e-06, "loss": 0.9684, "step": 687 }, { "epoch": 0.863237139272271, "grad_norm": 0.2451171875, "learning_rate": 5.683814303638645e-06, "loss": 0.9979, "step": 688 }, { "epoch": 0.8644918444165621, "grad_norm": 0.2275390625, "learning_rate": 5.67754077791719e-06, "loss": 0.9682, "step": 689 }, { "epoch": 0.8657465495608532, "grad_norm": 0.232421875, "learning_rate": 5.6712672521957345e-06, "loss": 0.9917, "step": 690 }, { "epoch": 0.8670012547051443, "grad_norm": 0.2392578125, "learning_rate": 5.6649937264742795e-06, "loss": 0.996, "step": 691 }, { "epoch": 0.8682559598494354, "grad_norm": 0.296875, "learning_rate": 5.658720200752823e-06, "loss": 1.0135, "step": 692 }, { "epoch": 0.8695106649937264, "grad_norm": 0.2353515625, "learning_rate": 5.652446675031368e-06, "loss": 0.9833, "step": 693 }, { "epoch": 0.8707653701380176, "grad_norm": 0.265625, "learning_rate": 5.6461731493099126e-06, "loss": 0.9651, "step": 694 }, { "epoch": 0.8720200752823086, "grad_norm": 0.232421875, "learning_rate": 5.6398996235884575e-06, "loss": 0.9382, "step": 695 }, { "epoch": 0.8732747804265998, "grad_norm": 0.2490234375, "learning_rate": 5.6336260978670024e-06, "loss": 0.9628, "step": 696 }, { "epoch": 0.8745294855708908, "grad_norm": 0.2412109375, "learning_rate": 5.627352572145546e-06, "loss": 0.9816, "step": 697 }, { "epoch": 0.875784190715182, "grad_norm": 0.2578125, "learning_rate": 5.621079046424091e-06, "loss": 1.0412, "step": 698 }, { "epoch": 0.877038895859473, "grad_norm": 0.26953125, "learning_rate": 5.6148055207026356e-06, "loss": 1.0063, "step": 699 }, { "epoch": 0.8782936010037641, "grad_norm": 0.2392578125, "learning_rate": 5.60853199498118e-06, "loss": 0.9486, "step": 700 }, { "epoch": 0.8795483061480552, "grad_norm": 0.232421875, "learning_rate": 5.602258469259725e-06, "loss": 0.9665, "step": 701 }, { "epoch": 0.8808030112923463, "grad_norm": 0.388671875, "learning_rate": 5.595984943538269e-06, "loss": 1.0313, "step": 702 }, { "epoch": 0.8820577164366374, "grad_norm": 0.244140625, "learning_rate": 5.589711417816814e-06, "loss": 0.9965, "step": 703 }, { "epoch": 0.8833124215809285, "grad_norm": 0.244140625, "learning_rate": 5.583437892095358e-06, "loss": 0.965, "step": 704 }, { "epoch": 0.8845671267252195, "grad_norm": 0.2314453125, "learning_rate": 5.577164366373903e-06, "loss": 0.929, "step": 705 }, { "epoch": 0.8858218318695107, "grad_norm": 0.26953125, "learning_rate": 5.5708908406524476e-06, "loss": 1.0154, "step": 706 }, { "epoch": 0.8870765370138017, "grad_norm": 0.2490234375, "learning_rate": 5.564617314930991e-06, "loss": 0.9772, "step": 707 }, { "epoch": 0.8883312421580929, "grad_norm": 0.27734375, "learning_rate": 5.558343789209536e-06, "loss": 1.0044, "step": 708 }, { "epoch": 0.8895859473023839, "grad_norm": 0.2578125, "learning_rate": 5.552070263488081e-06, "loss": 0.9345, "step": 709 }, { "epoch": 0.890840652446675, "grad_norm": 0.23828125, "learning_rate": 5.545796737766626e-06, "loss": 0.9892, "step": 710 }, { "epoch": 0.8920953575909661, "grad_norm": 0.267578125, "learning_rate": 5.5395232120451706e-06, "loss": 0.9661, "step": 711 }, { "epoch": 0.8933500627352572, "grad_norm": 0.248046875, "learning_rate": 5.533249686323714e-06, "loss": 0.9571, "step": 712 }, { "epoch": 0.8946047678795483, "grad_norm": 0.25390625, "learning_rate": 5.526976160602259e-06, "loss": 1.0126, "step": 713 }, { "epoch": 0.8958594730238394, "grad_norm": 0.2431640625, "learning_rate": 5.520702634880804e-06, "loss": 1.0071, "step": 714 }, { "epoch": 0.8971141781681304, "grad_norm": 0.251953125, "learning_rate": 5.514429109159349e-06, "loss": 0.9589, "step": 715 }, { "epoch": 0.8983688833124216, "grad_norm": 0.248046875, "learning_rate": 5.508155583437893e-06, "loss": 1.0016, "step": 716 }, { "epoch": 0.8996235884567126, "grad_norm": 0.232421875, "learning_rate": 5.501882057716437e-06, "loss": 1.0069, "step": 717 }, { "epoch": 0.9008782936010038, "grad_norm": 0.25390625, "learning_rate": 5.495608531994982e-06, "loss": 0.991, "step": 718 }, { "epoch": 0.9021329987452948, "grad_norm": 0.255859375, "learning_rate": 5.489335006273526e-06, "loss": 0.9608, "step": 719 }, { "epoch": 0.903387703889586, "grad_norm": 0.2314453125, "learning_rate": 5.483061480552071e-06, "loss": 0.9292, "step": 720 }, { "epoch": 0.904642409033877, "grad_norm": 0.240234375, "learning_rate": 5.476787954830616e-06, "loss": 0.9165, "step": 721 }, { "epoch": 0.9058971141781681, "grad_norm": 0.291015625, "learning_rate": 5.47051442910916e-06, "loss": 1.0174, "step": 722 }, { "epoch": 0.9071518193224593, "grad_norm": 0.2451171875, "learning_rate": 5.464240903387704e-06, "loss": 1.0248, "step": 723 }, { "epoch": 0.9084065244667503, "grad_norm": 0.267578125, "learning_rate": 5.457967377666249e-06, "loss": 0.9418, "step": 724 }, { "epoch": 0.9096612296110415, "grad_norm": 0.2333984375, "learning_rate": 5.451693851944794e-06, "loss": 0.9992, "step": 725 }, { "epoch": 0.9109159347553325, "grad_norm": 0.2470703125, "learning_rate": 5.445420326223339e-06, "loss": 0.9592, "step": 726 }, { "epoch": 0.9121706398996235, "grad_norm": 0.251953125, "learning_rate": 5.439146800501882e-06, "loss": 0.9948, "step": 727 }, { "epoch": 0.9134253450439147, "grad_norm": 0.26953125, "learning_rate": 5.432873274780427e-06, "loss": 1.0082, "step": 728 }, { "epoch": 0.9146800501882058, "grad_norm": 0.2734375, "learning_rate": 5.426599749058972e-06, "loss": 1.0209, "step": 729 }, { "epoch": 0.9159347553324969, "grad_norm": 0.33203125, "learning_rate": 5.420326223337517e-06, "loss": 0.9758, "step": 730 }, { "epoch": 0.917189460476788, "grad_norm": 0.2373046875, "learning_rate": 5.41405269761606e-06, "loss": 0.9645, "step": 731 }, { "epoch": 0.918444165621079, "grad_norm": 0.328125, "learning_rate": 5.407779171894605e-06, "loss": 0.9189, "step": 732 }, { "epoch": 0.9196988707653702, "grad_norm": 0.2392578125, "learning_rate": 5.40150564617315e-06, "loss": 1.0366, "step": 733 }, { "epoch": 0.9209535759096612, "grad_norm": 0.255859375, "learning_rate": 5.395232120451695e-06, "loss": 1.0319, "step": 734 }, { "epoch": 0.9222082810539524, "grad_norm": 0.265625, "learning_rate": 5.388958594730239e-06, "loss": 0.9712, "step": 735 }, { "epoch": 0.9234629861982434, "grad_norm": 0.2255859375, "learning_rate": 5.382685069008783e-06, "loss": 0.9951, "step": 736 }, { "epoch": 0.9247176913425345, "grad_norm": 0.267578125, "learning_rate": 5.376411543287328e-06, "loss": 1.0196, "step": 737 }, { "epoch": 0.9259723964868256, "grad_norm": 0.2490234375, "learning_rate": 5.370138017565872e-06, "loss": 0.9388, "step": 738 }, { "epoch": 0.9272271016311167, "grad_norm": 0.2421875, "learning_rate": 5.363864491844417e-06, "loss": 1.0241, "step": 739 }, { "epoch": 0.9284818067754078, "grad_norm": 0.228515625, "learning_rate": 5.357590966122962e-06, "loss": 0.9446, "step": 740 }, { "epoch": 0.9297365119196989, "grad_norm": 0.248046875, "learning_rate": 5.351317440401506e-06, "loss": 0.9939, "step": 741 }, { "epoch": 0.93099121706399, "grad_norm": 0.25, "learning_rate": 5.34504391468005e-06, "loss": 1.0261, "step": 742 }, { "epoch": 0.9322459222082811, "grad_norm": 0.2265625, "learning_rate": 5.338770388958595e-06, "loss": 0.9527, "step": 743 }, { "epoch": 0.9335006273525721, "grad_norm": 0.251953125, "learning_rate": 5.33249686323714e-06, "loss": 1.0055, "step": 744 }, { "epoch": 0.9347553324968633, "grad_norm": 0.244140625, "learning_rate": 5.326223337515685e-06, "loss": 0.9737, "step": 745 }, { "epoch": 0.9360100376411543, "grad_norm": 0.259765625, "learning_rate": 5.319949811794228e-06, "loss": 0.983, "step": 746 }, { "epoch": 0.9372647427854455, "grad_norm": 0.2353515625, "learning_rate": 5.313676286072773e-06, "loss": 0.9523, "step": 747 }, { "epoch": 0.9385194479297365, "grad_norm": 0.251953125, "learning_rate": 5.307402760351318e-06, "loss": 0.9419, "step": 748 }, { "epoch": 0.9397741530740276, "grad_norm": 0.298828125, "learning_rate": 5.301129234629863e-06, "loss": 0.9466, "step": 749 }, { "epoch": 0.9410288582183187, "grad_norm": 0.263671875, "learning_rate": 5.294855708908407e-06, "loss": 0.9727, "step": 750 }, { "epoch": 0.9422835633626098, "grad_norm": 0.248046875, "learning_rate": 5.288582183186951e-06, "loss": 0.9913, "step": 751 }, { "epoch": 0.9435382685069009, "grad_norm": 0.24609375, "learning_rate": 5.282308657465496e-06, "loss": 0.9766, "step": 752 }, { "epoch": 0.944792973651192, "grad_norm": 0.240234375, "learning_rate": 5.276035131744041e-06, "loss": 0.9515, "step": 753 }, { "epoch": 0.946047678795483, "grad_norm": 0.248046875, "learning_rate": 5.269761606022585e-06, "loss": 0.9839, "step": 754 }, { "epoch": 0.9473023839397742, "grad_norm": 0.248046875, "learning_rate": 5.26348808030113e-06, "loss": 1.0047, "step": 755 }, { "epoch": 0.9485570890840652, "grad_norm": 0.2578125, "learning_rate": 5.257214554579674e-06, "loss": 1.0116, "step": 756 }, { "epoch": 0.9498117942283564, "grad_norm": 0.2431640625, "learning_rate": 5.250941028858218e-06, "loss": 0.9394, "step": 757 }, { "epoch": 0.9510664993726474, "grad_norm": 0.2373046875, "learning_rate": 5.244667503136763e-06, "loss": 0.9357, "step": 758 }, { "epoch": 0.9523212045169385, "grad_norm": 0.27734375, "learning_rate": 5.238393977415308e-06, "loss": 1.0077, "step": 759 }, { "epoch": 0.9535759096612296, "grad_norm": 0.2412109375, "learning_rate": 5.232120451693853e-06, "loss": 0.9591, "step": 760 }, { "epoch": 0.9548306148055207, "grad_norm": 0.2451171875, "learning_rate": 5.225846925972396e-06, "loss": 1.0031, "step": 761 }, { "epoch": 0.9560853199498118, "grad_norm": 0.251953125, "learning_rate": 5.219573400250941e-06, "loss": 1.0178, "step": 762 }, { "epoch": 0.9573400250941029, "grad_norm": 0.251953125, "learning_rate": 5.213299874529486e-06, "loss": 0.9548, "step": 763 }, { "epoch": 0.958594730238394, "grad_norm": 0.263671875, "learning_rate": 5.207026348808031e-06, "loss": 0.9323, "step": 764 }, { "epoch": 0.9598494353826851, "grad_norm": 0.259765625, "learning_rate": 5.200752823086576e-06, "loss": 0.9438, "step": 765 }, { "epoch": 0.9611041405269761, "grad_norm": 0.25390625, "learning_rate": 5.194479297365119e-06, "loss": 0.9794, "step": 766 }, { "epoch": 0.9623588456712673, "grad_norm": 0.251953125, "learning_rate": 5.188205771643664e-06, "loss": 0.9809, "step": 767 }, { "epoch": 0.9636135508155583, "grad_norm": 0.2392578125, "learning_rate": 5.181932245922209e-06, "loss": 0.967, "step": 768 }, { "epoch": 0.9648682559598495, "grad_norm": 0.2373046875, "learning_rate": 5.175658720200753e-06, "loss": 0.9429, "step": 769 }, { "epoch": 0.9661229611041405, "grad_norm": 0.296875, "learning_rate": 5.169385194479298e-06, "loss": 0.9997, "step": 770 }, { "epoch": 0.9673776662484316, "grad_norm": 0.2451171875, "learning_rate": 5.163111668757842e-06, "loss": 0.959, "step": 771 }, { "epoch": 0.9686323713927227, "grad_norm": 0.234375, "learning_rate": 5.156838143036387e-06, "loss": 0.9715, "step": 772 }, { "epoch": 0.9698870765370138, "grad_norm": 0.2490234375, "learning_rate": 5.150564617314931e-06, "loss": 0.9637, "step": 773 }, { "epoch": 0.9711417816813049, "grad_norm": 0.2373046875, "learning_rate": 5.144291091593476e-06, "loss": 0.9583, "step": 774 }, { "epoch": 0.972396486825596, "grad_norm": 0.2412109375, "learning_rate": 5.13801756587202e-06, "loss": 0.994, "step": 775 }, { "epoch": 0.973651191969887, "grad_norm": 0.298828125, "learning_rate": 5.131744040150564e-06, "loss": 1.0683, "step": 776 }, { "epoch": 0.9749058971141782, "grad_norm": 0.263671875, "learning_rate": 5.125470514429109e-06, "loss": 0.9931, "step": 777 }, { "epoch": 0.9761606022584692, "grad_norm": 0.2353515625, "learning_rate": 5.119196988707654e-06, "loss": 0.9619, "step": 778 }, { "epoch": 0.9774153074027604, "grad_norm": 0.251953125, "learning_rate": 5.112923462986199e-06, "loss": 0.9305, "step": 779 }, { "epoch": 0.9786700125470514, "grad_norm": 0.2470703125, "learning_rate": 5.106649937264742e-06, "loss": 0.9744, "step": 780 }, { "epoch": 0.9799247176913425, "grad_norm": 0.2451171875, "learning_rate": 5.100376411543287e-06, "loss": 0.9268, "step": 781 }, { "epoch": 0.9811794228356336, "grad_norm": 0.25390625, "learning_rate": 5.094102885821832e-06, "loss": 1.0677, "step": 782 }, { "epoch": 0.9824341279799247, "grad_norm": 0.244140625, "learning_rate": 5.087829360100377e-06, "loss": 1.009, "step": 783 }, { "epoch": 0.9836888331242158, "grad_norm": 0.259765625, "learning_rate": 5.081555834378922e-06, "loss": 0.9853, "step": 784 }, { "epoch": 0.9849435382685069, "grad_norm": 0.2421875, "learning_rate": 5.075282308657465e-06, "loss": 0.9529, "step": 785 }, { "epoch": 0.986198243412798, "grad_norm": 0.244140625, "learning_rate": 5.06900878293601e-06, "loss": 0.9776, "step": 786 }, { "epoch": 0.9874529485570891, "grad_norm": 0.271484375, "learning_rate": 5.062735257214555e-06, "loss": 0.9594, "step": 787 }, { "epoch": 0.9887076537013801, "grad_norm": 0.2421875, "learning_rate": 5.056461731493099e-06, "loss": 0.9927, "step": 788 }, { "epoch": 0.9899623588456713, "grad_norm": 0.23046875, "learning_rate": 5.050188205771644e-06, "loss": 1.0209, "step": 789 }, { "epoch": 0.9912170639899623, "grad_norm": 0.255859375, "learning_rate": 5.043914680050188e-06, "loss": 1.0354, "step": 790 }, { "epoch": 0.9924717691342535, "grad_norm": 0.236328125, "learning_rate": 5.037641154328733e-06, "loss": 0.9642, "step": 791 }, { "epoch": 0.9937264742785445, "grad_norm": 0.2451171875, "learning_rate": 5.031367628607277e-06, "loss": 0.971, "step": 792 }, { "epoch": 0.9949811794228356, "grad_norm": 0.2353515625, "learning_rate": 5.025094102885822e-06, "loss": 1.0008, "step": 793 }, { "epoch": 0.9962358845671268, "grad_norm": 0.2294921875, "learning_rate": 5.018820577164367e-06, "loss": 0.9149, "step": 794 }, { "epoch": 0.9974905897114178, "grad_norm": 0.2265625, "learning_rate": 5.0125470514429105e-06, "loss": 0.9263, "step": 795 }, { "epoch": 0.998745294855709, "grad_norm": 0.263671875, "learning_rate": 5.0062735257214555e-06, "loss": 1.0114, "step": 796 }, { "epoch": 1.0, "grad_norm": 0.2353515625, "learning_rate": 5e-06, "loss": 1.0028, "step": 797 }, { "epoch": 1.001254705144291, "grad_norm": 0.255859375, "learning_rate": 4.993726474278545e-06, "loss": 1.0169, "step": 798 }, { "epoch": 1.002509410288582, "grad_norm": 0.232421875, "learning_rate": 4.9874529485570894e-06, "loss": 0.9793, "step": 799 }, { "epoch": 1.0037641154328734, "grad_norm": 0.267578125, "learning_rate": 4.981179422835634e-06, "loss": 0.9939, "step": 800 }, { "epoch": 1.0050188205771644, "grad_norm": 0.287109375, "learning_rate": 4.9749058971141785e-06, "loss": 0.9857, "step": 801 }, { "epoch": 1.0062735257214555, "grad_norm": 0.234375, "learning_rate": 4.968632371392723e-06, "loss": 0.9616, "step": 802 }, { "epoch": 1.0075282308657465, "grad_norm": 0.53515625, "learning_rate": 4.9623588456712675e-06, "loss": 1.0087, "step": 803 }, { "epoch": 1.0087829360100375, "grad_norm": 0.2578125, "learning_rate": 4.9560853199498124e-06, "loss": 0.9856, "step": 804 }, { "epoch": 1.0100376411543288, "grad_norm": 0.244140625, "learning_rate": 4.9498117942283565e-06, "loss": 0.9376, "step": 805 }, { "epoch": 1.0112923462986199, "grad_norm": 0.265625, "learning_rate": 4.9435382685069015e-06, "loss": 0.9482, "step": 806 }, { "epoch": 1.012547051442911, "grad_norm": 0.259765625, "learning_rate": 4.9372647427854455e-06, "loss": 1.0049, "step": 807 }, { "epoch": 1.013801756587202, "grad_norm": 0.236328125, "learning_rate": 4.9309912170639905e-06, "loss": 0.9482, "step": 808 }, { "epoch": 1.015056461731493, "grad_norm": 0.4296875, "learning_rate": 4.9247176913425346e-06, "loss": 0.9355, "step": 809 }, { "epoch": 1.0163111668757843, "grad_norm": 0.322265625, "learning_rate": 4.9184441656210795e-06, "loss": 1.0264, "step": 810 }, { "epoch": 1.0175658720200753, "grad_norm": 0.240234375, "learning_rate": 4.912170639899624e-06, "loss": 0.9649, "step": 811 }, { "epoch": 1.0188205771643664, "grad_norm": 0.2421875, "learning_rate": 4.9058971141781685e-06, "loss": 0.9816, "step": 812 }, { "epoch": 1.0200752823086574, "grad_norm": 0.2578125, "learning_rate": 4.899623588456713e-06, "loss": 0.9194, "step": 813 }, { "epoch": 1.0213299874529485, "grad_norm": 0.240234375, "learning_rate": 4.8933500627352576e-06, "loss": 0.9738, "step": 814 }, { "epoch": 1.0225846925972397, "grad_norm": 0.240234375, "learning_rate": 4.8870765370138025e-06, "loss": 0.9321, "step": 815 }, { "epoch": 1.0238393977415308, "grad_norm": 0.2578125, "learning_rate": 4.880803011292347e-06, "loss": 0.9479, "step": 816 }, { "epoch": 1.0250941028858218, "grad_norm": 0.234375, "learning_rate": 4.8745294855708915e-06, "loss": 0.9665, "step": 817 }, { "epoch": 1.0263488080301129, "grad_norm": 0.265625, "learning_rate": 4.868255959849436e-06, "loss": 1.0384, "step": 818 }, { "epoch": 1.027603513174404, "grad_norm": 0.26171875, "learning_rate": 4.8619824341279805e-06, "loss": 1.0045, "step": 819 }, { "epoch": 1.0288582183186952, "grad_norm": 0.251953125, "learning_rate": 4.8557089084065255e-06, "loss": 0.9594, "step": 820 }, { "epoch": 1.0301129234629862, "grad_norm": 0.244140625, "learning_rate": 4.84943538268507e-06, "loss": 0.9968, "step": 821 }, { "epoch": 1.0313676286072773, "grad_norm": 0.2373046875, "learning_rate": 4.843161856963614e-06, "loss": 0.9533, "step": 822 }, { "epoch": 1.0326223337515683, "grad_norm": 0.25390625, "learning_rate": 4.836888331242159e-06, "loss": 0.9983, "step": 823 }, { "epoch": 1.0338770388958596, "grad_norm": 0.24609375, "learning_rate": 4.830614805520703e-06, "loss": 0.9657, "step": 824 }, { "epoch": 1.0351317440401506, "grad_norm": 0.251953125, "learning_rate": 4.824341279799248e-06, "loss": 0.975, "step": 825 }, { "epoch": 1.0363864491844417, "grad_norm": 0.2421875, "learning_rate": 4.818067754077792e-06, "loss": 0.9593, "step": 826 }, { "epoch": 1.0376411543287327, "grad_norm": 0.26953125, "learning_rate": 4.811794228356337e-06, "loss": 0.9415, "step": 827 }, { "epoch": 1.0388958594730238, "grad_norm": 0.25390625, "learning_rate": 4.805520702634881e-06, "loss": 0.9527, "step": 828 }, { "epoch": 1.040150564617315, "grad_norm": 0.251953125, "learning_rate": 4.799247176913426e-06, "loss": 0.9545, "step": 829 }, { "epoch": 1.041405269761606, "grad_norm": 0.2431640625, "learning_rate": 4.792973651191971e-06, "loss": 0.9563, "step": 830 }, { "epoch": 1.0426599749058971, "grad_norm": 0.228515625, "learning_rate": 4.786700125470515e-06, "loss": 0.9412, "step": 831 }, { "epoch": 1.0439146800501882, "grad_norm": 0.2490234375, "learning_rate": 4.78042659974906e-06, "loss": 0.9827, "step": 832 }, { "epoch": 1.0451693851944792, "grad_norm": 0.275390625, "learning_rate": 4.774153074027604e-06, "loss": 1.0094, "step": 833 }, { "epoch": 1.0464240903387705, "grad_norm": 0.2578125, "learning_rate": 4.767879548306149e-06, "loss": 1.0049, "step": 834 }, { "epoch": 1.0476787954830615, "grad_norm": 0.25, "learning_rate": 4.761606022584693e-06, "loss": 0.9637, "step": 835 }, { "epoch": 1.0489335006273526, "grad_norm": 0.2421875, "learning_rate": 4.755332496863238e-06, "loss": 0.9611, "step": 836 }, { "epoch": 1.0501882057716436, "grad_norm": 0.23828125, "learning_rate": 4.749058971141783e-06, "loss": 0.933, "step": 837 }, { "epoch": 1.0514429109159347, "grad_norm": 0.251953125, "learning_rate": 4.742785445420327e-06, "loss": 0.9988, "step": 838 }, { "epoch": 1.052697616060226, "grad_norm": 0.259765625, "learning_rate": 4.736511919698872e-06, "loss": 0.9268, "step": 839 }, { "epoch": 1.053952321204517, "grad_norm": 0.2421875, "learning_rate": 4.730238393977416e-06, "loss": 0.9228, "step": 840 }, { "epoch": 1.055207026348808, "grad_norm": 0.255859375, "learning_rate": 4.72396486825596e-06, "loss": 0.9905, "step": 841 }, { "epoch": 1.056461731493099, "grad_norm": 0.25, "learning_rate": 4.717691342534505e-06, "loss": 0.9425, "step": 842 }, { "epoch": 1.0577164366373901, "grad_norm": 0.251953125, "learning_rate": 4.711417816813049e-06, "loss": 0.9765, "step": 843 }, { "epoch": 1.0589711417816814, "grad_norm": 0.240234375, "learning_rate": 4.705144291091594e-06, "loss": 0.9834, "step": 844 }, { "epoch": 1.0602258469259724, "grad_norm": 0.2353515625, "learning_rate": 4.698870765370138e-06, "loss": 0.9795, "step": 845 }, { "epoch": 1.0614805520702635, "grad_norm": 0.25, "learning_rate": 4.692597239648683e-06, "loss": 0.9632, "step": 846 }, { "epoch": 1.0627352572145545, "grad_norm": 0.265625, "learning_rate": 4.686323713927228e-06, "loss": 0.9818, "step": 847 }, { "epoch": 1.0639899623588456, "grad_norm": 0.234375, "learning_rate": 4.680050188205772e-06, "loss": 0.9652, "step": 848 }, { "epoch": 1.0652446675031368, "grad_norm": 0.2431640625, "learning_rate": 4.673776662484317e-06, "loss": 0.954, "step": 849 }, { "epoch": 1.066499372647428, "grad_norm": 0.244140625, "learning_rate": 4.667503136762861e-06, "loss": 0.9915, "step": 850 }, { "epoch": 1.067754077791719, "grad_norm": 0.244140625, "learning_rate": 4.661229611041406e-06, "loss": 0.955, "step": 851 }, { "epoch": 1.06900878293601, "grad_norm": 0.248046875, "learning_rate": 4.654956085319951e-06, "loss": 0.9386, "step": 852 }, { "epoch": 1.070263488080301, "grad_norm": 0.2578125, "learning_rate": 4.648682559598495e-06, "loss": 0.9751, "step": 853 }, { "epoch": 1.0715181932245923, "grad_norm": 0.275390625, "learning_rate": 4.64240903387704e-06, "loss": 0.9557, "step": 854 }, { "epoch": 1.0727728983688833, "grad_norm": 0.2412109375, "learning_rate": 4.636135508155584e-06, "loss": 0.9621, "step": 855 }, { "epoch": 1.0740276035131744, "grad_norm": 0.25390625, "learning_rate": 4.629861982434129e-06, "loss": 0.9607, "step": 856 }, { "epoch": 1.0752823086574654, "grad_norm": 0.34765625, "learning_rate": 4.623588456712673e-06, "loss": 1.0342, "step": 857 }, { "epoch": 1.0765370138017567, "grad_norm": 0.2373046875, "learning_rate": 4.617314930991217e-06, "loss": 0.9431, "step": 858 }, { "epoch": 1.0777917189460477, "grad_norm": 0.29296875, "learning_rate": 4.611041405269762e-06, "loss": 0.9379, "step": 859 }, { "epoch": 1.0790464240903388, "grad_norm": 0.240234375, "learning_rate": 4.604767879548306e-06, "loss": 0.94, "step": 860 }, { "epoch": 1.0803011292346298, "grad_norm": 0.31640625, "learning_rate": 4.598494353826851e-06, "loss": 0.9073, "step": 861 }, { "epoch": 1.0815558343789209, "grad_norm": 0.275390625, "learning_rate": 4.592220828105395e-06, "loss": 0.9015, "step": 862 }, { "epoch": 1.0828105395232122, "grad_norm": 0.263671875, "learning_rate": 4.58594730238394e-06, "loss": 0.987, "step": 863 }, { "epoch": 1.0840652446675032, "grad_norm": 0.263671875, "learning_rate": 4.579673776662485e-06, "loss": 0.968, "step": 864 }, { "epoch": 1.0853199498117942, "grad_norm": 0.25390625, "learning_rate": 4.573400250941029e-06, "loss": 0.9476, "step": 865 }, { "epoch": 1.0865746549560853, "grad_norm": 0.244140625, "learning_rate": 4.567126725219574e-06, "loss": 0.98, "step": 866 }, { "epoch": 1.0878293601003763, "grad_norm": 0.2451171875, "learning_rate": 4.560853199498118e-06, "loss": 0.9694, "step": 867 }, { "epoch": 1.0890840652446676, "grad_norm": 0.259765625, "learning_rate": 4.554579673776663e-06, "loss": 0.987, "step": 868 }, { "epoch": 1.0903387703889587, "grad_norm": 0.251953125, "learning_rate": 4.548306148055208e-06, "loss": 0.9491, "step": 869 }, { "epoch": 1.0915934755332497, "grad_norm": 0.240234375, "learning_rate": 4.542032622333752e-06, "loss": 0.9213, "step": 870 }, { "epoch": 1.0928481806775407, "grad_norm": 0.2578125, "learning_rate": 4.535759096612297e-06, "loss": 0.9718, "step": 871 }, { "epoch": 1.0941028858218318, "grad_norm": 0.234375, "learning_rate": 4.529485570890841e-06, "loss": 0.9289, "step": 872 }, { "epoch": 1.095357590966123, "grad_norm": 0.2890625, "learning_rate": 4.523212045169386e-06, "loss": 0.9691, "step": 873 }, { "epoch": 1.096612296110414, "grad_norm": 0.2421875, "learning_rate": 4.51693851944793e-06, "loss": 0.9874, "step": 874 }, { "epoch": 1.0978670012547052, "grad_norm": 0.2451171875, "learning_rate": 4.510664993726475e-06, "loss": 0.9902, "step": 875 }, { "epoch": 1.0991217063989962, "grad_norm": 0.236328125, "learning_rate": 4.504391468005019e-06, "loss": 0.9525, "step": 876 }, { "epoch": 1.1003764115432872, "grad_norm": 0.29296875, "learning_rate": 4.498117942283563e-06, "loss": 0.9714, "step": 877 }, { "epoch": 1.1016311166875785, "grad_norm": 0.25390625, "learning_rate": 4.491844416562108e-06, "loss": 0.9967, "step": 878 }, { "epoch": 1.1028858218318696, "grad_norm": 0.2578125, "learning_rate": 4.485570890840652e-06, "loss": 0.9643, "step": 879 }, { "epoch": 1.1041405269761606, "grad_norm": 0.255859375, "learning_rate": 4.479297365119197e-06, "loss": 0.9867, "step": 880 }, { "epoch": 1.1053952321204517, "grad_norm": 0.23828125, "learning_rate": 4.473023839397742e-06, "loss": 0.9647, "step": 881 }, { "epoch": 1.1066499372647427, "grad_norm": 0.318359375, "learning_rate": 4.466750313676286e-06, "loss": 0.9972, "step": 882 }, { "epoch": 1.107904642409034, "grad_norm": 0.2470703125, "learning_rate": 4.460476787954831e-06, "loss": 0.9866, "step": 883 }, { "epoch": 1.109159347553325, "grad_norm": 0.25390625, "learning_rate": 4.454203262233375e-06, "loss": 1.0036, "step": 884 }, { "epoch": 1.110414052697616, "grad_norm": 0.2451171875, "learning_rate": 4.44792973651192e-06, "loss": 0.9416, "step": 885 }, { "epoch": 1.111668757841907, "grad_norm": 0.2451171875, "learning_rate": 4.441656210790465e-06, "loss": 0.9422, "step": 886 }, { "epoch": 1.1129234629861982, "grad_norm": 0.2421875, "learning_rate": 4.435382685069009e-06, "loss": 0.944, "step": 887 }, { "epoch": 1.1141781681304894, "grad_norm": 0.248046875, "learning_rate": 4.429109159347554e-06, "loss": 0.9853, "step": 888 }, { "epoch": 1.1154328732747805, "grad_norm": 0.263671875, "learning_rate": 4.422835633626098e-06, "loss": 1.0227, "step": 889 }, { "epoch": 1.1166875784190715, "grad_norm": 0.2578125, "learning_rate": 4.416562107904643e-06, "loss": 0.9502, "step": 890 }, { "epoch": 1.1179422835633626, "grad_norm": 0.244140625, "learning_rate": 4.410288582183187e-06, "loss": 0.9464, "step": 891 }, { "epoch": 1.1191969887076536, "grad_norm": 0.3046875, "learning_rate": 4.404015056461732e-06, "loss": 1.0108, "step": 892 }, { "epoch": 1.1204516938519449, "grad_norm": 0.2392578125, "learning_rate": 4.397741530740276e-06, "loss": 1.0134, "step": 893 }, { "epoch": 1.121706398996236, "grad_norm": 0.240234375, "learning_rate": 4.391468005018821e-06, "loss": 0.9499, "step": 894 }, { "epoch": 1.122961104140527, "grad_norm": 0.251953125, "learning_rate": 4.385194479297365e-06, "loss": 0.9403, "step": 895 }, { "epoch": 1.124215809284818, "grad_norm": 0.251953125, "learning_rate": 4.37892095357591e-06, "loss": 1.0122, "step": 896 }, { "epoch": 1.125470514429109, "grad_norm": 0.2578125, "learning_rate": 4.372647427854454e-06, "loss": 0.9437, "step": 897 }, { "epoch": 1.1267252195734003, "grad_norm": 0.244140625, "learning_rate": 4.366373902132999e-06, "loss": 0.9695, "step": 898 }, { "epoch": 1.1279799247176914, "grad_norm": 0.26171875, "learning_rate": 4.360100376411543e-06, "loss": 0.9958, "step": 899 }, { "epoch": 1.1292346298619824, "grad_norm": 0.24609375, "learning_rate": 4.353826850690088e-06, "loss": 0.9655, "step": 900 }, { "epoch": 1.1304893350062735, "grad_norm": 0.2490234375, "learning_rate": 4.347553324968632e-06, "loss": 0.9744, "step": 901 }, { "epoch": 1.1317440401505645, "grad_norm": 0.51171875, "learning_rate": 4.341279799247177e-06, "loss": 0.9729, "step": 902 }, { "epoch": 1.1329987452948558, "grad_norm": 0.2421875, "learning_rate": 4.335006273525722e-06, "loss": 0.9441, "step": 903 }, { "epoch": 1.1342534504391468, "grad_norm": 0.34375, "learning_rate": 4.328732747804266e-06, "loss": 0.9836, "step": 904 }, { "epoch": 1.1355081555834379, "grad_norm": 0.25390625, "learning_rate": 4.322459222082811e-06, "loss": 0.9986, "step": 905 }, { "epoch": 1.136762860727729, "grad_norm": 0.25390625, "learning_rate": 4.316185696361355e-06, "loss": 1.0118, "step": 906 }, { "epoch": 1.13801756587202, "grad_norm": 0.2451171875, "learning_rate": 4.3099121706399e-06, "loss": 0.9114, "step": 907 }, { "epoch": 1.1392722710163112, "grad_norm": 0.2490234375, "learning_rate": 4.303638644918444e-06, "loss": 0.9837, "step": 908 }, { "epoch": 1.1405269761606023, "grad_norm": 0.2470703125, "learning_rate": 4.297365119196989e-06, "loss": 0.9587, "step": 909 }, { "epoch": 1.1417816813048933, "grad_norm": 0.23828125, "learning_rate": 4.291091593475533e-06, "loss": 0.9398, "step": 910 }, { "epoch": 1.1430363864491844, "grad_norm": 0.2421875, "learning_rate": 4.284818067754078e-06, "loss": 0.9205, "step": 911 }, { "epoch": 1.1442910915934754, "grad_norm": 0.41015625, "learning_rate": 4.278544542032622e-06, "loss": 1.0247, "step": 912 }, { "epoch": 1.1455457967377667, "grad_norm": 0.263671875, "learning_rate": 4.272271016311167e-06, "loss": 1.003, "step": 913 }, { "epoch": 1.1468005018820577, "grad_norm": 0.28125, "learning_rate": 4.2659974905897114e-06, "loss": 1.0015, "step": 914 }, { "epoch": 1.1480552070263488, "grad_norm": 0.32421875, "learning_rate": 4.259723964868256e-06, "loss": 1.0455, "step": 915 }, { "epoch": 1.1493099121706398, "grad_norm": 0.24609375, "learning_rate": 4.2534504391468005e-06, "loss": 0.9448, "step": 916 }, { "epoch": 1.1505646173149309, "grad_norm": 0.236328125, "learning_rate": 4.247176913425345e-06, "loss": 0.9234, "step": 917 }, { "epoch": 1.1518193224592221, "grad_norm": 0.25, "learning_rate": 4.24090338770389e-06, "loss": 0.9634, "step": 918 }, { "epoch": 1.1530740276035132, "grad_norm": 0.28515625, "learning_rate": 4.2346298619824344e-06, "loss": 0.9494, "step": 919 }, { "epoch": 1.1543287327478042, "grad_norm": 0.25390625, "learning_rate": 4.228356336260979e-06, "loss": 0.9815, "step": 920 }, { "epoch": 1.1555834378920953, "grad_norm": 0.251953125, "learning_rate": 4.2220828105395235e-06, "loss": 0.9971, "step": 921 }, { "epoch": 1.1568381430363865, "grad_norm": 0.2470703125, "learning_rate": 4.215809284818068e-06, "loss": 0.9297, "step": 922 }, { "epoch": 1.1580928481806776, "grad_norm": 0.244140625, "learning_rate": 4.2095357590966125e-06, "loss": 0.9684, "step": 923 }, { "epoch": 1.1593475533249686, "grad_norm": 0.26171875, "learning_rate": 4.203262233375157e-06, "loss": 0.9706, "step": 924 }, { "epoch": 1.1606022584692597, "grad_norm": 0.3046875, "learning_rate": 4.1969887076537015e-06, "loss": 0.9668, "step": 925 }, { "epoch": 1.1618569636135507, "grad_norm": 0.2578125, "learning_rate": 4.1907151819322464e-06, "loss": 0.9873, "step": 926 }, { "epoch": 1.163111668757842, "grad_norm": 0.2470703125, "learning_rate": 4.1844416562107905e-06, "loss": 0.9361, "step": 927 }, { "epoch": 1.164366373902133, "grad_norm": 0.259765625, "learning_rate": 4.1781681304893355e-06, "loss": 0.979, "step": 928 }, { "epoch": 1.165621079046424, "grad_norm": 0.294921875, "learning_rate": 4.1718946047678796e-06, "loss": 1.0341, "step": 929 }, { "epoch": 1.1668757841907151, "grad_norm": 0.2431640625, "learning_rate": 4.1656210790464245e-06, "loss": 0.9452, "step": 930 }, { "epoch": 1.1681304893350062, "grad_norm": 0.23828125, "learning_rate": 4.159347553324969e-06, "loss": 0.9652, "step": 931 }, { "epoch": 1.1693851944792975, "grad_norm": 0.251953125, "learning_rate": 4.1530740276035135e-06, "loss": 0.9587, "step": 932 }, { "epoch": 1.1706398996235885, "grad_norm": 0.2578125, "learning_rate": 4.146800501882058e-06, "loss": 0.9926, "step": 933 }, { "epoch": 1.1718946047678795, "grad_norm": 0.259765625, "learning_rate": 4.1405269761606026e-06, "loss": 1.0064, "step": 934 }, { "epoch": 1.1731493099121706, "grad_norm": 0.251953125, "learning_rate": 4.1342534504391475e-06, "loss": 0.9898, "step": 935 }, { "epoch": 1.1744040150564616, "grad_norm": 0.255859375, "learning_rate": 4.127979924717692e-06, "loss": 0.9641, "step": 936 }, { "epoch": 1.175658720200753, "grad_norm": 0.29296875, "learning_rate": 4.1217063989962365e-06, "loss": 0.946, "step": 937 }, { "epoch": 1.176913425345044, "grad_norm": 0.255859375, "learning_rate": 4.115432873274781e-06, "loss": 0.9484, "step": 938 }, { "epoch": 1.178168130489335, "grad_norm": 0.25390625, "learning_rate": 4.1091593475533255e-06, "loss": 0.9755, "step": 939 }, { "epoch": 1.179422835633626, "grad_norm": 0.251953125, "learning_rate": 4.1028858218318705e-06, "loss": 0.9571, "step": 940 }, { "epoch": 1.1806775407779173, "grad_norm": 0.23828125, "learning_rate": 4.0966122961104146e-06, "loss": 0.9611, "step": 941 }, { "epoch": 1.1819322459222084, "grad_norm": 0.240234375, "learning_rate": 4.0903387703889595e-06, "loss": 0.961, "step": 942 }, { "epoch": 1.1831869510664994, "grad_norm": 0.2431640625, "learning_rate": 4.084065244667504e-06, "loss": 0.9488, "step": 943 }, { "epoch": 1.1844416562107905, "grad_norm": 0.310546875, "learning_rate": 4.077791718946048e-06, "loss": 0.9263, "step": 944 }, { "epoch": 1.1856963613550815, "grad_norm": 0.369140625, "learning_rate": 4.071518193224593e-06, "loss": 0.958, "step": 945 }, { "epoch": 1.1869510664993728, "grad_norm": 0.267578125, "learning_rate": 4.065244667503137e-06, "loss": 0.9422, "step": 946 }, { "epoch": 1.1882057716436638, "grad_norm": 0.25, "learning_rate": 4.058971141781682e-06, "loss": 0.9805, "step": 947 }, { "epoch": 1.1894604767879549, "grad_norm": 0.259765625, "learning_rate": 4.052697616060226e-06, "loss": 0.9475, "step": 948 }, { "epoch": 1.190715181932246, "grad_norm": 0.275390625, "learning_rate": 4.046424090338771e-06, "loss": 0.99, "step": 949 }, { "epoch": 1.191969887076537, "grad_norm": 0.259765625, "learning_rate": 4.040150564617315e-06, "loss": 0.908, "step": 950 }, { "epoch": 1.1932245922208282, "grad_norm": 0.279296875, "learning_rate": 4.03387703889586e-06, "loss": 0.9607, "step": 951 }, { "epoch": 1.1944792973651193, "grad_norm": 0.2392578125, "learning_rate": 4.027603513174405e-06, "loss": 0.9432, "step": 952 }, { "epoch": 1.1957340025094103, "grad_norm": 0.259765625, "learning_rate": 4.021329987452949e-06, "loss": 1.01, "step": 953 }, { "epoch": 1.1969887076537014, "grad_norm": 0.2490234375, "learning_rate": 4.015056461731494e-06, "loss": 0.9903, "step": 954 }, { "epoch": 1.1982434127979924, "grad_norm": 0.384765625, "learning_rate": 4.008782936010038e-06, "loss": 0.9542, "step": 955 }, { "epoch": 1.1994981179422837, "grad_norm": 0.251953125, "learning_rate": 4.002509410288583e-06, "loss": 0.9573, "step": 956 }, { "epoch": 1.2007528230865747, "grad_norm": 0.25390625, "learning_rate": 3.996235884567128e-06, "loss": 0.9967, "step": 957 }, { "epoch": 1.2020075282308658, "grad_norm": 1.5546875, "learning_rate": 3.989962358845672e-06, "loss": 0.9705, "step": 958 }, { "epoch": 1.2032622333751568, "grad_norm": 0.2734375, "learning_rate": 3.983688833124217e-06, "loss": 1.0042, "step": 959 }, { "epoch": 1.2045169385194479, "grad_norm": 0.26953125, "learning_rate": 3.977415307402761e-06, "loss": 1.0116, "step": 960 }, { "epoch": 1.2057716436637391, "grad_norm": 0.25, "learning_rate": 3.971141781681306e-06, "loss": 0.9606, "step": 961 }, { "epoch": 1.2070263488080302, "grad_norm": 0.251953125, "learning_rate": 3.96486825595985e-06, "loss": 0.9663, "step": 962 }, { "epoch": 1.2082810539523212, "grad_norm": 0.24609375, "learning_rate": 3.958594730238394e-06, "loss": 0.9207, "step": 963 }, { "epoch": 1.2095357590966123, "grad_norm": 0.279296875, "learning_rate": 3.952321204516939e-06, "loss": 0.9735, "step": 964 }, { "epoch": 1.2107904642409033, "grad_norm": 0.25390625, "learning_rate": 3.946047678795483e-06, "loss": 0.9797, "step": 965 }, { "epoch": 1.2120451693851946, "grad_norm": 0.427734375, "learning_rate": 3.939774153074028e-06, "loss": 0.9495, "step": 966 }, { "epoch": 1.2132998745294856, "grad_norm": 0.2490234375, "learning_rate": 3.933500627352572e-06, "loss": 0.9823, "step": 967 }, { "epoch": 1.2145545796737767, "grad_norm": 0.240234375, "learning_rate": 3.927227101631117e-06, "loss": 0.988, "step": 968 }, { "epoch": 1.2158092848180677, "grad_norm": 0.2431640625, "learning_rate": 3.920953575909662e-06, "loss": 0.9397, "step": 969 }, { "epoch": 1.2170639899623588, "grad_norm": 0.24609375, "learning_rate": 3.914680050188206e-06, "loss": 0.9512, "step": 970 }, { "epoch": 1.21831869510665, "grad_norm": 0.275390625, "learning_rate": 3.908406524466751e-06, "loss": 0.9685, "step": 971 }, { "epoch": 1.219573400250941, "grad_norm": 0.251953125, "learning_rate": 3.902132998745295e-06, "loss": 0.9523, "step": 972 }, { "epoch": 1.2208281053952321, "grad_norm": 0.2392578125, "learning_rate": 3.89585947302384e-06, "loss": 0.9816, "step": 973 }, { "epoch": 1.2220828105395232, "grad_norm": 0.26953125, "learning_rate": 3.889585947302385e-06, "loss": 0.9482, "step": 974 }, { "epoch": 1.2233375156838142, "grad_norm": 0.2294921875, "learning_rate": 3.883312421580929e-06, "loss": 0.9453, "step": 975 }, { "epoch": 1.2245922208281055, "grad_norm": 0.2490234375, "learning_rate": 3.877038895859474e-06, "loss": 0.9528, "step": 976 }, { "epoch": 1.2258469259723965, "grad_norm": 0.2431640625, "learning_rate": 3.870765370138018e-06, "loss": 0.9128, "step": 977 }, { "epoch": 1.2271016311166876, "grad_norm": 0.2392578125, "learning_rate": 3.864491844416563e-06, "loss": 0.9655, "step": 978 }, { "epoch": 1.2283563362609786, "grad_norm": 0.25, "learning_rate": 3.858218318695107e-06, "loss": 0.9279, "step": 979 }, { "epoch": 1.2296110414052697, "grad_norm": 0.244140625, "learning_rate": 3.851944792973651e-06, "loss": 0.9709, "step": 980 }, { "epoch": 1.230865746549561, "grad_norm": 0.2421875, "learning_rate": 3.845671267252196e-06, "loss": 0.9353, "step": 981 }, { "epoch": 1.232120451693852, "grad_norm": 0.2412109375, "learning_rate": 3.83939774153074e-06, "loss": 0.9249, "step": 982 }, { "epoch": 1.233375156838143, "grad_norm": 0.2578125, "learning_rate": 3.833124215809285e-06, "loss": 0.9722, "step": 983 }, { "epoch": 1.234629861982434, "grad_norm": 0.2431640625, "learning_rate": 3.82685069008783e-06, "loss": 1.032, "step": 984 }, { "epoch": 1.2358845671267251, "grad_norm": 0.240234375, "learning_rate": 3.820577164366374e-06, "loss": 0.9713, "step": 985 }, { "epoch": 1.2371392722710164, "grad_norm": 0.283203125, "learning_rate": 3.814303638644919e-06, "loss": 0.9622, "step": 986 }, { "epoch": 1.2383939774153074, "grad_norm": 0.244140625, "learning_rate": 3.808030112923463e-06, "loss": 0.9594, "step": 987 }, { "epoch": 1.2396486825595985, "grad_norm": 0.2333984375, "learning_rate": 3.801756587202008e-06, "loss": 0.9673, "step": 988 }, { "epoch": 1.2409033877038895, "grad_norm": 0.244140625, "learning_rate": 3.795483061480552e-06, "loss": 0.9241, "step": 989 }, { "epoch": 1.2421580928481806, "grad_norm": 0.298828125, "learning_rate": 3.789209535759097e-06, "loss": 0.9682, "step": 990 }, { "epoch": 1.2434127979924718, "grad_norm": 0.25, "learning_rate": 3.7829360100376415e-06, "loss": 1.0046, "step": 991 }, { "epoch": 1.2446675031367629, "grad_norm": 0.255859375, "learning_rate": 3.776662484316186e-06, "loss": 0.9744, "step": 992 }, { "epoch": 1.245922208281054, "grad_norm": 0.234375, "learning_rate": 3.7703889585947305e-06, "loss": 0.9611, "step": 993 }, { "epoch": 1.247176913425345, "grad_norm": 0.251953125, "learning_rate": 3.764115432873275e-06, "loss": 1.0081, "step": 994 }, { "epoch": 1.248431618569636, "grad_norm": 0.25390625, "learning_rate": 3.7578419071518196e-06, "loss": 0.9994, "step": 995 }, { "epoch": 1.2496863237139273, "grad_norm": 0.267578125, "learning_rate": 3.7515683814303645e-06, "loss": 0.9351, "step": 996 }, { "epoch": 1.2509410288582183, "grad_norm": 0.25390625, "learning_rate": 3.7452948557089086e-06, "loss": 0.9905, "step": 997 }, { "epoch": 1.2521957340025094, "grad_norm": 0.2451171875, "learning_rate": 3.7390213299874535e-06, "loss": 0.9586, "step": 998 }, { "epoch": 1.2534504391468004, "grad_norm": 0.255859375, "learning_rate": 3.7327478042659976e-06, "loss": 0.9739, "step": 999 }, { "epoch": 1.2547051442910915, "grad_norm": 0.24609375, "learning_rate": 3.7264742785445425e-06, "loss": 0.9457, "step": 1000 }, { "epoch": 1.2559598494353827, "grad_norm": 0.2451171875, "learning_rate": 3.720200752823087e-06, "loss": 0.9679, "step": 1001 }, { "epoch": 1.2572145545796738, "grad_norm": 0.2451171875, "learning_rate": 3.713927227101631e-06, "loss": 1.0412, "step": 1002 }, { "epoch": 1.2584692597239648, "grad_norm": 0.25, "learning_rate": 3.707653701380176e-06, "loss": 0.9722, "step": 1003 }, { "epoch": 1.2597239648682559, "grad_norm": 0.25, "learning_rate": 3.70138017565872e-06, "loss": 1.0239, "step": 1004 }, { "epoch": 1.260978670012547, "grad_norm": 0.248046875, "learning_rate": 3.695106649937265e-06, "loss": 1.0087, "step": 1005 }, { "epoch": 1.2622333751568382, "grad_norm": 0.25, "learning_rate": 3.68883312421581e-06, "loss": 0.9596, "step": 1006 }, { "epoch": 1.2634880803011292, "grad_norm": 0.345703125, "learning_rate": 3.682559598494354e-06, "loss": 0.9607, "step": 1007 }, { "epoch": 1.2647427854454203, "grad_norm": 0.244140625, "learning_rate": 3.6762860727728987e-06, "loss": 0.9711, "step": 1008 }, { "epoch": 1.2659974905897113, "grad_norm": 0.2490234375, "learning_rate": 3.670012547051443e-06, "loss": 1.0, "step": 1009 }, { "epoch": 1.2672521957340024, "grad_norm": 0.248046875, "learning_rate": 3.6637390213299877e-06, "loss": 0.9939, "step": 1010 }, { "epoch": 1.2685069008782937, "grad_norm": 0.2578125, "learning_rate": 3.657465495608532e-06, "loss": 1.0312, "step": 1011 }, { "epoch": 1.2697616060225847, "grad_norm": 0.392578125, "learning_rate": 3.6511919698870767e-06, "loss": 0.9183, "step": 1012 }, { "epoch": 1.2710163111668757, "grad_norm": 0.29296875, "learning_rate": 3.6449184441656216e-06, "loss": 0.9571, "step": 1013 }, { "epoch": 1.272271016311167, "grad_norm": 0.28515625, "learning_rate": 3.6386449184441657e-06, "loss": 0.9335, "step": 1014 }, { "epoch": 1.2735257214554578, "grad_norm": 0.25390625, "learning_rate": 3.6323713927227107e-06, "loss": 0.9993, "step": 1015 }, { "epoch": 1.274780426599749, "grad_norm": 0.265625, "learning_rate": 3.6260978670012548e-06, "loss": 0.925, "step": 1016 }, { "epoch": 1.2760351317440402, "grad_norm": 0.26171875, "learning_rate": 3.6198243412797997e-06, "loss": 0.9771, "step": 1017 }, { "epoch": 1.2772898368883312, "grad_norm": 0.2431640625, "learning_rate": 3.6135508155583442e-06, "loss": 0.9446, "step": 1018 }, { "epoch": 1.2785445420326225, "grad_norm": 0.25, "learning_rate": 3.6072772898368887e-06, "loss": 0.902, "step": 1019 }, { "epoch": 1.2797992471769133, "grad_norm": 0.2470703125, "learning_rate": 3.6010037641154332e-06, "loss": 0.9355, "step": 1020 }, { "epoch": 1.2810539523212046, "grad_norm": 0.515625, "learning_rate": 3.5947302383939773e-06, "loss": 0.9073, "step": 1021 }, { "epoch": 1.2823086574654956, "grad_norm": 0.318359375, "learning_rate": 3.5884567126725223e-06, "loss": 0.9053, "step": 1022 }, { "epoch": 1.2835633626097867, "grad_norm": 0.248046875, "learning_rate": 3.582183186951067e-06, "loss": 0.9813, "step": 1023 }, { "epoch": 1.284818067754078, "grad_norm": 0.2451171875, "learning_rate": 3.5759096612296113e-06, "loss": 0.9838, "step": 1024 }, { "epoch": 1.286072772898369, "grad_norm": 0.24609375, "learning_rate": 3.5696361355081562e-06, "loss": 0.9267, "step": 1025 }, { "epoch": 1.28732747804266, "grad_norm": 0.251953125, "learning_rate": 3.5633626097867003e-06, "loss": 0.9749, "step": 1026 }, { "epoch": 1.288582183186951, "grad_norm": 0.2490234375, "learning_rate": 3.557089084065245e-06, "loss": 0.9172, "step": 1027 }, { "epoch": 1.289836888331242, "grad_norm": 0.24609375, "learning_rate": 3.5508155583437898e-06, "loss": 0.9319, "step": 1028 }, { "epoch": 1.2910915934755334, "grad_norm": 0.25, "learning_rate": 3.544542032622334e-06, "loss": 0.9767, "step": 1029 }, { "epoch": 1.2923462986198244, "grad_norm": 0.3671875, "learning_rate": 3.538268506900879e-06, "loss": 0.9783, "step": 1030 }, { "epoch": 1.2936010037641155, "grad_norm": 0.251953125, "learning_rate": 3.531994981179423e-06, "loss": 0.9616, "step": 1031 }, { "epoch": 1.2948557089084065, "grad_norm": 0.291015625, "learning_rate": 3.525721455457968e-06, "loss": 0.9911, "step": 1032 }, { "epoch": 1.2961104140526976, "grad_norm": 0.259765625, "learning_rate": 3.519447929736512e-06, "loss": 0.9884, "step": 1033 }, { "epoch": 1.2973651191969888, "grad_norm": 0.24609375, "learning_rate": 3.513174404015057e-06, "loss": 0.9305, "step": 1034 }, { "epoch": 1.2986198243412799, "grad_norm": 0.2451171875, "learning_rate": 3.5069008782936014e-06, "loss": 0.9713, "step": 1035 }, { "epoch": 1.299874529485571, "grad_norm": 0.26171875, "learning_rate": 3.500627352572146e-06, "loss": 0.945, "step": 1036 }, { "epoch": 1.301129234629862, "grad_norm": 0.32421875, "learning_rate": 3.4943538268506904e-06, "loss": 0.9704, "step": 1037 }, { "epoch": 1.302383939774153, "grad_norm": 0.2470703125, "learning_rate": 3.4880803011292345e-06, "loss": 0.9707, "step": 1038 }, { "epoch": 1.3036386449184443, "grad_norm": 0.255859375, "learning_rate": 3.4818067754077794e-06, "loss": 0.9814, "step": 1039 }, { "epoch": 1.3048933500627353, "grad_norm": 0.255859375, "learning_rate": 3.4755332496863244e-06, "loss": 0.9663, "step": 1040 }, { "epoch": 1.3061480552070264, "grad_norm": 0.248046875, "learning_rate": 3.4692597239648684e-06, "loss": 0.9575, "step": 1041 }, { "epoch": 1.3074027603513174, "grad_norm": 0.2490234375, "learning_rate": 3.4629861982434134e-06, "loss": 0.9844, "step": 1042 }, { "epoch": 1.3086574654956085, "grad_norm": 0.259765625, "learning_rate": 3.4567126725219575e-06, "loss": 0.9874, "step": 1043 }, { "epoch": 1.3099121706398997, "grad_norm": 0.259765625, "learning_rate": 3.450439146800502e-06, "loss": 0.9813, "step": 1044 }, { "epoch": 1.3111668757841908, "grad_norm": 0.26953125, "learning_rate": 3.444165621079047e-06, "loss": 0.9981, "step": 1045 }, { "epoch": 1.3124215809284818, "grad_norm": 0.25390625, "learning_rate": 3.437892095357591e-06, "loss": 0.9913, "step": 1046 }, { "epoch": 1.3136762860727729, "grad_norm": 0.251953125, "learning_rate": 3.431618569636136e-06, "loss": 0.9307, "step": 1047 }, { "epoch": 1.314930991217064, "grad_norm": 0.25390625, "learning_rate": 3.42534504391468e-06, "loss": 0.9607, "step": 1048 }, { "epoch": 1.3161856963613552, "grad_norm": 0.2333984375, "learning_rate": 3.419071518193225e-06, "loss": 0.9354, "step": 1049 }, { "epoch": 1.3174404015056462, "grad_norm": 0.25, "learning_rate": 3.4127979924717695e-06, "loss": 0.9736, "step": 1050 }, { "epoch": 1.3186951066499373, "grad_norm": 0.2470703125, "learning_rate": 3.406524466750314e-06, "loss": 0.9583, "step": 1051 }, { "epoch": 1.3199498117942283, "grad_norm": 0.244140625, "learning_rate": 3.4002509410288585e-06, "loss": 0.9409, "step": 1052 }, { "epoch": 1.3212045169385194, "grad_norm": 0.291015625, "learning_rate": 3.393977415307403e-06, "loss": 1.038, "step": 1053 }, { "epoch": 1.3224592220828106, "grad_norm": 0.2490234375, "learning_rate": 3.3877038895859475e-06, "loss": 1.0157, "step": 1054 }, { "epoch": 1.3237139272271017, "grad_norm": 0.2490234375, "learning_rate": 3.381430363864492e-06, "loss": 0.9303, "step": 1055 }, { "epoch": 1.3249686323713927, "grad_norm": 0.27734375, "learning_rate": 3.3751568381430366e-06, "loss": 0.9446, "step": 1056 }, { "epoch": 1.3262233375156838, "grad_norm": 0.25, "learning_rate": 3.3688833124215815e-06, "loss": 0.9686, "step": 1057 }, { "epoch": 1.3274780426599748, "grad_norm": 0.248046875, "learning_rate": 3.3626097867001256e-06, "loss": 0.9718, "step": 1058 }, { "epoch": 1.328732747804266, "grad_norm": 0.298828125, "learning_rate": 3.3563362609786705e-06, "loss": 0.99, "step": 1059 }, { "epoch": 1.3299874529485571, "grad_norm": 0.2578125, "learning_rate": 3.3500627352572146e-06, "loss": 0.9332, "step": 1060 }, { "epoch": 1.3312421580928482, "grad_norm": 0.2392578125, "learning_rate": 3.3437892095357596e-06, "loss": 0.9631, "step": 1061 }, { "epoch": 1.3324968632371392, "grad_norm": 0.2490234375, "learning_rate": 3.337515683814304e-06, "loss": 0.9952, "step": 1062 }, { "epoch": 1.3337515683814303, "grad_norm": 0.259765625, "learning_rate": 3.331242158092848e-06, "loss": 0.9517, "step": 1063 }, { "epoch": 1.3350062735257215, "grad_norm": 0.24609375, "learning_rate": 3.324968632371393e-06, "loss": 0.937, "step": 1064 }, { "epoch": 1.3362609786700126, "grad_norm": 0.25, "learning_rate": 3.318695106649937e-06, "loss": 0.9697, "step": 1065 }, { "epoch": 1.3375156838143036, "grad_norm": 0.2451171875, "learning_rate": 3.312421580928482e-06, "loss": 0.9468, "step": 1066 }, { "epoch": 1.3387703889585947, "grad_norm": 0.267578125, "learning_rate": 3.306148055207027e-06, "loss": 0.9503, "step": 1067 }, { "epoch": 1.3400250941028857, "grad_norm": 0.2412109375, "learning_rate": 3.299874529485571e-06, "loss": 0.9508, "step": 1068 }, { "epoch": 1.341279799247177, "grad_norm": 0.2490234375, "learning_rate": 3.2936010037641157e-06, "loss": 0.9755, "step": 1069 }, { "epoch": 1.342534504391468, "grad_norm": 0.259765625, "learning_rate": 3.28732747804266e-06, "loss": 0.9451, "step": 1070 }, { "epoch": 1.343789209535759, "grad_norm": 0.271484375, "learning_rate": 3.2810539523212047e-06, "loss": 0.9252, "step": 1071 }, { "epoch": 1.3450439146800501, "grad_norm": 0.265625, "learning_rate": 3.2747804265997496e-06, "loss": 0.9605, "step": 1072 }, { "epoch": 1.3462986198243412, "grad_norm": 0.279296875, "learning_rate": 3.2685069008782937e-06, "loss": 0.9478, "step": 1073 }, { "epoch": 1.3475533249686324, "grad_norm": 0.369140625, "learning_rate": 3.2622333751568387e-06, "loss": 0.9619, "step": 1074 }, { "epoch": 1.3488080301129235, "grad_norm": 0.2470703125, "learning_rate": 3.2559598494353827e-06, "loss": 0.9609, "step": 1075 }, { "epoch": 1.3500627352572145, "grad_norm": 0.2421875, "learning_rate": 3.2496863237139277e-06, "loss": 0.9667, "step": 1076 }, { "epoch": 1.3513174404015056, "grad_norm": 0.255859375, "learning_rate": 3.2434127979924718e-06, "loss": 0.9261, "step": 1077 }, { "epoch": 1.3525721455457966, "grad_norm": 0.259765625, "learning_rate": 3.2371392722710167e-06, "loss": 0.9483, "step": 1078 }, { "epoch": 1.353826850690088, "grad_norm": 0.25, "learning_rate": 3.2308657465495612e-06, "loss": 0.9661, "step": 1079 }, { "epoch": 1.355081555834379, "grad_norm": 0.251953125, "learning_rate": 3.2245922208281057e-06, "loss": 0.9574, "step": 1080 }, { "epoch": 1.35633626097867, "grad_norm": 0.259765625, "learning_rate": 3.2183186951066502e-06, "loss": 0.9856, "step": 1081 }, { "epoch": 1.357590966122961, "grad_norm": 0.259765625, "learning_rate": 3.2120451693851943e-06, "loss": 1.0208, "step": 1082 }, { "epoch": 1.358845671267252, "grad_norm": 0.2421875, "learning_rate": 3.2057716436637393e-06, "loss": 0.9416, "step": 1083 }, { "epoch": 1.3601003764115434, "grad_norm": 0.25390625, "learning_rate": 3.1994981179422842e-06, "loss": 1.0186, "step": 1084 }, { "epoch": 1.3613550815558344, "grad_norm": 0.2431640625, "learning_rate": 3.1932245922208283e-06, "loss": 0.9374, "step": 1085 }, { "epoch": 1.3626097867001254, "grad_norm": 0.2734375, "learning_rate": 3.186951066499373e-06, "loss": 0.9936, "step": 1086 }, { "epoch": 1.3638644918444165, "grad_norm": 0.25, "learning_rate": 3.1806775407779173e-06, "loss": 0.9828, "step": 1087 }, { "epoch": 1.3651191969887075, "grad_norm": 0.28515625, "learning_rate": 3.174404015056462e-06, "loss": 1.0003, "step": 1088 }, { "epoch": 1.3663739021329988, "grad_norm": 0.271484375, "learning_rate": 3.1681304893350068e-06, "loss": 0.9377, "step": 1089 }, { "epoch": 1.3676286072772899, "grad_norm": 0.248046875, "learning_rate": 3.161856963613551e-06, "loss": 0.9564, "step": 1090 }, { "epoch": 1.368883312421581, "grad_norm": 0.244140625, "learning_rate": 3.155583437892096e-06, "loss": 0.956, "step": 1091 }, { "epoch": 1.370138017565872, "grad_norm": 0.248046875, "learning_rate": 3.14930991217064e-06, "loss": 0.945, "step": 1092 }, { "epoch": 1.371392722710163, "grad_norm": 0.2578125, "learning_rate": 3.143036386449185e-06, "loss": 0.9223, "step": 1093 }, { "epoch": 1.3726474278544543, "grad_norm": 0.306640625, "learning_rate": 3.1367628607277293e-06, "loss": 0.9386, "step": 1094 }, { "epoch": 1.3739021329987453, "grad_norm": 0.255859375, "learning_rate": 3.130489335006274e-06, "loss": 0.9042, "step": 1095 }, { "epoch": 1.3751568381430364, "grad_norm": 0.255859375, "learning_rate": 3.1242158092848184e-06, "loss": 1.025, "step": 1096 }, { "epoch": 1.3764115432873274, "grad_norm": 0.244140625, "learning_rate": 3.117942283563363e-06, "loss": 0.9527, "step": 1097 }, { "epoch": 1.3776662484316184, "grad_norm": 0.2490234375, "learning_rate": 3.1116687578419074e-06, "loss": 0.9326, "step": 1098 }, { "epoch": 1.3789209535759097, "grad_norm": 0.2392578125, "learning_rate": 3.1053952321204515e-06, "loss": 0.9553, "step": 1099 }, { "epoch": 1.3801756587202008, "grad_norm": 0.2490234375, "learning_rate": 3.0991217063989964e-06, "loss": 0.9422, "step": 1100 }, { "epoch": 1.3814303638644918, "grad_norm": 0.296875, "learning_rate": 3.0928481806775414e-06, "loss": 0.9268, "step": 1101 }, { "epoch": 1.382685069008783, "grad_norm": 0.236328125, "learning_rate": 3.0865746549560855e-06, "loss": 0.9763, "step": 1102 }, { "epoch": 1.383939774153074, "grad_norm": 0.24609375, "learning_rate": 3.0803011292346304e-06, "loss": 0.9857, "step": 1103 }, { "epoch": 1.3851944792973652, "grad_norm": 0.25390625, "learning_rate": 3.0740276035131745e-06, "loss": 0.9781, "step": 1104 }, { "epoch": 1.3864491844416562, "grad_norm": 0.271484375, "learning_rate": 3.067754077791719e-06, "loss": 0.9473, "step": 1105 }, { "epoch": 1.3877038895859473, "grad_norm": 0.25390625, "learning_rate": 3.061480552070264e-06, "loss": 0.8954, "step": 1106 }, { "epoch": 1.3889585947302385, "grad_norm": 0.244140625, "learning_rate": 3.055207026348808e-06, "loss": 0.9166, "step": 1107 }, { "epoch": 1.3902132998745294, "grad_norm": 0.279296875, "learning_rate": 3.048933500627353e-06, "loss": 0.9358, "step": 1108 }, { "epoch": 1.3914680050188206, "grad_norm": 0.255859375, "learning_rate": 3.042659974905897e-06, "loss": 0.9622, "step": 1109 }, { "epoch": 1.3927227101631117, "grad_norm": 0.30078125, "learning_rate": 3.036386449184442e-06, "loss": 0.9681, "step": 1110 }, { "epoch": 1.3939774153074027, "grad_norm": 0.2431640625, "learning_rate": 3.0301129234629865e-06, "loss": 0.9602, "step": 1111 }, { "epoch": 1.395232120451694, "grad_norm": 0.248046875, "learning_rate": 3.023839397741531e-06, "loss": 0.9538, "step": 1112 }, { "epoch": 1.396486825595985, "grad_norm": 0.267578125, "learning_rate": 3.0175658720200755e-06, "loss": 0.962, "step": 1113 }, { "epoch": 1.397741530740276, "grad_norm": 0.2421875, "learning_rate": 3.01129234629862e-06, "loss": 0.9696, "step": 1114 }, { "epoch": 1.3989962358845671, "grad_norm": 0.24609375, "learning_rate": 3.0050188205771646e-06, "loss": 0.9559, "step": 1115 }, { "epoch": 1.4002509410288582, "grad_norm": 0.251953125, "learning_rate": 2.9987452948557095e-06, "loss": 0.9802, "step": 1116 }, { "epoch": 1.4015056461731494, "grad_norm": 0.2578125, "learning_rate": 2.9924717691342536e-06, "loss": 0.9942, "step": 1117 }, { "epoch": 1.4027603513174405, "grad_norm": 0.25390625, "learning_rate": 2.9861982434127985e-06, "loss": 1.0182, "step": 1118 }, { "epoch": 1.4040150564617315, "grad_norm": 0.2490234375, "learning_rate": 2.9799247176913426e-06, "loss": 0.9157, "step": 1119 }, { "epoch": 1.4052697616060226, "grad_norm": 0.298828125, "learning_rate": 2.9736511919698875e-06, "loss": 0.9836, "step": 1120 }, { "epoch": 1.4065244667503136, "grad_norm": 0.259765625, "learning_rate": 2.9673776662484316e-06, "loss": 0.9315, "step": 1121 }, { "epoch": 1.4077791718946049, "grad_norm": 0.2431640625, "learning_rate": 2.9611041405269766e-06, "loss": 0.9454, "step": 1122 }, { "epoch": 1.409033877038896, "grad_norm": 0.283203125, "learning_rate": 2.954830614805521e-06, "loss": 1.0034, "step": 1123 }, { "epoch": 1.410288582183187, "grad_norm": 0.26171875, "learning_rate": 2.948557089084065e-06, "loss": 0.9495, "step": 1124 }, { "epoch": 1.411543287327478, "grad_norm": 0.2734375, "learning_rate": 2.94228356336261e-06, "loss": 0.9241, "step": 1125 }, { "epoch": 1.412797992471769, "grad_norm": 1.2421875, "learning_rate": 2.936010037641154e-06, "loss": 0.9795, "step": 1126 }, { "epoch": 1.4140526976160603, "grad_norm": 0.25390625, "learning_rate": 2.929736511919699e-06, "loss": 0.978, "step": 1127 }, { "epoch": 1.4153074027603514, "grad_norm": 0.2373046875, "learning_rate": 2.923462986198244e-06, "loss": 0.9545, "step": 1128 }, { "epoch": 1.4165621079046424, "grad_norm": 0.263671875, "learning_rate": 2.917189460476788e-06, "loss": 0.9489, "step": 1129 }, { "epoch": 1.4178168130489335, "grad_norm": 0.302734375, "learning_rate": 2.9109159347553327e-06, "loss": 0.9899, "step": 1130 }, { "epoch": 1.4190715181932245, "grad_norm": 0.259765625, "learning_rate": 2.904642409033877e-06, "loss": 1.0053, "step": 1131 }, { "epoch": 1.4203262233375158, "grad_norm": 0.2578125, "learning_rate": 2.8983688833124217e-06, "loss": 0.9833, "step": 1132 }, { "epoch": 1.4215809284818068, "grad_norm": 0.2470703125, "learning_rate": 2.8920953575909666e-06, "loss": 0.9731, "step": 1133 }, { "epoch": 1.4228356336260979, "grad_norm": 0.255859375, "learning_rate": 2.8858218318695107e-06, "loss": 0.972, "step": 1134 }, { "epoch": 1.424090338770389, "grad_norm": 0.2421875, "learning_rate": 2.8795483061480557e-06, "loss": 1.0048, "step": 1135 }, { "epoch": 1.42534504391468, "grad_norm": 0.255859375, "learning_rate": 2.8732747804265998e-06, "loss": 1.0152, "step": 1136 }, { "epoch": 1.4265997490589712, "grad_norm": 0.251953125, "learning_rate": 2.8670012547051447e-06, "loss": 0.9834, "step": 1137 }, { "epoch": 1.4278544542032623, "grad_norm": 0.26953125, "learning_rate": 2.860727728983689e-06, "loss": 0.9915, "step": 1138 }, { "epoch": 1.4291091593475533, "grad_norm": 0.2421875, "learning_rate": 2.8544542032622337e-06, "loss": 0.9345, "step": 1139 }, { "epoch": 1.4303638644918444, "grad_norm": 0.259765625, "learning_rate": 2.8481806775407782e-06, "loss": 1.0115, "step": 1140 }, { "epoch": 1.4316185696361354, "grad_norm": 0.24609375, "learning_rate": 2.8419071518193223e-06, "loss": 0.9466, "step": 1141 }, { "epoch": 1.4328732747804267, "grad_norm": 0.25390625, "learning_rate": 2.8356336260978673e-06, "loss": 0.9742, "step": 1142 }, { "epoch": 1.4341279799247177, "grad_norm": 0.2578125, "learning_rate": 2.8293601003764113e-06, "loss": 0.922, "step": 1143 }, { "epoch": 1.4353826850690088, "grad_norm": 0.2578125, "learning_rate": 2.8230865746549563e-06, "loss": 0.9815, "step": 1144 }, { "epoch": 1.4366373902132998, "grad_norm": 0.2578125, "learning_rate": 2.8168130489335012e-06, "loss": 1.0049, "step": 1145 }, { "epoch": 1.4378920953575909, "grad_norm": 0.25, "learning_rate": 2.8105395232120453e-06, "loss": 0.9876, "step": 1146 }, { "epoch": 1.4391468005018822, "grad_norm": 0.25, "learning_rate": 2.80426599749059e-06, "loss": 0.8959, "step": 1147 }, { "epoch": 1.4404015056461732, "grad_norm": 0.251953125, "learning_rate": 2.7979924717691343e-06, "loss": 0.9703, "step": 1148 }, { "epoch": 1.4416562107904642, "grad_norm": 0.25390625, "learning_rate": 2.791718946047679e-06, "loss": 0.9734, "step": 1149 }, { "epoch": 1.4429109159347553, "grad_norm": 0.25, "learning_rate": 2.7854454203262238e-06, "loss": 1.0222, "step": 1150 }, { "epoch": 1.4441656210790463, "grad_norm": 0.3515625, "learning_rate": 2.779171894604768e-06, "loss": 0.9149, "step": 1151 }, { "epoch": 1.4454203262233376, "grad_norm": 0.2490234375, "learning_rate": 2.772898368883313e-06, "loss": 1.0035, "step": 1152 }, { "epoch": 1.4466750313676286, "grad_norm": 0.2578125, "learning_rate": 2.766624843161857e-06, "loss": 0.9868, "step": 1153 }, { "epoch": 1.4479297365119197, "grad_norm": 0.244140625, "learning_rate": 2.760351317440402e-06, "loss": 0.9263, "step": 1154 }, { "epoch": 1.4491844416562107, "grad_norm": 0.23828125, "learning_rate": 2.7540777917189464e-06, "loss": 0.9621, "step": 1155 }, { "epoch": 1.4504391468005018, "grad_norm": 0.2373046875, "learning_rate": 2.747804265997491e-06, "loss": 0.9059, "step": 1156 }, { "epoch": 1.451693851944793, "grad_norm": 0.259765625, "learning_rate": 2.7415307402760354e-06, "loss": 1.0311, "step": 1157 }, { "epoch": 1.452948557089084, "grad_norm": 0.244140625, "learning_rate": 2.73525721455458e-06, "loss": 0.9385, "step": 1158 }, { "epoch": 1.4542032622333751, "grad_norm": 0.23828125, "learning_rate": 2.7289836888331244e-06, "loss": 0.9537, "step": 1159 }, { "epoch": 1.4554579673776662, "grad_norm": 0.2490234375, "learning_rate": 2.7227101631116693e-06, "loss": 0.9951, "step": 1160 }, { "epoch": 1.4567126725219572, "grad_norm": 0.255859375, "learning_rate": 2.7164366373902134e-06, "loss": 1.0036, "step": 1161 }, { "epoch": 1.4579673776662485, "grad_norm": 0.23828125, "learning_rate": 2.7101631116687584e-06, "loss": 0.9326, "step": 1162 }, { "epoch": 1.4592220828105396, "grad_norm": 0.23828125, "learning_rate": 2.7038895859473025e-06, "loss": 0.9287, "step": 1163 }, { "epoch": 1.4604767879548306, "grad_norm": 0.2490234375, "learning_rate": 2.6976160602258474e-06, "loss": 0.9248, "step": 1164 }, { "epoch": 1.4617314930991216, "grad_norm": 0.2431640625, "learning_rate": 2.6913425345043915e-06, "loss": 0.9414, "step": 1165 }, { "epoch": 1.4629861982434127, "grad_norm": 0.2578125, "learning_rate": 2.685069008782936e-06, "loss": 0.9666, "step": 1166 }, { "epoch": 1.464240903387704, "grad_norm": 0.24609375, "learning_rate": 2.678795483061481e-06, "loss": 0.974, "step": 1167 }, { "epoch": 1.465495608531995, "grad_norm": 0.28125, "learning_rate": 2.672521957340025e-06, "loss": 0.949, "step": 1168 }, { "epoch": 1.466750313676286, "grad_norm": 0.330078125, "learning_rate": 2.66624843161857e-06, "loss": 0.9924, "step": 1169 }, { "epoch": 1.468005018820577, "grad_norm": 0.251953125, "learning_rate": 2.659974905897114e-06, "loss": 0.9851, "step": 1170 }, { "epoch": 1.4692597239648681, "grad_norm": 0.259765625, "learning_rate": 2.653701380175659e-06, "loss": 0.9361, "step": 1171 }, { "epoch": 1.4705144291091594, "grad_norm": 0.26171875, "learning_rate": 2.6474278544542035e-06, "loss": 0.9351, "step": 1172 }, { "epoch": 1.4717691342534505, "grad_norm": 0.251953125, "learning_rate": 2.641154328732748e-06, "loss": 0.9592, "step": 1173 }, { "epoch": 1.4730238393977415, "grad_norm": 0.2421875, "learning_rate": 2.6348808030112925e-06, "loss": 0.9526, "step": 1174 }, { "epoch": 1.4742785445420326, "grad_norm": 0.27734375, "learning_rate": 2.628607277289837e-06, "loss": 0.9207, "step": 1175 }, { "epoch": 1.4755332496863236, "grad_norm": 0.341796875, "learning_rate": 2.6223337515683816e-06, "loss": 0.9522, "step": 1176 }, { "epoch": 1.4767879548306149, "grad_norm": 0.259765625, "learning_rate": 2.6160602258469265e-06, "loss": 0.9745, "step": 1177 }, { "epoch": 1.478042659974906, "grad_norm": 0.2490234375, "learning_rate": 2.6097867001254706e-06, "loss": 0.9612, "step": 1178 }, { "epoch": 1.479297365119197, "grad_norm": 0.236328125, "learning_rate": 2.6035131744040155e-06, "loss": 0.9826, "step": 1179 }, { "epoch": 1.480552070263488, "grad_norm": 0.29296875, "learning_rate": 2.5972396486825596e-06, "loss": 0.9323, "step": 1180 }, { "epoch": 1.481806775407779, "grad_norm": 0.263671875, "learning_rate": 2.5909661229611045e-06, "loss": 0.9507, "step": 1181 }, { "epoch": 1.4830614805520703, "grad_norm": 0.255859375, "learning_rate": 2.584692597239649e-06, "loss": 0.943, "step": 1182 }, { "epoch": 1.4843161856963614, "grad_norm": 0.2470703125, "learning_rate": 2.5784190715181936e-06, "loss": 0.9428, "step": 1183 }, { "epoch": 1.4855708908406524, "grad_norm": 0.25, "learning_rate": 2.572145545796738e-06, "loss": 0.991, "step": 1184 }, { "epoch": 1.4868255959849435, "grad_norm": 0.2431640625, "learning_rate": 2.565872020075282e-06, "loss": 0.9269, "step": 1185 }, { "epoch": 1.4880803011292345, "grad_norm": 0.251953125, "learning_rate": 2.559598494353827e-06, "loss": 0.9356, "step": 1186 }, { "epoch": 1.4893350062735258, "grad_norm": 0.244140625, "learning_rate": 2.553324968632371e-06, "loss": 0.9683, "step": 1187 }, { "epoch": 1.4905897114178168, "grad_norm": 0.24609375, "learning_rate": 2.547051442910916e-06, "loss": 1.0149, "step": 1188 }, { "epoch": 1.4918444165621079, "grad_norm": 0.255859375, "learning_rate": 2.540777917189461e-06, "loss": 0.9582, "step": 1189 }, { "epoch": 1.4930991217063991, "grad_norm": 0.265625, "learning_rate": 2.534504391468005e-06, "loss": 0.9882, "step": 1190 }, { "epoch": 1.49435382685069, "grad_norm": 0.2734375, "learning_rate": 2.5282308657465497e-06, "loss": 0.9486, "step": 1191 }, { "epoch": 1.4956085319949812, "grad_norm": 0.294921875, "learning_rate": 2.521957340025094e-06, "loss": 0.9432, "step": 1192 }, { "epoch": 1.4968632371392723, "grad_norm": 0.26953125, "learning_rate": 2.5156838143036387e-06, "loss": 0.9537, "step": 1193 }, { "epoch": 1.4981179422835633, "grad_norm": 0.25390625, "learning_rate": 2.5094102885821836e-06, "loss": 0.9405, "step": 1194 }, { "epoch": 1.4993726474278546, "grad_norm": 0.2490234375, "learning_rate": 2.5031367628607277e-06, "loss": 0.9461, "step": 1195 }, { "epoch": 1.5006273525721454, "grad_norm": 0.265625, "learning_rate": 2.4968632371392727e-06, "loss": 0.9409, "step": 1196 }, { "epoch": 1.5018820577164367, "grad_norm": 0.2578125, "learning_rate": 2.490589711417817e-06, "loss": 0.9688, "step": 1197 }, { "epoch": 1.5031367628607277, "grad_norm": 0.255859375, "learning_rate": 2.4843161856963617e-06, "loss": 0.9342, "step": 1198 }, { "epoch": 1.5043914680050188, "grad_norm": 0.28515625, "learning_rate": 2.4780426599749062e-06, "loss": 1.0539, "step": 1199 }, { "epoch": 1.50564617314931, "grad_norm": 0.259765625, "learning_rate": 2.4717691342534507e-06, "loss": 1.0172, "step": 1200 }, { "epoch": 1.5069008782936009, "grad_norm": 0.25, "learning_rate": 2.4654956085319952e-06, "loss": 0.9835, "step": 1201 }, { "epoch": 1.5081555834378921, "grad_norm": 0.25390625, "learning_rate": 2.4592220828105398e-06, "loss": 0.9177, "step": 1202 }, { "epoch": 1.5094102885821832, "grad_norm": 0.2421875, "learning_rate": 2.4529485570890843e-06, "loss": 0.9269, "step": 1203 }, { "epoch": 1.5106649937264742, "grad_norm": 0.33203125, "learning_rate": 2.4466750313676288e-06, "loss": 0.9023, "step": 1204 }, { "epoch": 1.5119196988707655, "grad_norm": 0.251953125, "learning_rate": 2.4404015056461733e-06, "loss": 0.9453, "step": 1205 }, { "epoch": 1.5131744040150563, "grad_norm": 0.259765625, "learning_rate": 2.434127979924718e-06, "loss": 0.9538, "step": 1206 }, { "epoch": 1.5144291091593476, "grad_norm": 0.25, "learning_rate": 2.4278544542032627e-06, "loss": 1.0016, "step": 1207 }, { "epoch": 1.5156838143036386, "grad_norm": 0.2578125, "learning_rate": 2.421580928481807e-06, "loss": 1.0096, "step": 1208 }, { "epoch": 1.5169385194479297, "grad_norm": 0.2490234375, "learning_rate": 2.4153074027603513e-06, "loss": 0.9551, "step": 1209 }, { "epoch": 1.518193224592221, "grad_norm": 0.26953125, "learning_rate": 2.409033877038896e-06, "loss": 0.9705, "step": 1210 }, { "epoch": 1.5194479297365118, "grad_norm": 0.2431640625, "learning_rate": 2.4027603513174404e-06, "loss": 0.9716, "step": 1211 }, { "epoch": 1.520702634880803, "grad_norm": 0.244140625, "learning_rate": 2.3964868255959853e-06, "loss": 0.957, "step": 1212 }, { "epoch": 1.521957340025094, "grad_norm": 0.400390625, "learning_rate": 2.39021329987453e-06, "loss": 1.015, "step": 1213 }, { "epoch": 1.5232120451693851, "grad_norm": 0.2431640625, "learning_rate": 2.3839397741530743e-06, "loss": 0.9109, "step": 1214 }, { "epoch": 1.5244667503136764, "grad_norm": 0.2333984375, "learning_rate": 2.377666248431619e-06, "loss": 0.9318, "step": 1215 }, { "epoch": 1.5257214554579672, "grad_norm": 0.2421875, "learning_rate": 2.3713927227101634e-06, "loss": 0.9286, "step": 1216 }, { "epoch": 1.5269761606022585, "grad_norm": 0.255859375, "learning_rate": 2.365119196988708e-06, "loss": 0.9275, "step": 1217 }, { "epoch": 1.5282308657465495, "grad_norm": 0.2421875, "learning_rate": 2.3588456712672524e-06, "loss": 0.9726, "step": 1218 }, { "epoch": 1.5294855708908406, "grad_norm": 0.255859375, "learning_rate": 2.352572145545797e-06, "loss": 0.9369, "step": 1219 }, { "epoch": 1.5307402760351319, "grad_norm": 0.2578125, "learning_rate": 2.3462986198243414e-06, "loss": 0.9939, "step": 1220 }, { "epoch": 1.5319949811794227, "grad_norm": 0.25390625, "learning_rate": 2.340025094102886e-06, "loss": 1.0024, "step": 1221 }, { "epoch": 1.533249686323714, "grad_norm": 0.265625, "learning_rate": 2.3337515683814304e-06, "loss": 0.9675, "step": 1222 }, { "epoch": 1.534504391468005, "grad_norm": 0.2490234375, "learning_rate": 2.3274780426599754e-06, "loss": 0.9614, "step": 1223 }, { "epoch": 1.535759096612296, "grad_norm": 0.3046875, "learning_rate": 2.32120451693852e-06, "loss": 0.9769, "step": 1224 }, { "epoch": 1.5370138017565873, "grad_norm": 0.2431640625, "learning_rate": 2.3149309912170644e-06, "loss": 0.9687, "step": 1225 }, { "epoch": 1.5382685069008781, "grad_norm": 0.255859375, "learning_rate": 2.3086574654956085e-06, "loss": 0.8834, "step": 1226 }, { "epoch": 1.5395232120451694, "grad_norm": 0.28125, "learning_rate": 2.302383939774153e-06, "loss": 0.9589, "step": 1227 }, { "epoch": 1.5407779171894604, "grad_norm": 0.24609375, "learning_rate": 2.2961104140526975e-06, "loss": 0.9559, "step": 1228 }, { "epoch": 1.5420326223337515, "grad_norm": 0.2490234375, "learning_rate": 2.2898368883312425e-06, "loss": 0.9304, "step": 1229 }, { "epoch": 1.5432873274780428, "grad_norm": 0.2490234375, "learning_rate": 2.283563362609787e-06, "loss": 0.9788, "step": 1230 }, { "epoch": 1.5445420326223338, "grad_norm": 0.26171875, "learning_rate": 2.2772898368883315e-06, "loss": 0.9578, "step": 1231 }, { "epoch": 1.5457967377666249, "grad_norm": 0.25, "learning_rate": 2.271016311166876e-06, "loss": 1.0242, "step": 1232 }, { "epoch": 1.547051442910916, "grad_norm": 0.31640625, "learning_rate": 2.2647427854454205e-06, "loss": 0.9487, "step": 1233 }, { "epoch": 1.548306148055207, "grad_norm": 0.2353515625, "learning_rate": 2.258469259723965e-06, "loss": 0.965, "step": 1234 }, { "epoch": 1.5495608531994982, "grad_norm": 0.25390625, "learning_rate": 2.2521957340025095e-06, "loss": 0.9557, "step": 1235 }, { "epoch": 1.5508155583437893, "grad_norm": 0.2421875, "learning_rate": 2.245922208281054e-06, "loss": 0.9822, "step": 1236 }, { "epoch": 1.5520702634880803, "grad_norm": 0.2578125, "learning_rate": 2.2396486825595986e-06, "loss": 1.012, "step": 1237 }, { "epoch": 1.5533249686323714, "grad_norm": 0.2470703125, "learning_rate": 2.233375156838143e-06, "loss": 0.9867, "step": 1238 }, { "epoch": 1.5545796737766624, "grad_norm": 0.25, "learning_rate": 2.2271016311166876e-06, "loss": 0.9681, "step": 1239 }, { "epoch": 1.5558343789209537, "grad_norm": 0.259765625, "learning_rate": 2.2208281053952325e-06, "loss": 0.9965, "step": 1240 }, { "epoch": 1.5570890840652447, "grad_norm": 0.2470703125, "learning_rate": 2.214554579673777e-06, "loss": 0.9356, "step": 1241 }, { "epoch": 1.5583437892095358, "grad_norm": 0.2734375, "learning_rate": 2.2082810539523216e-06, "loss": 1.0049, "step": 1242 }, { "epoch": 1.5595984943538268, "grad_norm": 0.251953125, "learning_rate": 2.202007528230866e-06, "loss": 0.9493, "step": 1243 }, { "epoch": 1.5608531994981178, "grad_norm": 0.2490234375, "learning_rate": 2.1957340025094106e-06, "loss": 0.945, "step": 1244 }, { "epoch": 1.5621079046424091, "grad_norm": 0.255859375, "learning_rate": 2.189460476787955e-06, "loss": 0.9991, "step": 1245 }, { "epoch": 1.5633626097867002, "grad_norm": 0.2490234375, "learning_rate": 2.1831869510664996e-06, "loss": 0.994, "step": 1246 }, { "epoch": 1.5646173149309912, "grad_norm": 0.251953125, "learning_rate": 2.176913425345044e-06, "loss": 0.9864, "step": 1247 }, { "epoch": 1.5658720200752823, "grad_norm": 0.2421875, "learning_rate": 2.1706398996235886e-06, "loss": 0.9942, "step": 1248 }, { "epoch": 1.5671267252195733, "grad_norm": 0.248046875, "learning_rate": 2.164366373902133e-06, "loss": 1.0067, "step": 1249 }, { "epoch": 1.5683814303638646, "grad_norm": 0.255859375, "learning_rate": 2.1580928481806777e-06, "loss": 0.9382, "step": 1250 }, { "epoch": 1.5696361355081556, "grad_norm": 0.251953125, "learning_rate": 2.151819322459222e-06, "loss": 0.9519, "step": 1251 }, { "epoch": 1.5708908406524467, "grad_norm": 0.2470703125, "learning_rate": 2.1455457967377667e-06, "loss": 0.9622, "step": 1252 }, { "epoch": 1.572145545796738, "grad_norm": 0.265625, "learning_rate": 2.139272271016311e-06, "loss": 0.9792, "step": 1253 }, { "epoch": 1.5734002509410288, "grad_norm": 0.2490234375, "learning_rate": 2.1329987452948557e-06, "loss": 0.9214, "step": 1254 }, { "epoch": 1.57465495608532, "grad_norm": 0.248046875, "learning_rate": 2.1267252195734002e-06, "loss": 0.9165, "step": 1255 }, { "epoch": 1.575909661229611, "grad_norm": 0.251953125, "learning_rate": 2.120451693851945e-06, "loss": 1.0131, "step": 1256 }, { "epoch": 1.5771643663739021, "grad_norm": 0.25, "learning_rate": 2.1141781681304897e-06, "loss": 0.9996, "step": 1257 }, { "epoch": 1.5784190715181934, "grad_norm": 0.234375, "learning_rate": 2.107904642409034e-06, "loss": 0.9586, "step": 1258 }, { "epoch": 1.5796737766624842, "grad_norm": 0.24609375, "learning_rate": 2.1016311166875787e-06, "loss": 0.9526, "step": 1259 }, { "epoch": 1.5809284818067755, "grad_norm": 0.2412109375, "learning_rate": 2.0953575909661232e-06, "loss": 0.9601, "step": 1260 }, { "epoch": 1.5821831869510665, "grad_norm": 0.25390625, "learning_rate": 2.0890840652446677e-06, "loss": 0.92, "step": 1261 }, { "epoch": 1.5834378920953576, "grad_norm": 0.2578125, "learning_rate": 2.0828105395232122e-06, "loss": 0.9488, "step": 1262 }, { "epoch": 1.5846925972396488, "grad_norm": 0.25390625, "learning_rate": 2.0765370138017568e-06, "loss": 0.9395, "step": 1263 }, { "epoch": 1.5859473023839397, "grad_norm": 0.2451171875, "learning_rate": 2.0702634880803013e-06, "loss": 0.92, "step": 1264 }, { "epoch": 1.587202007528231, "grad_norm": 0.2451171875, "learning_rate": 2.063989962358846e-06, "loss": 0.9731, "step": 1265 }, { "epoch": 1.588456712672522, "grad_norm": 0.263671875, "learning_rate": 2.0577164366373903e-06, "loss": 0.9886, "step": 1266 }, { "epoch": 1.589711417816813, "grad_norm": 0.25390625, "learning_rate": 2.0514429109159352e-06, "loss": 0.9774, "step": 1267 }, { "epoch": 1.5909661229611043, "grad_norm": 0.333984375, "learning_rate": 2.0451693851944798e-06, "loss": 1.0486, "step": 1268 }, { "epoch": 1.5922208281053951, "grad_norm": 0.2412109375, "learning_rate": 2.038895859473024e-06, "loss": 0.9151, "step": 1269 }, { "epoch": 1.5934755332496864, "grad_norm": 0.2451171875, "learning_rate": 2.0326223337515684e-06, "loss": 0.9048, "step": 1270 }, { "epoch": 1.5947302383939774, "grad_norm": 0.251953125, "learning_rate": 2.026348808030113e-06, "loss": 0.913, "step": 1271 }, { "epoch": 1.5959849435382685, "grad_norm": 0.255859375, "learning_rate": 2.0200752823086574e-06, "loss": 0.9721, "step": 1272 }, { "epoch": 1.5972396486825597, "grad_norm": 0.25, "learning_rate": 2.0138017565872023e-06, "loss": 1.0206, "step": 1273 }, { "epoch": 1.5984943538268506, "grad_norm": 0.2421875, "learning_rate": 2.007528230865747e-06, "loss": 0.9724, "step": 1274 }, { "epoch": 1.5997490589711418, "grad_norm": 0.267578125, "learning_rate": 2.0012547051442913e-06, "loss": 0.959, "step": 1275 }, { "epoch": 1.6010037641154329, "grad_norm": 0.26171875, "learning_rate": 1.994981179422836e-06, "loss": 1.0229, "step": 1276 }, { "epoch": 1.602258469259724, "grad_norm": 0.26171875, "learning_rate": 1.9887076537013804e-06, "loss": 1.0605, "step": 1277 }, { "epoch": 1.6035131744040152, "grad_norm": 0.24609375, "learning_rate": 1.982434127979925e-06, "loss": 0.9007, "step": 1278 }, { "epoch": 1.604767879548306, "grad_norm": 0.24609375, "learning_rate": 1.9761606022584694e-06, "loss": 0.9723, "step": 1279 }, { "epoch": 1.6060225846925973, "grad_norm": 0.2490234375, "learning_rate": 1.969887076537014e-06, "loss": 0.9768, "step": 1280 }, { "epoch": 1.6072772898368883, "grad_norm": 0.251953125, "learning_rate": 1.9636135508155584e-06, "loss": 0.998, "step": 1281 }, { "epoch": 1.6085319949811794, "grad_norm": 0.251953125, "learning_rate": 1.957340025094103e-06, "loss": 0.9275, "step": 1282 }, { "epoch": 1.6097867001254706, "grad_norm": 0.251953125, "learning_rate": 1.9510664993726475e-06, "loss": 0.9514, "step": 1283 }, { "epoch": 1.6110414052697615, "grad_norm": 0.25, "learning_rate": 1.9447929736511924e-06, "loss": 0.977, "step": 1284 }, { "epoch": 1.6122961104140527, "grad_norm": 0.275390625, "learning_rate": 1.938519447929737e-06, "loss": 0.9785, "step": 1285 }, { "epoch": 1.6135508155583438, "grad_norm": 0.26953125, "learning_rate": 1.9322459222082814e-06, "loss": 0.9468, "step": 1286 }, { "epoch": 1.6148055207026348, "grad_norm": 0.34765625, "learning_rate": 1.9259723964868255e-06, "loss": 0.9193, "step": 1287 }, { "epoch": 1.616060225846926, "grad_norm": 0.2431640625, "learning_rate": 1.91969887076537e-06, "loss": 0.9781, "step": 1288 }, { "epoch": 1.617314930991217, "grad_norm": 0.2490234375, "learning_rate": 1.913425345043915e-06, "loss": 0.9489, "step": 1289 }, { "epoch": 1.6185696361355082, "grad_norm": 0.263671875, "learning_rate": 1.9071518193224595e-06, "loss": 0.9827, "step": 1290 }, { "epoch": 1.6198243412797992, "grad_norm": 0.2421875, "learning_rate": 1.900878293601004e-06, "loss": 0.9702, "step": 1291 }, { "epoch": 1.6210790464240903, "grad_norm": 0.2412109375, "learning_rate": 1.8946047678795485e-06, "loss": 0.9629, "step": 1292 }, { "epoch": 1.6223337515683816, "grad_norm": 0.2392578125, "learning_rate": 1.888331242158093e-06, "loss": 0.9717, "step": 1293 }, { "epoch": 1.6235884567126724, "grad_norm": 0.255859375, "learning_rate": 1.8820577164366375e-06, "loss": 1.0075, "step": 1294 }, { "epoch": 1.6248431618569636, "grad_norm": 0.25, "learning_rate": 1.8757841907151822e-06, "loss": 0.9794, "step": 1295 }, { "epoch": 1.6260978670012547, "grad_norm": 0.251953125, "learning_rate": 1.8695106649937268e-06, "loss": 0.9723, "step": 1296 }, { "epoch": 1.6273525721455457, "grad_norm": 0.240234375, "learning_rate": 1.8632371392722713e-06, "loss": 0.9223, "step": 1297 }, { "epoch": 1.628607277289837, "grad_norm": 0.24609375, "learning_rate": 1.8569636135508156e-06, "loss": 0.9182, "step": 1298 }, { "epoch": 1.6298619824341278, "grad_norm": 0.248046875, "learning_rate": 1.85069008782936e-06, "loss": 0.9764, "step": 1299 }, { "epoch": 1.631116687578419, "grad_norm": 0.25390625, "learning_rate": 1.844416562107905e-06, "loss": 0.9422, "step": 1300 }, { "epoch": 1.6323713927227101, "grad_norm": 0.244140625, "learning_rate": 1.8381430363864493e-06, "loss": 0.9543, "step": 1301 }, { "epoch": 1.6336260978670012, "grad_norm": 0.24609375, "learning_rate": 1.8318695106649938e-06, "loss": 0.9852, "step": 1302 }, { "epoch": 1.6348808030112925, "grad_norm": 0.24609375, "learning_rate": 1.8255959849435384e-06, "loss": 0.9611, "step": 1303 }, { "epoch": 1.6361355081555833, "grad_norm": 0.25390625, "learning_rate": 1.8193224592220829e-06, "loss": 0.9631, "step": 1304 }, { "epoch": 1.6373902132998746, "grad_norm": 0.380859375, "learning_rate": 1.8130489335006274e-06, "loss": 0.9309, "step": 1305 }, { "epoch": 1.6386449184441656, "grad_norm": 0.2490234375, "learning_rate": 1.8067754077791721e-06, "loss": 0.9827, "step": 1306 }, { "epoch": 1.6398996235884566, "grad_norm": 0.26171875, "learning_rate": 1.8005018820577166e-06, "loss": 0.9535, "step": 1307 }, { "epoch": 1.641154328732748, "grad_norm": 0.26171875, "learning_rate": 1.7942283563362611e-06, "loss": 0.9548, "step": 1308 }, { "epoch": 1.6424090338770387, "grad_norm": 0.3046875, "learning_rate": 1.7879548306148056e-06, "loss": 0.9546, "step": 1309 }, { "epoch": 1.64366373902133, "grad_norm": 0.25390625, "learning_rate": 1.7816813048933502e-06, "loss": 0.9278, "step": 1310 }, { "epoch": 1.644918444165621, "grad_norm": 0.255859375, "learning_rate": 1.7754077791718949e-06, "loss": 1.0356, "step": 1311 }, { "epoch": 1.646173149309912, "grad_norm": 0.310546875, "learning_rate": 1.7691342534504394e-06, "loss": 0.9434, "step": 1312 }, { "epoch": 1.6474278544542034, "grad_norm": 0.259765625, "learning_rate": 1.762860727728984e-06, "loss": 1.0403, "step": 1313 }, { "epoch": 1.6486825595984942, "grad_norm": 0.2490234375, "learning_rate": 1.7565872020075284e-06, "loss": 0.9372, "step": 1314 }, { "epoch": 1.6499372647427855, "grad_norm": 0.271484375, "learning_rate": 1.750313676286073e-06, "loss": 0.9652, "step": 1315 }, { "epoch": 1.6511919698870765, "grad_norm": 0.2412109375, "learning_rate": 1.7440401505646172e-06, "loss": 0.9588, "step": 1316 }, { "epoch": 1.6524466750313676, "grad_norm": 0.25, "learning_rate": 1.7377666248431622e-06, "loss": 0.9416, "step": 1317 }, { "epoch": 1.6537013801756588, "grad_norm": 0.2578125, "learning_rate": 1.7314930991217067e-06, "loss": 1.0272, "step": 1318 }, { "epoch": 1.6549560853199499, "grad_norm": 0.23828125, "learning_rate": 1.725219573400251e-06, "loss": 0.942, "step": 1319 }, { "epoch": 1.656210790464241, "grad_norm": 0.263671875, "learning_rate": 1.7189460476787955e-06, "loss": 1.0193, "step": 1320 }, { "epoch": 1.657465495608532, "grad_norm": 0.2490234375, "learning_rate": 1.71267252195734e-06, "loss": 0.9407, "step": 1321 }, { "epoch": 1.658720200752823, "grad_norm": 0.248046875, "learning_rate": 1.7063989962358847e-06, "loss": 0.9416, "step": 1322 }, { "epoch": 1.6599749058971143, "grad_norm": 0.2431640625, "learning_rate": 1.7001254705144293e-06, "loss": 0.9624, "step": 1323 }, { "epoch": 1.6612296110414053, "grad_norm": 0.2431640625, "learning_rate": 1.6938519447929738e-06, "loss": 0.9637, "step": 1324 }, { "epoch": 1.6624843161856964, "grad_norm": 0.259765625, "learning_rate": 1.6875784190715183e-06, "loss": 0.9952, "step": 1325 }, { "epoch": 1.6637390213299874, "grad_norm": 0.2451171875, "learning_rate": 1.6813048933500628e-06, "loss": 1.0053, "step": 1326 }, { "epoch": 1.6649937264742785, "grad_norm": 0.251953125, "learning_rate": 1.6750313676286073e-06, "loss": 0.9892, "step": 1327 }, { "epoch": 1.6662484316185697, "grad_norm": 0.271484375, "learning_rate": 1.668757841907152e-06, "loss": 0.9231, "step": 1328 }, { "epoch": 1.6675031367628608, "grad_norm": 0.25, "learning_rate": 1.6624843161856965e-06, "loss": 1.0228, "step": 1329 }, { "epoch": 1.6687578419071518, "grad_norm": 0.263671875, "learning_rate": 1.656210790464241e-06, "loss": 0.9472, "step": 1330 }, { "epoch": 1.6700125470514429, "grad_norm": 0.26171875, "learning_rate": 1.6499372647427856e-06, "loss": 0.9392, "step": 1331 }, { "epoch": 1.671267252195734, "grad_norm": 0.291015625, "learning_rate": 1.64366373902133e-06, "loss": 0.9525, "step": 1332 }, { "epoch": 1.6725219573400252, "grad_norm": 0.2451171875, "learning_rate": 1.6373902132998748e-06, "loss": 0.9818, "step": 1333 }, { "epoch": 1.6737766624843162, "grad_norm": 0.2451171875, "learning_rate": 1.6311166875784193e-06, "loss": 0.9388, "step": 1334 }, { "epoch": 1.6750313676286073, "grad_norm": 0.265625, "learning_rate": 1.6248431618569638e-06, "loss": 1.046, "step": 1335 }, { "epoch": 1.6762860727728983, "grad_norm": 0.4765625, "learning_rate": 1.6185696361355084e-06, "loss": 0.9539, "step": 1336 }, { "epoch": 1.6775407779171894, "grad_norm": 0.25390625, "learning_rate": 1.6122961104140529e-06, "loss": 0.9635, "step": 1337 }, { "epoch": 1.6787954830614806, "grad_norm": 0.2578125, "learning_rate": 1.6060225846925972e-06, "loss": 0.9832, "step": 1338 }, { "epoch": 1.6800501882057717, "grad_norm": 0.263671875, "learning_rate": 1.5997490589711421e-06, "loss": 0.9502, "step": 1339 }, { "epoch": 1.6813048933500627, "grad_norm": 0.265625, "learning_rate": 1.5934755332496864e-06, "loss": 0.9454, "step": 1340 }, { "epoch": 1.682559598494354, "grad_norm": 0.2890625, "learning_rate": 1.587202007528231e-06, "loss": 0.9606, "step": 1341 }, { "epoch": 1.6838143036386448, "grad_norm": 0.251953125, "learning_rate": 1.5809284818067754e-06, "loss": 1.0009, "step": 1342 }, { "epoch": 1.685069008782936, "grad_norm": 0.240234375, "learning_rate": 1.57465495608532e-06, "loss": 0.9829, "step": 1343 }, { "epoch": 1.6863237139272271, "grad_norm": 0.265625, "learning_rate": 1.5683814303638647e-06, "loss": 0.9571, "step": 1344 }, { "epoch": 1.6875784190715182, "grad_norm": 0.265625, "learning_rate": 1.5621079046424092e-06, "loss": 0.9688, "step": 1345 }, { "epoch": 1.6888331242158094, "grad_norm": 0.296875, "learning_rate": 1.5558343789209537e-06, "loss": 0.9731, "step": 1346 }, { "epoch": 1.6900878293601003, "grad_norm": 0.259765625, "learning_rate": 1.5495608531994982e-06, "loss": 0.9686, "step": 1347 }, { "epoch": 1.6913425345043915, "grad_norm": 0.39453125, "learning_rate": 1.5432873274780427e-06, "loss": 0.9737, "step": 1348 }, { "epoch": 1.6925972396486826, "grad_norm": 0.236328125, "learning_rate": 1.5370138017565872e-06, "loss": 0.9491, "step": 1349 }, { "epoch": 1.6938519447929736, "grad_norm": 0.2470703125, "learning_rate": 1.530740276035132e-06, "loss": 0.9703, "step": 1350 }, { "epoch": 1.695106649937265, "grad_norm": 0.33984375, "learning_rate": 1.5244667503136765e-06, "loss": 1.0391, "step": 1351 }, { "epoch": 1.6963613550815557, "grad_norm": 0.251953125, "learning_rate": 1.518193224592221e-06, "loss": 0.9895, "step": 1352 }, { "epoch": 1.697616060225847, "grad_norm": 0.2451171875, "learning_rate": 1.5119196988707655e-06, "loss": 0.9428, "step": 1353 }, { "epoch": 1.698870765370138, "grad_norm": 0.244140625, "learning_rate": 1.50564617314931e-06, "loss": 0.9609, "step": 1354 }, { "epoch": 1.700125470514429, "grad_norm": 0.24609375, "learning_rate": 1.4993726474278547e-06, "loss": 0.9676, "step": 1355 }, { "epoch": 1.7013801756587204, "grad_norm": 0.244140625, "learning_rate": 1.4930991217063993e-06, "loss": 0.9772, "step": 1356 }, { "epoch": 1.7026348808030112, "grad_norm": 0.255859375, "learning_rate": 1.4868255959849438e-06, "loss": 0.9564, "step": 1357 }, { "epoch": 1.7038895859473024, "grad_norm": 0.248046875, "learning_rate": 1.4805520702634883e-06, "loss": 0.9178, "step": 1358 }, { "epoch": 1.7051442910915935, "grad_norm": 0.298828125, "learning_rate": 1.4742785445420326e-06, "loss": 1.0764, "step": 1359 }, { "epoch": 1.7063989962358845, "grad_norm": 0.244140625, "learning_rate": 1.468005018820577e-06, "loss": 0.9683, "step": 1360 }, { "epoch": 1.7076537013801758, "grad_norm": 0.2421875, "learning_rate": 1.461731493099122e-06, "loss": 0.9466, "step": 1361 }, { "epoch": 1.7089084065244666, "grad_norm": 0.244140625, "learning_rate": 1.4554579673776663e-06, "loss": 0.9904, "step": 1362 }, { "epoch": 1.710163111668758, "grad_norm": 0.2431640625, "learning_rate": 1.4491844416562109e-06, "loss": 0.9713, "step": 1363 }, { "epoch": 1.711417816813049, "grad_norm": 0.251953125, "learning_rate": 1.4429109159347554e-06, "loss": 0.9457, "step": 1364 }, { "epoch": 1.71267252195734, "grad_norm": 0.337890625, "learning_rate": 1.4366373902132999e-06, "loss": 0.9741, "step": 1365 }, { "epoch": 1.7139272271016313, "grad_norm": 0.255859375, "learning_rate": 1.4303638644918446e-06, "loss": 0.9508, "step": 1366 }, { "epoch": 1.715181932245922, "grad_norm": 0.251953125, "learning_rate": 1.4240903387703891e-06, "loss": 0.9839, "step": 1367 }, { "epoch": 1.7164366373902133, "grad_norm": 0.24609375, "learning_rate": 1.4178168130489336e-06, "loss": 0.9751, "step": 1368 }, { "epoch": 1.7176913425345044, "grad_norm": 0.244140625, "learning_rate": 1.4115432873274781e-06, "loss": 0.9425, "step": 1369 }, { "epoch": 1.7189460476787954, "grad_norm": 0.24609375, "learning_rate": 1.4052697616060227e-06, "loss": 0.9595, "step": 1370 }, { "epoch": 1.7202007528230867, "grad_norm": 0.259765625, "learning_rate": 1.3989962358845672e-06, "loss": 0.9384, "step": 1371 }, { "epoch": 1.7214554579673775, "grad_norm": 0.318359375, "learning_rate": 1.3927227101631119e-06, "loss": 0.9116, "step": 1372 }, { "epoch": 1.7227101631116688, "grad_norm": 0.26171875, "learning_rate": 1.3864491844416564e-06, "loss": 0.9959, "step": 1373 }, { "epoch": 1.7239648682559598, "grad_norm": 0.2451171875, "learning_rate": 1.380175658720201e-06, "loss": 0.9347, "step": 1374 }, { "epoch": 1.725219573400251, "grad_norm": 0.2373046875, "learning_rate": 1.3739021329987454e-06, "loss": 0.9887, "step": 1375 }, { "epoch": 1.7264742785445422, "grad_norm": 0.251953125, "learning_rate": 1.36762860727729e-06, "loss": 0.9294, "step": 1376 }, { "epoch": 1.727728983688833, "grad_norm": 0.296875, "learning_rate": 1.3613550815558347e-06, "loss": 0.9859, "step": 1377 }, { "epoch": 1.7289836888331243, "grad_norm": 0.24609375, "learning_rate": 1.3550815558343792e-06, "loss": 0.9703, "step": 1378 }, { "epoch": 1.7302383939774153, "grad_norm": 0.302734375, "learning_rate": 1.3488080301129237e-06, "loss": 0.9181, "step": 1379 }, { "epoch": 1.7314930991217063, "grad_norm": 0.2490234375, "learning_rate": 1.342534504391468e-06, "loss": 0.9885, "step": 1380 }, { "epoch": 1.7327478042659976, "grad_norm": 0.248046875, "learning_rate": 1.3362609786700125e-06, "loss": 0.9762, "step": 1381 }, { "epoch": 1.7340025094102884, "grad_norm": 0.255859375, "learning_rate": 1.329987452948557e-06, "loss": 0.9711, "step": 1382 }, { "epoch": 1.7352572145545797, "grad_norm": 0.2412109375, "learning_rate": 1.3237139272271018e-06, "loss": 0.9975, "step": 1383 }, { "epoch": 1.7365119196988708, "grad_norm": 0.322265625, "learning_rate": 1.3174404015056463e-06, "loss": 0.9726, "step": 1384 }, { "epoch": 1.7377666248431618, "grad_norm": 0.244140625, "learning_rate": 1.3111668757841908e-06, "loss": 0.9854, "step": 1385 }, { "epoch": 1.739021329987453, "grad_norm": 0.2412109375, "learning_rate": 1.3048933500627353e-06, "loss": 0.9514, "step": 1386 }, { "epoch": 1.740276035131744, "grad_norm": 0.240234375, "learning_rate": 1.2986198243412798e-06, "loss": 0.927, "step": 1387 }, { "epoch": 1.7415307402760352, "grad_norm": 0.25390625, "learning_rate": 1.2923462986198245e-06, "loss": 0.9718, "step": 1388 }, { "epoch": 1.7427854454203262, "grad_norm": 0.24609375, "learning_rate": 1.286072772898369e-06, "loss": 0.9792, "step": 1389 }, { "epoch": 1.7440401505646173, "grad_norm": 0.310546875, "learning_rate": 1.2797992471769136e-06, "loss": 0.9467, "step": 1390 }, { "epoch": 1.7452948557089085, "grad_norm": 0.2890625, "learning_rate": 1.273525721455458e-06, "loss": 1.0075, "step": 1391 }, { "epoch": 1.7465495608531993, "grad_norm": 0.248046875, "learning_rate": 1.2672521957340026e-06, "loss": 0.9811, "step": 1392 }, { "epoch": 1.7478042659974906, "grad_norm": 0.244140625, "learning_rate": 1.260978670012547e-06, "loss": 0.9395, "step": 1393 }, { "epoch": 1.7490589711417817, "grad_norm": 0.2490234375, "learning_rate": 1.2547051442910918e-06, "loss": 0.9687, "step": 1394 }, { "epoch": 1.7503136762860727, "grad_norm": 0.25, "learning_rate": 1.2484316185696363e-06, "loss": 0.9478, "step": 1395 }, { "epoch": 1.751568381430364, "grad_norm": 0.255859375, "learning_rate": 1.2421580928481808e-06, "loss": 0.993, "step": 1396 }, { "epoch": 1.7528230865746548, "grad_norm": 0.2578125, "learning_rate": 1.2358845671267254e-06, "loss": 0.963, "step": 1397 }, { "epoch": 1.754077791718946, "grad_norm": 0.26171875, "learning_rate": 1.2296110414052699e-06, "loss": 0.9892, "step": 1398 }, { "epoch": 1.7553324968632371, "grad_norm": 0.25390625, "learning_rate": 1.2233375156838144e-06, "loss": 0.958, "step": 1399 }, { "epoch": 1.7565872020075282, "grad_norm": 0.2578125, "learning_rate": 1.217063989962359e-06, "loss": 0.9885, "step": 1400 }, { "epoch": 1.7578419071518194, "grad_norm": 0.25390625, "learning_rate": 1.2107904642409034e-06, "loss": 0.9824, "step": 1401 }, { "epoch": 1.7590966122961103, "grad_norm": 0.2421875, "learning_rate": 1.204516938519448e-06, "loss": 0.9638, "step": 1402 }, { "epoch": 1.7603513174404015, "grad_norm": 0.2373046875, "learning_rate": 1.1982434127979927e-06, "loss": 0.9526, "step": 1403 }, { "epoch": 1.7616060225846926, "grad_norm": 0.244140625, "learning_rate": 1.1919698870765372e-06, "loss": 0.9569, "step": 1404 }, { "epoch": 1.7628607277289836, "grad_norm": 0.2451171875, "learning_rate": 1.1856963613550817e-06, "loss": 0.9358, "step": 1405 }, { "epoch": 1.7641154328732749, "grad_norm": 0.2451171875, "learning_rate": 1.1794228356336262e-06, "loss": 0.9574, "step": 1406 }, { "epoch": 1.765370138017566, "grad_norm": 0.2578125, "learning_rate": 1.1731493099121707e-06, "loss": 0.9357, "step": 1407 }, { "epoch": 1.766624843161857, "grad_norm": 0.248046875, "learning_rate": 1.1668757841907152e-06, "loss": 0.9273, "step": 1408 }, { "epoch": 1.767879548306148, "grad_norm": 0.255859375, "learning_rate": 1.16060225846926e-06, "loss": 0.9987, "step": 1409 }, { "epoch": 1.769134253450439, "grad_norm": 0.302734375, "learning_rate": 1.1543287327478042e-06, "loss": 0.9274, "step": 1410 }, { "epoch": 1.7703889585947303, "grad_norm": 0.251953125, "learning_rate": 1.1480552070263488e-06, "loss": 0.9541, "step": 1411 }, { "epoch": 1.7716436637390214, "grad_norm": 0.255859375, "learning_rate": 1.1417816813048935e-06, "loss": 1.0194, "step": 1412 }, { "epoch": 1.7728983688833124, "grad_norm": 0.275390625, "learning_rate": 1.135508155583438e-06, "loss": 0.983, "step": 1413 }, { "epoch": 1.7741530740276035, "grad_norm": 0.248046875, "learning_rate": 1.1292346298619825e-06, "loss": 0.9283, "step": 1414 }, { "epoch": 1.7754077791718945, "grad_norm": 0.2451171875, "learning_rate": 1.122961104140527e-06, "loss": 0.9955, "step": 1415 }, { "epoch": 1.7766624843161858, "grad_norm": 0.25, "learning_rate": 1.1166875784190715e-06, "loss": 0.9544, "step": 1416 }, { "epoch": 1.7779171894604768, "grad_norm": 0.265625, "learning_rate": 1.1104140526976163e-06, "loss": 0.9414, "step": 1417 }, { "epoch": 1.7791718946047679, "grad_norm": 0.25, "learning_rate": 1.1041405269761608e-06, "loss": 0.9918, "step": 1418 }, { "epoch": 1.780426599749059, "grad_norm": 0.2451171875, "learning_rate": 1.0978670012547053e-06, "loss": 0.9335, "step": 1419 }, { "epoch": 1.78168130489335, "grad_norm": 0.25, "learning_rate": 1.0915934755332498e-06, "loss": 0.923, "step": 1420 }, { "epoch": 1.7829360100376412, "grad_norm": 0.2578125, "learning_rate": 1.0853199498117943e-06, "loss": 0.9601, "step": 1421 }, { "epoch": 1.7841907151819323, "grad_norm": 0.244140625, "learning_rate": 1.0790464240903388e-06, "loss": 0.9612, "step": 1422 }, { "epoch": 1.7854454203262233, "grad_norm": 0.2451171875, "learning_rate": 1.0727728983688833e-06, "loss": 0.9488, "step": 1423 }, { "epoch": 1.7867001254705144, "grad_norm": 0.248046875, "learning_rate": 1.0664993726474279e-06, "loss": 0.9709, "step": 1424 }, { "epoch": 1.7879548306148054, "grad_norm": 0.24609375, "learning_rate": 1.0602258469259726e-06, "loss": 0.9302, "step": 1425 }, { "epoch": 1.7892095357590967, "grad_norm": 0.2451171875, "learning_rate": 1.053952321204517e-06, "loss": 0.9405, "step": 1426 }, { "epoch": 1.7904642409033877, "grad_norm": 0.279296875, "learning_rate": 1.0476787954830616e-06, "loss": 0.937, "step": 1427 }, { "epoch": 1.7917189460476788, "grad_norm": 0.25390625, "learning_rate": 1.0414052697616061e-06, "loss": 0.9827, "step": 1428 }, { "epoch": 1.79297365119197, "grad_norm": 0.279296875, "learning_rate": 1.0351317440401506e-06, "loss": 0.9758, "step": 1429 }, { "epoch": 1.7942283563362609, "grad_norm": 0.263671875, "learning_rate": 1.0288582183186952e-06, "loss": 0.9321, "step": 1430 }, { "epoch": 1.7954830614805521, "grad_norm": 0.2470703125, "learning_rate": 1.0225846925972399e-06, "loss": 0.9952, "step": 1431 }, { "epoch": 1.7967377666248432, "grad_norm": 0.251953125, "learning_rate": 1.0163111668757842e-06, "loss": 0.9242, "step": 1432 }, { "epoch": 1.7979924717691342, "grad_norm": 0.251953125, "learning_rate": 1.0100376411543287e-06, "loss": 0.9662, "step": 1433 }, { "epoch": 1.7992471769134255, "grad_norm": 0.2412109375, "learning_rate": 1.0037641154328734e-06, "loss": 0.9351, "step": 1434 }, { "epoch": 1.8005018820577163, "grad_norm": 0.29296875, "learning_rate": 9.97490589711418e-07, "loss": 0.955, "step": 1435 }, { "epoch": 1.8017565872020076, "grad_norm": 0.26171875, "learning_rate": 9.912170639899624e-07, "loss": 0.9089, "step": 1436 }, { "epoch": 1.8030112923462986, "grad_norm": 0.248046875, "learning_rate": 9.84943538268507e-07, "loss": 0.928, "step": 1437 }, { "epoch": 1.8042659974905897, "grad_norm": 0.3515625, "learning_rate": 9.786700125470515e-07, "loss": 0.9456, "step": 1438 }, { "epoch": 1.805520702634881, "grad_norm": 0.24609375, "learning_rate": 9.723964868255962e-07, "loss": 0.9363, "step": 1439 }, { "epoch": 1.8067754077791718, "grad_norm": 0.255859375, "learning_rate": 9.661229611041407e-07, "loss": 0.9461, "step": 1440 }, { "epoch": 1.808030112923463, "grad_norm": 0.2470703125, "learning_rate": 9.59849435382685e-07, "loss": 1.0064, "step": 1441 }, { "epoch": 1.809284818067754, "grad_norm": 0.2431640625, "learning_rate": 9.535759096612297e-07, "loss": 0.9719, "step": 1442 }, { "epoch": 1.8105395232120451, "grad_norm": 0.2890625, "learning_rate": 9.473023839397742e-07, "loss": 0.9648, "step": 1443 }, { "epoch": 1.8117942283563364, "grad_norm": 0.28125, "learning_rate": 9.410288582183188e-07, "loss": 0.9784, "step": 1444 }, { "epoch": 1.8130489335006272, "grad_norm": 0.267578125, "learning_rate": 9.347553324968634e-07, "loss": 0.9391, "step": 1445 }, { "epoch": 1.8143036386449185, "grad_norm": 0.251953125, "learning_rate": 9.284818067754078e-07, "loss": 0.9698, "step": 1446 }, { "epoch": 1.8155583437892095, "grad_norm": 0.271484375, "learning_rate": 9.222082810539525e-07, "loss": 0.9857, "step": 1447 }, { "epoch": 1.8168130489335006, "grad_norm": 0.267578125, "learning_rate": 9.159347553324969e-07, "loss": 0.9808, "step": 1448 }, { "epoch": 1.8180677540777919, "grad_norm": 0.296875, "learning_rate": 9.096612296110414e-07, "loss": 0.9199, "step": 1449 }, { "epoch": 1.8193224592220827, "grad_norm": 0.255859375, "learning_rate": 9.033877038895861e-07, "loss": 0.9769, "step": 1450 }, { "epoch": 1.820577164366374, "grad_norm": 0.2412109375, "learning_rate": 8.971141781681306e-07, "loss": 0.9409, "step": 1451 }, { "epoch": 1.821831869510665, "grad_norm": 0.24609375, "learning_rate": 8.908406524466751e-07, "loss": 0.9738, "step": 1452 }, { "epoch": 1.823086574654956, "grad_norm": 0.255859375, "learning_rate": 8.845671267252197e-07, "loss": 0.9735, "step": 1453 }, { "epoch": 1.8243412797992473, "grad_norm": 0.2578125, "learning_rate": 8.782936010037642e-07, "loss": 0.9347, "step": 1454 }, { "epoch": 1.8255959849435381, "grad_norm": 0.248046875, "learning_rate": 8.720200752823086e-07, "loss": 0.9594, "step": 1455 }, { "epoch": 1.8268506900878294, "grad_norm": 0.2392578125, "learning_rate": 8.657465495608533e-07, "loss": 0.8945, "step": 1456 }, { "epoch": 1.8281053952321205, "grad_norm": 0.287109375, "learning_rate": 8.594730238393978e-07, "loss": 0.9788, "step": 1457 }, { "epoch": 1.8293601003764115, "grad_norm": 0.24609375, "learning_rate": 8.531994981179424e-07, "loss": 1.0104, "step": 1458 }, { "epoch": 1.8306148055207028, "grad_norm": 0.26171875, "learning_rate": 8.469259723964869e-07, "loss": 1.0467, "step": 1459 }, { "epoch": 1.8318695106649936, "grad_norm": 0.26171875, "learning_rate": 8.406524466750314e-07, "loss": 0.9581, "step": 1460 }, { "epoch": 1.8331242158092849, "grad_norm": 0.251953125, "learning_rate": 8.34378920953576e-07, "loss": 0.9658, "step": 1461 }, { "epoch": 1.834378920953576, "grad_norm": 0.244140625, "learning_rate": 8.281053952321205e-07, "loss": 0.9759, "step": 1462 }, { "epoch": 1.835633626097867, "grad_norm": 0.240234375, "learning_rate": 8.21831869510665e-07, "loss": 0.9718, "step": 1463 }, { "epoch": 1.8368883312421582, "grad_norm": 0.240234375, "learning_rate": 8.155583437892097e-07, "loss": 0.9721, "step": 1464 }, { "epoch": 1.838143036386449, "grad_norm": 0.330078125, "learning_rate": 8.092848180677542e-07, "loss": 0.9604, "step": 1465 }, { "epoch": 1.8393977415307403, "grad_norm": 0.25, "learning_rate": 8.030112923462986e-07, "loss": 1.0038, "step": 1466 }, { "epoch": 1.8406524466750314, "grad_norm": 0.251953125, "learning_rate": 7.967377666248432e-07, "loss": 0.9694, "step": 1467 }, { "epoch": 1.8419071518193224, "grad_norm": 0.2431640625, "learning_rate": 7.904642409033877e-07, "loss": 0.9637, "step": 1468 }, { "epoch": 1.8431618569636137, "grad_norm": 0.263671875, "learning_rate": 7.841907151819323e-07, "loss": 0.98, "step": 1469 }, { "epoch": 1.8444165621079045, "grad_norm": 0.244140625, "learning_rate": 7.779171894604768e-07, "loss": 0.9617, "step": 1470 }, { "epoch": 1.8456712672521958, "grad_norm": 0.25390625, "learning_rate": 7.716436637390214e-07, "loss": 0.9761, "step": 1471 }, { "epoch": 1.8469259723964868, "grad_norm": 0.244140625, "learning_rate": 7.65370138017566e-07, "loss": 0.943, "step": 1472 }, { "epoch": 1.8481806775407779, "grad_norm": 0.2578125, "learning_rate": 7.590966122961105e-07, "loss": 0.9487, "step": 1473 }, { "epoch": 1.8494353826850691, "grad_norm": 0.263671875, "learning_rate": 7.52823086574655e-07, "loss": 0.9754, "step": 1474 }, { "epoch": 1.85069008782936, "grad_norm": 0.3515625, "learning_rate": 7.465495608531996e-07, "loss": 0.8855, "step": 1475 }, { "epoch": 1.8519447929736512, "grad_norm": 0.248046875, "learning_rate": 7.402760351317441e-07, "loss": 0.9648, "step": 1476 }, { "epoch": 1.8531994981179423, "grad_norm": 0.2734375, "learning_rate": 7.340025094102885e-07, "loss": 0.9814, "step": 1477 }, { "epoch": 1.8544542032622333, "grad_norm": 0.240234375, "learning_rate": 7.277289836888332e-07, "loss": 0.9578, "step": 1478 }, { "epoch": 1.8557089084065246, "grad_norm": 0.255859375, "learning_rate": 7.214554579673777e-07, "loss": 0.9594, "step": 1479 }, { "epoch": 1.8569636135508154, "grad_norm": 0.26953125, "learning_rate": 7.151819322459223e-07, "loss": 0.9936, "step": 1480 }, { "epoch": 1.8582183186951067, "grad_norm": 0.265625, "learning_rate": 7.089084065244668e-07, "loss": 0.9137, "step": 1481 }, { "epoch": 1.8594730238393977, "grad_norm": 0.23828125, "learning_rate": 7.026348808030113e-07, "loss": 0.9791, "step": 1482 }, { "epoch": 1.8607277289836888, "grad_norm": 0.25390625, "learning_rate": 6.963613550815559e-07, "loss": 0.964, "step": 1483 }, { "epoch": 1.86198243412798, "grad_norm": 0.25, "learning_rate": 6.900878293601005e-07, "loss": 0.9375, "step": 1484 }, { "epoch": 1.8632371392722709, "grad_norm": 0.244140625, "learning_rate": 6.83814303638645e-07, "loss": 0.9997, "step": 1485 }, { "epoch": 1.8644918444165621, "grad_norm": 0.265625, "learning_rate": 6.775407779171896e-07, "loss": 1.0039, "step": 1486 }, { "epoch": 1.8657465495608532, "grad_norm": 0.3125, "learning_rate": 6.71267252195734e-07, "loss": 0.9573, "step": 1487 }, { "epoch": 1.8670012547051442, "grad_norm": 0.2392578125, "learning_rate": 6.649937264742785e-07, "loss": 0.9593, "step": 1488 }, { "epoch": 1.8682559598494355, "grad_norm": 0.259765625, "learning_rate": 6.587202007528231e-07, "loss": 0.9548, "step": 1489 }, { "epoch": 1.8695106649937263, "grad_norm": 0.244140625, "learning_rate": 6.524466750313676e-07, "loss": 0.9257, "step": 1490 }, { "epoch": 1.8707653701380176, "grad_norm": 0.25, "learning_rate": 6.461731493099123e-07, "loss": 0.9773, "step": 1491 }, { "epoch": 1.8720200752823086, "grad_norm": 0.2578125, "learning_rate": 6.398996235884568e-07, "loss": 0.9517, "step": 1492 }, { "epoch": 1.8732747804265997, "grad_norm": 0.2578125, "learning_rate": 6.336260978670013e-07, "loss": 1.001, "step": 1493 }, { "epoch": 1.874529485570891, "grad_norm": 0.24609375, "learning_rate": 6.273525721455459e-07, "loss": 0.9219, "step": 1494 }, { "epoch": 1.875784190715182, "grad_norm": 0.2470703125, "learning_rate": 6.210790464240904e-07, "loss": 0.9905, "step": 1495 }, { "epoch": 1.877038895859473, "grad_norm": 0.2578125, "learning_rate": 6.148055207026349e-07, "loss": 1.0018, "step": 1496 }, { "epoch": 1.878293601003764, "grad_norm": 0.3203125, "learning_rate": 6.085319949811795e-07, "loss": 0.9407, "step": 1497 }, { "epoch": 1.8795483061480551, "grad_norm": 0.2490234375, "learning_rate": 6.02258469259724e-07, "loss": 1.0004, "step": 1498 }, { "epoch": 1.8808030112923464, "grad_norm": 0.25, "learning_rate": 5.959849435382686e-07, "loss": 0.9433, "step": 1499 }, { "epoch": 1.8820577164366374, "grad_norm": 0.26171875, "learning_rate": 5.897114178168131e-07, "loss": 0.963, "step": 1500 }, { "epoch": 1.8833124215809285, "grad_norm": 0.259765625, "learning_rate": 5.834378920953576e-07, "loss": 0.9867, "step": 1501 }, { "epoch": 1.8845671267252195, "grad_norm": 0.2451171875, "learning_rate": 5.771643663739021e-07, "loss": 0.9332, "step": 1502 }, { "epoch": 1.8858218318695106, "grad_norm": 0.2431640625, "learning_rate": 5.708908406524467e-07, "loss": 0.9577, "step": 1503 }, { "epoch": 1.8870765370138018, "grad_norm": 0.2451171875, "learning_rate": 5.646173149309913e-07, "loss": 0.9242, "step": 1504 }, { "epoch": 1.888331242158093, "grad_norm": 0.24609375, "learning_rate": 5.583437892095358e-07, "loss": 0.9942, "step": 1505 }, { "epoch": 1.889585947302384, "grad_norm": 0.2431640625, "learning_rate": 5.520702634880804e-07, "loss": 0.9791, "step": 1506 }, { "epoch": 1.890840652446675, "grad_norm": 0.24609375, "learning_rate": 5.457967377666249e-07, "loss": 0.9819, "step": 1507 }, { "epoch": 1.892095357590966, "grad_norm": 0.25, "learning_rate": 5.395232120451694e-07, "loss": 0.9905, "step": 1508 }, { "epoch": 1.8933500627352573, "grad_norm": 0.310546875, "learning_rate": 5.332496863237139e-07, "loss": 0.9457, "step": 1509 }, { "epoch": 1.8946047678795483, "grad_norm": 0.2392578125, "learning_rate": 5.269761606022585e-07, "loss": 0.9336, "step": 1510 }, { "epoch": 1.8958594730238394, "grad_norm": 0.25390625, "learning_rate": 5.207026348808031e-07, "loss": 0.9916, "step": 1511 }, { "epoch": 1.8971141781681304, "grad_norm": 0.2431640625, "learning_rate": 5.144291091593476e-07, "loss": 0.9674, "step": 1512 }, { "epoch": 1.8983688833124215, "grad_norm": 0.3515625, "learning_rate": 5.081555834378921e-07, "loss": 0.9725, "step": 1513 }, { "epoch": 1.8996235884567128, "grad_norm": 0.2490234375, "learning_rate": 5.018820577164367e-07, "loss": 0.9968, "step": 1514 }, { "epoch": 1.9008782936010038, "grad_norm": 0.2451171875, "learning_rate": 4.956085319949812e-07, "loss": 0.953, "step": 1515 }, { "epoch": 1.9021329987452948, "grad_norm": 0.271484375, "learning_rate": 4.893350062735257e-07, "loss": 1.0023, "step": 1516 }, { "epoch": 1.9033877038895861, "grad_norm": 0.240234375, "learning_rate": 4.830614805520704e-07, "loss": 0.95, "step": 1517 }, { "epoch": 1.904642409033877, "grad_norm": 0.2470703125, "learning_rate": 4.7678795483061487e-07, "loss": 0.9752, "step": 1518 }, { "epoch": 1.9058971141781682, "grad_norm": 0.326171875, "learning_rate": 4.705144291091594e-07, "loss": 0.9564, "step": 1519 }, { "epoch": 1.9071518193224593, "grad_norm": 0.2421875, "learning_rate": 4.642409033877039e-07, "loss": 0.9455, "step": 1520 }, { "epoch": 1.9084065244667503, "grad_norm": 0.2451171875, "learning_rate": 4.5796737766624846e-07, "loss": 0.9335, "step": 1521 }, { "epoch": 1.9096612296110416, "grad_norm": 0.255859375, "learning_rate": 4.5169385194479303e-07, "loss": 0.9364, "step": 1522 }, { "epoch": 1.9109159347553324, "grad_norm": 0.302734375, "learning_rate": 4.4542032622333754e-07, "loss": 0.9572, "step": 1523 }, { "epoch": 1.9121706398996237, "grad_norm": 0.259765625, "learning_rate": 4.391468005018821e-07, "loss": 0.9979, "step": 1524 }, { "epoch": 1.9134253450439147, "grad_norm": 0.275390625, "learning_rate": 4.3287327478042667e-07, "loss": 0.9491, "step": 1525 }, { "epoch": 1.9146800501882058, "grad_norm": 0.302734375, "learning_rate": 4.265997490589712e-07, "loss": 0.9725, "step": 1526 }, { "epoch": 1.915934755332497, "grad_norm": 0.2451171875, "learning_rate": 4.203262233375157e-07, "loss": 0.9314, "step": 1527 }, { "epoch": 1.9171894604767878, "grad_norm": 0.255859375, "learning_rate": 4.1405269761606027e-07, "loss": 0.9985, "step": 1528 }, { "epoch": 1.9184441656210791, "grad_norm": 0.2392578125, "learning_rate": 4.0777917189460483e-07, "loss": 0.9471, "step": 1529 }, { "epoch": 1.9196988707653702, "grad_norm": 0.2373046875, "learning_rate": 4.015056461731493e-07, "loss": 0.9332, "step": 1530 }, { "epoch": 1.9209535759096612, "grad_norm": 0.2451171875, "learning_rate": 3.9523212045169386e-07, "loss": 0.9438, "step": 1531 }, { "epoch": 1.9222082810539525, "grad_norm": 0.25, "learning_rate": 3.889585947302384e-07, "loss": 0.9459, "step": 1532 }, { "epoch": 1.9234629861982433, "grad_norm": 0.259765625, "learning_rate": 3.82685069008783e-07, "loss": 0.9999, "step": 1533 }, { "epoch": 1.9247176913425346, "grad_norm": 0.259765625, "learning_rate": 3.764115432873275e-07, "loss": 1.0136, "step": 1534 }, { "epoch": 1.9259723964868256, "grad_norm": 0.298828125, "learning_rate": 3.7013801756587207e-07, "loss": 0.9667, "step": 1535 }, { "epoch": 1.9272271016311167, "grad_norm": 0.23828125, "learning_rate": 3.638644918444166e-07, "loss": 0.9592, "step": 1536 }, { "epoch": 1.928481806775408, "grad_norm": 0.2470703125, "learning_rate": 3.5759096612296115e-07, "loss": 0.9759, "step": 1537 }, { "epoch": 1.9297365119196987, "grad_norm": 0.2412109375, "learning_rate": 3.5131744040150566e-07, "loss": 0.9508, "step": 1538 }, { "epoch": 1.93099121706399, "grad_norm": 0.2578125, "learning_rate": 3.4504391468005023e-07, "loss": 0.9976, "step": 1539 }, { "epoch": 1.932245922208281, "grad_norm": 0.24609375, "learning_rate": 3.387703889585948e-07, "loss": 1.0088, "step": 1540 }, { "epoch": 1.933500627352572, "grad_norm": 0.251953125, "learning_rate": 3.3249686323713926e-07, "loss": 1.0041, "step": 1541 }, { "epoch": 1.9347553324968634, "grad_norm": 0.296875, "learning_rate": 3.262233375156838e-07, "loss": 0.9552, "step": 1542 }, { "epoch": 1.9360100376411542, "grad_norm": 0.2490234375, "learning_rate": 3.199498117942284e-07, "loss": 0.9948, "step": 1543 }, { "epoch": 1.9372647427854455, "grad_norm": 0.2412109375, "learning_rate": 3.1367628607277296e-07, "loss": 0.91, "step": 1544 }, { "epoch": 1.9385194479297365, "grad_norm": 0.267578125, "learning_rate": 3.0740276035131747e-07, "loss": 0.9383, "step": 1545 }, { "epoch": 1.9397741530740276, "grad_norm": 0.263671875, "learning_rate": 3.01129234629862e-07, "loss": 0.9658, "step": 1546 }, { "epoch": 1.9410288582183188, "grad_norm": 0.259765625, "learning_rate": 2.9485570890840655e-07, "loss": 0.9627, "step": 1547 }, { "epoch": 1.9422835633626097, "grad_norm": 0.2451171875, "learning_rate": 2.8858218318695106e-07, "loss": 0.9506, "step": 1548 }, { "epoch": 1.943538268506901, "grad_norm": 0.263671875, "learning_rate": 2.8230865746549563e-07, "loss": 0.9855, "step": 1549 }, { "epoch": 1.944792973651192, "grad_norm": 0.251953125, "learning_rate": 2.760351317440402e-07, "loss": 0.9382, "step": 1550 }, { "epoch": 1.946047678795483, "grad_norm": 0.2490234375, "learning_rate": 2.697616060225847e-07, "loss": 0.9745, "step": 1551 }, { "epoch": 1.9473023839397743, "grad_norm": 0.28125, "learning_rate": 2.634880803011293e-07, "loss": 0.89, "step": 1552 }, { "epoch": 1.948557089084065, "grad_norm": 0.267578125, "learning_rate": 2.572145545796738e-07, "loss": 0.9664, "step": 1553 }, { "epoch": 1.9498117942283564, "grad_norm": 0.6328125, "learning_rate": 2.5094102885821835e-07, "loss": 0.9924, "step": 1554 }, { "epoch": 1.9510664993726474, "grad_norm": 0.267578125, "learning_rate": 2.4466750313676287e-07, "loss": 1.001, "step": 1555 }, { "epoch": 1.9523212045169385, "grad_norm": 0.2490234375, "learning_rate": 2.3839397741530743e-07, "loss": 0.9597, "step": 1556 }, { "epoch": 1.9535759096612297, "grad_norm": 0.251953125, "learning_rate": 2.3212045169385195e-07, "loss": 0.9707, "step": 1557 }, { "epoch": 1.9548306148055206, "grad_norm": 0.26171875, "learning_rate": 2.2584692597239651e-07, "loss": 1.0041, "step": 1558 }, { "epoch": 1.9560853199498118, "grad_norm": 0.244140625, "learning_rate": 2.1957340025094105e-07, "loss": 0.9564, "step": 1559 }, { "epoch": 1.9573400250941029, "grad_norm": 0.25390625, "learning_rate": 2.132998745294856e-07, "loss": 0.9825, "step": 1560 }, { "epoch": 1.958594730238394, "grad_norm": 0.251953125, "learning_rate": 2.0702634880803013e-07, "loss": 0.9746, "step": 1561 }, { "epoch": 1.9598494353826852, "grad_norm": 0.2431640625, "learning_rate": 2.0075282308657465e-07, "loss": 0.9816, "step": 1562 }, { "epoch": 1.961104140526976, "grad_norm": 0.248046875, "learning_rate": 1.944792973651192e-07, "loss": 0.941, "step": 1563 }, { "epoch": 1.9623588456712673, "grad_norm": 0.259765625, "learning_rate": 1.8820577164366375e-07, "loss": 0.9033, "step": 1564 }, { "epoch": 1.9636135508155583, "grad_norm": 0.2490234375, "learning_rate": 1.819322459222083e-07, "loss": 0.9563, "step": 1565 }, { "epoch": 1.9648682559598494, "grad_norm": 0.291015625, "learning_rate": 1.7565872020075283e-07, "loss": 0.9217, "step": 1566 }, { "epoch": 1.9661229611041406, "grad_norm": 0.248046875, "learning_rate": 1.693851944792974e-07, "loss": 0.9565, "step": 1567 }, { "epoch": 1.9673776662484315, "grad_norm": 0.25390625, "learning_rate": 1.631116687578419e-07, "loss": 0.9773, "step": 1568 }, { "epoch": 1.9686323713927227, "grad_norm": 0.25390625, "learning_rate": 1.5683814303638648e-07, "loss": 0.9868, "step": 1569 }, { "epoch": 1.9698870765370138, "grad_norm": 0.26953125, "learning_rate": 1.50564617314931e-07, "loss": 0.9415, "step": 1570 }, { "epoch": 1.9711417816813048, "grad_norm": 0.248046875, "learning_rate": 1.4429109159347553e-07, "loss": 0.9734, "step": 1571 }, { "epoch": 1.972396486825596, "grad_norm": 0.318359375, "learning_rate": 1.380175658720201e-07, "loss": 0.9412, "step": 1572 }, { "epoch": 1.973651191969887, "grad_norm": 0.2451171875, "learning_rate": 1.3174404015056464e-07, "loss": 0.9194, "step": 1573 }, { "epoch": 1.9749058971141782, "grad_norm": 0.251953125, "learning_rate": 1.2547051442910918e-07, "loss": 0.938, "step": 1574 }, { "epoch": 1.9761606022584692, "grad_norm": 0.2451171875, "learning_rate": 1.1919698870765372e-07, "loss": 0.9942, "step": 1575 }, { "epoch": 1.9774153074027603, "grad_norm": 0.244140625, "learning_rate": 1.1292346298619826e-07, "loss": 0.9509, "step": 1576 }, { "epoch": 1.9786700125470515, "grad_norm": 0.259765625, "learning_rate": 1.066499372647428e-07, "loss": 0.9805, "step": 1577 }, { "epoch": 1.9799247176913424, "grad_norm": 0.2451171875, "learning_rate": 1.0037641154328732e-07, "loss": 0.9868, "step": 1578 }, { "epoch": 1.9811794228356336, "grad_norm": 0.248046875, "learning_rate": 9.410288582183188e-08, "loss": 0.9469, "step": 1579 }, { "epoch": 1.9824341279799247, "grad_norm": 0.255859375, "learning_rate": 8.782936010037642e-08, "loss": 0.9957, "step": 1580 }, { "epoch": 1.9836888331242157, "grad_norm": 0.244140625, "learning_rate": 8.155583437892096e-08, "loss": 0.9547, "step": 1581 }, { "epoch": 1.984943538268507, "grad_norm": 0.265625, "learning_rate": 7.52823086574655e-08, "loss": 1.0141, "step": 1582 }, { "epoch": 1.986198243412798, "grad_norm": 0.251953125, "learning_rate": 6.900878293601005e-08, "loss": 0.9889, "step": 1583 }, { "epoch": 1.987452948557089, "grad_norm": 0.259765625, "learning_rate": 6.273525721455459e-08, "loss": 0.9383, "step": 1584 }, { "epoch": 1.9887076537013801, "grad_norm": 0.2392578125, "learning_rate": 5.646173149309913e-08, "loss": 0.968, "step": 1585 }, { "epoch": 1.9899623588456712, "grad_norm": 0.25, "learning_rate": 5.018820577164366e-08, "loss": 0.9575, "step": 1586 }, { "epoch": 1.9912170639899625, "grad_norm": 0.2451171875, "learning_rate": 4.391468005018821e-08, "loss": 0.9828, "step": 1587 }, { "epoch": 1.9924717691342535, "grad_norm": 0.26953125, "learning_rate": 3.764115432873275e-08, "loss": 0.9665, "step": 1588 }, { "epoch": 1.9937264742785445, "grad_norm": 0.3515625, "learning_rate": 3.1367628607277294e-08, "loss": 0.9836, "step": 1589 }, { "epoch": 1.9949811794228356, "grad_norm": 0.283203125, "learning_rate": 2.509410288582183e-08, "loss": 0.9912, "step": 1590 }, { "epoch": 1.9962358845671266, "grad_norm": 0.2431640625, "learning_rate": 1.8820577164366374e-08, "loss": 0.9454, "step": 1591 }, { "epoch": 1.997490589711418, "grad_norm": 0.283203125, "learning_rate": 1.2547051442910915e-08, "loss": 0.9569, "step": 1592 }, { "epoch": 1.998745294855709, "grad_norm": 0.25, "learning_rate": 6.273525721455458e-09, "loss": 0.9325, "step": 1593 }, { "epoch": 2.0, "grad_norm": 0.26953125, "learning_rate": 0.0, "loss": 0.9637, "step": 1594 } ], "logging_steps": 1.0, "max_steps": 1594, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.11793679377734e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }