diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,11207 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.99608457321848, + "eval_steps": 500, + "global_step": 1595, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0031323414252153485, + "grad_norm": 11.050841614106629, + "learning_rate": 1.2500000000000002e-07, + "loss": 2.1157, + "step": 1 + }, + { + "epoch": 0.006264682850430697, + "grad_norm": 10.770240996509447, + "learning_rate": 2.5000000000000004e-07, + "loss": 2.1788, + "step": 2 + }, + { + "epoch": 0.009397024275646046, + "grad_norm": 10.853179484768518, + "learning_rate": 3.75e-07, + "loss": 2.109, + "step": 3 + }, + { + "epoch": 0.012529365700861394, + "grad_norm": 10.639536499246885, + "learning_rate": 5.000000000000001e-07, + "loss": 2.1164, + "step": 4 + }, + { + "epoch": 0.015661707126076743, + "grad_norm": 10.748209885589805, + "learning_rate": 6.25e-07, + "loss": 2.1148, + "step": 5 + }, + { + "epoch": 0.018794048551292093, + "grad_norm": 10.463406145556146, + "learning_rate": 7.5e-07, + "loss": 2.1268, + "step": 6 + }, + { + "epoch": 0.02192638997650744, + "grad_norm": 10.729571035762252, + "learning_rate": 8.75e-07, + "loss": 2.1396, + "step": 7 + }, + { + "epoch": 0.025058731401722788, + "grad_norm": 9.834267944614176, + "learning_rate": 1.0000000000000002e-06, + "loss": 2.1216, + "step": 8 + }, + { + "epoch": 0.028191072826938137, + "grad_norm": 9.6285022315209, + "learning_rate": 1.125e-06, + "loss": 2.0811, + "step": 9 + }, + { + "epoch": 0.031323414252153486, + "grad_norm": 8.068778598644403, + "learning_rate": 1.25e-06, + "loss": 2.0954, + "step": 10 + }, + { + "epoch": 0.03445575567736883, + "grad_norm": 7.900246512788569, + "learning_rate": 1.3750000000000002e-06, + "loss": 2.093, + "step": 11 + }, + { + "epoch": 0.037588097102584185, + "grad_norm": 7.672686278535105, + "learning_rate": 1.5e-06, + "loss": 2.035, + "step": 12 + }, + { + "epoch": 0.04072043852779953, + "grad_norm": 4.005341384803414, + "learning_rate": 1.6250000000000001e-06, + "loss": 1.9538, + "step": 13 + }, + { + "epoch": 0.04385277995301488, + "grad_norm": 4.033643898372103, + "learning_rate": 1.75e-06, + "loss": 1.9371, + "step": 14 + }, + { + "epoch": 0.04698512137823023, + "grad_norm": 3.9058441402584543, + "learning_rate": 1.8750000000000003e-06, + "loss": 1.9222, + "step": 15 + }, + { + "epoch": 0.050117462803445575, + "grad_norm": 3.93510615716937, + "learning_rate": 2.0000000000000003e-06, + "loss": 1.9425, + "step": 16 + }, + { + "epoch": 0.05324980422866092, + "grad_norm": 3.9972887378242508, + "learning_rate": 2.125e-06, + "loss": 1.871, + "step": 17 + }, + { + "epoch": 0.056382145653876274, + "grad_norm": 4.637696367766985, + "learning_rate": 2.25e-06, + "loss": 1.704, + "step": 18 + }, + { + "epoch": 0.05951448707909162, + "grad_norm": 4.895075769171093, + "learning_rate": 2.375e-06, + "loss": 1.7081, + "step": 19 + }, + { + "epoch": 0.06264682850430697, + "grad_norm": 4.887182700750148, + "learning_rate": 2.5e-06, + "loss": 1.6922, + "step": 20 + }, + { + "epoch": 0.06577916992952232, + "grad_norm": 4.387829116707637, + "learning_rate": 2.6250000000000003e-06, + "loss": 1.6223, + "step": 21 + }, + { + "epoch": 0.06891151135473766, + "grad_norm": 3.9332482151095056, + "learning_rate": 2.7500000000000004e-06, + "loss": 1.6429, + "step": 22 + }, + { + "epoch": 0.07204385277995301, + "grad_norm": 3.621520125738942, + "learning_rate": 2.875e-06, + "loss": 1.5692, + "step": 23 + }, + { + "epoch": 0.07517619420516837, + "grad_norm": 4.869727787056477, + "learning_rate": 3e-06, + "loss": 1.4277, + "step": 24 + }, + { + "epoch": 0.07830853563038372, + "grad_norm": 2.760787151921623, + "learning_rate": 3.125e-06, + "loss": 1.343, + "step": 25 + }, + { + "epoch": 0.08144087705559906, + "grad_norm": 2.158543950171073, + "learning_rate": 3.2500000000000002e-06, + "loss": 1.3304, + "step": 26 + }, + { + "epoch": 0.08457321848081441, + "grad_norm": 1.8629759121807041, + "learning_rate": 3.3750000000000003e-06, + "loss": 1.3279, + "step": 27 + }, + { + "epoch": 0.08770555990602975, + "grad_norm": 1.4660115041698683, + "learning_rate": 3.5e-06, + "loss": 1.3604, + "step": 28 + }, + { + "epoch": 0.0908379013312451, + "grad_norm": 1.2902107389850677, + "learning_rate": 3.625e-06, + "loss": 1.2715, + "step": 29 + }, + { + "epoch": 0.09397024275646046, + "grad_norm": 1.201092304355642, + "learning_rate": 3.7500000000000005e-06, + "loss": 1.3227, + "step": 30 + }, + { + "epoch": 0.0971025841816758, + "grad_norm": 1.1572726408932534, + "learning_rate": 3.875e-06, + "loss": 1.3139, + "step": 31 + }, + { + "epoch": 0.10023492560689115, + "grad_norm": 1.0860599652926863, + "learning_rate": 4.000000000000001e-06, + "loss": 1.277, + "step": 32 + }, + { + "epoch": 0.1033672670321065, + "grad_norm": 1.1601900235559528, + "learning_rate": 4.125e-06, + "loss": 1.3422, + "step": 33 + }, + { + "epoch": 0.10649960845732184, + "grad_norm": 1.0702922916707374, + "learning_rate": 4.25e-06, + "loss": 1.2635, + "step": 34 + }, + { + "epoch": 0.1096319498825372, + "grad_norm": 1.1224186551004272, + "learning_rate": 4.3750000000000005e-06, + "loss": 1.3686, + "step": 35 + }, + { + "epoch": 0.11276429130775255, + "grad_norm": 1.0316309615389836, + "learning_rate": 4.5e-06, + "loss": 1.2546, + "step": 36 + }, + { + "epoch": 0.1158966327329679, + "grad_norm": 1.0350470961145155, + "learning_rate": 4.625000000000001e-06, + "loss": 1.2617, + "step": 37 + }, + { + "epoch": 0.11902897415818324, + "grad_norm": 0.9499273751228338, + "learning_rate": 4.75e-06, + "loss": 1.2218, + "step": 38 + }, + { + "epoch": 0.12216131558339859, + "grad_norm": 0.9633898650879267, + "learning_rate": 4.875e-06, + "loss": 1.2513, + "step": 39 + }, + { + "epoch": 0.12529365700861395, + "grad_norm": 0.9820874085749698, + "learning_rate": 5e-06, + "loss": 1.2438, + "step": 40 + }, + { + "epoch": 0.1284259984338293, + "grad_norm": 0.9949392511683506, + "learning_rate": 5.125e-06, + "loss": 1.2127, + "step": 41 + }, + { + "epoch": 0.13155833985904464, + "grad_norm": 0.963927255316921, + "learning_rate": 5.2500000000000006e-06, + "loss": 1.1986, + "step": 42 + }, + { + "epoch": 0.13469068128425998, + "grad_norm": 0.9265554217258913, + "learning_rate": 5.375e-06, + "loss": 1.2378, + "step": 43 + }, + { + "epoch": 0.13782302270947533, + "grad_norm": 0.895459875699844, + "learning_rate": 5.500000000000001e-06, + "loss": 1.2106, + "step": 44 + }, + { + "epoch": 0.14095536413469067, + "grad_norm": 0.8697327947413821, + "learning_rate": 5.625e-06, + "loss": 1.1968, + "step": 45 + }, + { + "epoch": 0.14408770555990602, + "grad_norm": 0.8663662696743084, + "learning_rate": 5.75e-06, + "loss": 1.1794, + "step": 46 + }, + { + "epoch": 0.14722004698512137, + "grad_norm": 0.925686203130037, + "learning_rate": 5.8750000000000005e-06, + "loss": 1.2847, + "step": 47 + }, + { + "epoch": 0.15035238841033674, + "grad_norm": 0.9075811490967844, + "learning_rate": 6e-06, + "loss": 1.2584, + "step": 48 + }, + { + "epoch": 0.1534847298355521, + "grad_norm": 0.8432956780425832, + "learning_rate": 6.125000000000001e-06, + "loss": 1.1543, + "step": 49 + }, + { + "epoch": 0.15661707126076743, + "grad_norm": 0.8713896231009577, + "learning_rate": 6.25e-06, + "loss": 1.2038, + "step": 50 + }, + { + "epoch": 0.15974941268598278, + "grad_norm": 0.7897618524028729, + "learning_rate": 6.375e-06, + "loss": 1.1681, + "step": 51 + }, + { + "epoch": 0.16288175411119812, + "grad_norm": 0.8144877727194363, + "learning_rate": 6.5000000000000004e-06, + "loss": 1.1969, + "step": 52 + }, + { + "epoch": 0.16601409553641347, + "grad_norm": 0.8337995472721724, + "learning_rate": 6.625e-06, + "loss": 1.1631, + "step": 53 + }, + { + "epoch": 0.16914643696162882, + "grad_norm": 0.8286193073403529, + "learning_rate": 6.750000000000001e-06, + "loss": 1.231, + "step": 54 + }, + { + "epoch": 0.17227877838684416, + "grad_norm": 0.8272909936450883, + "learning_rate": 6.875e-06, + "loss": 1.1908, + "step": 55 + }, + { + "epoch": 0.1754111198120595, + "grad_norm": 0.8159834359002948, + "learning_rate": 7e-06, + "loss": 1.1909, + "step": 56 + }, + { + "epoch": 0.17854346123727485, + "grad_norm": 0.8327400175086643, + "learning_rate": 7.125e-06, + "loss": 1.2279, + "step": 57 + }, + { + "epoch": 0.1816758026624902, + "grad_norm": 0.8345916890457833, + "learning_rate": 7.25e-06, + "loss": 1.2322, + "step": 58 + }, + { + "epoch": 0.18480814408770557, + "grad_norm": 0.8009822359933073, + "learning_rate": 7.375000000000001e-06, + "loss": 1.1961, + "step": 59 + }, + { + "epoch": 0.18794048551292092, + "grad_norm": 0.8174246707753029, + "learning_rate": 7.500000000000001e-06, + "loss": 1.159, + "step": 60 + }, + { + "epoch": 0.19107282693813626, + "grad_norm": 0.8166829885787309, + "learning_rate": 7.625e-06, + "loss": 1.2641, + "step": 61 + }, + { + "epoch": 0.1942051683633516, + "grad_norm": 0.8047242090989581, + "learning_rate": 7.75e-06, + "loss": 1.1689, + "step": 62 + }, + { + "epoch": 0.19733750978856696, + "grad_norm": 0.7730776885197735, + "learning_rate": 7.875e-06, + "loss": 1.1269, + "step": 63 + }, + { + "epoch": 0.2004698512137823, + "grad_norm": 0.7779957947469543, + "learning_rate": 8.000000000000001e-06, + "loss": 1.1762, + "step": 64 + }, + { + "epoch": 0.20360219263899765, + "grad_norm": 0.8000417296639418, + "learning_rate": 8.125000000000001e-06, + "loss": 1.2025, + "step": 65 + }, + { + "epoch": 0.206734534064213, + "grad_norm": 0.8222150496696322, + "learning_rate": 8.25e-06, + "loss": 1.2, + "step": 66 + }, + { + "epoch": 0.20986687548942834, + "grad_norm": 0.7895026739737903, + "learning_rate": 8.375e-06, + "loss": 1.1623, + "step": 67 + }, + { + "epoch": 0.21299921691464369, + "grad_norm": 0.7784224076989291, + "learning_rate": 8.5e-06, + "loss": 1.1554, + "step": 68 + }, + { + "epoch": 0.21613155833985903, + "grad_norm": 0.798211358388717, + "learning_rate": 8.625000000000001e-06, + "loss": 1.1389, + "step": 69 + }, + { + "epoch": 0.2192638997650744, + "grad_norm": 0.8027997517397614, + "learning_rate": 8.750000000000001e-06, + "loss": 1.1578, + "step": 70 + }, + { + "epoch": 0.22239624119028975, + "grad_norm": 0.7884275020306974, + "learning_rate": 8.875e-06, + "loss": 1.1476, + "step": 71 + }, + { + "epoch": 0.2255285826155051, + "grad_norm": 0.8162843570408501, + "learning_rate": 9e-06, + "loss": 1.197, + "step": 72 + }, + { + "epoch": 0.22866092404072044, + "grad_norm": 0.8195467948994536, + "learning_rate": 9.125e-06, + "loss": 1.1463, + "step": 73 + }, + { + "epoch": 0.2317932654659358, + "grad_norm": 0.785823621627204, + "learning_rate": 9.250000000000001e-06, + "loss": 1.13, + "step": 74 + }, + { + "epoch": 0.23492560689115113, + "grad_norm": 0.7858506830531828, + "learning_rate": 9.375000000000001e-06, + "loss": 1.1392, + "step": 75 + }, + { + "epoch": 0.23805794831636648, + "grad_norm": 0.8126411047255567, + "learning_rate": 9.5e-06, + "loss": 1.174, + "step": 76 + }, + { + "epoch": 0.24119028974158183, + "grad_norm": 0.8134240118545614, + "learning_rate": 9.625e-06, + "loss": 1.149, + "step": 77 + }, + { + "epoch": 0.24432263116679717, + "grad_norm": 2.0900195366866186, + "learning_rate": 9.75e-06, + "loss": 1.1669, + "step": 78 + }, + { + "epoch": 0.24745497259201252, + "grad_norm": 0.8049430353655749, + "learning_rate": 9.875000000000001e-06, + "loss": 1.1632, + "step": 79 + }, + { + "epoch": 0.2505873140172279, + "grad_norm": 0.8154348013283664, + "learning_rate": 1e-05, + "loss": 1.1752, + "step": 80 + }, + { + "epoch": 0.25371965544244324, + "grad_norm": 0.7766532231446668, + "learning_rate": 1.0125e-05, + "loss": 1.1562, + "step": 81 + }, + { + "epoch": 0.2568519968676586, + "grad_norm": 0.805338167968417, + "learning_rate": 1.025e-05, + "loss": 1.1875, + "step": 82 + }, + { + "epoch": 0.25998433829287393, + "grad_norm": 0.7634601102161442, + "learning_rate": 1.0375000000000001e-05, + "loss": 1.1216, + "step": 83 + }, + { + "epoch": 0.2631166797180893, + "grad_norm": 0.7954986672515693, + "learning_rate": 1.0500000000000001e-05, + "loss": 1.1681, + "step": 84 + }, + { + "epoch": 0.2662490211433046, + "grad_norm": 0.7666715336890474, + "learning_rate": 1.0625e-05, + "loss": 1.1383, + "step": 85 + }, + { + "epoch": 0.26938136256851997, + "grad_norm": 0.7841383740610983, + "learning_rate": 1.075e-05, + "loss": 1.1956, + "step": 86 + }, + { + "epoch": 0.2725137039937353, + "grad_norm": 0.7593684758700783, + "learning_rate": 1.0875e-05, + "loss": 1.1199, + "step": 87 + }, + { + "epoch": 0.27564604541895066, + "grad_norm": 0.7913739392990039, + "learning_rate": 1.1000000000000001e-05, + "loss": 1.1665, + "step": 88 + }, + { + "epoch": 0.278778386844166, + "grad_norm": 0.8205958812550826, + "learning_rate": 1.1125000000000001e-05, + "loss": 1.1756, + "step": 89 + }, + { + "epoch": 0.28191072826938135, + "grad_norm": 0.7967279563962926, + "learning_rate": 1.125e-05, + "loss": 1.1559, + "step": 90 + }, + { + "epoch": 0.2850430696945967, + "grad_norm": 0.7673412238486342, + "learning_rate": 1.1375e-05, + "loss": 1.1744, + "step": 91 + }, + { + "epoch": 0.28817541111981204, + "grad_norm": 0.805175551828789, + "learning_rate": 1.15e-05, + "loss": 1.1629, + "step": 92 + }, + { + "epoch": 0.2913077525450274, + "grad_norm": 0.7505564301768121, + "learning_rate": 1.1625000000000001e-05, + "loss": 1.1384, + "step": 93 + }, + { + "epoch": 0.29444009397024273, + "grad_norm": 0.8039055214719127, + "learning_rate": 1.1750000000000001e-05, + "loss": 1.146, + "step": 94 + }, + { + "epoch": 0.2975724353954581, + "grad_norm": 0.7825064856728625, + "learning_rate": 1.1875e-05, + "loss": 1.1644, + "step": 95 + }, + { + "epoch": 0.3007047768206735, + "grad_norm": 0.7935646900032081, + "learning_rate": 1.2e-05, + "loss": 1.1469, + "step": 96 + }, + { + "epoch": 0.3038371182458888, + "grad_norm": 0.8024905927017434, + "learning_rate": 1.2125e-05, + "loss": 1.1402, + "step": 97 + }, + { + "epoch": 0.3069694596711042, + "grad_norm": 0.7830126081823029, + "learning_rate": 1.2250000000000001e-05, + "loss": 1.1733, + "step": 98 + }, + { + "epoch": 0.3101018010963195, + "grad_norm": 0.7997702404780472, + "learning_rate": 1.2375000000000001e-05, + "loss": 1.1991, + "step": 99 + }, + { + "epoch": 0.31323414252153486, + "grad_norm": 0.7557112262293625, + "learning_rate": 1.25e-05, + "loss": 1.1463, + "step": 100 + }, + { + "epoch": 0.3163664839467502, + "grad_norm": 0.7794955385488898, + "learning_rate": 1.2625e-05, + "loss": 1.192, + "step": 101 + }, + { + "epoch": 0.31949882537196556, + "grad_norm": 0.8098258584079975, + "learning_rate": 1.275e-05, + "loss": 1.213, + "step": 102 + }, + { + "epoch": 0.3226311667971809, + "grad_norm": 0.7578592334982148, + "learning_rate": 1.2875000000000001e-05, + "loss": 1.1538, + "step": 103 + }, + { + "epoch": 0.32576350822239625, + "grad_norm": 0.7801947192543522, + "learning_rate": 1.3000000000000001e-05, + "loss": 1.1362, + "step": 104 + }, + { + "epoch": 0.3288958496476116, + "grad_norm": 0.803547746608655, + "learning_rate": 1.3125e-05, + "loss": 1.1337, + "step": 105 + }, + { + "epoch": 0.33202819107282694, + "grad_norm": 0.749613358405691, + "learning_rate": 1.325e-05, + "loss": 1.1038, + "step": 106 + }, + { + "epoch": 0.3351605324980423, + "grad_norm": 0.8427226009611682, + "learning_rate": 1.3375e-05, + "loss": 1.1611, + "step": 107 + }, + { + "epoch": 0.33829287392325763, + "grad_norm": 0.8147379441469031, + "learning_rate": 1.3500000000000001e-05, + "loss": 1.158, + "step": 108 + }, + { + "epoch": 0.341425215348473, + "grad_norm": 0.7724455205425795, + "learning_rate": 1.3625e-05, + "loss": 1.1123, + "step": 109 + }, + { + "epoch": 0.3445575567736883, + "grad_norm": 0.7752938561359628, + "learning_rate": 1.375e-05, + "loss": 1.1516, + "step": 110 + }, + { + "epoch": 0.34768989819890367, + "grad_norm": 0.7714492124885363, + "learning_rate": 1.3875e-05, + "loss": 1.1251, + "step": 111 + }, + { + "epoch": 0.350822239624119, + "grad_norm": 0.7951581981638454, + "learning_rate": 1.4e-05, + "loss": 1.1362, + "step": 112 + }, + { + "epoch": 0.35395458104933436, + "grad_norm": 0.7654142695855861, + "learning_rate": 1.4125000000000003e-05, + "loss": 1.1494, + "step": 113 + }, + { + "epoch": 0.3570869224745497, + "grad_norm": 0.7809679002598582, + "learning_rate": 1.425e-05, + "loss": 1.1533, + "step": 114 + }, + { + "epoch": 0.36021926389976505, + "grad_norm": 0.8081986701754312, + "learning_rate": 1.4375e-05, + "loss": 1.1654, + "step": 115 + }, + { + "epoch": 0.3633516053249804, + "grad_norm": 0.8085004016002348, + "learning_rate": 1.45e-05, + "loss": 1.1343, + "step": 116 + }, + { + "epoch": 0.36648394675019574, + "grad_norm": 0.8329554845008073, + "learning_rate": 1.4625e-05, + "loss": 1.1569, + "step": 117 + }, + { + "epoch": 0.36961628817541115, + "grad_norm": 0.783780585930527, + "learning_rate": 1.4750000000000003e-05, + "loss": 1.125, + "step": 118 + }, + { + "epoch": 0.3727486296006265, + "grad_norm": 0.7912033183545758, + "learning_rate": 1.4875000000000002e-05, + "loss": 1.1073, + "step": 119 + }, + { + "epoch": 0.37588097102584184, + "grad_norm": 0.821268337393993, + "learning_rate": 1.5000000000000002e-05, + "loss": 1.1386, + "step": 120 + }, + { + "epoch": 0.3790133124510572, + "grad_norm": 0.832261258255235, + "learning_rate": 1.5125e-05, + "loss": 1.1683, + "step": 121 + }, + { + "epoch": 0.38214565387627253, + "grad_norm": 0.8094345539354894, + "learning_rate": 1.525e-05, + "loss": 1.1508, + "step": 122 + }, + { + "epoch": 0.3852779953014879, + "grad_norm": 0.8008883805537554, + "learning_rate": 1.5375e-05, + "loss": 1.1111, + "step": 123 + }, + { + "epoch": 0.3884103367267032, + "grad_norm": 0.7915578589687696, + "learning_rate": 1.55e-05, + "loss": 1.1479, + "step": 124 + }, + { + "epoch": 0.39154267815191857, + "grad_norm": 0.8327731000166747, + "learning_rate": 1.5625e-05, + "loss": 1.1777, + "step": 125 + }, + { + "epoch": 0.3946750195771339, + "grad_norm": 0.8189737080837218, + "learning_rate": 1.575e-05, + "loss": 1.1558, + "step": 126 + }, + { + "epoch": 0.39780736100234926, + "grad_norm": 0.7783531314158876, + "learning_rate": 1.5875e-05, + "loss": 1.1095, + "step": 127 + }, + { + "epoch": 0.4009397024275646, + "grad_norm": 0.8543936473595438, + "learning_rate": 1.6000000000000003e-05, + "loss": 1.1427, + "step": 128 + }, + { + "epoch": 0.40407204385277995, + "grad_norm": 0.8243727971086049, + "learning_rate": 1.6125000000000002e-05, + "loss": 1.1636, + "step": 129 + }, + { + "epoch": 0.4072043852779953, + "grad_norm": 0.7919846838112398, + "learning_rate": 1.6250000000000002e-05, + "loss": 1.1753, + "step": 130 + }, + { + "epoch": 0.41033672670321064, + "grad_norm": 0.7953800117937856, + "learning_rate": 1.6375e-05, + "loss": 1.1121, + "step": 131 + }, + { + "epoch": 0.413469068128426, + "grad_norm": 0.7941547055326952, + "learning_rate": 1.65e-05, + "loss": 1.1759, + "step": 132 + }, + { + "epoch": 0.41660140955364133, + "grad_norm": 0.751465418050413, + "learning_rate": 1.6625e-05, + "loss": 1.1131, + "step": 133 + }, + { + "epoch": 0.4197337509788567, + "grad_norm": 0.782469109387437, + "learning_rate": 1.675e-05, + "loss": 1.1319, + "step": 134 + }, + { + "epoch": 0.422866092404072, + "grad_norm": 0.7537743415586728, + "learning_rate": 1.6875e-05, + "loss": 1.123, + "step": 135 + }, + { + "epoch": 0.42599843382928737, + "grad_norm": 0.8266848129941319, + "learning_rate": 1.7e-05, + "loss": 1.2021, + "step": 136 + }, + { + "epoch": 0.4291307752545027, + "grad_norm": 0.7907070577194092, + "learning_rate": 1.7125e-05, + "loss": 1.1311, + "step": 137 + }, + { + "epoch": 0.43226311667971806, + "grad_norm": 0.770956600368574, + "learning_rate": 1.7250000000000003e-05, + "loss": 1.1406, + "step": 138 + }, + { + "epoch": 0.43539545810493346, + "grad_norm": 0.7942699632538319, + "learning_rate": 1.7375000000000002e-05, + "loss": 1.121, + "step": 139 + }, + { + "epoch": 0.4385277995301488, + "grad_norm": 0.7669673284070234, + "learning_rate": 1.7500000000000002e-05, + "loss": 1.1206, + "step": 140 + }, + { + "epoch": 0.44166014095536416, + "grad_norm": 0.7910094531836759, + "learning_rate": 1.7625e-05, + "loss": 1.1069, + "step": 141 + }, + { + "epoch": 0.4447924823805795, + "grad_norm": 0.7711860139608495, + "learning_rate": 1.775e-05, + "loss": 1.1422, + "step": 142 + }, + { + "epoch": 0.44792482380579485, + "grad_norm": 0.7957935969277278, + "learning_rate": 1.7875e-05, + "loss": 1.1059, + "step": 143 + }, + { + "epoch": 0.4510571652310102, + "grad_norm": 0.7632895910705135, + "learning_rate": 1.8e-05, + "loss": 1.1327, + "step": 144 + }, + { + "epoch": 0.45418950665622554, + "grad_norm": 0.7972939663996078, + "learning_rate": 1.8125e-05, + "loss": 1.1898, + "step": 145 + }, + { + "epoch": 0.4573218480814409, + "grad_norm": 0.768609653869839, + "learning_rate": 1.825e-05, + "loss": 1.1361, + "step": 146 + }, + { + "epoch": 0.46045418950665623, + "grad_norm": 0.74095309811706, + "learning_rate": 1.8375e-05, + "loss": 1.0726, + "step": 147 + }, + { + "epoch": 0.4635865309318716, + "grad_norm": 0.7937353734822679, + "learning_rate": 1.8500000000000002e-05, + "loss": 1.1412, + "step": 148 + }, + { + "epoch": 0.4667188723570869, + "grad_norm": 0.8231685857372454, + "learning_rate": 1.8625000000000002e-05, + "loss": 1.1928, + "step": 149 + }, + { + "epoch": 0.46985121378230227, + "grad_norm": 0.7582039516040869, + "learning_rate": 1.8750000000000002e-05, + "loss": 1.1157, + "step": 150 + }, + { + "epoch": 0.4729835552075176, + "grad_norm": 0.7945667336329015, + "learning_rate": 1.8875e-05, + "loss": 1.1339, + "step": 151 + }, + { + "epoch": 0.47611589663273296, + "grad_norm": 0.7511917754873464, + "learning_rate": 1.9e-05, + "loss": 1.115, + "step": 152 + }, + { + "epoch": 0.4792482380579483, + "grad_norm": 0.7567481788127869, + "learning_rate": 1.9125000000000004e-05, + "loss": 1.1438, + "step": 153 + }, + { + "epoch": 0.48238057948316365, + "grad_norm": 0.8035923511844805, + "learning_rate": 1.925e-05, + "loss": 1.1488, + "step": 154 + }, + { + "epoch": 0.485512920908379, + "grad_norm": 0.7832113472335267, + "learning_rate": 1.9375e-05, + "loss": 1.1473, + "step": 155 + }, + { + "epoch": 0.48864526233359434, + "grad_norm": 0.7835793180590945, + "learning_rate": 1.95e-05, + "loss": 1.2008, + "step": 156 + }, + { + "epoch": 0.4917776037588097, + "grad_norm": 0.8058256022750308, + "learning_rate": 1.9625e-05, + "loss": 1.148, + "step": 157 + }, + { + "epoch": 0.49490994518402504, + "grad_norm": 0.7847452922181013, + "learning_rate": 1.9750000000000002e-05, + "loss": 1.1196, + "step": 158 + }, + { + "epoch": 0.4980422866092404, + "grad_norm": 0.7812512360657365, + "learning_rate": 1.9875000000000002e-05, + "loss": 1.1419, + "step": 159 + }, + { + "epoch": 0.5011746280344558, + "grad_norm": 0.8824042145521714, + "learning_rate": 2e-05, + "loss": 1.0868, + "step": 160 + }, + { + "epoch": 0.5043069694596711, + "grad_norm": 0.7567570291946419, + "learning_rate": 1.9999976035643366e-05, + "loss": 1.1777, + "step": 161 + }, + { + "epoch": 0.5074393108848865, + "grad_norm": 0.7952722559164349, + "learning_rate": 1.9999904142688313e-05, + "loss": 1.1345, + "step": 162 + }, + { + "epoch": 0.5105716523101018, + "grad_norm": 0.7743964880190665, + "learning_rate": 1.9999784321479416e-05, + "loss": 1.1041, + "step": 163 + }, + { + "epoch": 0.5137039937353172, + "grad_norm": 0.8498468306677794, + "learning_rate": 1.999961657259097e-05, + "loss": 1.1567, + "step": 164 + }, + { + "epoch": 0.5168363351605325, + "grad_norm": 0.7947824846601378, + "learning_rate": 1.9999400896826965e-05, + "loss": 1.1733, + "step": 165 + }, + { + "epoch": 0.5199686765857479, + "grad_norm": 0.7487074186272767, + "learning_rate": 1.999913729522111e-05, + "loss": 1.161, + "step": 166 + }, + { + "epoch": 0.5231010180109632, + "grad_norm": 0.8037325407031123, + "learning_rate": 1.999882576903682e-05, + "loss": 1.1646, + "step": 167 + }, + { + "epoch": 0.5262333594361785, + "grad_norm": 0.7799604031117069, + "learning_rate": 1.9998466319767188e-05, + "loss": 1.153, + "step": 168 + }, + { + "epoch": 0.5293657008613939, + "grad_norm": 0.8012827006886666, + "learning_rate": 1.9998058949135016e-05, + "loss": 1.2302, + "step": 169 + }, + { + "epoch": 0.5324980422866092, + "grad_norm": 0.7752213646522256, + "learning_rate": 1.9997603659092773e-05, + "loss": 1.1444, + "step": 170 + }, + { + "epoch": 0.5356303837118246, + "grad_norm": 0.7880510647431511, + "learning_rate": 1.9997100451822616e-05, + "loss": 1.1522, + "step": 171 + }, + { + "epoch": 0.5387627251370399, + "grad_norm": 0.802804888286548, + "learning_rate": 1.9996549329736345e-05, + "loss": 1.1926, + "step": 172 + }, + { + "epoch": 0.5418950665622553, + "grad_norm": 0.7425430954083608, + "learning_rate": 1.9995950295475422e-05, + "loss": 1.1118, + "step": 173 + }, + { + "epoch": 0.5450274079874706, + "grad_norm": 0.794380669252506, + "learning_rate": 1.9995303351910934e-05, + "loss": 1.1748, + "step": 174 + }, + { + "epoch": 0.548159749412686, + "grad_norm": 0.7533240888688524, + "learning_rate": 1.9994608502143603e-05, + "loss": 1.1243, + "step": 175 + }, + { + "epoch": 0.5512920908379013, + "grad_norm": 0.7646481885397147, + "learning_rate": 1.9993865749503754e-05, + "loss": 1.1415, + "step": 176 + }, + { + "epoch": 0.5544244322631167, + "grad_norm": 0.7915751390688428, + "learning_rate": 1.9993075097551302e-05, + "loss": 1.1005, + "step": 177 + }, + { + "epoch": 0.557556773688332, + "grad_norm": 0.7921308350778985, + "learning_rate": 1.9992236550075746e-05, + "loss": 1.1215, + "step": 178 + }, + { + "epoch": 0.5606891151135474, + "grad_norm": 0.7469409103210944, + "learning_rate": 1.9991350111096133e-05, + "loss": 1.1023, + "step": 179 + }, + { + "epoch": 0.5638214565387627, + "grad_norm": 0.7382348346776821, + "learning_rate": 1.999041578486105e-05, + "loss": 1.1578, + "step": 180 + }, + { + "epoch": 0.566953797963978, + "grad_norm": 0.7926745472354245, + "learning_rate": 1.99894335758486e-05, + "loss": 1.1692, + "step": 181 + }, + { + "epoch": 0.5700861393891934, + "grad_norm": 0.7569926636991706, + "learning_rate": 1.998840348876639e-05, + "loss": 1.1385, + "step": 182 + }, + { + "epoch": 0.5732184808144087, + "grad_norm": 0.7458272413839812, + "learning_rate": 1.9987325528551492e-05, + "loss": 1.1332, + "step": 183 + }, + { + "epoch": 0.5763508222396241, + "grad_norm": 0.7842596742794642, + "learning_rate": 1.9986199700370435e-05, + "loss": 1.1599, + "step": 184 + }, + { + "epoch": 0.5794831636648394, + "grad_norm": 0.7560805393380029, + "learning_rate": 1.998502600961916e-05, + "loss": 1.0842, + "step": 185 + }, + { + "epoch": 0.5826155050900548, + "grad_norm": 0.7498681844673027, + "learning_rate": 1.9983804461923022e-05, + "loss": 1.1539, + "step": 186 + }, + { + "epoch": 0.5857478465152701, + "grad_norm": 0.7698674252021521, + "learning_rate": 1.998253506313674e-05, + "loss": 1.1244, + "step": 187 + }, + { + "epoch": 0.5888801879404855, + "grad_norm": 0.7668778266672499, + "learning_rate": 1.998121781934438e-05, + "loss": 1.1268, + "step": 188 + }, + { + "epoch": 0.5920125293657008, + "grad_norm": 0.7473755501240149, + "learning_rate": 1.997985273685932e-05, + "loss": 1.0964, + "step": 189 + }, + { + "epoch": 0.5951448707909162, + "grad_norm": 0.7456163683549248, + "learning_rate": 1.9978439822224228e-05, + "loss": 1.1468, + "step": 190 + }, + { + "epoch": 0.5982772122161315, + "grad_norm": 0.7586612489506767, + "learning_rate": 1.997697908221102e-05, + "loss": 1.1148, + "step": 191 + }, + { + "epoch": 0.601409553641347, + "grad_norm": 0.7526265106209291, + "learning_rate": 1.9975470523820835e-05, + "loss": 1.1596, + "step": 192 + }, + { + "epoch": 0.6045418950665623, + "grad_norm": 0.7515061295922136, + "learning_rate": 1.9973914154284e-05, + "loss": 1.132, + "step": 193 + }, + { + "epoch": 0.6076742364917777, + "grad_norm": 0.7628083735034853, + "learning_rate": 1.997230998105999e-05, + "loss": 1.0992, + "step": 194 + }, + { + "epoch": 0.610806577916993, + "grad_norm": 0.7372667581302207, + "learning_rate": 1.9970658011837404e-05, + "loss": 1.1549, + "step": 195 + }, + { + "epoch": 0.6139389193422083, + "grad_norm": 0.7570962352925633, + "learning_rate": 1.9968958254533922e-05, + "loss": 1.1437, + "step": 196 + }, + { + "epoch": 0.6170712607674237, + "grad_norm": 0.7760513349389108, + "learning_rate": 1.9967210717296256e-05, + "loss": 1.1143, + "step": 197 + }, + { + "epoch": 0.620203602192639, + "grad_norm": 0.7710793602174225, + "learning_rate": 1.996541540850013e-05, + "loss": 1.1611, + "step": 198 + }, + { + "epoch": 0.6233359436178544, + "grad_norm": 0.7497240836215765, + "learning_rate": 1.9963572336750227e-05, + "loss": 1.0894, + "step": 199 + }, + { + "epoch": 0.6264682850430697, + "grad_norm": 0.769485721788448, + "learning_rate": 1.996168151088015e-05, + "loss": 1.1326, + "step": 200 + }, + { + "epoch": 0.6296006264682851, + "grad_norm": 0.7914744857793413, + "learning_rate": 1.9959742939952393e-05, + "loss": 1.1642, + "step": 201 + }, + { + "epoch": 0.6327329678935004, + "grad_norm": 0.7749252132709005, + "learning_rate": 1.9957756633258264e-05, + "loss": 1.1097, + "step": 202 + }, + { + "epoch": 0.6358653093187158, + "grad_norm": 0.7248448375095949, + "learning_rate": 1.9955722600317887e-05, + "loss": 1.1149, + "step": 203 + }, + { + "epoch": 0.6389976507439311, + "grad_norm": 0.7893182002778344, + "learning_rate": 1.9953640850880118e-05, + "loss": 1.1343, + "step": 204 + }, + { + "epoch": 0.6421299921691465, + "grad_norm": 0.752880032513866, + "learning_rate": 1.995151139492251e-05, + "loss": 1.1393, + "step": 205 + }, + { + "epoch": 0.6452623335943618, + "grad_norm": 0.7715420215410178, + "learning_rate": 1.994933424265127e-05, + "loss": 1.1498, + "step": 206 + }, + { + "epoch": 0.6483946750195771, + "grad_norm": 0.7359564724287257, + "learning_rate": 1.9947109404501216e-05, + "loss": 1.1758, + "step": 207 + }, + { + "epoch": 0.6515270164447925, + "grad_norm": 0.7587433580271375, + "learning_rate": 1.99448368911357e-05, + "loss": 1.1346, + "step": 208 + }, + { + "epoch": 0.6546593578700078, + "grad_norm": 0.7720483403367886, + "learning_rate": 1.99425167134466e-05, + "loss": 1.1527, + "step": 209 + }, + { + "epoch": 0.6577916992952232, + "grad_norm": 0.7504516082730889, + "learning_rate": 1.9940148882554223e-05, + "loss": 1.0908, + "step": 210 + }, + { + "epoch": 0.6609240407204385, + "grad_norm": 0.7583232566276451, + "learning_rate": 1.9937733409807273e-05, + "loss": 1.1387, + "step": 211 + }, + { + "epoch": 0.6640563821456539, + "grad_norm": 0.7275565863134074, + "learning_rate": 1.9935270306782804e-05, + "loss": 1.1061, + "step": 212 + }, + { + "epoch": 0.6671887235708692, + "grad_norm": 0.7308212360196517, + "learning_rate": 1.9932759585286154e-05, + "loss": 1.14, + "step": 213 + }, + { + "epoch": 0.6703210649960846, + "grad_norm": 0.7395935935887239, + "learning_rate": 1.9930201257350883e-05, + "loss": 1.1356, + "step": 214 + }, + { + "epoch": 0.6734534064212999, + "grad_norm": 0.7318764377322893, + "learning_rate": 1.9927595335238736e-05, + "loss": 1.1468, + "step": 215 + }, + { + "epoch": 0.6765857478465153, + "grad_norm": 0.716132066894738, + "learning_rate": 1.992494183143955e-05, + "loss": 1.1329, + "step": 216 + }, + { + "epoch": 0.6797180892717306, + "grad_norm": 0.783963003986274, + "learning_rate": 1.992224075867124e-05, + "loss": 1.134, + "step": 217 + }, + { + "epoch": 0.682850430696946, + "grad_norm": 0.7765840844001187, + "learning_rate": 1.991949212987969e-05, + "loss": 1.1525, + "step": 218 + }, + { + "epoch": 0.6859827721221613, + "grad_norm": 0.7664402025464453, + "learning_rate": 1.9916695958238736e-05, + "loss": 1.1363, + "step": 219 + }, + { + "epoch": 0.6891151135473766, + "grad_norm": 0.7619234106278223, + "learning_rate": 1.9913852257150053e-05, + "loss": 1.1043, + "step": 220 + }, + { + "epoch": 0.692247454972592, + "grad_norm": 0.7503302852552577, + "learning_rate": 1.9910961040243148e-05, + "loss": 1.1503, + "step": 221 + }, + { + "epoch": 0.6953797963978073, + "grad_norm": 0.7386647741445785, + "learning_rate": 1.9908022321375244e-05, + "loss": 1.126, + "step": 222 + }, + { + "epoch": 0.6985121378230227, + "grad_norm": 0.772006501836167, + "learning_rate": 1.9905036114631247e-05, + "loss": 1.1324, + "step": 223 + }, + { + "epoch": 0.701644479248238, + "grad_norm": 0.7498396218310553, + "learning_rate": 1.9902002434323658e-05, + "loss": 1.13, + "step": 224 + }, + { + "epoch": 0.7047768206734534, + "grad_norm": 0.7499406276159518, + "learning_rate": 1.9898921294992516e-05, + "loss": 1.136, + "step": 225 + }, + { + "epoch": 0.7079091620986687, + "grad_norm": 0.8050655298488321, + "learning_rate": 1.9895792711405332e-05, + "loss": 1.1725, + "step": 226 + }, + { + "epoch": 0.7110415035238841, + "grad_norm": 0.785084787806491, + "learning_rate": 1.9892616698556996e-05, + "loss": 1.1521, + "step": 227 + }, + { + "epoch": 0.7141738449490994, + "grad_norm": 0.7548865828004917, + "learning_rate": 1.9889393271669735e-05, + "loss": 1.1182, + "step": 228 + }, + { + "epoch": 0.7173061863743148, + "grad_norm": 0.7609216107863833, + "learning_rate": 1.9886122446193015e-05, + "loss": 1.1345, + "step": 229 + }, + { + "epoch": 0.7204385277995301, + "grad_norm": 0.7547249839650533, + "learning_rate": 1.9882804237803487e-05, + "loss": 1.1427, + "step": 230 + }, + { + "epoch": 0.7235708692247454, + "grad_norm": 0.7632257000147761, + "learning_rate": 1.9879438662404893e-05, + "loss": 1.1235, + "step": 231 + }, + { + "epoch": 0.7267032106499608, + "grad_norm": 0.7568637017446223, + "learning_rate": 1.9876025736128e-05, + "loss": 1.1301, + "step": 232 + }, + { + "epoch": 0.7298355520751761, + "grad_norm": 0.7420518311600437, + "learning_rate": 1.987256547533053e-05, + "loss": 1.1342, + "step": 233 + }, + { + "epoch": 0.7329678935003915, + "grad_norm": 0.7646439659344628, + "learning_rate": 1.9869057896597068e-05, + "loss": 1.1352, + "step": 234 + }, + { + "epoch": 0.7361002349256069, + "grad_norm": 0.7312821097164371, + "learning_rate": 1.9865503016738983e-05, + "loss": 1.0321, + "step": 235 + }, + { + "epoch": 0.7392325763508223, + "grad_norm": 0.7546957399103286, + "learning_rate": 1.9861900852794357e-05, + "loss": 1.1377, + "step": 236 + }, + { + "epoch": 0.7423649177760376, + "grad_norm": 0.7194994232421758, + "learning_rate": 1.9858251422027903e-05, + "loss": 1.1237, + "step": 237 + }, + { + "epoch": 0.745497259201253, + "grad_norm": 0.7785862815076194, + "learning_rate": 1.985455474193087e-05, + "loss": 1.1609, + "step": 238 + }, + { + "epoch": 0.7486296006264683, + "grad_norm": 0.740903933948647, + "learning_rate": 1.985081083022097e-05, + "loss": 1.0945, + "step": 239 + }, + { + "epoch": 0.7517619420516837, + "grad_norm": 0.7291780203060889, + "learning_rate": 1.984701970484229e-05, + "loss": 1.0856, + "step": 240 + }, + { + "epoch": 0.754894283476899, + "grad_norm": 0.7587452855951344, + "learning_rate": 1.9843181383965207e-05, + "loss": 1.1665, + "step": 241 + }, + { + "epoch": 0.7580266249021144, + "grad_norm": 0.7407242126385574, + "learning_rate": 1.98392958859863e-05, + "loss": 1.1194, + "step": 242 + }, + { + "epoch": 0.7611589663273297, + "grad_norm": 0.7518927137010917, + "learning_rate": 1.9835363229528255e-05, + "loss": 1.1213, + "step": 243 + }, + { + "epoch": 0.7642913077525451, + "grad_norm": 0.749707981917893, + "learning_rate": 1.9831383433439798e-05, + "loss": 1.1007, + "step": 244 + }, + { + "epoch": 0.7674236491777604, + "grad_norm": 0.7695157385802572, + "learning_rate": 1.982735651679557e-05, + "loss": 1.1203, + "step": 245 + }, + { + "epoch": 0.7705559906029757, + "grad_norm": 0.7438784077225071, + "learning_rate": 1.982328249889607e-05, + "loss": 1.135, + "step": 246 + }, + { + "epoch": 0.7736883320281911, + "grad_norm": 0.7415443518523247, + "learning_rate": 1.9819161399267545e-05, + "loss": 1.076, + "step": 247 + }, + { + "epoch": 0.7768206734534064, + "grad_norm": 0.7482623409973128, + "learning_rate": 1.9814993237661887e-05, + "loss": 1.1561, + "step": 248 + }, + { + "epoch": 0.7799530148786218, + "grad_norm": 0.7460111329694793, + "learning_rate": 1.9810778034056562e-05, + "loss": 1.1132, + "step": 249 + }, + { + "epoch": 0.7830853563038371, + "grad_norm": 0.7288330170842255, + "learning_rate": 1.98065158086545e-05, + "loss": 1.0535, + "step": 250 + }, + { + "epoch": 0.7862176977290525, + "grad_norm": 0.7314199802562354, + "learning_rate": 1.9802206581883992e-05, + "loss": 1.1316, + "step": 251 + }, + { + "epoch": 0.7893500391542678, + "grad_norm": 0.7804910252834903, + "learning_rate": 1.9797850374398617e-05, + "loss": 1.1219, + "step": 252 + }, + { + "epoch": 0.7924823805794832, + "grad_norm": 0.7298549749672437, + "learning_rate": 1.9793447207077114e-05, + "loss": 1.1124, + "step": 253 + }, + { + "epoch": 0.7956147220046985, + "grad_norm": 0.7551562547765204, + "learning_rate": 1.9788997101023294e-05, + "loss": 1.1458, + "step": 254 + }, + { + "epoch": 0.7987470634299139, + "grad_norm": 0.7641456778799793, + "learning_rate": 1.9784500077565944e-05, + "loss": 1.1743, + "step": 255 + }, + { + "epoch": 0.8018794048551292, + "grad_norm": 0.7350148397112547, + "learning_rate": 1.977995615825872e-05, + "loss": 1.137, + "step": 256 + }, + { + "epoch": 0.8050117462803446, + "grad_norm": 0.742949716719734, + "learning_rate": 1.9775365364880043e-05, + "loss": 1.1294, + "step": 257 + }, + { + "epoch": 0.8081440877055599, + "grad_norm": 0.7336184274603342, + "learning_rate": 1.9770727719432994e-05, + "loss": 1.13, + "step": 258 + }, + { + "epoch": 0.8112764291307752, + "grad_norm": 0.7150049627370941, + "learning_rate": 1.976604324414521e-05, + "loss": 1.079, + "step": 259 + }, + { + "epoch": 0.8144087705559906, + "grad_norm": 0.729004714169207, + "learning_rate": 1.9761311961468782e-05, + "loss": 1.0903, + "step": 260 + }, + { + "epoch": 0.8175411119812059, + "grad_norm": 0.7432345451419001, + "learning_rate": 1.9756533894080132e-05, + "loss": 1.1718, + "step": 261 + }, + { + "epoch": 0.8206734534064213, + "grad_norm": 0.7485136400635154, + "learning_rate": 1.975170906487993e-05, + "loss": 1.123, + "step": 262 + }, + { + "epoch": 0.8238057948316366, + "grad_norm": 0.7792494517105039, + "learning_rate": 1.9746837496992953e-05, + "loss": 1.1018, + "step": 263 + }, + { + "epoch": 0.826938136256852, + "grad_norm": 0.7534542131696997, + "learning_rate": 1.974191921376801e-05, + "loss": 1.1376, + "step": 264 + }, + { + "epoch": 0.8300704776820673, + "grad_norm": 0.7486965183472643, + "learning_rate": 1.9736954238777793e-05, + "loss": 1.1112, + "step": 265 + }, + { + "epoch": 0.8332028191072827, + "grad_norm": 0.7675798633230227, + "learning_rate": 1.9731942595818786e-05, + "loss": 1.1506, + "step": 266 + }, + { + "epoch": 0.836335160532498, + "grad_norm": 0.7361146624834083, + "learning_rate": 1.9726884308911153e-05, + "loss": 1.1072, + "step": 267 + }, + { + "epoch": 0.8394675019577134, + "grad_norm": 0.7546626067336956, + "learning_rate": 1.9721779402298614e-05, + "loss": 1.105, + "step": 268 + }, + { + "epoch": 0.8425998433829287, + "grad_norm": 0.7592163829713124, + "learning_rate": 1.9716627900448323e-05, + "loss": 1.1045, + "step": 269 + }, + { + "epoch": 0.845732184808144, + "grad_norm": 0.7263452707909194, + "learning_rate": 1.971142982805077e-05, + "loss": 1.1131, + "step": 270 + }, + { + "epoch": 0.8488645262333594, + "grad_norm": 0.7354286459707577, + "learning_rate": 1.9706185210019648e-05, + "loss": 1.1676, + "step": 271 + }, + { + "epoch": 0.8519968676585747, + "grad_norm": 0.7423168880868267, + "learning_rate": 1.9700894071491736e-05, + "loss": 1.1529, + "step": 272 + }, + { + "epoch": 0.8551292090837901, + "grad_norm": 0.7524433622381994, + "learning_rate": 1.969555643782677e-05, + "loss": 1.11, + "step": 273 + }, + { + "epoch": 0.8582615505090054, + "grad_norm": 0.7553141589784657, + "learning_rate": 1.969017233460736e-05, + "loss": 1.1777, + "step": 274 + }, + { + "epoch": 0.8613938919342208, + "grad_norm": 0.7431732388608101, + "learning_rate": 1.9684741787638808e-05, + "loss": 1.1588, + "step": 275 + }, + { + "epoch": 0.8645262333594361, + "grad_norm": 0.7563523934795757, + "learning_rate": 1.9679264822949027e-05, + "loss": 1.1277, + "step": 276 + }, + { + "epoch": 0.8676585747846516, + "grad_norm": 0.743913029928575, + "learning_rate": 1.967374146678841e-05, + "loss": 1.0578, + "step": 277 + }, + { + "epoch": 0.8707909162098669, + "grad_norm": 0.76235114176879, + "learning_rate": 1.966817174562968e-05, + "loss": 1.1283, + "step": 278 + }, + { + "epoch": 0.8739232576350823, + "grad_norm": 0.7537117820426102, + "learning_rate": 1.9662555686167808e-05, + "loss": 1.1312, + "step": 279 + }, + { + "epoch": 0.8770555990602976, + "grad_norm": 0.7425818024035084, + "learning_rate": 1.9656893315319838e-05, + "loss": 1.1127, + "step": 280 + }, + { + "epoch": 0.880187940485513, + "grad_norm": 0.7902486001191769, + "learning_rate": 1.9651184660224786e-05, + "loss": 1.1369, + "step": 281 + }, + { + "epoch": 0.8833202819107283, + "grad_norm": 0.7580680784170756, + "learning_rate": 1.96454297482435e-05, + "loss": 1.0847, + "step": 282 + }, + { + "epoch": 0.8864526233359437, + "grad_norm": 0.7813627798317231, + "learning_rate": 1.9639628606958535e-05, + "loss": 1.0851, + "step": 283 + }, + { + "epoch": 0.889584964761159, + "grad_norm": 0.7613328270635236, + "learning_rate": 1.963378126417401e-05, + "loss": 1.077, + "step": 284 + }, + { + "epoch": 0.8927173061863743, + "grad_norm": 0.7773885553361843, + "learning_rate": 1.9627887747915496e-05, + "loss": 1.0678, + "step": 285 + }, + { + "epoch": 0.8958496476115897, + "grad_norm": 0.7650287423748288, + "learning_rate": 1.9621948086429847e-05, + "loss": 1.1636, + "step": 286 + }, + { + "epoch": 0.898981989036805, + "grad_norm": 0.7817877274905041, + "learning_rate": 1.96159623081851e-05, + "loss": 1.1262, + "step": 287 + }, + { + "epoch": 0.9021143304620204, + "grad_norm": 0.7527999272006599, + "learning_rate": 1.960993044187032e-05, + "loss": 1.1082, + "step": 288 + }, + { + "epoch": 0.9052466718872357, + "grad_norm": 0.7350684686021072, + "learning_rate": 1.9603852516395472e-05, + "loss": 1.1019, + "step": 289 + }, + { + "epoch": 0.9083790133124511, + "grad_norm": 0.7222753898804246, + "learning_rate": 1.9597728560891266e-05, + "loss": 1.1204, + "step": 290 + }, + { + "epoch": 0.9115113547376664, + "grad_norm": 0.7894532542241856, + "learning_rate": 1.9591558604709028e-05, + "loss": 1.1405, + "step": 291 + }, + { + "epoch": 0.9146436961628818, + "grad_norm": 0.7634432547529683, + "learning_rate": 1.958534267742057e-05, + "loss": 1.1368, + "step": 292 + }, + { + "epoch": 0.9177760375880971, + "grad_norm": 0.763149271096417, + "learning_rate": 1.9579080808818035e-05, + "loss": 1.122, + "step": 293 + }, + { + "epoch": 0.9209083790133125, + "grad_norm": 0.7380685951966892, + "learning_rate": 1.9572773028913744e-05, + "loss": 1.1044, + "step": 294 + }, + { + "epoch": 0.9240407204385278, + "grad_norm": 0.7455823416370709, + "learning_rate": 1.956641936794008e-05, + "loss": 1.1205, + "step": 295 + }, + { + "epoch": 0.9271730618637432, + "grad_norm": 0.7364688923639685, + "learning_rate": 1.956001985634932e-05, + "loss": 1.1512, + "step": 296 + }, + { + "epoch": 0.9303054032889585, + "grad_norm": 0.7134786654834753, + "learning_rate": 1.9553574524813507e-05, + "loss": 1.1056, + "step": 297 + }, + { + "epoch": 0.9334377447141738, + "grad_norm": 0.7170844865094715, + "learning_rate": 1.9547083404224273e-05, + "loss": 1.1325, + "step": 298 + }, + { + "epoch": 0.9365700861393892, + "grad_norm": 0.6823076058596949, + "learning_rate": 1.9540546525692736e-05, + "loss": 1.0178, + "step": 299 + }, + { + "epoch": 0.9397024275646045, + "grad_norm": 0.7416982640172877, + "learning_rate": 1.9533963920549307e-05, + "loss": 1.1293, + "step": 300 + }, + { + "epoch": 0.9428347689898199, + "grad_norm": 0.6949473720891305, + "learning_rate": 1.9527335620343567e-05, + "loss": 1.1152, + "step": 301 + }, + { + "epoch": 0.9459671104150352, + "grad_norm": 0.7300442646550352, + "learning_rate": 1.9520661656844107e-05, + "loss": 1.1324, + "step": 302 + }, + { + "epoch": 0.9490994518402506, + "grad_norm": 0.7453171429773002, + "learning_rate": 1.9513942062038375e-05, + "loss": 1.1833, + "step": 303 + }, + { + "epoch": 0.9522317932654659, + "grad_norm": 0.7387019005056833, + "learning_rate": 1.9507176868132517e-05, + "loss": 1.101, + "step": 304 + }, + { + "epoch": 0.9553641346906813, + "grad_norm": 0.7116175190537836, + "learning_rate": 1.950036610755125e-05, + "loss": 1.1204, + "step": 305 + }, + { + "epoch": 0.9584964761158966, + "grad_norm": 0.7466335049949244, + "learning_rate": 1.9493509812937666e-05, + "loss": 1.1376, + "step": 306 + }, + { + "epoch": 0.961628817541112, + "grad_norm": 0.7380075174756867, + "learning_rate": 1.94866080171531e-05, + "loss": 1.1225, + "step": 307 + }, + { + "epoch": 0.9647611589663273, + "grad_norm": 0.747788910002534, + "learning_rate": 1.9479660753276978e-05, + "loss": 1.1651, + "step": 308 + }, + { + "epoch": 0.9678935003915426, + "grad_norm": 0.7277939144026558, + "learning_rate": 1.947266805460664e-05, + "loss": 1.1005, + "step": 309 + }, + { + "epoch": 0.971025841816758, + "grad_norm": 0.7594540196531739, + "learning_rate": 1.9465629954657185e-05, + "loss": 1.131, + "step": 310 + }, + { + "epoch": 0.9741581832419733, + "grad_norm": 0.7685734409142018, + "learning_rate": 1.9458546487161326e-05, + "loss": 1.1645, + "step": 311 + }, + { + "epoch": 0.9772905246671887, + "grad_norm": 0.7221986487220532, + "learning_rate": 1.945141768606921e-05, + "loss": 1.0698, + "step": 312 + }, + { + "epoch": 0.980422866092404, + "grad_norm": 0.7136839321857378, + "learning_rate": 1.9444243585548266e-05, + "loss": 1.1085, + "step": 313 + }, + { + "epoch": 0.9835552075176194, + "grad_norm": 0.7257358526029265, + "learning_rate": 1.943702421998303e-05, + "loss": 1.088, + "step": 314 + }, + { + "epoch": 0.9866875489428347, + "grad_norm": 0.7283314263257034, + "learning_rate": 1.9429759623974992e-05, + "loss": 1.1275, + "step": 315 + }, + { + "epoch": 0.9898198903680501, + "grad_norm": 0.7150867610674063, + "learning_rate": 1.942244983234243e-05, + "loss": 1.1191, + "step": 316 + }, + { + "epoch": 0.9929522317932654, + "grad_norm": 0.7256772289345883, + "learning_rate": 1.9415094880120236e-05, + "loss": 1.1267, + "step": 317 + }, + { + "epoch": 0.9960845732184808, + "grad_norm": 0.7160733246362201, + "learning_rate": 1.940769480255974e-05, + "loss": 1.0924, + "step": 318 + }, + { + "epoch": 0.9992169146436961, + "grad_norm": 0.6968857896809572, + "learning_rate": 1.9400249635128575e-05, + "loss": 1.0343, + "step": 319 + }, + { + "epoch": 1.0, + "grad_norm": 0.6968857896809572, + "learning_rate": 1.939275941351046e-05, + "loss": 1.1068, + "step": 320 + }, + { + "epoch": 1.0031323414252153, + "grad_norm": 1.6156520613544338, + "learning_rate": 1.9385224173605072e-05, + "loss": 1.0187, + "step": 321 + }, + { + "epoch": 1.0062646828504307, + "grad_norm": 0.8615572642140844, + "learning_rate": 1.9377643951527834e-05, + "loss": 0.9501, + "step": 322 + }, + { + "epoch": 1.009397024275646, + "grad_norm": 0.8236605914660184, + "learning_rate": 1.937001878360979e-05, + "loss": 0.9672, + "step": 323 + }, + { + "epoch": 1.0125293657008614, + "grad_norm": 0.81342630340967, + "learning_rate": 1.9362348706397374e-05, + "loss": 0.9559, + "step": 324 + }, + { + "epoch": 1.0156617071260767, + "grad_norm": 0.8581319434910718, + "learning_rate": 1.935463375665229e-05, + "loss": 0.9808, + "step": 325 + }, + { + "epoch": 1.018794048551292, + "grad_norm": 0.8129587194790275, + "learning_rate": 1.934687397135129e-05, + "loss": 0.9512, + "step": 326 + }, + { + "epoch": 1.0219263899765074, + "grad_norm": 0.8327468346046575, + "learning_rate": 1.933906938768604e-05, + "loss": 0.9776, + "step": 327 + }, + { + "epoch": 1.0250587314017228, + "grad_norm": 0.8873213443616459, + "learning_rate": 1.9331220043062894e-05, + "loss": 0.9568, + "step": 328 + }, + { + "epoch": 1.0281910728269381, + "grad_norm": 0.8763873438765251, + "learning_rate": 1.9323325975102757e-05, + "loss": 0.9854, + "step": 329 + }, + { + "epoch": 1.0313234142521535, + "grad_norm": 0.8737645235450247, + "learning_rate": 1.9315387221640874e-05, + "loss": 0.9766, + "step": 330 + }, + { + "epoch": 1.0344557556773688, + "grad_norm": 0.8226677263540446, + "learning_rate": 1.9307403820726675e-05, + "loss": 0.9439, + "step": 331 + }, + { + "epoch": 1.0375880971025842, + "grad_norm": 0.8561637749159361, + "learning_rate": 1.929937581062357e-05, + "loss": 0.962, + "step": 332 + }, + { + "epoch": 1.0407204385277995, + "grad_norm": 0.8096315231114, + "learning_rate": 1.929130322980878e-05, + "loss": 0.9454, + "step": 333 + }, + { + "epoch": 1.0438527799530148, + "grad_norm": 0.8289922628570354, + "learning_rate": 1.9283186116973146e-05, + "loss": 0.9844, + "step": 334 + }, + { + "epoch": 1.0469851213782302, + "grad_norm": 0.7901830667535128, + "learning_rate": 1.927502451102095e-05, + "loss": 0.9493, + "step": 335 + }, + { + "epoch": 1.0501174628034455, + "grad_norm": 0.7761466037885246, + "learning_rate": 1.926681845106971e-05, + "loss": 0.9465, + "step": 336 + }, + { + "epoch": 1.0532498042286609, + "grad_norm": 0.8141355109509392, + "learning_rate": 1.925856797645002e-05, + "loss": 0.95, + "step": 337 + }, + { + "epoch": 1.0563821456538762, + "grad_norm": 0.8074412075844187, + "learning_rate": 1.925027312670534e-05, + "loss": 0.9436, + "step": 338 + }, + { + "epoch": 1.0595144870790916, + "grad_norm": 0.8272827849228525, + "learning_rate": 1.9241933941591826e-05, + "loss": 0.9703, + "step": 339 + }, + { + "epoch": 1.062646828504307, + "grad_norm": 0.7914982733889313, + "learning_rate": 1.9233550461078114e-05, + "loss": 0.9512, + "step": 340 + }, + { + "epoch": 1.0657791699295223, + "grad_norm": 0.7994475682723842, + "learning_rate": 1.9225122725345145e-05, + "loss": 0.9456, + "step": 341 + }, + { + "epoch": 1.0689115113547376, + "grad_norm": 0.8465580424768789, + "learning_rate": 1.9216650774785975e-05, + "loss": 0.9432, + "step": 342 + }, + { + "epoch": 1.072043852779953, + "grad_norm": 0.8466406131696012, + "learning_rate": 1.920813465000557e-05, + "loss": 0.9697, + "step": 343 + }, + { + "epoch": 1.0751761942051683, + "grad_norm": 0.8033646805720518, + "learning_rate": 1.9199574391820627e-05, + "loss": 0.9541, + "step": 344 + }, + { + "epoch": 1.0783085356303836, + "grad_norm": 0.8224052298742338, + "learning_rate": 1.9190970041259352e-05, + "loss": 1.0003, + "step": 345 + }, + { + "epoch": 1.081440877055599, + "grad_norm": 0.7997901412752633, + "learning_rate": 1.91823216395613e-05, + "loss": 0.9903, + "step": 346 + }, + { + "epoch": 1.0845732184808143, + "grad_norm": 0.8140939297756549, + "learning_rate": 1.917362922817714e-05, + "loss": 0.9549, + "step": 347 + }, + { + "epoch": 1.0877055599060297, + "grad_norm": 0.7824285972233995, + "learning_rate": 1.9164892848768485e-05, + "loss": 0.9593, + "step": 348 + }, + { + "epoch": 1.090837901331245, + "grad_norm": 0.7836791598591205, + "learning_rate": 1.9156112543207674e-05, + "loss": 0.9582, + "step": 349 + }, + { + "epoch": 1.0939702427564604, + "grad_norm": 0.7911753479916747, + "learning_rate": 1.914728835357759e-05, + "loss": 0.9348, + "step": 350 + }, + { + "epoch": 1.0971025841816757, + "grad_norm": 0.7782991822882203, + "learning_rate": 1.913842032217143e-05, + "loss": 0.9631, + "step": 351 + }, + { + "epoch": 1.100234925606891, + "grad_norm": 0.8014360461302352, + "learning_rate": 1.912950849149253e-05, + "loss": 0.9351, + "step": 352 + }, + { + "epoch": 1.1033672670321064, + "grad_norm": 0.8312030203793829, + "learning_rate": 1.912055290425415e-05, + "loss": 0.9668, + "step": 353 + }, + { + "epoch": 1.1064996084573218, + "grad_norm": 0.7891597416989514, + "learning_rate": 1.911155360337926e-05, + "loss": 0.9432, + "step": 354 + }, + { + "epoch": 1.109631949882537, + "grad_norm": 0.7878291420752258, + "learning_rate": 1.9102510632000365e-05, + "loss": 0.9521, + "step": 355 + }, + { + "epoch": 1.1127642913077525, + "grad_norm": 0.8120607285651226, + "learning_rate": 1.909342403345925e-05, + "loss": 0.9425, + "step": 356 + }, + { + "epoch": 1.1158966327329678, + "grad_norm": 0.7856758097045018, + "learning_rate": 1.9084293851306815e-05, + "loss": 0.9542, + "step": 357 + }, + { + "epoch": 1.1190289741581831, + "grad_norm": 0.7684160480129255, + "learning_rate": 1.9075120129302857e-05, + "loss": 0.9454, + "step": 358 + }, + { + "epoch": 1.1221613155833985, + "grad_norm": 0.8316731012269218, + "learning_rate": 1.9065902911415836e-05, + "loss": 0.9919, + "step": 359 + }, + { + "epoch": 1.1252936570086138, + "grad_norm": 0.822006226986475, + "learning_rate": 1.905664224182269e-05, + "loss": 0.9751, + "step": 360 + }, + { + "epoch": 1.1284259984338294, + "grad_norm": 0.7696798368031216, + "learning_rate": 1.9047338164908628e-05, + "loss": 0.9717, + "step": 361 + }, + { + "epoch": 1.1315583398590445, + "grad_norm": 0.8046009384769924, + "learning_rate": 1.903799072526688e-05, + "loss": 0.8908, + "step": 362 + }, + { + "epoch": 1.13469068128426, + "grad_norm": 0.7979550376850155, + "learning_rate": 1.9028599967698533e-05, + "loss": 0.9904, + "step": 363 + }, + { + "epoch": 1.1378230227094752, + "grad_norm": 0.8107407192621531, + "learning_rate": 1.9019165937212273e-05, + "loss": 0.9528, + "step": 364 + }, + { + "epoch": 1.1409553641346908, + "grad_norm": 0.7767024196508194, + "learning_rate": 1.900968867902419e-05, + "loss": 0.9159, + "step": 365 + }, + { + "epoch": 1.144087705559906, + "grad_norm": 0.7927579094268248, + "learning_rate": 1.9000168238557574e-05, + "loss": 0.9252, + "step": 366 + }, + { + "epoch": 1.1472200469851215, + "grad_norm": 0.8295313527055139, + "learning_rate": 1.899060466144267e-05, + "loss": 0.941, + "step": 367 + }, + { + "epoch": 1.1503523884103368, + "grad_norm": 0.7746260340598791, + "learning_rate": 1.8980997993516462e-05, + "loss": 0.9237, + "step": 368 + }, + { + "epoch": 1.1534847298355522, + "grad_norm": 0.7971878626092886, + "learning_rate": 1.897134828082248e-05, + "loss": 0.9387, + "step": 369 + }, + { + "epoch": 1.1566170712607675, + "grad_norm": 0.807852831490928, + "learning_rate": 1.8961655569610557e-05, + "loss": 0.9958, + "step": 370 + }, + { + "epoch": 1.1597494126859829, + "grad_norm": 0.8147989237592862, + "learning_rate": 1.895191990633661e-05, + "loss": 0.953, + "step": 371 + }, + { + "epoch": 1.1628817541111982, + "grad_norm": 0.7959883702024442, + "learning_rate": 1.8942141337662415e-05, + "loss": 0.9419, + "step": 372 + }, + { + "epoch": 1.1660140955364136, + "grad_norm": 0.7895917093808684, + "learning_rate": 1.89323199104554e-05, + "loss": 1.0, + "step": 373 + }, + { + "epoch": 1.169146436961629, + "grad_norm": 0.7788706927325837, + "learning_rate": 1.89224556717884e-05, + "loss": 0.9641, + "step": 374 + }, + { + "epoch": 1.1722787783868442, + "grad_norm": 0.8275735256549743, + "learning_rate": 1.891254866893944e-05, + "loss": 0.9596, + "step": 375 + }, + { + "epoch": 1.1754111198120596, + "grad_norm": 0.7901905557676397, + "learning_rate": 1.8902598949391507e-05, + "loss": 0.9881, + "step": 376 + }, + { + "epoch": 1.178543461237275, + "grad_norm": 0.7938940344571249, + "learning_rate": 1.8892606560832335e-05, + "loss": 0.9056, + "step": 377 + }, + { + "epoch": 1.1816758026624903, + "grad_norm": 0.7774862380646365, + "learning_rate": 1.8882571551154152e-05, + "loss": 0.9408, + "step": 378 + }, + { + "epoch": 1.1848081440877056, + "grad_norm": 0.8020328348235701, + "learning_rate": 1.8872493968453467e-05, + "loss": 0.9251, + "step": 379 + }, + { + "epoch": 1.187940485512921, + "grad_norm": 0.806279534293864, + "learning_rate": 1.8862373861030838e-05, + "loss": 0.948, + "step": 380 + }, + { + "epoch": 1.1910728269381363, + "grad_norm": 0.8025985732944699, + "learning_rate": 1.885221127739064e-05, + "loss": 0.9588, + "step": 381 + }, + { + "epoch": 1.1942051683633517, + "grad_norm": 0.7681611064314322, + "learning_rate": 1.8842006266240826e-05, + "loss": 0.9286, + "step": 382 + }, + { + "epoch": 1.197337509788567, + "grad_norm": 0.7830003805481669, + "learning_rate": 1.88317588764927e-05, + "loss": 0.9641, + "step": 383 + }, + { + "epoch": 1.2004698512137824, + "grad_norm": 0.7682889170068867, + "learning_rate": 1.8821469157260687e-05, + "loss": 0.9293, + "step": 384 + }, + { + "epoch": 1.2036021926389977, + "grad_norm": 0.8006318522209638, + "learning_rate": 1.8811137157862084e-05, + "loss": 0.9283, + "step": 385 + }, + { + "epoch": 1.206734534064213, + "grad_norm": 0.8064409365222602, + "learning_rate": 1.8800762927816835e-05, + "loss": 0.9974, + "step": 386 + }, + { + "epoch": 1.2098668754894284, + "grad_norm": 0.7771414813466695, + "learning_rate": 1.8790346516847288e-05, + "loss": 0.9414, + "step": 387 + }, + { + "epoch": 1.2129992169146437, + "grad_norm": 0.8516980588228736, + "learning_rate": 1.8779887974877965e-05, + "loss": 0.9708, + "step": 388 + }, + { + "epoch": 1.216131558339859, + "grad_norm": 0.8233081545979598, + "learning_rate": 1.876938735203531e-05, + "loss": 0.9856, + "step": 389 + }, + { + "epoch": 1.2192638997650744, + "grad_norm": 0.8148185121214713, + "learning_rate": 1.8758844698647457e-05, + "loss": 0.9441, + "step": 390 + }, + { + "epoch": 1.2223962411902898, + "grad_norm": 0.8121759062847519, + "learning_rate": 1.8748260065243985e-05, + "loss": 0.9865, + "step": 391 + }, + { + "epoch": 1.2255285826155051, + "grad_norm": 0.8027197651573186, + "learning_rate": 1.8737633502555685e-05, + "loss": 0.9809, + "step": 392 + }, + { + "epoch": 1.2286609240407205, + "grad_norm": 0.8058527578911673, + "learning_rate": 1.8726965061514297e-05, + "loss": 0.9611, + "step": 393 + }, + { + "epoch": 1.2317932654659358, + "grad_norm": 0.7952817742989116, + "learning_rate": 1.8716254793252294e-05, + "loss": 0.9786, + "step": 394 + }, + { + "epoch": 1.2349256068911512, + "grad_norm": 0.7839895990296804, + "learning_rate": 1.870550274910261e-05, + "loss": 0.9601, + "step": 395 + }, + { + "epoch": 1.2380579483163665, + "grad_norm": 0.8549762083059709, + "learning_rate": 1.869470898059841e-05, + "loss": 1.0219, + "step": 396 + }, + { + "epoch": 1.2411902897415819, + "grad_norm": 0.788953768471007, + "learning_rate": 1.868387353947284e-05, + "loss": 1.0143, + "step": 397 + }, + { + "epoch": 1.2443226311667972, + "grad_norm": 0.8217786523602275, + "learning_rate": 1.8672996477658767e-05, + "loss": 0.9523, + "step": 398 + }, + { + "epoch": 1.2474549725920125, + "grad_norm": 0.7704666004345789, + "learning_rate": 1.8662077847288554e-05, + "loss": 0.9207, + "step": 399 + }, + { + "epoch": 1.250587314017228, + "grad_norm": 0.7525516825767109, + "learning_rate": 1.8651117700693794e-05, + "loss": 0.9477, + "step": 400 + }, + { + "epoch": 1.2537196554424432, + "grad_norm": 0.8253629375930066, + "learning_rate": 1.8640116090405055e-05, + "loss": 0.9731, + "step": 401 + }, + { + "epoch": 1.2568519968676586, + "grad_norm": 0.7915558764489208, + "learning_rate": 1.8629073069151638e-05, + "loss": 0.9199, + "step": 402 + }, + { + "epoch": 1.259984338292874, + "grad_norm": 0.7782487189596384, + "learning_rate": 1.8617988689861333e-05, + "loss": 0.9199, + "step": 403 + }, + { + "epoch": 1.2631166797180893, + "grad_norm": 0.7980723005204162, + "learning_rate": 1.8606863005660133e-05, + "loss": 0.9212, + "step": 404 + }, + { + "epoch": 1.2662490211433046, + "grad_norm": 0.8048484955000166, + "learning_rate": 1.8595696069872013e-05, + "loss": 0.9581, + "step": 405 + }, + { + "epoch": 1.26938136256852, + "grad_norm": 0.8032453880110384, + "learning_rate": 1.8584487936018663e-05, + "loss": 0.9821, + "step": 406 + }, + { + "epoch": 1.2725137039937353, + "grad_norm": 0.8005507805303662, + "learning_rate": 1.8573238657819225e-05, + "loss": 0.9411, + "step": 407 + }, + { + "epoch": 1.2756460454189507, + "grad_norm": 0.7916521575263914, + "learning_rate": 1.8561948289190035e-05, + "loss": 0.9357, + "step": 408 + }, + { + "epoch": 1.278778386844166, + "grad_norm": 0.7979831477850623, + "learning_rate": 1.855061688424439e-05, + "loss": 0.9679, + "step": 409 + }, + { + "epoch": 1.2819107282693813, + "grad_norm": 0.7843000585547435, + "learning_rate": 1.8539244497292248e-05, + "loss": 0.9636, + "step": 410 + }, + { + "epoch": 1.2850430696945967, + "grad_norm": 0.784278644818112, + "learning_rate": 1.8527831182839997e-05, + "loss": 1.0007, + "step": 411 + }, + { + "epoch": 1.288175411119812, + "grad_norm": 0.7946831238148018, + "learning_rate": 1.8516376995590185e-05, + "loss": 0.9813, + "step": 412 + }, + { + "epoch": 1.2913077525450274, + "grad_norm": 0.8109210495548596, + "learning_rate": 1.850488199044126e-05, + "loss": 0.9349, + "step": 413 + }, + { + "epoch": 1.2944400939702427, + "grad_norm": 0.7934617847842895, + "learning_rate": 1.84933462224873e-05, + "loss": 0.9416, + "step": 414 + }, + { + "epoch": 1.297572435395458, + "grad_norm": 0.7902841843838354, + "learning_rate": 1.848176974701775e-05, + "loss": 0.9346, + "step": 415 + }, + { + "epoch": 1.3007047768206734, + "grad_norm": 0.779580351091375, + "learning_rate": 1.847015261951718e-05, + "loss": 0.9739, + "step": 416 + }, + { + "epoch": 1.3038371182458888, + "grad_norm": 0.7785529627916624, + "learning_rate": 1.845849489566498e-05, + "loss": 0.9583, + "step": 417 + }, + { + "epoch": 1.3069694596711041, + "grad_norm": 0.8436224516793075, + "learning_rate": 1.8446796631335123e-05, + "loss": 0.9681, + "step": 418 + }, + { + "epoch": 1.3101018010963195, + "grad_norm": 0.8155304400749016, + "learning_rate": 1.8435057882595885e-05, + "loss": 0.9455, + "step": 419 + }, + { + "epoch": 1.3132341425215348, + "grad_norm": 0.8154776896881121, + "learning_rate": 1.8423278705709573e-05, + "loss": 0.9319, + "step": 420 + }, + { + "epoch": 1.3163664839467502, + "grad_norm": 0.7875316068008983, + "learning_rate": 1.841145915713227e-05, + "loss": 0.9839, + "step": 421 + }, + { + "epoch": 1.3194988253719655, + "grad_norm": 0.7878642203922693, + "learning_rate": 1.8399599293513555e-05, + "loss": 0.9611, + "step": 422 + }, + { + "epoch": 1.3226311667971808, + "grad_norm": 0.8272412655909572, + "learning_rate": 1.8387699171696222e-05, + "loss": 0.9676, + "step": 423 + }, + { + "epoch": 1.3257635082223962, + "grad_norm": 0.8036668057730373, + "learning_rate": 1.837575884871603e-05, + "loss": 0.9585, + "step": 424 + }, + { + "epoch": 1.3288958496476115, + "grad_norm": 0.7745104838154117, + "learning_rate": 1.8363778381801404e-05, + "loss": 0.9586, + "step": 425 + }, + { + "epoch": 1.3320281910728269, + "grad_norm": 0.8044310075899959, + "learning_rate": 1.8351757828373183e-05, + "loss": 0.9773, + "step": 426 + }, + { + "epoch": 1.3351605324980422, + "grad_norm": 0.7816652761458432, + "learning_rate": 1.8339697246044333e-05, + "loss": 0.9483, + "step": 427 + }, + { + "epoch": 1.3382928739232576, + "grad_norm": 0.805576362415665, + "learning_rate": 1.8327596692619674e-05, + "loss": 1.0151, + "step": 428 + }, + { + "epoch": 1.341425215348473, + "grad_norm": 0.7789246864462362, + "learning_rate": 1.83154562260956e-05, + "loss": 0.9395, + "step": 429 + }, + { + "epoch": 1.3445575567736883, + "grad_norm": 0.7843379086528441, + "learning_rate": 1.8303275904659807e-05, + "loss": 0.9632, + "step": 430 + }, + { + "epoch": 1.3476898981989036, + "grad_norm": 0.7738505946833321, + "learning_rate": 1.829105578669101e-05, + "loss": 0.898, + "step": 431 + }, + { + "epoch": 1.350822239624119, + "grad_norm": 0.773409651313221, + "learning_rate": 1.8278795930758656e-05, + "loss": 0.925, + "step": 432 + }, + { + "epoch": 1.3539545810493343, + "grad_norm": 0.775638063691069, + "learning_rate": 1.826649639562266e-05, + "loss": 0.9296, + "step": 433 + }, + { + "epoch": 1.3570869224745497, + "grad_norm": 0.8255837346883753, + "learning_rate": 1.8254157240233114e-05, + "loss": 0.9578, + "step": 434 + }, + { + "epoch": 1.360219263899765, + "grad_norm": 0.8225702778393079, + "learning_rate": 1.8241778523729997e-05, + "loss": 0.9938, + "step": 435 + }, + { + "epoch": 1.3633516053249803, + "grad_norm": 0.8024681843740595, + "learning_rate": 1.8229360305442906e-05, + "loss": 0.9898, + "step": 436 + }, + { + "epoch": 1.3664839467501957, + "grad_norm": 0.7659268418287953, + "learning_rate": 1.8216902644890768e-05, + "loss": 0.9582, + "step": 437 + }, + { + "epoch": 1.3696162881754113, + "grad_norm": 0.7979578032515822, + "learning_rate": 1.8204405601781542e-05, + "loss": 0.9539, + "step": 438 + }, + { + "epoch": 1.3727486296006264, + "grad_norm": 0.7840678597712154, + "learning_rate": 1.819186923601195e-05, + "loss": 0.9662, + "step": 439 + }, + { + "epoch": 1.375880971025842, + "grad_norm": 0.7951677564440199, + "learning_rate": 1.8179293607667177e-05, + "loss": 0.963, + "step": 440 + }, + { + "epoch": 1.379013312451057, + "grad_norm": 0.7824747662278654, + "learning_rate": 1.8166678777020595e-05, + "loss": 0.9646, + "step": 441 + }, + { + "epoch": 1.3821456538762726, + "grad_norm": 0.7889899716482371, + "learning_rate": 1.8154024804533464e-05, + "loss": 0.9949, + "step": 442 + }, + { + "epoch": 1.3852779953014878, + "grad_norm": 0.8294307839008276, + "learning_rate": 1.8141331750854647e-05, + "loss": 0.9745, + "step": 443 + }, + { + "epoch": 1.3884103367267033, + "grad_norm": 0.8026550217216318, + "learning_rate": 1.8128599676820306e-05, + "loss": 0.9605, + "step": 444 + }, + { + "epoch": 1.3915426781519185, + "grad_norm": 0.8043494358409277, + "learning_rate": 1.8115828643453647e-05, + "loss": 0.9419, + "step": 445 + }, + { + "epoch": 1.394675019577134, + "grad_norm": 0.7904914993084787, + "learning_rate": 1.8103018711964588e-05, + "loss": 0.9636, + "step": 446 + }, + { + "epoch": 1.3978073610023491, + "grad_norm": 0.833878643616376, + "learning_rate": 1.8090169943749477e-05, + "loss": 0.9799, + "step": 447 + }, + { + "epoch": 1.4009397024275647, + "grad_norm": 0.8351491373093172, + "learning_rate": 1.807728240039081e-05, + "loss": 0.954, + "step": 448 + }, + { + "epoch": 1.4040720438527798, + "grad_norm": 0.8413539861179071, + "learning_rate": 1.806435614365692e-05, + "loss": 0.9939, + "step": 449 + }, + { + "epoch": 1.4072043852779954, + "grad_norm": 0.8339283054897446, + "learning_rate": 1.8051391235501695e-05, + "loss": 0.9699, + "step": 450 + }, + { + "epoch": 1.4103367267032105, + "grad_norm": 0.8013532736907198, + "learning_rate": 1.8038387738064274e-05, + "loss": 0.9581, + "step": 451 + }, + { + "epoch": 1.413469068128426, + "grad_norm": 0.8255816811624092, + "learning_rate": 1.8025345713668746e-05, + "loss": 1.0, + "step": 452 + }, + { + "epoch": 1.4166014095536412, + "grad_norm": 0.8242603985956736, + "learning_rate": 1.8012265224823853e-05, + "loss": 0.9364, + "step": 453 + }, + { + "epoch": 1.4197337509788568, + "grad_norm": 0.8322278182809092, + "learning_rate": 1.79991463342227e-05, + "loss": 0.9851, + "step": 454 + }, + { + "epoch": 1.422866092404072, + "grad_norm": 0.8163638205578437, + "learning_rate": 1.7985989104742434e-05, + "loss": 0.9499, + "step": 455 + }, + { + "epoch": 1.4259984338292875, + "grad_norm": 0.8442359156970612, + "learning_rate": 1.7972793599443966e-05, + "loss": 0.9491, + "step": 456 + }, + { + "epoch": 1.4291307752545026, + "grad_norm": 0.7893355911335334, + "learning_rate": 1.795955988157166e-05, + "loss": 0.9816, + "step": 457 + }, + { + "epoch": 1.4322631166797182, + "grad_norm": 0.806471339752882, + "learning_rate": 1.7946288014553017e-05, + "loss": 0.9549, + "step": 458 + }, + { + "epoch": 1.4353954581049335, + "grad_norm": 0.8364165434031879, + "learning_rate": 1.7932978061998393e-05, + "loss": 0.9836, + "step": 459 + }, + { + "epoch": 1.4385277995301489, + "grad_norm": 0.7859146658803116, + "learning_rate": 1.7919630087700672e-05, + "loss": 0.9842, + "step": 460 + }, + { + "epoch": 1.4416601409553642, + "grad_norm": 0.7690910792708224, + "learning_rate": 1.790624415563498e-05, + "loss": 0.9396, + "step": 461 + }, + { + "epoch": 1.4447924823805796, + "grad_norm": 0.8496502770784555, + "learning_rate": 1.7892820329958368e-05, + "loss": 0.9579, + "step": 462 + }, + { + "epoch": 1.447924823805795, + "grad_norm": 0.8471130092608972, + "learning_rate": 1.7879358675009506e-05, + "loss": 0.9973, + "step": 463 + }, + { + "epoch": 1.4510571652310102, + "grad_norm": 0.784501551931219, + "learning_rate": 1.786585925530837e-05, + "loss": 0.9641, + "step": 464 + }, + { + "epoch": 1.4541895066562256, + "grad_norm": 0.836729539160783, + "learning_rate": 1.7852322135555946e-05, + "loss": 0.9319, + "step": 465 + }, + { + "epoch": 1.457321848081441, + "grad_norm": 0.8356458620469223, + "learning_rate": 1.7838747380633904e-05, + "loss": 0.9571, + "step": 466 + }, + { + "epoch": 1.4604541895066563, + "grad_norm": 0.8216191110677196, + "learning_rate": 1.78251350556043e-05, + "loss": 0.9553, + "step": 467 + }, + { + "epoch": 1.4635865309318716, + "grad_norm": 0.7966731084754197, + "learning_rate": 1.7811485225709255e-05, + "loss": 0.9519, + "step": 468 + }, + { + "epoch": 1.466718872357087, + "grad_norm": 0.869796047610942, + "learning_rate": 1.7797797956370646e-05, + "loss": 0.9836, + "step": 469 + }, + { + "epoch": 1.4698512137823023, + "grad_norm": 0.8527775209514664, + "learning_rate": 1.7784073313189796e-05, + "loss": 0.9559, + "step": 470 + }, + { + "epoch": 1.4729835552075177, + "grad_norm": 0.7722201957501883, + "learning_rate": 1.7770311361947155e-05, + "loss": 0.9424, + "step": 471 + }, + { + "epoch": 1.476115896632733, + "grad_norm": 0.8231174830960291, + "learning_rate": 1.775651216860198e-05, + "loss": 1.0208, + "step": 472 + }, + { + "epoch": 1.4792482380579484, + "grad_norm": 0.8323084233244761, + "learning_rate": 1.7742675799292033e-05, + "loss": 1.0034, + "step": 473 + }, + { + "epoch": 1.4823805794831637, + "grad_norm": 0.8153252414548278, + "learning_rate": 1.772880232033325e-05, + "loss": 0.9844, + "step": 474 + }, + { + "epoch": 1.485512920908379, + "grad_norm": 0.8189526410041873, + "learning_rate": 1.7714891798219432e-05, + "loss": 0.9666, + "step": 475 + }, + { + "epoch": 1.4886452623335944, + "grad_norm": 0.8078003198916532, + "learning_rate": 1.770094429962192e-05, + "loss": 0.944, + "step": 476 + }, + { + "epoch": 1.4917776037588097, + "grad_norm": 0.8110193552079895, + "learning_rate": 1.768695989138928e-05, + "loss": 0.9544, + "step": 477 + }, + { + "epoch": 1.494909945184025, + "grad_norm": 0.8312565557198122, + "learning_rate": 1.767293864054698e-05, + "loss": 0.9701, + "step": 478 + }, + { + "epoch": 1.4980422866092404, + "grad_norm": 0.8061090156479217, + "learning_rate": 1.7658880614297075e-05, + "loss": 0.9687, + "step": 479 + }, + { + "epoch": 1.5011746280344558, + "grad_norm": 0.823951800509433, + "learning_rate": 1.7644785880017874e-05, + "loss": 0.9723, + "step": 480 + }, + { + "epoch": 1.5043069694596711, + "grad_norm": 0.7965679907922073, + "learning_rate": 1.7630654505263628e-05, + "loss": 0.9073, + "step": 481 + }, + { + "epoch": 1.5074393108848865, + "grad_norm": 0.9569281276943931, + "learning_rate": 1.7616486557764187e-05, + "loss": 0.9808, + "step": 482 + }, + { + "epoch": 1.5105716523101018, + "grad_norm": 0.8085640633910255, + "learning_rate": 1.7602282105424712e-05, + "loss": 0.9732, + "step": 483 + }, + { + "epoch": 1.5137039937353172, + "grad_norm": 0.8532004513694563, + "learning_rate": 1.7588041216325314e-05, + "loss": 0.9507, + "step": 484 + }, + { + "epoch": 1.5168363351605325, + "grad_norm": 0.7833644189128819, + "learning_rate": 1.7573763958720736e-05, + "loss": 0.9411, + "step": 485 + }, + { + "epoch": 1.5199686765857479, + "grad_norm": 0.8189040727283652, + "learning_rate": 1.755945040104004e-05, + "loss": 0.9655, + "step": 486 + }, + { + "epoch": 1.5231010180109632, + "grad_norm": 0.7722121152681558, + "learning_rate": 1.754510061188627e-05, + "loss": 0.9185, + "step": 487 + }, + { + "epoch": 1.5262333594361785, + "grad_norm": 0.8037152368297265, + "learning_rate": 1.7530714660036112e-05, + "loss": 0.9541, + "step": 488 + }, + { + "epoch": 1.529365700861394, + "grad_norm": 0.8119393404590528, + "learning_rate": 1.7516292614439586e-05, + "loss": 0.9657, + "step": 489 + }, + { + "epoch": 1.5324980422866092, + "grad_norm": 0.805216081294044, + "learning_rate": 1.7501834544219697e-05, + "loss": 0.9448, + "step": 490 + }, + { + "epoch": 1.5356303837118246, + "grad_norm": 0.7949712346927176, + "learning_rate": 1.748734051867212e-05, + "loss": 0.9463, + "step": 491 + }, + { + "epoch": 1.53876272513704, + "grad_norm": 0.77990875986472, + "learning_rate": 1.7472810607264853e-05, + "loss": 0.9422, + "step": 492 + }, + { + "epoch": 1.5418950665622553, + "grad_norm": 0.7886283708612627, + "learning_rate": 1.7458244879637897e-05, + "loss": 0.9487, + "step": 493 + }, + { + "epoch": 1.5450274079874706, + "grad_norm": 0.7958418765522267, + "learning_rate": 1.74436434056029e-05, + "loss": 0.9714, + "step": 494 + }, + { + "epoch": 1.548159749412686, + "grad_norm": 0.8069518461832199, + "learning_rate": 1.742900625514285e-05, + "loss": 0.9829, + "step": 495 + }, + { + "epoch": 1.5512920908379013, + "grad_norm": 0.8073763682254401, + "learning_rate": 1.7414333498411734e-05, + "loss": 0.9569, + "step": 496 + }, + { + "epoch": 1.5544244322631167, + "grad_norm": 0.8342340049061451, + "learning_rate": 1.739962520573418e-05, + "loss": 1.0084, + "step": 497 + }, + { + "epoch": 1.557556773688332, + "grad_norm": 0.8120689308519743, + "learning_rate": 1.7384881447605144e-05, + "loss": 1.0443, + "step": 498 + }, + { + "epoch": 1.5606891151135474, + "grad_norm": 0.8010972622074474, + "learning_rate": 1.7370102294689563e-05, + "loss": 0.959, + "step": 499 + }, + { + "epoch": 1.5638214565387627, + "grad_norm": 0.7709293012877854, + "learning_rate": 1.7355287817822014e-05, + "loss": 0.9864, + "step": 500 + }, + { + "epoch": 1.566953797963978, + "grad_norm": 0.799753436392405, + "learning_rate": 1.7340438088006376e-05, + "loss": 0.9981, + "step": 501 + }, + { + "epoch": 1.5700861393891934, + "grad_norm": 0.7779275987751509, + "learning_rate": 1.7325553176415496e-05, + "loss": 0.9644, + "step": 502 + }, + { + "epoch": 1.5732184808144087, + "grad_norm": 0.8008846365371122, + "learning_rate": 1.731063315439084e-05, + "loss": 0.9368, + "step": 503 + }, + { + "epoch": 1.576350822239624, + "grad_norm": 0.774005109549041, + "learning_rate": 1.7295678093442155e-05, + "loss": 0.9435, + "step": 504 + }, + { + "epoch": 1.5794831636648394, + "grad_norm": 0.795159470143028, + "learning_rate": 1.728068806524712e-05, + "loss": 1.0071, + "step": 505 + }, + { + "epoch": 1.5826155050900548, + "grad_norm": 0.7963785520950486, + "learning_rate": 1.7265663141651016e-05, + "loss": 0.9922, + "step": 506 + }, + { + "epoch": 1.5857478465152701, + "grad_norm": 0.7877046535940556, + "learning_rate": 1.7250603394666364e-05, + "loss": 0.9514, + "step": 507 + }, + { + "epoch": 1.5888801879404855, + "grad_norm": 0.7715293903451117, + "learning_rate": 1.7235508896472593e-05, + "loss": 0.933, + "step": 508 + }, + { + "epoch": 1.5920125293657008, + "grad_norm": 0.7934117596728014, + "learning_rate": 1.72203797194157e-05, + "loss": 0.973, + "step": 509 + }, + { + "epoch": 1.5951448707909162, + "grad_norm": 0.8181710440792368, + "learning_rate": 1.720521593600787e-05, + "loss": 0.986, + "step": 510 + }, + { + "epoch": 1.5982772122161315, + "grad_norm": 0.7837650575509727, + "learning_rate": 1.7190017618927177e-05, + "loss": 0.9612, + "step": 511 + }, + { + "epoch": 1.6014095536413469, + "grad_norm": 0.8412641440444018, + "learning_rate": 1.7174784841017197e-05, + "loss": 0.9745, + "step": 512 + }, + { + "epoch": 1.6045418950665624, + "grad_norm": 0.7881270804166439, + "learning_rate": 1.7159517675286667e-05, + "loss": 0.9644, + "step": 513 + }, + { + "epoch": 1.6076742364917775, + "grad_norm": 0.8119820827446707, + "learning_rate": 1.7144216194909157e-05, + "loss": 0.9886, + "step": 514 + }, + { + "epoch": 1.610806577916993, + "grad_norm": 0.7998253925780444, + "learning_rate": 1.7128880473222688e-05, + "loss": 0.9815, + "step": 515 + }, + { + "epoch": 1.6139389193422082, + "grad_norm": 0.7739056976249491, + "learning_rate": 1.7113510583729403e-05, + "loss": 0.9766, + "step": 516 + }, + { + "epoch": 1.6170712607674238, + "grad_norm": 0.8069370636653986, + "learning_rate": 1.7098106600095204e-05, + "loss": 0.9496, + "step": 517 + }, + { + "epoch": 1.620203602192639, + "grad_norm": 0.8037770500974222, + "learning_rate": 1.7082668596149402e-05, + "loss": 0.9646, + "step": 518 + }, + { + "epoch": 1.6233359436178545, + "grad_norm": 0.7878273006407442, + "learning_rate": 1.706719664588437e-05, + "loss": 0.9569, + "step": 519 + }, + { + "epoch": 1.6264682850430696, + "grad_norm": 0.7827154823147211, + "learning_rate": 1.7051690823455162e-05, + "loss": 0.9405, + "step": 520 + }, + { + "epoch": 1.6296006264682852, + "grad_norm": 0.8114142283057644, + "learning_rate": 1.70361512031792e-05, + "loss": 0.9804, + "step": 521 + }, + { + "epoch": 1.6327329678935003, + "grad_norm": 0.7889387459740889, + "learning_rate": 1.702057785953588e-05, + "loss": 0.9405, + "step": 522 + }, + { + "epoch": 1.6358653093187159, + "grad_norm": 0.8311030879338528, + "learning_rate": 1.7004970867166238e-05, + "loss": 0.9593, + "step": 523 + }, + { + "epoch": 1.638997650743931, + "grad_norm": 0.7984284134204775, + "learning_rate": 1.6989330300872576e-05, + "loss": 0.951, + "step": 524 + }, + { + "epoch": 1.6421299921691466, + "grad_norm": 0.7997161389779351, + "learning_rate": 1.6973656235618114e-05, + "loss": 0.9379, + "step": 525 + }, + { + "epoch": 1.6452623335943617, + "grad_norm": 0.7928450659072758, + "learning_rate": 1.6957948746526637e-05, + "loss": 0.9796, + "step": 526 + }, + { + "epoch": 1.6483946750195773, + "grad_norm": 0.7864284356410199, + "learning_rate": 1.6942207908882115e-05, + "loss": 0.9595, + "step": 527 + }, + { + "epoch": 1.6515270164447924, + "grad_norm": 0.7748245207452046, + "learning_rate": 1.692643379812836e-05, + "loss": 0.9363, + "step": 528 + }, + { + "epoch": 1.654659357870008, + "grad_norm": 0.7904035921159812, + "learning_rate": 1.691062648986865e-05, + "loss": 0.9698, + "step": 529 + }, + { + "epoch": 1.657791699295223, + "grad_norm": 0.7699380547476845, + "learning_rate": 1.6894786059865383e-05, + "loss": 0.9229, + "step": 530 + }, + { + "epoch": 1.6609240407204386, + "grad_norm": 0.756885007762418, + "learning_rate": 1.6878912584039698e-05, + "loss": 0.9664, + "step": 531 + }, + { + "epoch": 1.6640563821456538, + "grad_norm": 0.8314070737417165, + "learning_rate": 1.686300613847113e-05, + "loss": 0.9685, + "step": 532 + }, + { + "epoch": 1.6671887235708693, + "grad_norm": 0.8339161427130307, + "learning_rate": 1.6847066799397224e-05, + "loss": 0.9272, + "step": 533 + }, + { + "epoch": 1.6703210649960845, + "grad_norm": 0.8349954422927233, + "learning_rate": 1.683109464321317e-05, + "loss": 0.9764, + "step": 534 + }, + { + "epoch": 1.6734534064213, + "grad_norm": 0.7964469038876641, + "learning_rate": 1.6815089746471472e-05, + "loss": 1.0066, + "step": 535 + }, + { + "epoch": 1.6765857478465152, + "grad_norm": 0.7854281643639673, + "learning_rate": 1.6799052185881535e-05, + "loss": 0.9901, + "step": 536 + }, + { + "epoch": 1.6797180892717307, + "grad_norm": 0.7974377816831979, + "learning_rate": 1.678298203830932e-05, + "loss": 0.969, + "step": 537 + }, + { + "epoch": 1.6828504306969458, + "grad_norm": 0.7734247111191668, + "learning_rate": 1.6766879380776983e-05, + "loss": 0.9183, + "step": 538 + }, + { + "epoch": 1.6859827721221614, + "grad_norm": 0.7724972901824376, + "learning_rate": 1.6750744290462484e-05, + "loss": 0.9186, + "step": 539 + }, + { + "epoch": 1.6891151135473765, + "grad_norm": 0.7968886343195833, + "learning_rate": 1.6734576844699234e-05, + "loss": 0.9409, + "step": 540 + }, + { + "epoch": 1.692247454972592, + "grad_norm": 0.8009596190830679, + "learning_rate": 1.671837712097573e-05, + "loss": 0.9639, + "step": 541 + }, + { + "epoch": 1.6953797963978072, + "grad_norm": 0.8367539665803249, + "learning_rate": 1.670214519693515e-05, + "loss": 0.9655, + "step": 542 + }, + { + "epoch": 1.6985121378230228, + "grad_norm": 0.8071935125435302, + "learning_rate": 1.6685881150375027e-05, + "loss": 0.9635, + "step": 543 + }, + { + "epoch": 1.701644479248238, + "grad_norm": 0.8293405021567984, + "learning_rate": 1.6669585059246837e-05, + "loss": 0.9681, + "step": 544 + }, + { + "epoch": 1.7047768206734535, + "grad_norm": 0.8081858624977237, + "learning_rate": 1.6653257001655652e-05, + "loss": 0.9918, + "step": 545 + }, + { + "epoch": 1.7079091620986686, + "grad_norm": 0.8057738404404446, + "learning_rate": 1.6636897055859754e-05, + "loss": 0.9727, + "step": 546 + }, + { + "epoch": 1.7110415035238842, + "grad_norm": 0.7493081094786209, + "learning_rate": 1.6620505300270254e-05, + "loss": 0.923, + "step": 547 + }, + { + "epoch": 1.7141738449490993, + "grad_norm": 0.7989097981932453, + "learning_rate": 1.6604081813450724e-05, + "loss": 0.9529, + "step": 548 + }, + { + "epoch": 1.7173061863743149, + "grad_norm": 0.7713974329610848, + "learning_rate": 1.6587626674116827e-05, + "loss": 0.962, + "step": 549 + }, + { + "epoch": 1.72043852779953, + "grad_norm": 0.8015438391238155, + "learning_rate": 1.657113996113593e-05, + "loss": 0.9886, + "step": 550 + }, + { + "epoch": 1.7235708692247456, + "grad_norm": 0.8065266518656921, + "learning_rate": 1.6554621753526724e-05, + "loss": 0.9868, + "step": 551 + }, + { + "epoch": 1.7267032106499607, + "grad_norm": 0.8310150242959837, + "learning_rate": 1.6538072130458853e-05, + "loss": 0.9989, + "step": 552 + }, + { + "epoch": 1.7298355520751763, + "grad_norm": 0.7832374226495794, + "learning_rate": 1.652149117125253e-05, + "loss": 0.9472, + "step": 553 + }, + { + "epoch": 1.7329678935003914, + "grad_norm": 0.7885410215851165, + "learning_rate": 1.6504878955378166e-05, + "loss": 0.9987, + "step": 554 + }, + { + "epoch": 1.736100234925607, + "grad_norm": 0.8206660199757921, + "learning_rate": 1.6488235562455964e-05, + "loss": 0.9761, + "step": 555 + }, + { + "epoch": 1.7392325763508223, + "grad_norm": 0.7997233241816074, + "learning_rate": 1.6471561072255576e-05, + "loss": 0.9768, + "step": 556 + }, + { + "epoch": 1.7423649177760376, + "grad_norm": 0.7936840487628978, + "learning_rate": 1.6454855564695674e-05, + "loss": 0.9434, + "step": 557 + }, + { + "epoch": 1.745497259201253, + "grad_norm": 0.7829379543649524, + "learning_rate": 1.643811911984362e-05, + "loss": 0.951, + "step": 558 + }, + { + "epoch": 1.7486296006264683, + "grad_norm": 0.7854616494314223, + "learning_rate": 1.6421351817915025e-05, + "loss": 0.9417, + "step": 559 + }, + { + "epoch": 1.7517619420516837, + "grad_norm": 0.7885222818513928, + "learning_rate": 1.6404553739273426e-05, + "loss": 0.9678, + "step": 560 + }, + { + "epoch": 1.754894283476899, + "grad_norm": 0.7691482528488672, + "learning_rate": 1.6387724964429845e-05, + "loss": 0.9235, + "step": 561 + }, + { + "epoch": 1.7580266249021144, + "grad_norm": 0.7730839050185739, + "learning_rate": 1.6370865574042435e-05, + "loss": 0.9821, + "step": 562 + }, + { + "epoch": 1.7611589663273297, + "grad_norm": 0.7879677074806722, + "learning_rate": 1.6353975648916086e-05, + "loss": 0.9699, + "step": 563 + }, + { + "epoch": 1.764291307752545, + "grad_norm": 0.7858596514917898, + "learning_rate": 1.633705527000203e-05, + "loss": 0.994, + "step": 564 + }, + { + "epoch": 1.7674236491777604, + "grad_norm": 0.8012498378641922, + "learning_rate": 1.6320104518397473e-05, + "loss": 0.967, + "step": 565 + }, + { + "epoch": 1.7705559906029757, + "grad_norm": 0.8188666888354701, + "learning_rate": 1.6303123475345182e-05, + "loss": 0.9785, + "step": 566 + }, + { + "epoch": 1.773688332028191, + "grad_norm": 0.7981343235293162, + "learning_rate": 1.6286112222233113e-05, + "loss": 0.975, + "step": 567 + }, + { + "epoch": 1.7768206734534064, + "grad_norm": 0.7838444823903047, + "learning_rate": 1.6269070840594017e-05, + "loss": 0.9458, + "step": 568 + }, + { + "epoch": 1.7799530148786218, + "grad_norm": 0.7846905745122678, + "learning_rate": 1.6251999412105036e-05, + "loss": 0.9298, + "step": 569 + }, + { + "epoch": 1.7830853563038371, + "grad_norm": 0.8149058395623431, + "learning_rate": 1.6234898018587336e-05, + "loss": 0.9436, + "step": 570 + }, + { + "epoch": 1.7862176977290525, + "grad_norm": 0.7941672166648026, + "learning_rate": 1.6217766742005693e-05, + "loss": 0.9985, + "step": 571 + }, + { + "epoch": 1.7893500391542678, + "grad_norm": 0.7710930788046828, + "learning_rate": 1.6200605664468113e-05, + "loss": 0.97, + "step": 572 + }, + { + "epoch": 1.7924823805794832, + "grad_norm": 0.7957133349582347, + "learning_rate": 1.6183414868225434e-05, + "loss": 0.9959, + "step": 573 + }, + { + "epoch": 1.7956147220046985, + "grad_norm": 0.79291459979055, + "learning_rate": 1.616619443567092e-05, + "loss": 0.9442, + "step": 574 + }, + { + "epoch": 1.7987470634299139, + "grad_norm": 0.785169568406983, + "learning_rate": 1.6148944449339904e-05, + "loss": 0.9639, + "step": 575 + }, + { + "epoch": 1.8018794048551292, + "grad_norm": 0.7997066561831516, + "learning_rate": 1.6131664991909335e-05, + "loss": 1.0127, + "step": 576 + }, + { + "epoch": 1.8050117462803446, + "grad_norm": 0.8141399391865273, + "learning_rate": 1.6114356146197445e-05, + "loss": 0.9386, + "step": 577 + }, + { + "epoch": 1.80814408770556, + "grad_norm": 0.8097117540881008, + "learning_rate": 1.6097017995163294e-05, + "loss": 0.9774, + "step": 578 + }, + { + "epoch": 1.8112764291307752, + "grad_norm": 0.7730781490152179, + "learning_rate": 1.607965062190641e-05, + "loss": 0.9453, + "step": 579 + }, + { + "epoch": 1.8144087705559906, + "grad_norm": 0.7838923521873067, + "learning_rate": 1.6062254109666383e-05, + "loss": 0.9628, + "step": 580 + }, + { + "epoch": 1.817541111981206, + "grad_norm": 0.790301862199868, + "learning_rate": 1.604482854182245e-05, + "loss": 0.9068, + "step": 581 + }, + { + "epoch": 1.8206734534064213, + "grad_norm": 0.7750930497911785, + "learning_rate": 1.6027374001893123e-05, + "loss": 0.9927, + "step": 582 + }, + { + "epoch": 1.8238057948316366, + "grad_norm": 0.7779386940305499, + "learning_rate": 1.6009890573535763e-05, + "loss": 0.919, + "step": 583 + }, + { + "epoch": 1.826938136256852, + "grad_norm": 0.7992055706023535, + "learning_rate": 1.5992378340546194e-05, + "loss": 0.9624, + "step": 584 + }, + { + "epoch": 1.8300704776820673, + "grad_norm": 0.8153176494752826, + "learning_rate": 1.597483738685829e-05, + "loss": 1.0051, + "step": 585 + }, + { + "epoch": 1.8332028191072827, + "grad_norm": 0.7936500253751112, + "learning_rate": 1.595726779654359e-05, + "loss": 0.9692, + "step": 586 + }, + { + "epoch": 1.836335160532498, + "grad_norm": 0.8140127238131111, + "learning_rate": 1.5939669653810882e-05, + "loss": 0.9763, + "step": 587 + }, + { + "epoch": 1.8394675019577134, + "grad_norm": 0.8084565243659966, + "learning_rate": 1.592204304300579e-05, + "loss": 0.9402, + "step": 588 + }, + { + "epoch": 1.8425998433829287, + "grad_norm": 0.7527144155069814, + "learning_rate": 1.59043880486104e-05, + "loss": 0.9255, + "step": 589 + }, + { + "epoch": 1.845732184808144, + "grad_norm": 0.7985587310153884, + "learning_rate": 1.588670475524283e-05, + "loss": 0.9465, + "step": 590 + }, + { + "epoch": 1.8488645262333594, + "grad_norm": 0.7524132760020612, + "learning_rate": 1.5868993247656822e-05, + "loss": 0.9865, + "step": 591 + }, + { + "epoch": 1.8519968676585747, + "grad_norm": 0.7823183770981433, + "learning_rate": 1.5851253610741357e-05, + "loss": 0.8974, + "step": 592 + }, + { + "epoch": 1.85512920908379, + "grad_norm": 0.769457387825749, + "learning_rate": 1.583348592952023e-05, + "loss": 0.9776, + "step": 593 + }, + { + "epoch": 1.8582615505090054, + "grad_norm": 0.7714656018935003, + "learning_rate": 1.581569028915166e-05, + "loss": 0.9399, + "step": 594 + }, + { + "epoch": 1.8613938919342208, + "grad_norm": 0.7784139001699675, + "learning_rate": 1.5797866774927848e-05, + "loss": 0.9699, + "step": 595 + }, + { + "epoch": 1.8645262333594361, + "grad_norm": 0.7797773255228968, + "learning_rate": 1.5780015472274613e-05, + "loss": 0.9404, + "step": 596 + }, + { + "epoch": 1.8676585747846515, + "grad_norm": 0.7758341272190327, + "learning_rate": 1.5762136466750947e-05, + "loss": 0.9719, + "step": 597 + }, + { + "epoch": 1.870790916209867, + "grad_norm": 0.7786162669677469, + "learning_rate": 1.5744229844048627e-05, + "loss": 0.9436, + "step": 598 + }, + { + "epoch": 1.8739232576350822, + "grad_norm": 0.7543252031375369, + "learning_rate": 1.5726295689991787e-05, + "loss": 0.9784, + "step": 599 + }, + { + "epoch": 1.8770555990602977, + "grad_norm": 0.7757844187122288, + "learning_rate": 1.570833409053653e-05, + "loss": 0.9713, + "step": 600 + }, + { + "epoch": 1.8801879404855129, + "grad_norm": 0.7653394710213172, + "learning_rate": 1.5690345131770474e-05, + "loss": 0.8965, + "step": 601 + }, + { + "epoch": 1.8833202819107284, + "grad_norm": 0.7896033443700016, + "learning_rate": 1.5672328899912397e-05, + "loss": 1.0193, + "step": 602 + }, + { + "epoch": 1.8864526233359435, + "grad_norm": 0.7846649604330512, + "learning_rate": 1.5654285481311777e-05, + "loss": 0.9474, + "step": 603 + }, + { + "epoch": 1.8895849647611591, + "grad_norm": 0.8135389235047997, + "learning_rate": 1.5636214962448396e-05, + "loss": 0.9618, + "step": 604 + }, + { + "epoch": 1.8927173061863742, + "grad_norm": 0.7824047803609824, + "learning_rate": 1.5618117429931923e-05, + "loss": 0.9408, + "step": 605 + }, + { + "epoch": 1.8958496476115898, + "grad_norm": 0.8254165694685803, + "learning_rate": 1.5599992970501512e-05, + "loss": 0.9767, + "step": 606 + }, + { + "epoch": 1.898981989036805, + "grad_norm": 0.7942279881179067, + "learning_rate": 1.5581841671025358e-05, + "loss": 0.9657, + "step": 607 + }, + { + "epoch": 1.9021143304620205, + "grad_norm": 0.7607717329156005, + "learning_rate": 1.55636636185003e-05, + "loss": 0.9641, + "step": 608 + }, + { + "epoch": 1.9052466718872356, + "grad_norm": 0.7992715938950239, + "learning_rate": 1.5545458900051417e-05, + "loss": 0.9466, + "step": 609 + }, + { + "epoch": 1.9083790133124512, + "grad_norm": 0.7710025331360174, + "learning_rate": 1.552722760293157e-05, + "loss": 0.9349, + "step": 610 + }, + { + "epoch": 1.9115113547376663, + "grad_norm": 0.7995999712279581, + "learning_rate": 1.5508969814521026e-05, + "loss": 0.9784, + "step": 611 + }, + { + "epoch": 1.9146436961628819, + "grad_norm": 0.7972698916237193, + "learning_rate": 1.5490685622327015e-05, + "loss": 0.9954, + "step": 612 + }, + { + "epoch": 1.917776037588097, + "grad_norm": 0.7658223645888151, + "learning_rate": 1.547237511398332e-05, + "loss": 0.931, + "step": 613 + }, + { + "epoch": 1.9209083790133126, + "grad_norm": 0.7891075170093855, + "learning_rate": 1.5454038377249845e-05, + "loss": 0.9559, + "step": 614 + }, + { + "epoch": 1.9240407204385277, + "grad_norm": 0.8134205306266102, + "learning_rate": 1.5435675500012212e-05, + "loss": 0.9907, + "step": 615 + }, + { + "epoch": 1.9271730618637433, + "grad_norm": 0.803815630288296, + "learning_rate": 1.5417286570281334e-05, + "loss": 0.9834, + "step": 616 + }, + { + "epoch": 1.9303054032889584, + "grad_norm": 0.771918493512692, + "learning_rate": 1.5398871676192978e-05, + "loss": 1.0094, + "step": 617 + }, + { + "epoch": 1.933437744714174, + "grad_norm": 0.793008120980106, + "learning_rate": 1.538043090600737e-05, + "loss": 0.9266, + "step": 618 + }, + { + "epoch": 1.936570086139389, + "grad_norm": 0.7778816113808421, + "learning_rate": 1.5361964348108736e-05, + "loss": 0.9665, + "step": 619 + }, + { + "epoch": 1.9397024275646046, + "grad_norm": 0.7989827850535249, + "learning_rate": 1.5343472091004925e-05, + "loss": 0.9524, + "step": 620 + }, + { + "epoch": 1.9428347689898198, + "grad_norm": 0.813278631018117, + "learning_rate": 1.532495422332694e-05, + "loss": 1.0166, + "step": 621 + }, + { + "epoch": 1.9459671104150353, + "grad_norm": 0.7855425028524953, + "learning_rate": 1.5306410833828534e-05, + "loss": 0.9816, + "step": 622 + }, + { + "epoch": 1.9490994518402505, + "grad_norm": 0.7943986605440319, + "learning_rate": 1.5287842011385798e-05, + "loss": 0.9852, + "step": 623 + }, + { + "epoch": 1.952231793265466, + "grad_norm": 0.8316767537841315, + "learning_rate": 1.52692478449967e-05, + "loss": 0.9811, + "step": 624 + }, + { + "epoch": 1.9553641346906812, + "grad_norm": 0.8079701202216203, + "learning_rate": 1.5250628423780684e-05, + "loss": 0.989, + "step": 625 + }, + { + "epoch": 1.9584964761158967, + "grad_norm": 0.8001611635685686, + "learning_rate": 1.523198383697825e-05, + "loss": 0.9859, + "step": 626 + }, + { + "epoch": 1.9616288175411118, + "grad_norm": 0.7922734841320317, + "learning_rate": 1.5213314173950498e-05, + "loss": 0.9785, + "step": 627 + }, + { + "epoch": 1.9647611589663274, + "grad_norm": 0.7833384810882262, + "learning_rate": 1.5194619524178718e-05, + "loss": 0.9407, + "step": 628 + }, + { + "epoch": 1.9678935003915425, + "grad_norm": 0.8098242578519478, + "learning_rate": 1.5175899977263963e-05, + "loss": 0.9224, + "step": 629 + }, + { + "epoch": 1.971025841816758, + "grad_norm": 0.8182996603065131, + "learning_rate": 1.515715562292662e-05, + "loss": 0.9622, + "step": 630 + }, + { + "epoch": 1.9741581832419732, + "grad_norm": 0.7901742990362269, + "learning_rate": 1.513838655100596e-05, + "loss": 0.9337, + "step": 631 + }, + { + "epoch": 1.9772905246671888, + "grad_norm": 0.7615944653532644, + "learning_rate": 1.5119592851459726e-05, + "loss": 0.9414, + "step": 632 + }, + { + "epoch": 1.980422866092404, + "grad_norm": 0.7946500220579569, + "learning_rate": 1.5100774614363708e-05, + "loss": 0.9512, + "step": 633 + }, + { + "epoch": 1.9835552075176195, + "grad_norm": 0.8076792545370913, + "learning_rate": 1.5081931929911298e-05, + "loss": 0.973, + "step": 634 + }, + { + "epoch": 1.9866875489428346, + "grad_norm": 0.7861723773682193, + "learning_rate": 1.5063064888413048e-05, + "loss": 0.9842, + "step": 635 + }, + { + "epoch": 1.9898198903680502, + "grad_norm": 0.7955732049686428, + "learning_rate": 1.5044173580296267e-05, + "loss": 0.9272, + "step": 636 + }, + { + "epoch": 1.9929522317932653, + "grad_norm": 0.8037074259104876, + "learning_rate": 1.5025258096104563e-05, + "loss": 0.9697, + "step": 637 + }, + { + "epoch": 1.9960845732184809, + "grad_norm": 0.8103505085995998, + "learning_rate": 1.500631852649742e-05, + "loss": 0.948, + "step": 638 + }, + { + "epoch": 1.999216914643696, + "grad_norm": 0.7802553935331468, + "learning_rate": 1.4987354962249749e-05, + "loss": 0.9517, + "step": 639 + }, + { + "epoch": 2.0, + "grad_norm": 0.7802553935331468, + "learning_rate": 1.4968367494251486e-05, + "loss": 0.9289, + "step": 640 + }, + { + "epoch": 2.0093970242756463, + "grad_norm": 1.3210123581817999, + "learning_rate": 1.4949356213507113e-05, + "loss": 0.7754, + "step": 641 + }, + { + "epoch": 2.0125293657008614, + "grad_norm": 1.3238346751721688, + "learning_rate": 1.4930321211135258e-05, + "loss": 0.7855, + "step": 642 + }, + { + "epoch": 2.015661707126077, + "grad_norm": 1.1616452207714454, + "learning_rate": 1.4911262578368233e-05, + "loss": 0.7959, + "step": 643 + }, + { + "epoch": 2.018794048551292, + "grad_norm": 0.9412974130528684, + "learning_rate": 1.4892180406551618e-05, + "loss": 0.7173, + "step": 644 + }, + { + "epoch": 2.0219263899765076, + "grad_norm": 1.3175139050237945, + "learning_rate": 1.48730747871438e-05, + "loss": 0.7316, + "step": 645 + }, + { + "epoch": 2.0250587314017228, + "grad_norm": 1.5954314046677502, + "learning_rate": 1.4853945811715563e-05, + "loss": 0.7542, + "step": 646 + }, + { + "epoch": 2.0281910728269383, + "grad_norm": 1.37008550444395, + "learning_rate": 1.4834793571949617e-05, + "loss": 0.7587, + "step": 647 + }, + { + "epoch": 2.0313234142521535, + "grad_norm": 1.169303068811387, + "learning_rate": 1.4815618159640188e-05, + "loss": 0.7195, + "step": 648 + }, + { + "epoch": 2.034455755677369, + "grad_norm": 1.120757117831837, + "learning_rate": 1.4796419666692555e-05, + "loss": 0.6884, + "step": 649 + }, + { + "epoch": 2.037588097102584, + "grad_norm": 1.1374034296234556, + "learning_rate": 1.477719818512263e-05, + "loss": 0.7608, + "step": 650 + }, + { + "epoch": 2.0407204385277997, + "grad_norm": 1.018857278719166, + "learning_rate": 1.47579538070565e-05, + "loss": 0.7136, + "step": 651 + }, + { + "epoch": 2.043852779953015, + "grad_norm": 0.991608010520256, + "learning_rate": 1.4738686624729987e-05, + "loss": 0.7105, + "step": 652 + }, + { + "epoch": 2.0469851213782304, + "grad_norm": 1.0388281087329523, + "learning_rate": 1.4719396730488225e-05, + "loss": 0.7348, + "step": 653 + }, + { + "epoch": 2.0501174628034455, + "grad_norm": 1.0171773999697262, + "learning_rate": 1.4700084216785189e-05, + "loss": 0.7212, + "step": 654 + }, + { + "epoch": 2.053249804228661, + "grad_norm": 0.9712484652548604, + "learning_rate": 1.4680749176183275e-05, + "loss": 0.7152, + "step": 655 + }, + { + "epoch": 2.0563821456538762, + "grad_norm": 0.9445446614613634, + "learning_rate": 1.4661391701352844e-05, + "loss": 0.7144, + "step": 656 + }, + { + "epoch": 2.059514487079092, + "grad_norm": 0.9945223666447938, + "learning_rate": 1.464201188507178e-05, + "loss": 0.7119, + "step": 657 + }, + { + "epoch": 2.062646828504307, + "grad_norm": 0.9677117020420452, + "learning_rate": 1.4622609820225052e-05, + "loss": 0.7148, + "step": 658 + }, + { + "epoch": 2.0657791699295225, + "grad_norm": 1.0137018942940528, + "learning_rate": 1.4603185599804259e-05, + "loss": 0.7435, + "step": 659 + }, + { + "epoch": 2.0689115113547376, + "grad_norm": 1.02511003538228, + "learning_rate": 1.4583739316907188e-05, + "loss": 0.6999, + "step": 660 + }, + { + "epoch": 2.072043852779953, + "grad_norm": 0.9904700091988614, + "learning_rate": 1.4564271064737376e-05, + "loss": 0.7449, + "step": 661 + }, + { + "epoch": 2.0751761942051683, + "grad_norm": 0.9279680631763316, + "learning_rate": 1.4544780936603644e-05, + "loss": 0.7023, + "step": 662 + }, + { + "epoch": 2.078308535630384, + "grad_norm": 0.9287114111608203, + "learning_rate": 1.4525269025919677e-05, + "loss": 0.7163, + "step": 663 + }, + { + "epoch": 2.081440877055599, + "grad_norm": 0.943750195528307, + "learning_rate": 1.4505735426203545e-05, + "loss": 0.7377, + "step": 664 + }, + { + "epoch": 2.0845732184808146, + "grad_norm": 0.9053784915303666, + "learning_rate": 1.4486180231077278e-05, + "loss": 0.7282, + "step": 665 + }, + { + "epoch": 2.0877055599060297, + "grad_norm": 0.955797509257669, + "learning_rate": 1.446660353426641e-05, + "loss": 0.7668, + "step": 666 + }, + { + "epoch": 2.0908379013312453, + "grad_norm": 0.9130054155780589, + "learning_rate": 1.4447005429599537e-05, + "loss": 0.695, + "step": 667 + }, + { + "epoch": 2.0939702427564604, + "grad_norm": 0.9516666064010463, + "learning_rate": 1.4427386011007847e-05, + "loss": 0.7255, + "step": 668 + }, + { + "epoch": 2.097102584181676, + "grad_norm": 0.9546422449504612, + "learning_rate": 1.4407745372524686e-05, + "loss": 0.7224, + "step": 669 + }, + { + "epoch": 2.100234925606891, + "grad_norm": 0.9448475040935651, + "learning_rate": 1.4388083608285112e-05, + "loss": 0.7222, + "step": 670 + }, + { + "epoch": 2.1033672670321066, + "grad_norm": 0.9222655058113182, + "learning_rate": 1.4368400812525434e-05, + "loss": 0.7251, + "step": 671 + }, + { + "epoch": 2.1064996084573218, + "grad_norm": 0.9777941043274174, + "learning_rate": 1.4348697079582752e-05, + "loss": 0.7703, + "step": 672 + }, + { + "epoch": 2.1096319498825373, + "grad_norm": 0.9749998119435571, + "learning_rate": 1.4328972503894525e-05, + "loss": 0.7574, + "step": 673 + }, + { + "epoch": 2.1127642913077525, + "grad_norm": 0.9756161103337109, + "learning_rate": 1.4309227179998109e-05, + "loss": 0.7646, + "step": 674 + }, + { + "epoch": 2.115896632732968, + "grad_norm": 0.9561100629580807, + "learning_rate": 1.4289461202530297e-05, + "loss": 0.7504, + "step": 675 + }, + { + "epoch": 2.119028974158183, + "grad_norm": 0.9458044599682323, + "learning_rate": 1.4269674666226879e-05, + "loss": 0.7406, + "step": 676 + }, + { + "epoch": 2.1221613155833987, + "grad_norm": 0.9193297731491735, + "learning_rate": 1.4249867665922175e-05, + "loss": 0.6953, + "step": 677 + }, + { + "epoch": 2.125293657008614, + "grad_norm": 0.938231383030172, + "learning_rate": 1.4230040296548588e-05, + "loss": 0.7742, + "step": 678 + }, + { + "epoch": 2.1284259984338294, + "grad_norm": 0.9608218431580744, + "learning_rate": 1.4210192653136151e-05, + "loss": 0.745, + "step": 679 + }, + { + "epoch": 2.1315583398590445, + "grad_norm": 0.9131562406248528, + "learning_rate": 1.4190324830812067e-05, + "loss": 0.706, + "step": 680 + }, + { + "epoch": 2.13469068128426, + "grad_norm": 0.9330499851170364, + "learning_rate": 1.4170436924800251e-05, + "loss": 0.7198, + "step": 681 + }, + { + "epoch": 2.137823022709475, + "grad_norm": 0.9744409362566172, + "learning_rate": 1.415052903042087e-05, + "loss": 0.748, + "step": 682 + }, + { + "epoch": 2.140955364134691, + "grad_norm": 0.956855447487281, + "learning_rate": 1.4130601243089905e-05, + "loss": 0.772, + "step": 683 + }, + { + "epoch": 2.144087705559906, + "grad_norm": 0.9105854308119294, + "learning_rate": 1.4110653658318682e-05, + "loss": 0.7066, + "step": 684 + }, + { + "epoch": 2.1472200469851215, + "grad_norm": 0.9437505928196139, + "learning_rate": 1.4090686371713403e-05, + "loss": 0.7511, + "step": 685 + }, + { + "epoch": 2.1503523884103366, + "grad_norm": 0.9754488006918184, + "learning_rate": 1.4070699478974698e-05, + "loss": 0.7273, + "step": 686 + }, + { + "epoch": 2.153484729835552, + "grad_norm": 0.9660297851499952, + "learning_rate": 1.4050693075897181e-05, + "loss": 0.7423, + "step": 687 + }, + { + "epoch": 2.1566170712607673, + "grad_norm": 0.9710350268049464, + "learning_rate": 1.4030667258368968e-05, + "loss": 0.7325, + "step": 688 + }, + { + "epoch": 2.159749412685983, + "grad_norm": 0.9168340841543151, + "learning_rate": 1.4010622122371219e-05, + "loss": 0.7178, + "step": 689 + }, + { + "epoch": 2.162881754111198, + "grad_norm": 0.967756101002905, + "learning_rate": 1.3990557763977694e-05, + "loss": 0.7614, + "step": 690 + }, + { + "epoch": 2.1660140955364136, + "grad_norm": 0.9340155531023159, + "learning_rate": 1.3970474279354283e-05, + "loss": 0.7106, + "step": 691 + }, + { + "epoch": 2.1691464369616287, + "grad_norm": 0.9505636046987749, + "learning_rate": 1.3950371764758543e-05, + "loss": 0.7243, + "step": 692 + }, + { + "epoch": 2.1722787783868442, + "grad_norm": 0.9197842895474289, + "learning_rate": 1.3930250316539237e-05, + "loss": 0.7501, + "step": 693 + }, + { + "epoch": 2.1754111198120594, + "grad_norm": 0.9424997582167158, + "learning_rate": 1.3910110031135884e-05, + "loss": 0.7472, + "step": 694 + }, + { + "epoch": 2.178543461237275, + "grad_norm": 0.9606400745011864, + "learning_rate": 1.3889951005078271e-05, + "loss": 0.7475, + "step": 695 + }, + { + "epoch": 2.18167580266249, + "grad_norm": 0.9659045083450463, + "learning_rate": 1.3869773334986023e-05, + "loss": 0.7488, + "step": 696 + }, + { + "epoch": 2.1848081440877056, + "grad_norm": 0.984473222173499, + "learning_rate": 1.3849577117568115e-05, + "loss": 0.7502, + "step": 697 + }, + { + "epoch": 2.1879404855129208, + "grad_norm": 0.9738775239386929, + "learning_rate": 1.3829362449622416e-05, + "loss": 0.7482, + "step": 698 + }, + { + "epoch": 2.1910728269381363, + "grad_norm": 0.9315292619814982, + "learning_rate": 1.3809129428035229e-05, + "loss": 0.7511, + "step": 699 + }, + { + "epoch": 2.1942051683633514, + "grad_norm": 0.9347672091944302, + "learning_rate": 1.3788878149780827e-05, + "loss": 0.7436, + "step": 700 + }, + { + "epoch": 2.197337509788567, + "grad_norm": 0.9359481225994939, + "learning_rate": 1.376860871192098e-05, + "loss": 0.7509, + "step": 701 + }, + { + "epoch": 2.200469851213782, + "grad_norm": 0.9219014695571008, + "learning_rate": 1.3748321211604488e-05, + "loss": 0.7389, + "step": 702 + }, + { + "epoch": 2.2036021926389977, + "grad_norm": 0.9880848258280983, + "learning_rate": 1.3728015746066736e-05, + "loss": 0.7475, + "step": 703 + }, + { + "epoch": 2.206734534064213, + "grad_norm": 0.9408764680190602, + "learning_rate": 1.370769241262921e-05, + "loss": 0.7402, + "step": 704 + }, + { + "epoch": 2.2098668754894284, + "grad_norm": 0.9301989809594925, + "learning_rate": 1.3687351308699026e-05, + "loss": 0.716, + "step": 705 + }, + { + "epoch": 2.2129992169146435, + "grad_norm": 0.9179501547426313, + "learning_rate": 1.3666992531768482e-05, + "loss": 0.7059, + "step": 706 + }, + { + "epoch": 2.216131558339859, + "grad_norm": 0.9666221682985012, + "learning_rate": 1.3646616179414573e-05, + "loss": 0.7599, + "step": 707 + }, + { + "epoch": 2.219263899765074, + "grad_norm": 0.9675432440063244, + "learning_rate": 1.3626222349298539e-05, + "loss": 0.7518, + "step": 708 + }, + { + "epoch": 2.22239624119029, + "grad_norm": 0.9390656597784647, + "learning_rate": 1.360581113916538e-05, + "loss": 0.7252, + "step": 709 + }, + { + "epoch": 2.225528582615505, + "grad_norm": 0.9581962913041326, + "learning_rate": 1.3585382646843396e-05, + "loss": 0.7605, + "step": 710 + }, + { + "epoch": 2.2286609240407205, + "grad_norm": 0.9610602302318673, + "learning_rate": 1.356493697024373e-05, + "loss": 0.7811, + "step": 711 + }, + { + "epoch": 2.2317932654659356, + "grad_norm": 0.9867556442165821, + "learning_rate": 1.3544474207359875e-05, + "loss": 0.7349, + "step": 712 + }, + { + "epoch": 2.234925606891151, + "grad_norm": 0.9425764813872476, + "learning_rate": 1.352399445626722e-05, + "loss": 0.7226, + "step": 713 + }, + { + "epoch": 2.2380579483163663, + "grad_norm": 1.0102139516905289, + "learning_rate": 1.3503497815122577e-05, + "loss": 0.7671, + "step": 714 + }, + { + "epoch": 2.241190289741582, + "grad_norm": 0.965501835494664, + "learning_rate": 1.3482984382163713e-05, + "loss": 0.7449, + "step": 715 + }, + { + "epoch": 2.244322631166797, + "grad_norm": 0.9131072345452055, + "learning_rate": 1.3462454255708868e-05, + "loss": 0.714, + "step": 716 + }, + { + "epoch": 2.2474549725920125, + "grad_norm": 0.9498450701506534, + "learning_rate": 1.3441907534156298e-05, + "loss": 0.7484, + "step": 717 + }, + { + "epoch": 2.2505873140172277, + "grad_norm": 0.9689310041592919, + "learning_rate": 1.34213443159838e-05, + "loss": 0.7772, + "step": 718 + }, + { + "epoch": 2.2537196554424432, + "grad_norm": 0.9414801485286473, + "learning_rate": 1.3400764699748223e-05, + "loss": 0.7307, + "step": 719 + }, + { + "epoch": 2.256851996867659, + "grad_norm": 0.918278021033119, + "learning_rate": 1.3380168784085028e-05, + "loss": 0.7179, + "step": 720 + }, + { + "epoch": 2.259984338292874, + "grad_norm": 0.9166637511796559, + "learning_rate": 1.3359556667707786e-05, + "loss": 0.7142, + "step": 721 + }, + { + "epoch": 2.263116679718089, + "grad_norm": 0.9495008728633978, + "learning_rate": 1.3338928449407721e-05, + "loss": 0.7138, + "step": 722 + }, + { + "epoch": 2.2662490211433046, + "grad_norm": 0.9787409627000769, + "learning_rate": 1.3318284228053225e-05, + "loss": 0.7126, + "step": 723 + }, + { + "epoch": 2.26938136256852, + "grad_norm": 0.9627840952028726, + "learning_rate": 1.3297624102589395e-05, + "loss": 0.7219, + "step": 724 + }, + { + "epoch": 2.2725137039937353, + "grad_norm": 0.9476059282284458, + "learning_rate": 1.3276948172037556e-05, + "loss": 0.7392, + "step": 725 + }, + { + "epoch": 2.2756460454189504, + "grad_norm": 0.9906364986347657, + "learning_rate": 1.3256256535494783e-05, + "loss": 0.7349, + "step": 726 + }, + { + "epoch": 2.278778386844166, + "grad_norm": 0.9675118908082809, + "learning_rate": 1.3235549292133425e-05, + "loss": 0.7364, + "step": 727 + }, + { + "epoch": 2.2819107282693816, + "grad_norm": 0.9429167511984878, + "learning_rate": 1.3214826541200638e-05, + "loss": 0.7514, + "step": 728 + }, + { + "epoch": 2.2850430696945967, + "grad_norm": 0.9847232006632042, + "learning_rate": 1.3194088382017897e-05, + "loss": 0.7606, + "step": 729 + }, + { + "epoch": 2.288175411119812, + "grad_norm": 0.9563800977188294, + "learning_rate": 1.3173334913980535e-05, + "loss": 0.7388, + "step": 730 + }, + { + "epoch": 2.2913077525450274, + "grad_norm": 0.9735616571758612, + "learning_rate": 1.3152566236557251e-05, + "loss": 0.7312, + "step": 731 + }, + { + "epoch": 2.294440093970243, + "grad_norm": 0.9639178348678931, + "learning_rate": 1.3131782449289646e-05, + "loss": 0.7385, + "step": 732 + }, + { + "epoch": 2.297572435395458, + "grad_norm": 0.9529934639226041, + "learning_rate": 1.3110983651791731e-05, + "loss": 0.7194, + "step": 733 + }, + { + "epoch": 2.3007047768206736, + "grad_norm": 0.9678440459923399, + "learning_rate": 1.3090169943749475e-05, + "loss": 0.7589, + "step": 734 + }, + { + "epoch": 2.3038371182458888, + "grad_norm": 0.9540843435038134, + "learning_rate": 1.3069341424920301e-05, + "loss": 0.7811, + "step": 735 + }, + { + "epoch": 2.3069694596711043, + "grad_norm": 1.0181162510512147, + "learning_rate": 1.3048498195132613e-05, + "loss": 0.7492, + "step": 736 + }, + { + "epoch": 2.3101018010963195, + "grad_norm": 0.9572210519063549, + "learning_rate": 1.3027640354285333e-05, + "loss": 0.788, + "step": 737 + }, + { + "epoch": 2.313234142521535, + "grad_norm": 0.9538184834756422, + "learning_rate": 1.3006768002347414e-05, + "loss": 0.7343, + "step": 738 + }, + { + "epoch": 2.31636648394675, + "grad_norm": 0.9765399045822022, + "learning_rate": 1.298588123935735e-05, + "loss": 0.7252, + "step": 739 + }, + { + "epoch": 2.3194988253719657, + "grad_norm": 0.9618045560134357, + "learning_rate": 1.2964980165422701e-05, + "loss": 0.7373, + "step": 740 + }, + { + "epoch": 2.322631166797181, + "grad_norm": 0.9565307835074357, + "learning_rate": 1.2944064880719634e-05, + "loss": 0.7335, + "step": 741 + }, + { + "epoch": 2.3257635082223964, + "grad_norm": 0.967282888806741, + "learning_rate": 1.2923135485492413e-05, + "loss": 0.7845, + "step": 742 + }, + { + "epoch": 2.3288958496476115, + "grad_norm": 0.9686861328866166, + "learning_rate": 1.290219208005294e-05, + "loss": 0.7312, + "step": 743 + }, + { + "epoch": 2.332028191072827, + "grad_norm": 0.9726533064758265, + "learning_rate": 1.2881234764780257e-05, + "loss": 0.7484, + "step": 744 + }, + { + "epoch": 2.3351605324980422, + "grad_norm": 0.957502804625888, + "learning_rate": 1.2860263640120086e-05, + "loss": 0.7295, + "step": 745 + }, + { + "epoch": 2.338292873923258, + "grad_norm": 1.0246439398068528, + "learning_rate": 1.2839278806584323e-05, + "loss": 0.7583, + "step": 746 + }, + { + "epoch": 2.341425215348473, + "grad_norm": 0.9881765469071092, + "learning_rate": 1.2818280364750577e-05, + "loss": 0.7605, + "step": 747 + }, + { + "epoch": 2.3445575567736885, + "grad_norm": 0.9226451453922399, + "learning_rate": 1.2797268415261681e-05, + "loss": 0.7449, + "step": 748 + }, + { + "epoch": 2.3476898981989036, + "grad_norm": 0.9389466675244833, + "learning_rate": 1.2776243058825199e-05, + "loss": 0.7508, + "step": 749 + }, + { + "epoch": 2.350822239624119, + "grad_norm": 0.9680649310436178, + "learning_rate": 1.2755204396212965e-05, + "loss": 0.7514, + "step": 750 + }, + { + "epoch": 2.3539545810493343, + "grad_norm": 0.9508971737498871, + "learning_rate": 1.273415252826058e-05, + "loss": 0.7163, + "step": 751 + }, + { + "epoch": 2.35708692247455, + "grad_norm": 1.0037413765304937, + "learning_rate": 1.2713087555866937e-05, + "loss": 0.7523, + "step": 752 + }, + { + "epoch": 2.360219263899765, + "grad_norm": 0.9675958858875301, + "learning_rate": 1.2692009579993736e-05, + "loss": 0.752, + "step": 753 + }, + { + "epoch": 2.3633516053249806, + "grad_norm": 0.9426804383890924, + "learning_rate": 1.2670918701665007e-05, + "loss": 0.6878, + "step": 754 + }, + { + "epoch": 2.3664839467501957, + "grad_norm": 0.9505901023922874, + "learning_rate": 1.264981502196662e-05, + "loss": 0.73, + "step": 755 + }, + { + "epoch": 2.3696162881754113, + "grad_norm": 0.982586546277607, + "learning_rate": 1.2628698642045786e-05, + "loss": 0.7187, + "step": 756 + }, + { + "epoch": 2.3727486296006264, + "grad_norm": 1.0016723707151867, + "learning_rate": 1.2607569663110603e-05, + "loss": 0.7703, + "step": 757 + }, + { + "epoch": 2.375880971025842, + "grad_norm": 0.9650854481892198, + "learning_rate": 1.258642818642955e-05, + "loss": 0.722, + "step": 758 + }, + { + "epoch": 2.379013312451057, + "grad_norm": 0.9838012057082094, + "learning_rate": 1.2565274313331001e-05, + "loss": 0.7776, + "step": 759 + }, + { + "epoch": 2.3821456538762726, + "grad_norm": 0.9546713861914818, + "learning_rate": 1.2544108145202748e-05, + "loss": 0.7277, + "step": 760 + }, + { + "epoch": 2.3852779953014878, + "grad_norm": 0.9520183658938295, + "learning_rate": 1.2522929783491508e-05, + "loss": 0.7334, + "step": 761 + }, + { + "epoch": 2.3884103367267033, + "grad_norm": 0.9540342124185963, + "learning_rate": 1.2501739329702453e-05, + "loss": 0.7291, + "step": 762 + }, + { + "epoch": 2.3915426781519185, + "grad_norm": 0.9549925390587612, + "learning_rate": 1.2480536885398697e-05, + "loss": 0.7329, + "step": 763 + }, + { + "epoch": 2.394675019577134, + "grad_norm": 0.9728937913697199, + "learning_rate": 1.2459322552200824e-05, + "loss": 0.7802, + "step": 764 + }, + { + "epoch": 2.397807361002349, + "grad_norm": 0.9640580751608233, + "learning_rate": 1.2438096431786408e-05, + "loss": 0.7606, + "step": 765 + }, + { + "epoch": 2.4009397024275647, + "grad_norm": 0.9304598478726142, + "learning_rate": 1.2416858625889507e-05, + "loss": 0.7361, + "step": 766 + }, + { + "epoch": 2.40407204385278, + "grad_norm": 0.9806923344541586, + "learning_rate": 1.2395609236300198e-05, + "loss": 0.7155, + "step": 767 + }, + { + "epoch": 2.4072043852779954, + "grad_norm": 0.969342428109374, + "learning_rate": 1.2374348364864068e-05, + "loss": 0.7559, + "step": 768 + }, + { + "epoch": 2.4103367267032105, + "grad_norm": 0.9350655072927523, + "learning_rate": 1.2353076113481742e-05, + "loss": 0.7544, + "step": 769 + }, + { + "epoch": 2.413469068128426, + "grad_norm": 0.9447261809188683, + "learning_rate": 1.2331792584108375e-05, + "loss": 0.7191, + "step": 770 + }, + { + "epoch": 2.4166014095536412, + "grad_norm": 0.9674909744485723, + "learning_rate": 1.2310497878753196e-05, + "loss": 0.7686, + "step": 771 + }, + { + "epoch": 2.419733750978857, + "grad_norm": 0.9854774668922955, + "learning_rate": 1.2289192099478979e-05, + "loss": 0.7595, + "step": 772 + }, + { + "epoch": 2.422866092404072, + "grad_norm": 0.9841981999228102, + "learning_rate": 1.2267875348401585e-05, + "loss": 0.7653, + "step": 773 + }, + { + "epoch": 2.4259984338292875, + "grad_norm": 0.9641773998632913, + "learning_rate": 1.224654772768946e-05, + "loss": 0.7338, + "step": 774 + }, + { + "epoch": 2.4291307752545026, + "grad_norm": 0.947511728550051, + "learning_rate": 1.2225209339563144e-05, + "loss": 0.7583, + "step": 775 + }, + { + "epoch": 2.432263116679718, + "grad_norm": 0.9870470819669465, + "learning_rate": 1.220386028629479e-05, + "loss": 0.781, + "step": 776 + }, + { + "epoch": 2.4353954581049333, + "grad_norm": 0.9540467312167563, + "learning_rate": 1.218250067020766e-05, + "loss": 0.732, + "step": 777 + }, + { + "epoch": 2.438527799530149, + "grad_norm": 0.9742439189723893, + "learning_rate": 1.2161130593675643e-05, + "loss": 0.7654, + "step": 778 + }, + { + "epoch": 2.441660140955364, + "grad_norm": 0.947883079443643, + "learning_rate": 1.2139750159122771e-05, + "loss": 0.7291, + "step": 779 + }, + { + "epoch": 2.4447924823805796, + "grad_norm": 1.0108972783151455, + "learning_rate": 1.2118359469022714e-05, + "loss": 0.7343, + "step": 780 + }, + { + "epoch": 2.4479248238057947, + "grad_norm": 0.9909962900496487, + "learning_rate": 1.2096958625898294e-05, + "loss": 0.75, + "step": 781 + }, + { + "epoch": 2.4510571652310102, + "grad_norm": 0.979708350323752, + "learning_rate": 1.2075547732321005e-05, + "loss": 0.7502, + "step": 782 + }, + { + "epoch": 2.4541895066562254, + "grad_norm": 0.9605132744547762, + "learning_rate": 1.2054126890910499e-05, + "loss": 0.7519, + "step": 783 + }, + { + "epoch": 2.457321848081441, + "grad_norm": 0.9692514022395361, + "learning_rate": 1.2032696204334115e-05, + "loss": 0.7273, + "step": 784 + }, + { + "epoch": 2.460454189506656, + "grad_norm": 0.9449589861483914, + "learning_rate": 1.2011255775306378e-05, + "loss": 0.7511, + "step": 785 + }, + { + "epoch": 2.4635865309318716, + "grad_norm": 0.9471704960405938, + "learning_rate": 1.1989805706588502e-05, + "loss": 0.7347, + "step": 786 + }, + { + "epoch": 2.4667188723570868, + "grad_norm": 0.9320311390343631, + "learning_rate": 1.1968346100987906e-05, + "loss": 0.7082, + "step": 787 + }, + { + "epoch": 2.4698512137823023, + "grad_norm": 0.9162015711675013, + "learning_rate": 1.1946877061357723e-05, + "loss": 0.7287, + "step": 788 + }, + { + "epoch": 2.4729835552075174, + "grad_norm": 0.9676114383404737, + "learning_rate": 1.1925398690596293e-05, + "loss": 0.753, + "step": 789 + }, + { + "epoch": 2.476115896632733, + "grad_norm": 0.9877343226967886, + "learning_rate": 1.1903911091646684e-05, + "loss": 0.7758, + "step": 790 + }, + { + "epoch": 2.479248238057948, + "grad_norm": 1.023377404363016, + "learning_rate": 1.1882414367496196e-05, + "loss": 0.7392, + "step": 791 + }, + { + "epoch": 2.4823805794831637, + "grad_norm": 0.9945895799174997, + "learning_rate": 1.186090862117586e-05, + "loss": 0.7502, + "step": 792 + }, + { + "epoch": 2.485512920908379, + "grad_norm": 1.0038795381065821, + "learning_rate": 1.183939395575995e-05, + "loss": 0.806, + "step": 793 + }, + { + "epoch": 2.4886452623335944, + "grad_norm": 0.9926041291737507, + "learning_rate": 1.181787047436549e-05, + "loss": 0.7617, + "step": 794 + }, + { + "epoch": 2.4917776037588095, + "grad_norm": 0.9747916863077687, + "learning_rate": 1.1796338280151756e-05, + "loss": 0.7506, + "step": 795 + }, + { + "epoch": 2.494909945184025, + "grad_norm": 0.928076659108732, + "learning_rate": 1.1774797476319785e-05, + "loss": 0.7306, + "step": 796 + }, + { + "epoch": 2.49804228660924, + "grad_norm": 1.0261524870519174, + "learning_rate": 1.175324816611188e-05, + "loss": 0.7463, + "step": 797 + }, + { + "epoch": 2.501174628034456, + "grad_norm": 0.9847678627771624, + "learning_rate": 1.1731690452811107e-05, + "loss": 0.7281, + "step": 798 + }, + { + "epoch": 2.5043069694596714, + "grad_norm": 0.9643463190600754, + "learning_rate": 1.1710124439740816e-05, + "loss": 0.76, + "step": 799 + }, + { + "epoch": 2.5074393108848865, + "grad_norm": 0.969773920531062, + "learning_rate": 1.1688550230264128e-05, + "loss": 0.7658, + "step": 800 + }, + { + "epoch": 2.5105716523101016, + "grad_norm": 0.9626642855194483, + "learning_rate": 1.1666967927783461e-05, + "loss": 0.7478, + "step": 801 + }, + { + "epoch": 2.513703993735317, + "grad_norm": 0.9884939163865005, + "learning_rate": 1.164537763574001e-05, + "loss": 0.7153, + "step": 802 + }, + { + "epoch": 2.5168363351605327, + "grad_norm": 0.9485195981984704, + "learning_rate": 1.1623779457613263e-05, + "loss": 0.7493, + "step": 803 + }, + { + "epoch": 2.519968676585748, + "grad_norm": 0.9438738641259264, + "learning_rate": 1.160217349692051e-05, + "loss": 0.7592, + "step": 804 + }, + { + "epoch": 2.523101018010963, + "grad_norm": 0.9775231067597289, + "learning_rate": 1.1580559857216347e-05, + "loss": 0.7669, + "step": 805 + }, + { + "epoch": 2.5262333594361785, + "grad_norm": 0.9964813304861763, + "learning_rate": 1.1558938642092161e-05, + "loss": 0.75, + "step": 806 + }, + { + "epoch": 2.529365700861394, + "grad_norm": 0.9604452115462686, + "learning_rate": 1.1537309955175655e-05, + "loss": 0.7367, + "step": 807 + }, + { + "epoch": 2.5324980422866092, + "grad_norm": 0.9243296564169551, + "learning_rate": 1.1515673900130344e-05, + "loss": 0.7135, + "step": 808 + }, + { + "epoch": 2.5356303837118244, + "grad_norm": 0.9840647742781437, + "learning_rate": 1.149403058065506e-05, + "loss": 0.7744, + "step": 809 + }, + { + "epoch": 2.53876272513704, + "grad_norm": 0.9363389375104039, + "learning_rate": 1.1472380100483438e-05, + "loss": 0.7333, + "step": 810 + }, + { + "epoch": 2.5418950665622555, + "grad_norm": 1.005204535054278, + "learning_rate": 1.145072256338345e-05, + "loss": 0.7535, + "step": 811 + }, + { + "epoch": 2.5450274079874706, + "grad_norm": 0.9828566960512756, + "learning_rate": 1.1429058073156886e-05, + "loss": 0.8036, + "step": 812 + }, + { + "epoch": 2.5481597494126857, + "grad_norm": 0.9671196176463093, + "learning_rate": 1.140738673363886e-05, + "loss": 0.7781, + "step": 813 + }, + { + "epoch": 2.5512920908379013, + "grad_norm": 0.9532345512189722, + "learning_rate": 1.1385708648697306e-05, + "loss": 0.7126, + "step": 814 + }, + { + "epoch": 2.554424432263117, + "grad_norm": 0.9251640219717523, + "learning_rate": 1.1364023922232503e-05, + "loss": 0.7222, + "step": 815 + }, + { + "epoch": 2.557556773688332, + "grad_norm": 0.9869679713832883, + "learning_rate": 1.1342332658176556e-05, + "loss": 0.7058, + "step": 816 + }, + { + "epoch": 2.560689115113547, + "grad_norm": 0.9930580206564761, + "learning_rate": 1.13206349604929e-05, + "loss": 0.7729, + "step": 817 + }, + { + "epoch": 2.5638214565387627, + "grad_norm": 0.9420081126831853, + "learning_rate": 1.1298930933175805e-05, + "loss": 0.7284, + "step": 818 + }, + { + "epoch": 2.5669537979639783, + "grad_norm": 1.0000782701863635, + "learning_rate": 1.1277220680249887e-05, + "loss": 0.7776, + "step": 819 + }, + { + "epoch": 2.5700861393891934, + "grad_norm": 0.9515958392786626, + "learning_rate": 1.1255504305769589e-05, + "loss": 0.7402, + "step": 820 + }, + { + "epoch": 2.5732184808144085, + "grad_norm": 0.9398121571554576, + "learning_rate": 1.1233781913818707e-05, + "loss": 0.7255, + "step": 821 + }, + { + "epoch": 2.576350822239624, + "grad_norm": 0.9937093752648678, + "learning_rate": 1.1212053608509862e-05, + "loss": 0.7627, + "step": 822 + }, + { + "epoch": 2.5794831636648397, + "grad_norm": 0.9791666530858968, + "learning_rate": 1.119031949398403e-05, + "loss": 0.7689, + "step": 823 + }, + { + "epoch": 2.5826155050900548, + "grad_norm": 0.9496103182684476, + "learning_rate": 1.1168579674410026e-05, + "loss": 0.7458, + "step": 824 + }, + { + "epoch": 2.58574784651527, + "grad_norm": 0.9935430896103922, + "learning_rate": 1.1146834253984008e-05, + "loss": 0.7673, + "step": 825 + }, + { + "epoch": 2.5888801879404855, + "grad_norm": 0.9901446257978567, + "learning_rate": 1.1125083336928974e-05, + "loss": 0.767, + "step": 826 + }, + { + "epoch": 2.592012529365701, + "grad_norm": 0.9886430099783694, + "learning_rate": 1.1103327027494272e-05, + "loss": 0.7649, + "step": 827 + }, + { + "epoch": 2.595144870790916, + "grad_norm": 0.957433731138305, + "learning_rate": 1.10815654299551e-05, + "loss": 0.7496, + "step": 828 + }, + { + "epoch": 2.5982772122161313, + "grad_norm": 0.9549512909328034, + "learning_rate": 1.1059798648611988e-05, + "loss": 0.7479, + "step": 829 + }, + { + "epoch": 2.601409553641347, + "grad_norm": 0.9729410513377862, + "learning_rate": 1.103802678779032e-05, + "loss": 0.7215, + "step": 830 + }, + { + "epoch": 2.6045418950665624, + "grad_norm": 0.9739903952440084, + "learning_rate": 1.1016249951839824e-05, + "loss": 0.7481, + "step": 831 + }, + { + "epoch": 2.6076742364917775, + "grad_norm": 0.9413669343604363, + "learning_rate": 1.0994468245134071e-05, + "loss": 0.7567, + "step": 832 + }, + { + "epoch": 2.610806577916993, + "grad_norm": 0.9742157973187829, + "learning_rate": 1.0972681772069981e-05, + "loss": 0.7463, + "step": 833 + }, + { + "epoch": 2.6139389193422082, + "grad_norm": 0.9778627829815928, + "learning_rate": 1.0950890637067317e-05, + "loss": 0.7633, + "step": 834 + }, + { + "epoch": 2.617071260767424, + "grad_norm": 0.9429091429491205, + "learning_rate": 1.0929094944568182e-05, + "loss": 0.7321, + "step": 835 + }, + { + "epoch": 2.620203602192639, + "grad_norm": 0.9807225580513256, + "learning_rate": 1.0907294799036525e-05, + "loss": 0.7599, + "step": 836 + }, + { + "epoch": 2.6233359436178545, + "grad_norm": 0.9466168414336543, + "learning_rate": 1.0885490304957642e-05, + "loss": 0.7111, + "step": 837 + }, + { + "epoch": 2.6264682850430696, + "grad_norm": 0.9473666583781692, + "learning_rate": 1.0863681566837666e-05, + "loss": 0.7333, + "step": 838 + }, + { + "epoch": 2.629600626468285, + "grad_norm": 0.9524909535219792, + "learning_rate": 1.0841868689203072e-05, + "loss": 0.7625, + "step": 839 + }, + { + "epoch": 2.6327329678935003, + "grad_norm": 0.9432869054961515, + "learning_rate": 1.0820051776600175e-05, + "loss": 0.745, + "step": 840 + }, + { + "epoch": 2.635865309318716, + "grad_norm": 0.940106657073856, + "learning_rate": 1.0798230933594629e-05, + "loss": 0.7438, + "step": 841 + }, + { + "epoch": 2.638997650743931, + "grad_norm": 0.959855959982091, + "learning_rate": 1.0776406264770933e-05, + "loss": 0.7728, + "step": 842 + }, + { + "epoch": 2.6421299921691466, + "grad_norm": 1.0045635045124952, + "learning_rate": 1.0754577874731911e-05, + "loss": 0.774, + "step": 843 + }, + { + "epoch": 2.6452623335943617, + "grad_norm": 0.9757139770765741, + "learning_rate": 1.073274586809823e-05, + "loss": 0.7312, + "step": 844 + }, + { + "epoch": 2.6483946750195773, + "grad_norm": 0.9438533061715757, + "learning_rate": 1.0710910349507881e-05, + "loss": 0.7237, + "step": 845 + }, + { + "epoch": 2.6515270164447924, + "grad_norm": 0.9531758260863951, + "learning_rate": 1.0689071423615708e-05, + "loss": 0.7348, + "step": 846 + }, + { + "epoch": 2.654659357870008, + "grad_norm": 0.9446644812284712, + "learning_rate": 1.0667229195092868e-05, + "loss": 0.7602, + "step": 847 + }, + { + "epoch": 2.657791699295223, + "grad_norm": 0.9782412701275756, + "learning_rate": 1.0645383768626345e-05, + "loss": 0.7765, + "step": 848 + }, + { + "epoch": 2.6609240407204386, + "grad_norm": 1.007139575996801, + "learning_rate": 1.0623535248918463e-05, + "loss": 0.7751, + "step": 849 + }, + { + "epoch": 2.6640563821456538, + "grad_norm": 0.972578870009894, + "learning_rate": 1.0601683740686366e-05, + "loss": 0.8046, + "step": 850 + }, + { + "epoch": 2.6671887235708693, + "grad_norm": 0.9533200080779424, + "learning_rate": 1.0579829348661522e-05, + "loss": 0.7268, + "step": 851 + }, + { + "epoch": 2.6703210649960845, + "grad_norm": 0.9569784524332555, + "learning_rate": 1.0557972177589214e-05, + "loss": 0.7914, + "step": 852 + }, + { + "epoch": 2.6734534064213, + "grad_norm": 0.9554744612301984, + "learning_rate": 1.0536112332228057e-05, + "loss": 0.7656, + "step": 853 + }, + { + "epoch": 2.676585747846515, + "grad_norm": 0.9379633252744265, + "learning_rate": 1.0514249917349473e-05, + "loss": 0.7174, + "step": 854 + }, + { + "epoch": 2.6797180892717307, + "grad_norm": 0.9498518801324679, + "learning_rate": 1.0492385037737207e-05, + "loss": 0.7359, + "step": 855 + }, + { + "epoch": 2.682850430696946, + "grad_norm": 0.9323391831700777, + "learning_rate": 1.0470517798186813e-05, + "loss": 0.7516, + "step": 856 + }, + { + "epoch": 2.6859827721221614, + "grad_norm": 0.977808841738496, + "learning_rate": 1.044864830350515e-05, + "loss": 0.7588, + "step": 857 + }, + { + "epoch": 2.6891151135473765, + "grad_norm": 0.9607140331334884, + "learning_rate": 1.0426776658509899e-05, + "loss": 0.7297, + "step": 858 + }, + { + "epoch": 2.692247454972592, + "grad_norm": 0.9644638205133018, + "learning_rate": 1.040490296802904e-05, + "loss": 0.7602, + "step": 859 + }, + { + "epoch": 2.6953797963978072, + "grad_norm": 0.9541062791921975, + "learning_rate": 1.0383027336900356e-05, + "loss": 0.747, + "step": 860 + }, + { + "epoch": 2.698512137823023, + "grad_norm": 0.9713788632902975, + "learning_rate": 1.0361149869970929e-05, + "loss": 0.7486, + "step": 861 + }, + { + "epoch": 2.701644479248238, + "grad_norm": 0.9599164742932305, + "learning_rate": 1.0339270672096645e-05, + "loss": 0.7677, + "step": 862 + }, + { + "epoch": 2.7047768206734535, + "grad_norm": 0.9851639258139112, + "learning_rate": 1.0317389848141685e-05, + "loss": 0.735, + "step": 863 + }, + { + "epoch": 2.7079091620986686, + "grad_norm": 0.9627844834333039, + "learning_rate": 1.0295507502978024e-05, + "loss": 0.7423, + "step": 864 + }, + { + "epoch": 2.711041503523884, + "grad_norm": 0.9528844028329699, + "learning_rate": 1.0273623741484924e-05, + "loss": 0.7384, + "step": 865 + }, + { + "epoch": 2.7141738449490993, + "grad_norm": 1.0340380298089817, + "learning_rate": 1.0251738668548439e-05, + "loss": 0.8031, + "step": 866 + }, + { + "epoch": 2.717306186374315, + "grad_norm": 0.9538815854838193, + "learning_rate": 1.0229852389060905e-05, + "loss": 0.7339, + "step": 867 + }, + { + "epoch": 2.72043852779953, + "grad_norm": 0.9723625755825231, + "learning_rate": 1.0207965007920448e-05, + "loss": 0.7837, + "step": 868 + }, + { + "epoch": 2.7235708692247456, + "grad_norm": 0.967760456332842, + "learning_rate": 1.0186076630030469e-05, + "loss": 0.7522, + "step": 869 + }, + { + "epoch": 2.7267032106499607, + "grad_norm": 0.956046387699816, + "learning_rate": 1.0164187360299143e-05, + "loss": 0.7055, + "step": 870 + }, + { + "epoch": 2.7298355520751763, + "grad_norm": 0.9568472870023063, + "learning_rate": 1.0142297303638926e-05, + "loss": 0.7284, + "step": 871 + }, + { + "epoch": 2.7329678935003914, + "grad_norm": 0.988027374434136, + "learning_rate": 1.0120406564966043e-05, + "loss": 0.7714, + "step": 872 + }, + { + "epoch": 2.736100234925607, + "grad_norm": 1.028366948188634, + "learning_rate": 1.0098515249199988e-05, + "loss": 0.8163, + "step": 873 + }, + { + "epoch": 2.7392325763508225, + "grad_norm": 1.4617689871519928, + "learning_rate": 1.0076623461263017e-05, + "loss": 0.7347, + "step": 874 + }, + { + "epoch": 2.7423649177760376, + "grad_norm": 0.9546506131293968, + "learning_rate": 1.0054731306079657e-05, + "loss": 0.7131, + "step": 875 + }, + { + "epoch": 2.7454972592012528, + "grad_norm": 0.9591903904399064, + "learning_rate": 1.0032838888576194e-05, + "loss": 0.72, + "step": 876 + }, + { + "epoch": 2.7486296006264683, + "grad_norm": 0.9704760247984783, + "learning_rate": 1.0010946313680156e-05, + "loss": 0.7519, + "step": 877 + }, + { + "epoch": 2.751761942051684, + "grad_norm": 0.9496820302125051, + "learning_rate": 9.989053686319849e-06, + "loss": 0.7435, + "step": 878 + }, + { + "epoch": 2.754894283476899, + "grad_norm": 0.9773806934541943, + "learning_rate": 9.967161111423813e-06, + "loss": 0.7795, + "step": 879 + }, + { + "epoch": 2.758026624902114, + "grad_norm": 0.9202467047384671, + "learning_rate": 9.945268693920346e-06, + "loss": 0.7292, + "step": 880 + }, + { + "epoch": 2.7611589663273297, + "grad_norm": 0.974687628354556, + "learning_rate": 9.923376538736985e-06, + "loss": 0.7761, + "step": 881 + }, + { + "epoch": 2.7642913077525453, + "grad_norm": 0.9294513579702289, + "learning_rate": 9.901484750800014e-06, + "loss": 0.7235, + "step": 882 + }, + { + "epoch": 2.7674236491777604, + "grad_norm": 0.9850151969699499, + "learning_rate": 9.879593435033957e-06, + "loss": 0.7845, + "step": 883 + }, + { + "epoch": 2.7705559906029755, + "grad_norm": 0.929292922655392, + "learning_rate": 9.857702696361076e-06, + "loss": 0.75, + "step": 884 + }, + { + "epoch": 2.773688332028191, + "grad_norm": 0.9256442001669153, + "learning_rate": 9.835812639700862e-06, + "loss": 0.7215, + "step": 885 + }, + { + "epoch": 2.7768206734534067, + "grad_norm": 0.9481844945658222, + "learning_rate": 9.813923369969535e-06, + "loss": 0.7381, + "step": 886 + }, + { + "epoch": 2.779953014878622, + "grad_norm": 0.9481037841878706, + "learning_rate": 9.792034992079554e-06, + "loss": 0.7556, + "step": 887 + }, + { + "epoch": 2.783085356303837, + "grad_norm": 0.952957986686458, + "learning_rate": 9.770147610939098e-06, + "loss": 0.7339, + "step": 888 + }, + { + "epoch": 2.7862176977290525, + "grad_norm": 0.9549009956792547, + "learning_rate": 9.748261331451566e-06, + "loss": 0.7476, + "step": 889 + }, + { + "epoch": 2.789350039154268, + "grad_norm": 0.9782654041320651, + "learning_rate": 9.726376258515077e-06, + "loss": 0.7421, + "step": 890 + }, + { + "epoch": 2.792482380579483, + "grad_norm": 0.9900817275321431, + "learning_rate": 9.70449249702198e-06, + "loss": 0.7703, + "step": 891 + }, + { + "epoch": 2.7956147220046983, + "grad_norm": 0.9684826883994128, + "learning_rate": 9.682610151858317e-06, + "loss": 0.7874, + "step": 892 + }, + { + "epoch": 2.798747063429914, + "grad_norm": 0.9509743876454371, + "learning_rate": 9.660729327903359e-06, + "loss": 0.7554, + "step": 893 + }, + { + "epoch": 2.8018794048551294, + "grad_norm": 0.9651752336989294, + "learning_rate": 9.638850130029075e-06, + "loss": 0.7081, + "step": 894 + }, + { + "epoch": 2.8050117462803446, + "grad_norm": 0.9947227809634293, + "learning_rate": 9.616972663099648e-06, + "loss": 0.7533, + "step": 895 + }, + { + "epoch": 2.8081440877055597, + "grad_norm": 0.9575600590215773, + "learning_rate": 9.595097031970963e-06, + "loss": 0.7567, + "step": 896 + }, + { + "epoch": 2.8112764291307752, + "grad_norm": 0.9658030860447785, + "learning_rate": 9.5732233414901e-06, + "loss": 0.747, + "step": 897 + }, + { + "epoch": 2.814408770555991, + "grad_norm": 0.995325139613679, + "learning_rate": 9.551351696494854e-06, + "loss": 0.7438, + "step": 898 + }, + { + "epoch": 2.817541111981206, + "grad_norm": 0.9712772975214642, + "learning_rate": 9.529482201813192e-06, + "loss": 0.7293, + "step": 899 + }, + { + "epoch": 2.820673453406421, + "grad_norm": 0.9546909848683623, + "learning_rate": 9.507614962262796e-06, + "loss": 0.768, + "step": 900 + }, + { + "epoch": 2.8238057948316366, + "grad_norm": 0.9329589715311891, + "learning_rate": 9.48575008265053e-06, + "loss": 0.7341, + "step": 901 + }, + { + "epoch": 2.826938136256852, + "grad_norm": 0.9357071849591595, + "learning_rate": 9.463887667771946e-06, + "loss": 0.7195, + "step": 902 + }, + { + "epoch": 2.8300704776820673, + "grad_norm": 0.933792474330246, + "learning_rate": 9.442027822410786e-06, + "loss": 0.7563, + "step": 903 + }, + { + "epoch": 2.8332028191072824, + "grad_norm": 0.9936060316623524, + "learning_rate": 9.420170651338482e-06, + "loss": 0.7546, + "step": 904 + }, + { + "epoch": 2.836335160532498, + "grad_norm": 0.9522662188147043, + "learning_rate": 9.398316259313636e-06, + "loss": 0.744, + "step": 905 + }, + { + "epoch": 2.8394675019577136, + "grad_norm": 0.9768599306563925, + "learning_rate": 9.376464751081542e-06, + "loss": 0.7547, + "step": 906 + }, + { + "epoch": 2.8425998433829287, + "grad_norm": 0.9612851410888434, + "learning_rate": 9.354616231373659e-06, + "loss": 0.7662, + "step": 907 + }, + { + "epoch": 2.845732184808144, + "grad_norm": 0.9419299539508802, + "learning_rate": 9.332770804907137e-06, + "loss": 0.7293, + "step": 908 + }, + { + "epoch": 2.8488645262333594, + "grad_norm": 0.9437811849610814, + "learning_rate": 9.310928576384293e-06, + "loss": 0.7258, + "step": 909 + }, + { + "epoch": 2.851996867658575, + "grad_norm": 0.9714173836845524, + "learning_rate": 9.289089650492119e-06, + "loss": 0.732, + "step": 910 + }, + { + "epoch": 2.85512920908379, + "grad_norm": 0.9482886998971121, + "learning_rate": 9.267254131901772e-06, + "loss": 0.7273, + "step": 911 + }, + { + "epoch": 2.858261550509005, + "grad_norm": 0.9868925213214547, + "learning_rate": 9.245422125268094e-06, + "loss": 0.7215, + "step": 912 + }, + { + "epoch": 2.861393891934221, + "grad_norm": 0.9529681945509453, + "learning_rate": 9.223593735229069e-06, + "loss": 0.7498, + "step": 913 + }, + { + "epoch": 2.8645262333594363, + "grad_norm": 0.9383207113099425, + "learning_rate": 9.201769066405373e-06, + "loss": 0.7302, + "step": 914 + }, + { + "epoch": 2.8676585747846515, + "grad_norm": 0.9565533325468113, + "learning_rate": 9.179948223399828e-06, + "loss": 0.7352, + "step": 915 + }, + { + "epoch": 2.870790916209867, + "grad_norm": 0.973007632374376, + "learning_rate": 9.15813131079693e-06, + "loss": 0.7646, + "step": 916 + }, + { + "epoch": 2.873923257635082, + "grad_norm": 0.9966309662857158, + "learning_rate": 9.136318433162336e-06, + "loss": 0.724, + "step": 917 + }, + { + "epoch": 2.8770555990602977, + "grad_norm": 0.9853350271329723, + "learning_rate": 9.114509695042362e-06, + "loss": 0.7615, + "step": 918 + }, + { + "epoch": 2.880187940485513, + "grad_norm": 0.9504300421132944, + "learning_rate": 9.092705200963478e-06, + "loss": 0.7663, + "step": 919 + }, + { + "epoch": 2.8833202819107284, + "grad_norm": 0.9477723894786849, + "learning_rate": 9.070905055431822e-06, + "loss": 0.7372, + "step": 920 + }, + { + "epoch": 2.8864526233359435, + "grad_norm": 0.9495506324641605, + "learning_rate": 9.049109362932685e-06, + "loss": 0.7323, + "step": 921 + }, + { + "epoch": 2.889584964761159, + "grad_norm": 0.9315436534258359, + "learning_rate": 9.02731822793002e-06, + "loss": 0.7298, + "step": 922 + }, + { + "epoch": 2.8927173061863742, + "grad_norm": 0.9633745464994017, + "learning_rate": 9.005531754865929e-06, + "loss": 0.7329, + "step": 923 + }, + { + "epoch": 2.89584964761159, + "grad_norm": 0.9520011329637497, + "learning_rate": 8.983750048160178e-06, + "loss": 0.7461, + "step": 924 + }, + { + "epoch": 2.898981989036805, + "grad_norm": 0.947202489605334, + "learning_rate": 8.961973212209684e-06, + "loss": 0.7353, + "step": 925 + }, + { + "epoch": 2.9021143304620205, + "grad_norm": 0.9564588694025306, + "learning_rate": 8.940201351388016e-06, + "loss": 0.7499, + "step": 926 + }, + { + "epoch": 2.9052466718872356, + "grad_norm": 0.9660352117956935, + "learning_rate": 8.918434570044904e-06, + "loss": 0.774, + "step": 927 + }, + { + "epoch": 2.908379013312451, + "grad_norm": 0.9516602740353381, + "learning_rate": 8.89667297250573e-06, + "loss": 0.7196, + "step": 928 + }, + { + "epoch": 2.9115113547376663, + "grad_norm": 0.9503514513189356, + "learning_rate": 8.874916663071027e-06, + "loss": 0.7621, + "step": 929 + }, + { + "epoch": 2.914643696162882, + "grad_norm": 0.9732615080165525, + "learning_rate": 8.853165746015997e-06, + "loss": 0.7596, + "step": 930 + }, + { + "epoch": 2.917776037588097, + "grad_norm": 1.010101952673623, + "learning_rate": 8.831420325589976e-06, + "loss": 0.7654, + "step": 931 + }, + { + "epoch": 2.9209083790133126, + "grad_norm": 0.9740311574523944, + "learning_rate": 8.809680506015973e-06, + "loss": 0.7513, + "step": 932 + }, + { + "epoch": 2.9240407204385277, + "grad_norm": 0.9960714830033841, + "learning_rate": 8.787946391490141e-06, + "loss": 0.7656, + "step": 933 + }, + { + "epoch": 2.9271730618637433, + "grad_norm": 0.9638028878471048, + "learning_rate": 8.766218086181298e-06, + "loss": 0.7776, + "step": 934 + }, + { + "epoch": 2.9303054032889584, + "grad_norm": 0.9722845022040784, + "learning_rate": 8.744495694230413e-06, + "loss": 0.7301, + "step": 935 + }, + { + "epoch": 2.933437744714174, + "grad_norm": 0.9628137063571912, + "learning_rate": 8.722779319750116e-06, + "loss": 0.7206, + "step": 936 + }, + { + "epoch": 2.936570086139389, + "grad_norm": 0.9465035800698545, + "learning_rate": 8.701069066824195e-06, + "loss": 0.722, + "step": 937 + }, + { + "epoch": 2.9397024275646046, + "grad_norm": 0.9902093031760724, + "learning_rate": 8.679365039507104e-06, + "loss": 0.7857, + "step": 938 + }, + { + "epoch": 2.9428347689898198, + "grad_norm": 0.948550028383559, + "learning_rate": 8.657667341823449e-06, + "loss": 0.7501, + "step": 939 + }, + { + "epoch": 2.9459671104150353, + "grad_norm": 0.9610537515494351, + "learning_rate": 8.6359760777675e-06, + "loss": 0.7242, + "step": 940 + }, + { + "epoch": 2.9490994518402505, + "grad_norm": 0.9553995440705682, + "learning_rate": 8.614291351302697e-06, + "loss": 0.7344, + "step": 941 + }, + { + "epoch": 2.952231793265466, + "grad_norm": 0.9521585819489644, + "learning_rate": 8.592613266361145e-06, + "loss": 0.7399, + "step": 942 + }, + { + "epoch": 2.955364134690681, + "grad_norm": 0.9675254410966843, + "learning_rate": 8.570941926843117e-06, + "loss": 0.7474, + "step": 943 + }, + { + "epoch": 2.9584964761158967, + "grad_norm": 0.949925314091798, + "learning_rate": 8.549277436616551e-06, + "loss": 0.7398, + "step": 944 + }, + { + "epoch": 2.961628817541112, + "grad_norm": 0.9743585973394909, + "learning_rate": 8.527619899516567e-06, + "loss": 0.7281, + "step": 945 + }, + { + "epoch": 2.9647611589663274, + "grad_norm": 0.9721394272700253, + "learning_rate": 8.505969419344946e-06, + "loss": 0.7691, + "step": 946 + }, + { + "epoch": 2.9678935003915425, + "grad_norm": 0.9756436120696973, + "learning_rate": 8.484326099869659e-06, + "loss": 0.7346, + "step": 947 + }, + { + "epoch": 2.971025841816758, + "grad_norm": 0.963968460621726, + "learning_rate": 8.462690044824347e-06, + "loss": 0.7632, + "step": 948 + }, + { + "epoch": 2.9741581832419732, + "grad_norm": 0.9886664730071104, + "learning_rate": 8.441061357907842e-06, + "loss": 0.7795, + "step": 949 + }, + { + "epoch": 2.977290524667189, + "grad_norm": 0.9402625444423229, + "learning_rate": 8.419440142783653e-06, + "loss": 0.7237, + "step": 950 + }, + { + "epoch": 2.980422866092404, + "grad_norm": 0.9602243212134389, + "learning_rate": 8.397826503079489e-06, + "loss": 0.7561, + "step": 951 + }, + { + "epoch": 2.9835552075176195, + "grad_norm": 0.9617618663543123, + "learning_rate": 8.376220542386742e-06, + "loss": 0.7799, + "step": 952 + }, + { + "epoch": 2.9866875489428346, + "grad_norm": 0.9190755212609129, + "learning_rate": 8.354622364259995e-06, + "loss": 0.7042, + "step": 953 + }, + { + "epoch": 2.98981989036805, + "grad_norm": 0.9586872985486566, + "learning_rate": 8.333032072216542e-06, + "loss": 0.7411, + "step": 954 + }, + { + "epoch": 2.9929522317932653, + "grad_norm": 0.9490934425198838, + "learning_rate": 8.311449769735874e-06, + "loss": 0.7699, + "step": 955 + }, + { + "epoch": 2.996084573218481, + "grad_norm": 0.9483587215466358, + "learning_rate": 8.289875560259189e-06, + "loss": 0.729, + "step": 956 + }, + { + "epoch": 2.999216914643696, + "grad_norm": 0.9426645402788387, + "learning_rate": 8.268309547188894e-06, + "loss": 0.7605, + "step": 957 + }, + { + "epoch": 3.0031323414252156, + "grad_norm": 2.4449431031791447, + "learning_rate": 8.246751833888122e-06, + "loss": 1.363, + "step": 958 + }, + { + "epoch": 3.0062646828504307, + "grad_norm": 1.6451995720214123, + "learning_rate": 8.225202523680218e-06, + "loss": 0.5471, + "step": 959 + }, + { + "epoch": 3.0093970242756463, + "grad_norm": 1.6727678935052752, + "learning_rate": 8.203661719848249e-06, + "loss": 0.5595, + "step": 960 + }, + { + "epoch": 3.0125293657008614, + "grad_norm": 1.5460240913854986, + "learning_rate": 8.182129525634513e-06, + "loss": 0.5325, + "step": 961 + }, + { + "epoch": 3.015661707126077, + "grad_norm": 1.2460927908669523, + "learning_rate": 8.160606044240052e-06, + "loss": 0.5299, + "step": 962 + }, + { + "epoch": 3.018794048551292, + "grad_norm": 1.2859700150226807, + "learning_rate": 8.139091378824144e-06, + "loss": 0.526, + "step": 963 + }, + { + "epoch": 3.0219263899765076, + "grad_norm": 1.7164212978411557, + "learning_rate": 8.117585632503805e-06, + "loss": 0.5641, + "step": 964 + }, + { + "epoch": 3.0250587314017228, + "grad_norm": 1.7851156197762952, + "learning_rate": 8.096088908353316e-06, + "loss": 0.5354, + "step": 965 + }, + { + "epoch": 3.0281910728269383, + "grad_norm": 1.7956222015211165, + "learning_rate": 8.07460130940371e-06, + "loss": 0.5468, + "step": 966 + }, + { + "epoch": 3.0313234142521535, + "grad_norm": 1.7931934712777515, + "learning_rate": 8.05312293864228e-06, + "loss": 0.5571, + "step": 967 + }, + { + "epoch": 3.034455755677369, + "grad_norm": 1.6976805325976927, + "learning_rate": 8.031653899012098e-06, + "loss": 0.5492, + "step": 968 + }, + { + "epoch": 3.037588097102584, + "grad_norm": 1.596398489281078, + "learning_rate": 8.010194293411502e-06, + "loss": 0.4879, + "step": 969 + }, + { + "epoch": 3.0407204385277997, + "grad_norm": 1.4311365997826093, + "learning_rate": 7.988744224693625e-06, + "loss": 0.5483, + "step": 970 + }, + { + "epoch": 3.043852779953015, + "grad_norm": 1.3005396695284464, + "learning_rate": 7.967303795665887e-06, + "loss": 0.5155, + "step": 971 + }, + { + "epoch": 3.0469851213782304, + "grad_norm": 1.1670816271780478, + "learning_rate": 7.945873109089503e-06, + "loss": 0.5483, + "step": 972 + }, + { + "epoch": 3.0501174628034455, + "grad_norm": 1.2526413892546973, + "learning_rate": 7.924452267678998e-06, + "loss": 0.5208, + "step": 973 + }, + { + "epoch": 3.053249804228661, + "grad_norm": 1.3258191230420822, + "learning_rate": 7.903041374101707e-06, + "loss": 0.5597, + "step": 974 + }, + { + "epoch": 3.0563821456538762, + "grad_norm": 1.3179145976075899, + "learning_rate": 7.881640530977287e-06, + "loss": 0.5544, + "step": 975 + }, + { + "epoch": 3.059514487079092, + "grad_norm": 1.2661427204898215, + "learning_rate": 7.860249840877232e-06, + "loss": 0.5306, + "step": 976 + }, + { + "epoch": 3.062646828504307, + "grad_norm": 1.20947370272958, + "learning_rate": 7.838869406324358e-06, + "loss": 0.529, + "step": 977 + }, + { + "epoch": 3.0657791699295225, + "grad_norm": 1.1968126239778358, + "learning_rate": 7.817499329792342e-06, + "loss": 0.5685, + "step": 978 + }, + { + "epoch": 3.0689115113547376, + "grad_norm": 1.1994215893244842, + "learning_rate": 7.796139713705214e-06, + "loss": 0.5157, + "step": 979 + }, + { + "epoch": 3.072043852779953, + "grad_norm": 1.231635060309257, + "learning_rate": 7.774790660436857e-06, + "loss": 0.5195, + "step": 980 + }, + { + "epoch": 3.0751761942051683, + "grad_norm": 1.2984303455103845, + "learning_rate": 7.753452272310544e-06, + "loss": 0.5169, + "step": 981 + }, + { + "epoch": 3.078308535630384, + "grad_norm": 1.2495741983021167, + "learning_rate": 7.732124651598418e-06, + "loss": 0.5286, + "step": 982 + }, + { + "epoch": 3.081440877055599, + "grad_norm": 1.2371927309933792, + "learning_rate": 7.710807900521023e-06, + "loss": 0.5285, + "step": 983 + }, + { + "epoch": 3.0845732184808146, + "grad_norm": 1.1598033338515672, + "learning_rate": 7.689502121246809e-06, + "loss": 0.5285, + "step": 984 + }, + { + "epoch": 3.0877055599060297, + "grad_norm": 1.137936793677818, + "learning_rate": 7.668207415891625e-06, + "loss": 0.5257, + "step": 985 + }, + { + "epoch": 3.0908379013312453, + "grad_norm": 1.2251396369541545, + "learning_rate": 7.646923886518263e-06, + "loss": 0.5563, + "step": 986 + }, + { + "epoch": 3.0939702427564604, + "grad_norm": 1.1991699852871776, + "learning_rate": 7.625651635135933e-06, + "loss": 0.527, + "step": 987 + }, + { + "epoch": 3.097102584181676, + "grad_norm": 1.1439621826805295, + "learning_rate": 7.6043907636998026e-06, + "loss": 0.5163, + "step": 988 + }, + { + "epoch": 3.100234925606891, + "grad_norm": 1.123851307419761, + "learning_rate": 7.583141374110495e-06, + "loss": 0.5206, + "step": 989 + }, + { + "epoch": 3.1033672670321066, + "grad_norm": 1.1309411775644675, + "learning_rate": 7.561903568213595e-06, + "loss": 0.5416, + "step": 990 + }, + { + "epoch": 3.1064996084573218, + "grad_norm": 1.164015600425183, + "learning_rate": 7.540677447799176e-06, + "loss": 0.5062, + "step": 991 + }, + { + "epoch": 3.1096319498825373, + "grad_norm": 1.1836169096484468, + "learning_rate": 7.519463114601304e-06, + "loss": 0.5476, + "step": 992 + }, + { + "epoch": 3.1127642913077525, + "grad_norm": 1.142816670241574, + "learning_rate": 7.4982606702975505e-06, + "loss": 0.5342, + "step": 993 + }, + { + "epoch": 3.115896632732968, + "grad_norm": 1.1503408738938847, + "learning_rate": 7.477070216508493e-06, + "loss": 0.5646, + "step": 994 + }, + { + "epoch": 3.119028974158183, + "grad_norm": 1.150741615449422, + "learning_rate": 7.455891854797256e-06, + "loss": 0.5175, + "step": 995 + }, + { + "epoch": 3.1221613155833987, + "grad_norm": 1.2034319285665727, + "learning_rate": 7.434725686669002e-06, + "loss": 0.5765, + "step": 996 + }, + { + "epoch": 3.125293657008614, + "grad_norm": 1.1915855507050621, + "learning_rate": 7.413571813570453e-06, + "loss": 0.5323, + "step": 997 + }, + { + "epoch": 3.1284259984338294, + "grad_norm": 1.2081976282063787, + "learning_rate": 7.392430336889398e-06, + "loss": 0.5266, + "step": 998 + }, + { + "epoch": 3.1315583398590445, + "grad_norm": 1.1322504322898508, + "learning_rate": 7.371301357954219e-06, + "loss": 0.5129, + "step": 999 + }, + { + "epoch": 3.13469068128426, + "grad_norm": 1.1548883914063983, + "learning_rate": 7.350184978033386e-06, + "loss": 0.5193, + "step": 1000 + }, + { + "epoch": 3.137823022709475, + "grad_norm": 1.1526932585853438, + "learning_rate": 7.329081298334994e-06, + "loss": 0.5102, + "step": 1001 + }, + { + "epoch": 3.140955364134691, + "grad_norm": 1.1697102123089387, + "learning_rate": 7.307990420006266e-06, + "loss": 0.5326, + "step": 1002 + }, + { + "epoch": 3.144087705559906, + "grad_norm": 1.1265238542719669, + "learning_rate": 7.286912444133067e-06, + "loss": 0.5299, + "step": 1003 + }, + { + "epoch": 3.1472200469851215, + "grad_norm": 1.1537612326667075, + "learning_rate": 7.265847471739422e-06, + "loss": 0.5151, + "step": 1004 + }, + { + "epoch": 3.1503523884103366, + "grad_norm": 1.1284758438997318, + "learning_rate": 7.244795603787036e-06, + "loss": 0.5215, + "step": 1005 + }, + { + "epoch": 3.153484729835552, + "grad_norm": 1.1977561671812984, + "learning_rate": 7.2237569411748045e-06, + "loss": 0.5126, + "step": 1006 + }, + { + "epoch": 3.1566170712607673, + "grad_norm": 1.1649308019431355, + "learning_rate": 7.202731584738323e-06, + "loss": 0.5289, + "step": 1007 + }, + { + "epoch": 3.159749412685983, + "grad_norm": 1.1377687577701208, + "learning_rate": 7.1817196352494245e-06, + "loss": 0.5191, + "step": 1008 + }, + { + "epoch": 3.162881754111198, + "grad_norm": 1.1332419084126149, + "learning_rate": 7.160721193415681e-06, + "loss": 0.4928, + "step": 1009 + }, + { + "epoch": 3.1660140955364136, + "grad_norm": 1.1255882422146672, + "learning_rate": 7.139736359879916e-06, + "loss": 0.505, + "step": 1010 + }, + { + "epoch": 3.1691464369616287, + "grad_norm": 1.1396125393838465, + "learning_rate": 7.118765235219742e-06, + "loss": 0.5262, + "step": 1011 + }, + { + "epoch": 3.1722787783868442, + "grad_norm": 1.221304685970849, + "learning_rate": 7.0978079199470596e-06, + "loss": 0.5129, + "step": 1012 + }, + { + "epoch": 3.1754111198120594, + "grad_norm": 1.1646706730711807, + "learning_rate": 7.0768645145075885e-06, + "loss": 0.5509, + "step": 1013 + }, + { + "epoch": 3.178543461237275, + "grad_norm": 1.1909834940309783, + "learning_rate": 7.055935119280369e-06, + "loss": 0.5352, + "step": 1014 + }, + { + "epoch": 3.18167580266249, + "grad_norm": 1.1275894474381307, + "learning_rate": 7.035019834577301e-06, + "loss": 0.4987, + "step": 1015 + }, + { + "epoch": 3.1848081440877056, + "grad_norm": 1.1098330354725712, + "learning_rate": 7.014118760642654e-06, + "loss": 0.5429, + "step": 1016 + }, + { + "epoch": 3.1879404855129208, + "grad_norm": 1.1380777641842976, + "learning_rate": 6.993231997652587e-06, + "loss": 0.516, + "step": 1017 + }, + { + "epoch": 3.1910728269381363, + "grad_norm": 1.1693763120706757, + "learning_rate": 6.972359645714666e-06, + "loss": 0.5276, + "step": 1018 + }, + { + "epoch": 3.1942051683633514, + "grad_norm": 1.0981642819607513, + "learning_rate": 6.951501804867392e-06, + "loss": 0.5042, + "step": 1019 + }, + { + "epoch": 3.197337509788567, + "grad_norm": 1.1293301389682295, + "learning_rate": 6.930658575079705e-06, + "loss": 0.5456, + "step": 1020 + }, + { + "epoch": 3.200469851213782, + "grad_norm": 1.0968445808904248, + "learning_rate": 6.909830056250527e-06, + "loss": 0.5079, + "step": 1021 + }, + { + "epoch": 3.2036021926389977, + "grad_norm": 1.1042880780334279, + "learning_rate": 6.889016348208272e-06, + "loss": 0.5247, + "step": 1022 + }, + { + "epoch": 3.206734534064213, + "grad_norm": 1.1473219744960321, + "learning_rate": 6.868217550710359e-06, + "loss": 0.529, + "step": 1023 + }, + { + "epoch": 3.2098668754894284, + "grad_norm": 1.1820536568504343, + "learning_rate": 6.8474337634427505e-06, + "loss": 0.5653, + "step": 1024 + }, + { + "epoch": 3.2129992169146435, + "grad_norm": 1.1788578418720204, + "learning_rate": 6.826665086019467e-06, + "loss": 0.556, + "step": 1025 + }, + { + "epoch": 3.216131558339859, + "grad_norm": 1.1275364008712732, + "learning_rate": 6.805911617982105e-06, + "loss": 0.5211, + "step": 1026 + }, + { + "epoch": 3.219263899765074, + "grad_norm": 1.1788948899624612, + "learning_rate": 6.785173458799365e-06, + "loss": 0.5527, + "step": 1027 + }, + { + "epoch": 3.22239624119029, + "grad_norm": 1.190831430517554, + "learning_rate": 6.764450707866577e-06, + "loss": 0.5569, + "step": 1028 + }, + { + "epoch": 3.225528582615505, + "grad_norm": 1.1631264177590235, + "learning_rate": 6.743743464505219e-06, + "loss": 0.5389, + "step": 1029 + }, + { + "epoch": 3.2286609240407205, + "grad_norm": 1.141781399383516, + "learning_rate": 6.723051827962445e-06, + "loss": 0.5309, + "step": 1030 + }, + { + "epoch": 3.2317932654659356, + "grad_norm": 1.179512415938292, + "learning_rate": 6.7023758974106065e-06, + "loss": 0.525, + "step": 1031 + }, + { + "epoch": 3.234925606891151, + "grad_norm": 1.1944956246302587, + "learning_rate": 6.681715771946777e-06, + "loss": 0.5541, + "step": 1032 + }, + { + "epoch": 3.2380579483163663, + "grad_norm": 1.1591633344391639, + "learning_rate": 6.661071550592284e-06, + "loss": 0.5347, + "step": 1033 + }, + { + "epoch": 3.241190289741582, + "grad_norm": 1.12772338651315, + "learning_rate": 6.6404433322922146e-06, + "loss": 0.5155, + "step": 1034 + }, + { + "epoch": 3.244322631166797, + "grad_norm": 1.165589037264499, + "learning_rate": 6.619831215914974e-06, + "loss": 0.5346, + "step": 1035 + }, + { + "epoch": 3.2474549725920125, + "grad_norm": 1.103553776475765, + "learning_rate": 6.599235300251778e-06, + "loss": 0.5066, + "step": 1036 + }, + { + "epoch": 3.2505873140172277, + "grad_norm": 1.151353363986568, + "learning_rate": 6.578655684016204e-06, + "loss": 0.5456, + "step": 1037 + }, + { + "epoch": 3.2537196554424432, + "grad_norm": 1.159828047686325, + "learning_rate": 6.558092465843704e-06, + "loss": 0.5341, + "step": 1038 + }, + { + "epoch": 3.256851996867659, + "grad_norm": 1.1724554056458623, + "learning_rate": 6.537545744291134e-06, + "loss": 0.5598, + "step": 1039 + }, + { + "epoch": 3.259984338292874, + "grad_norm": 1.1868205375108176, + "learning_rate": 6.517015617836292e-06, + "loss": 0.5404, + "step": 1040 + }, + { + "epoch": 3.263116679718089, + "grad_norm": 1.180384610418905, + "learning_rate": 6.496502184877426e-06, + "loss": 0.5215, + "step": 1041 + }, + { + "epoch": 3.2662490211433046, + "grad_norm": 1.205146545691501, + "learning_rate": 6.476005543732783e-06, + "loss": 0.5314, + "step": 1042 + }, + { + "epoch": 3.26938136256852, + "grad_norm": 1.1244830342386452, + "learning_rate": 6.45552579264013e-06, + "loss": 0.5061, + "step": 1043 + }, + { + "epoch": 3.2725137039937353, + "grad_norm": 1.1324469436671378, + "learning_rate": 6.4350630297562726e-06, + "loss": 0.5411, + "step": 1044 + }, + { + "epoch": 3.2756460454189504, + "grad_norm": 1.170084873645457, + "learning_rate": 6.414617353156605e-06, + "loss": 0.5256, + "step": 1045 + }, + { + "epoch": 3.278778386844166, + "grad_norm": 1.154539238112036, + "learning_rate": 6.394188860834626e-06, + "loss": 0.5353, + "step": 1046 + }, + { + "epoch": 3.2819107282693816, + "grad_norm": 1.145934214062239, + "learning_rate": 6.373777650701465e-06, + "loss": 0.4949, + "step": 1047 + }, + { + "epoch": 3.2850430696945967, + "grad_norm": 1.1758118112093967, + "learning_rate": 6.35338382058543e-06, + "loss": 0.5388, + "step": 1048 + }, + { + "epoch": 3.288175411119812, + "grad_norm": 1.1930083102260012, + "learning_rate": 6.333007468231521e-06, + "loss": 0.5357, + "step": 1049 + }, + { + "epoch": 3.2913077525450274, + "grad_norm": 1.2215776268642229, + "learning_rate": 6.312648691300975e-06, + "loss": 0.5159, + "step": 1050 + }, + { + "epoch": 3.294440093970243, + "grad_norm": 1.149266078166264, + "learning_rate": 6.292307587370794e-06, + "loss": 0.5329, + "step": 1051 + }, + { + "epoch": 3.297572435395458, + "grad_norm": 1.188650823733469, + "learning_rate": 6.271984253933266e-06, + "loss": 0.5607, + "step": 1052 + }, + { + "epoch": 3.3007047768206736, + "grad_norm": 1.1743831994156586, + "learning_rate": 6.2516787883955165e-06, + "loss": 0.5574, + "step": 1053 + }, + { + "epoch": 3.3038371182458888, + "grad_norm": 1.1970079452073716, + "learning_rate": 6.231391288079027e-06, + "loss": 0.5173, + "step": 1054 + }, + { + "epoch": 3.3069694596711043, + "grad_norm": 1.122279685545005, + "learning_rate": 6.211121850219175e-06, + "loss": 0.5361, + "step": 1055 + }, + { + "epoch": 3.3101018010963195, + "grad_norm": 1.225293554009653, + "learning_rate": 6.1908705719647735e-06, + "loss": 0.555, + "step": 1056 + }, + { + "epoch": 3.313234142521535, + "grad_norm": 1.1655884451182597, + "learning_rate": 6.170637550377587e-06, + "loss": 0.5321, + "step": 1057 + }, + { + "epoch": 3.31636648394675, + "grad_norm": 1.1850726593613186, + "learning_rate": 6.150422882431888e-06, + "loss": 0.5158, + "step": 1058 + }, + { + "epoch": 3.3194988253719657, + "grad_norm": 1.1741932650006341, + "learning_rate": 6.130226665013979e-06, + "loss": 0.5278, + "step": 1059 + }, + { + "epoch": 3.322631166797181, + "grad_norm": 1.1637151007879059, + "learning_rate": 6.110048994921735e-06, + "loss": 0.5147, + "step": 1060 + }, + { + "epoch": 3.3257635082223964, + "grad_norm": 1.19810414141727, + "learning_rate": 6.089889968864122e-06, + "loss": 0.5523, + "step": 1061 + }, + { + "epoch": 3.3288958496476115, + "grad_norm": 1.1747671263259907, + "learning_rate": 6.069749683460765e-06, + "loss": 0.5259, + "step": 1062 + }, + { + "epoch": 3.332028191072827, + "grad_norm": 1.189733533665569, + "learning_rate": 6.049628235241459e-06, + "loss": 0.5727, + "step": 1063 + }, + { + "epoch": 3.3351605324980422, + "grad_norm": 1.1907858250963501, + "learning_rate": 6.029525720645719e-06, + "loss": 0.5581, + "step": 1064 + }, + { + "epoch": 3.338292873923258, + "grad_norm": 1.1727785316101889, + "learning_rate": 6.009442236022307e-06, + "loss": 0.5527, + "step": 1065 + }, + { + "epoch": 3.341425215348473, + "grad_norm": 1.1473257646366246, + "learning_rate": 5.989377877628783e-06, + "loss": 0.5354, + "step": 1066 + }, + { + "epoch": 3.3445575567736885, + "grad_norm": 1.1904303579624806, + "learning_rate": 5.969332741631037e-06, + "loss": 0.5299, + "step": 1067 + }, + { + "epoch": 3.3476898981989036, + "grad_norm": 1.1467680391867652, + "learning_rate": 5.949306924102822e-06, + "loss": 0.5347, + "step": 1068 + }, + { + "epoch": 3.350822239624119, + "grad_norm": 1.164592432138905, + "learning_rate": 5.929300521025304e-06, + "loss": 0.5295, + "step": 1069 + }, + { + "epoch": 3.3539545810493343, + "grad_norm": 1.183763669072413, + "learning_rate": 5.9093136282866014e-06, + "loss": 0.5309, + "step": 1070 + }, + { + "epoch": 3.35708692247455, + "grad_norm": 1.1341809428622889, + "learning_rate": 5.889346341681319e-06, + "loss": 0.5057, + "step": 1071 + }, + { + "epoch": 3.360219263899765, + "grad_norm": 1.2114438092248463, + "learning_rate": 5.869398756910095e-06, + "loss": 0.5285, + "step": 1072 + }, + { + "epoch": 3.3633516053249806, + "grad_norm": 1.1899159606732028, + "learning_rate": 5.8494709695791364e-06, + "loss": 0.515, + "step": 1073 + }, + { + "epoch": 3.3664839467501957, + "grad_norm": 1.1585091379070802, + "learning_rate": 5.829563075199753e-06, + "loss": 0.5215, + "step": 1074 + }, + { + "epoch": 3.3696162881754113, + "grad_norm": 1.1554751170108555, + "learning_rate": 5.809675169187936e-06, + "loss": 0.5276, + "step": 1075 + }, + { + "epoch": 3.3727486296006264, + "grad_norm": 1.1531221192759264, + "learning_rate": 5.789807346863849e-06, + "loss": 0.5343, + "step": 1076 + }, + { + "epoch": 3.375880971025842, + "grad_norm": 1.2264179028180235, + "learning_rate": 5.769959703451413e-06, + "loss": 0.5331, + "step": 1077 + }, + { + "epoch": 3.379013312451057, + "grad_norm": 1.1909513224731456, + "learning_rate": 5.750132334077827e-06, + "loss": 0.5221, + "step": 1078 + }, + { + "epoch": 3.3821456538762726, + "grad_norm": 1.1410796381664094, + "learning_rate": 5.730325333773122e-06, + "loss": 0.5236, + "step": 1079 + }, + { + "epoch": 3.3852779953014878, + "grad_norm": 1.165393643113191, + "learning_rate": 5.710538797469707e-06, + "loss": 0.5363, + "step": 1080 + }, + { + "epoch": 3.3884103367267033, + "grad_norm": 1.109034591095622, + "learning_rate": 5.690772820001894e-06, + "loss": 0.5129, + "step": 1081 + }, + { + "epoch": 3.3915426781519185, + "grad_norm": 1.1479442023962463, + "learning_rate": 5.671027496105477e-06, + "loss": 0.5767, + "step": 1082 + }, + { + "epoch": 3.394675019577134, + "grad_norm": 1.1983194565401882, + "learning_rate": 5.651302920417252e-06, + "loss": 0.5233, + "step": 1083 + }, + { + "epoch": 3.397807361002349, + "grad_norm": 1.1247461153791098, + "learning_rate": 5.63159918747457e-06, + "loss": 0.531, + "step": 1084 + }, + { + "epoch": 3.4009397024275647, + "grad_norm": 1.1645414490351773, + "learning_rate": 5.611916391714887e-06, + "loss": 0.5356, + "step": 1085 + }, + { + "epoch": 3.40407204385278, + "grad_norm": 1.1955131839630717, + "learning_rate": 5.592254627475313e-06, + "loss": 0.5181, + "step": 1086 + }, + { + "epoch": 3.4072043852779954, + "grad_norm": 1.1669277156801512, + "learning_rate": 5.572613988992157e-06, + "loss": 0.5505, + "step": 1087 + }, + { + "epoch": 3.4103367267032105, + "grad_norm": 1.1969803225344355, + "learning_rate": 5.552994570400465e-06, + "loss": 0.5279, + "step": 1088 + }, + { + "epoch": 3.413469068128426, + "grad_norm": 1.149976867281222, + "learning_rate": 5.5333964657335894e-06, + "loss": 0.5176, + "step": 1089 + }, + { + "epoch": 3.4166014095536412, + "grad_norm": 1.1530538912791253, + "learning_rate": 5.513819768922723e-06, + "loss": 0.529, + "step": 1090 + }, + { + "epoch": 3.419733750978857, + "grad_norm": 1.1611367505386145, + "learning_rate": 5.49426457379646e-06, + "loss": 0.5045, + "step": 1091 + }, + { + "epoch": 3.422866092404072, + "grad_norm": 1.157629907018517, + "learning_rate": 5.474730974080327e-06, + "loss": 0.5384, + "step": 1092 + }, + { + "epoch": 3.4259984338292875, + "grad_norm": 1.1436433927685985, + "learning_rate": 5.455219063396356e-06, + "loss": 0.5164, + "step": 1093 + }, + { + "epoch": 3.4291307752545026, + "grad_norm": 1.1624946647803467, + "learning_rate": 5.4357289352626284e-06, + "loss": 0.518, + "step": 1094 + }, + { + "epoch": 3.432263116679718, + "grad_norm": 1.2059544843365895, + "learning_rate": 5.416260683092814e-06, + "loss": 0.5413, + "step": 1095 + }, + { + "epoch": 3.4353954581049333, + "grad_norm": 1.1848557963338198, + "learning_rate": 5.3968144001957445e-06, + "loss": 0.556, + "step": 1096 + }, + { + "epoch": 3.438527799530149, + "grad_norm": 1.2267248653375662, + "learning_rate": 5.377390179774951e-06, + "loss": 0.5735, + "step": 1097 + }, + { + "epoch": 3.441660140955364, + "grad_norm": 1.219382842483644, + "learning_rate": 5.357988114928221e-06, + "loss": 0.5224, + "step": 1098 + }, + { + "epoch": 3.4447924823805796, + "grad_norm": 1.2019925545507635, + "learning_rate": 5.338608298647161e-06, + "loss": 0.5188, + "step": 1099 + }, + { + "epoch": 3.4479248238057947, + "grad_norm": 1.2054098999261758, + "learning_rate": 5.319250823816732e-06, + "loss": 0.5603, + "step": 1100 + }, + { + "epoch": 3.4510571652310102, + "grad_norm": 1.149167269524483, + "learning_rate": 5.299915783214816e-06, + "loss": 0.5103, + "step": 1101 + }, + { + "epoch": 3.4541895066562254, + "grad_norm": 1.190224838500585, + "learning_rate": 5.28060326951178e-06, + "loss": 0.5445, + "step": 1102 + }, + { + "epoch": 3.457321848081441, + "grad_norm": 1.145339136200069, + "learning_rate": 5.2613133752700145e-06, + "loss": 0.5432, + "step": 1103 + }, + { + "epoch": 3.460454189506656, + "grad_norm": 1.1801121893769682, + "learning_rate": 5.242046192943503e-06, + "loss": 0.5321, + "step": 1104 + }, + { + "epoch": 3.4635865309318716, + "grad_norm": 1.196242124148136, + "learning_rate": 5.22280181487737e-06, + "loss": 0.5097, + "step": 1105 + }, + { + "epoch": 3.4667188723570868, + "grad_norm": 1.1558254840616595, + "learning_rate": 5.2035803333074435e-06, + "loss": 0.515, + "step": 1106 + }, + { + "epoch": 3.4698512137823023, + "grad_norm": 1.1436458411967738, + "learning_rate": 5.184381840359815e-06, + "loss": 0.5004, + "step": 1107 + }, + { + "epoch": 3.4729835552075174, + "grad_norm": 1.137825928298088, + "learning_rate": 5.1652064280503884e-06, + "loss": 0.5164, + "step": 1108 + }, + { + "epoch": 3.476115896632733, + "grad_norm": 1.1431285887634195, + "learning_rate": 5.146054188284443e-06, + "loss": 0.5324, + "step": 1109 + }, + { + "epoch": 3.479248238057948, + "grad_norm": 1.1848905862756827, + "learning_rate": 5.126925212856202e-06, + "loss": 0.541, + "step": 1110 + }, + { + "epoch": 3.4823805794831637, + "grad_norm": 1.15107617387345, + "learning_rate": 5.107819593448385e-06, + "loss": 0.5049, + "step": 1111 + }, + { + "epoch": 3.485512920908379, + "grad_norm": 1.1324253201763683, + "learning_rate": 5.088737421631767e-06, + "loss": 0.5263, + "step": 1112 + }, + { + "epoch": 3.4886452623335944, + "grad_norm": 1.1791937767344005, + "learning_rate": 5.069678788864742e-06, + "loss": 0.5301, + "step": 1113 + }, + { + "epoch": 3.4917776037588095, + "grad_norm": 1.177125639354513, + "learning_rate": 5.050643786492888e-06, + "loss": 0.5299, + "step": 1114 + }, + { + "epoch": 3.494909945184025, + "grad_norm": 1.12975187321893, + "learning_rate": 5.031632505748516e-06, + "loss": 0.5191, + "step": 1115 + }, + { + "epoch": 3.49804228660924, + "grad_norm": 1.1702301585282073, + "learning_rate": 5.012645037750254e-06, + "loss": 0.4911, + "step": 1116 + }, + { + "epoch": 3.501174628034456, + "grad_norm": 1.1898809996788224, + "learning_rate": 4.993681473502586e-06, + "loss": 0.5846, + "step": 1117 + }, + { + "epoch": 3.5043069694596714, + "grad_norm": 1.1772803798080353, + "learning_rate": 4.97474190389544e-06, + "loss": 0.5449, + "step": 1118 + }, + { + "epoch": 3.5074393108848865, + "grad_norm": 1.1458679793457238, + "learning_rate": 4.955826419703736e-06, + "loss": 0.527, + "step": 1119 + }, + { + "epoch": 3.5105716523101016, + "grad_norm": 1.197898331863354, + "learning_rate": 4.936935111586953e-06, + "loss": 0.5279, + "step": 1120 + }, + { + "epoch": 3.513703993735317, + "grad_norm": 1.1490696601995432, + "learning_rate": 4.918068070088708e-06, + "loss": 0.5479, + "step": 1121 + }, + { + "epoch": 3.5168363351605327, + "grad_norm": 1.1338863485542614, + "learning_rate": 4.899225385636293e-06, + "loss": 0.5143, + "step": 1122 + }, + { + "epoch": 3.519968676585748, + "grad_norm": 1.1180638940506065, + "learning_rate": 4.880407148540275e-06, + "loss": 0.5018, + "step": 1123 + }, + { + "epoch": 3.523101018010963, + "grad_norm": 1.1649197234392008, + "learning_rate": 4.8616134489940455e-06, + "loss": 0.5086, + "step": 1124 + }, + { + "epoch": 3.5262333594361785, + "grad_norm": 1.175823733278329, + "learning_rate": 4.8428443770733835e-06, + "loss": 0.5128, + "step": 1125 + }, + { + "epoch": 3.529365700861394, + "grad_norm": 1.1338414552899265, + "learning_rate": 4.824100022736036e-06, + "loss": 0.4854, + "step": 1126 + }, + { + "epoch": 3.5324980422866092, + "grad_norm": 1.208715108777811, + "learning_rate": 4.805380475821287e-06, + "loss": 0.562, + "step": 1127 + }, + { + "epoch": 3.5356303837118244, + "grad_norm": 1.2832853162474576, + "learning_rate": 4.786685826049508e-06, + "loss": 0.5227, + "step": 1128 + }, + { + "epoch": 3.53876272513704, + "grad_norm": 1.20463188289643, + "learning_rate": 4.768016163021754e-06, + "loss": 0.5343, + "step": 1129 + }, + { + "epoch": 3.5418950665622555, + "grad_norm": 1.1932857110540964, + "learning_rate": 4.749371576219317e-06, + "loss": 0.4933, + "step": 1130 + }, + { + "epoch": 3.5450274079874706, + "grad_norm": 1.1785435772368695, + "learning_rate": 4.730752155003303e-06, + "loss": 0.5262, + "step": 1131 + }, + { + "epoch": 3.5481597494126857, + "grad_norm": 1.1881352282816962, + "learning_rate": 4.7121579886142065e-06, + "loss": 0.5564, + "step": 1132 + }, + { + "epoch": 3.5512920908379013, + "grad_norm": 1.1608651788367013, + "learning_rate": 4.693589166171466e-06, + "loss": 0.5287, + "step": 1133 + }, + { + "epoch": 3.554424432263117, + "grad_norm": 1.2094527888449869, + "learning_rate": 4.675045776673067e-06, + "loss": 0.5463, + "step": 1134 + }, + { + "epoch": 3.557556773688332, + "grad_norm": 1.1971780880370575, + "learning_rate": 4.65652790899508e-06, + "loss": 0.5117, + "step": 1135 + }, + { + "epoch": 3.560689115113547, + "grad_norm": 1.1896379190718653, + "learning_rate": 4.638035651891266e-06, + "loss": 0.543, + "step": 1136 + }, + { + "epoch": 3.5638214565387627, + "grad_norm": 1.123251870010005, + "learning_rate": 4.619569093992636e-06, + "loss": 0.5263, + "step": 1137 + }, + { + "epoch": 3.5669537979639783, + "grad_norm": 1.1685045536154566, + "learning_rate": 4.601128323807023e-06, + "loss": 0.5341, + "step": 1138 + }, + { + "epoch": 3.5700861393891934, + "grad_norm": 1.1479704004071765, + "learning_rate": 4.582713429718667e-06, + "loss": 0.5503, + "step": 1139 + }, + { + "epoch": 3.5732184808144085, + "grad_norm": 1.138427168032927, + "learning_rate": 4.56432449998779e-06, + "loss": 0.492, + "step": 1140 + }, + { + "epoch": 3.576350822239624, + "grad_norm": 1.2372766348911877, + "learning_rate": 4.545961622750158e-06, + "loss": 0.5512, + "step": 1141 + }, + { + "epoch": 3.5794831636648397, + "grad_norm": 1.1774919589788904, + "learning_rate": 4.527624886016686e-06, + "loss": 0.5398, + "step": 1142 + }, + { + "epoch": 3.5826155050900548, + "grad_norm": 1.1693438413445578, + "learning_rate": 4.509314377672987e-06, + "loss": 0.5444, + "step": 1143 + }, + { + "epoch": 3.58574784651527, + "grad_norm": 1.206456317247258, + "learning_rate": 4.491030185478976e-06, + "loss": 0.5362, + "step": 1144 + }, + { + "epoch": 3.5888801879404855, + "grad_norm": 1.1483671630028152, + "learning_rate": 4.472772397068431e-06, + "loss": 0.5115, + "step": 1145 + }, + { + "epoch": 3.592012529365701, + "grad_norm": 1.2246520853364462, + "learning_rate": 4.454541099948585e-06, + "loss": 0.5305, + "step": 1146 + }, + { + "epoch": 3.595144870790916, + "grad_norm": 1.1831900976113294, + "learning_rate": 4.436336381499701e-06, + "loss": 0.5413, + "step": 1147 + }, + { + "epoch": 3.5982772122161313, + "grad_norm": 1.2043230578018518, + "learning_rate": 4.418158328974647e-06, + "loss": 0.5196, + "step": 1148 + }, + { + "epoch": 3.601409553641347, + "grad_norm": 1.1773163602914556, + "learning_rate": 4.40000702949849e-06, + "loss": 0.5001, + "step": 1149 + }, + { + "epoch": 3.6045418950665624, + "grad_norm": 1.1478622164586656, + "learning_rate": 4.381882570068079e-06, + "loss": 0.5303, + "step": 1150 + }, + { + "epoch": 3.6076742364917775, + "grad_norm": 1.191685305824786, + "learning_rate": 4.3637850375516085e-06, + "loss": 0.5137, + "step": 1151 + }, + { + "epoch": 3.610806577916993, + "grad_norm": 1.181732648263047, + "learning_rate": 4.345714518688226e-06, + "loss": 0.5246, + "step": 1152 + }, + { + "epoch": 3.6139389193422082, + "grad_norm": 1.1923764986284755, + "learning_rate": 4.327671100087605e-06, + "loss": 0.5327, + "step": 1153 + }, + { + "epoch": 3.617071260767424, + "grad_norm": 1.20809690015103, + "learning_rate": 4.3096548682295304e-06, + "loss": 0.5498, + "step": 1154 + }, + { + "epoch": 3.620203602192639, + "grad_norm": 1.1491784685338666, + "learning_rate": 4.291665909463477e-06, + "loss": 0.5041, + "step": 1155 + }, + { + "epoch": 3.6233359436178545, + "grad_norm": 1.1751280305249536, + "learning_rate": 4.273704310008213e-06, + "loss": 0.5229, + "step": 1156 + }, + { + "epoch": 3.6264682850430696, + "grad_norm": 1.1175615334495055, + "learning_rate": 4.255770155951374e-06, + "loss": 0.5215, + "step": 1157 + }, + { + "epoch": 3.629600626468285, + "grad_norm": 1.155032398758409, + "learning_rate": 4.237863533249057e-06, + "loss": 0.5344, + "step": 1158 + }, + { + "epoch": 3.6327329678935003, + "grad_norm": 1.1988336359099494, + "learning_rate": 4.21998452772539e-06, + "loss": 0.5579, + "step": 1159 + }, + { + "epoch": 3.635865309318716, + "grad_norm": 1.1698366279218717, + "learning_rate": 4.202133225072153e-06, + "loss": 0.5288, + "step": 1160 + }, + { + "epoch": 3.638997650743931, + "grad_norm": 1.1286869148862564, + "learning_rate": 4.1843097108483465e-06, + "loss": 0.5245, + "step": 1161 + }, + { + "epoch": 3.6421299921691466, + "grad_norm": 1.1861981743925876, + "learning_rate": 4.16651407047977e-06, + "loss": 0.5391, + "step": 1162 + }, + { + "epoch": 3.6452623335943617, + "grad_norm": 1.2269024042403185, + "learning_rate": 4.148746389258645e-06, + "loss": 0.5417, + "step": 1163 + }, + { + "epoch": 3.6483946750195773, + "grad_norm": 1.2098949537700534, + "learning_rate": 4.1310067523431795e-06, + "loss": 0.5187, + "step": 1164 + }, + { + "epoch": 3.6515270164447924, + "grad_norm": 1.2311237257384957, + "learning_rate": 4.113295244757171e-06, + "loss": 0.5112, + "step": 1165 + }, + { + "epoch": 3.654659357870008, + "grad_norm": 1.1381539159403147, + "learning_rate": 4.0956119513896005e-06, + "loss": 0.5065, + "step": 1166 + }, + { + "epoch": 3.657791699295223, + "grad_norm": 1.1621884247268668, + "learning_rate": 4.077956956994212e-06, + "loss": 0.511, + "step": 1167 + }, + { + "epoch": 3.6609240407204386, + "grad_norm": 1.168551137549956, + "learning_rate": 4.060330346189125e-06, + "loss": 0.5177, + "step": 1168 + }, + { + "epoch": 3.6640563821456538, + "grad_norm": 1.210641000943774, + "learning_rate": 4.042732203456413e-06, + "loss": 0.5197, + "step": 1169 + }, + { + "epoch": 3.6671887235708693, + "grad_norm": 1.1705201776756187, + "learning_rate": 4.0251626131417135e-06, + "loss": 0.5467, + "step": 1170 + }, + { + "epoch": 3.6703210649960845, + "grad_norm": 1.1690543008503043, + "learning_rate": 4.00762165945381e-06, + "loss": 0.5404, + "step": 1171 + }, + { + "epoch": 3.6734534064213, + "grad_norm": 1.1598971397306825, + "learning_rate": 3.990109426464237e-06, + "loss": 0.5178, + "step": 1172 + }, + { + "epoch": 3.676585747846515, + "grad_norm": 1.1616293085138196, + "learning_rate": 3.972625998106876e-06, + "loss": 0.5435, + "step": 1173 + }, + { + "epoch": 3.6797180892717307, + "grad_norm": 1.1407718103960551, + "learning_rate": 3.955171458177551e-06, + "loss": 0.5406, + "step": 1174 + }, + { + "epoch": 3.682850430696946, + "grad_norm": 1.1669175281894106, + "learning_rate": 3.937745890333623e-06, + "loss": 0.5251, + "step": 1175 + }, + { + "epoch": 3.6859827721221614, + "grad_norm": 1.1257651461718745, + "learning_rate": 3.9203493780935935e-06, + "loss": 0.5345, + "step": 1176 + }, + { + "epoch": 3.6891151135473765, + "grad_norm": 1.1540314221528836, + "learning_rate": 3.902982004836709e-06, + "loss": 0.5351, + "step": 1177 + }, + { + "epoch": 3.692247454972592, + "grad_norm": 1.1524838421819237, + "learning_rate": 3.885643853802557e-06, + "loss": 0.5488, + "step": 1178 + }, + { + "epoch": 3.6953797963978072, + "grad_norm": 1.188466802503505, + "learning_rate": 3.868335008090664e-06, + "loss": 0.5142, + "step": 1179 + }, + { + "epoch": 3.698512137823023, + "grad_norm": 1.138912921077403, + "learning_rate": 3.851055550660098e-06, + "loss": 0.5269, + "step": 1180 + }, + { + "epoch": 3.701644479248238, + "grad_norm": 1.1870550813670917, + "learning_rate": 3.8338055643290815e-06, + "loss": 0.527, + "step": 1181 + }, + { + "epoch": 3.7047768206734535, + "grad_norm": 1.1829270162712773, + "learning_rate": 3.8165851317745705e-06, + "loss": 0.526, + "step": 1182 + }, + { + "epoch": 3.7079091620986686, + "grad_norm": 1.2162157673701168, + "learning_rate": 3.799394335531891e-06, + "loss": 0.5644, + "step": 1183 + }, + { + "epoch": 3.711041503523884, + "grad_norm": 1.1789615458756362, + "learning_rate": 3.782233257994309e-06, + "loss": 0.5369, + "step": 1184 + }, + { + "epoch": 3.7141738449490993, + "grad_norm": 1.1676281601812897, + "learning_rate": 3.7651019814126656e-06, + "loss": 0.5243, + "step": 1185 + }, + { + "epoch": 3.717306186374315, + "grad_norm": 1.1571405119696987, + "learning_rate": 3.7480005878949645e-06, + "loss": 0.5584, + "step": 1186 + }, + { + "epoch": 3.72043852779953, + "grad_norm": 1.1512138914116103, + "learning_rate": 3.730929159405984e-06, + "loss": 0.5272, + "step": 1187 + }, + { + "epoch": 3.7235708692247456, + "grad_norm": 1.1290612876998436, + "learning_rate": 3.713887777766888e-06, + "loss": 0.5212, + "step": 1188 + }, + { + "epoch": 3.7267032106499607, + "grad_norm": 1.1615587465761752, + "learning_rate": 3.69687652465482e-06, + "loss": 0.5271, + "step": 1189 + }, + { + "epoch": 3.7298355520751763, + "grad_norm": 1.165644217808434, + "learning_rate": 3.679895481602529e-06, + "loss": 0.5268, + "step": 1190 + }, + { + "epoch": 3.7329678935003914, + "grad_norm": 1.1460726933346908, + "learning_rate": 3.6629447299979747e-06, + "loss": 0.5345, + "step": 1191 + }, + { + "epoch": 3.736100234925607, + "grad_norm": 1.1701380925076978, + "learning_rate": 3.64602435108392e-06, + "loss": 0.5482, + "step": 1192 + }, + { + "epoch": 3.7392325763508225, + "grad_norm": 1.1069644314839941, + "learning_rate": 3.629134425957567e-06, + "loss": 0.4975, + "step": 1193 + }, + { + "epoch": 3.7423649177760376, + "grad_norm": 1.1742604377155452, + "learning_rate": 3.612275035570155e-06, + "loss": 0.5382, + "step": 1194 + }, + { + "epoch": 3.7454972592012528, + "grad_norm": 1.1594884695567513, + "learning_rate": 3.595446260726576e-06, + "loss": 0.5276, + "step": 1195 + }, + { + "epoch": 3.7486296006264683, + "grad_norm": 1.2009203643761532, + "learning_rate": 3.578648182084975e-06, + "loss": 0.5377, + "step": 1196 + }, + { + "epoch": 3.751761942051684, + "grad_norm": 1.170884036670357, + "learning_rate": 3.561880880156384e-06, + "loss": 0.5576, + "step": 1197 + }, + { + "epoch": 3.754894283476899, + "grad_norm": 1.1959487905145205, + "learning_rate": 3.5451444353043263e-06, + "loss": 0.5373, + "step": 1198 + }, + { + "epoch": 3.758026624902114, + "grad_norm": 1.1796481257439184, + "learning_rate": 3.5284389277444296e-06, + "loss": 0.5313, + "step": 1199 + }, + { + "epoch": 3.7611589663273297, + "grad_norm": 1.2094185606216385, + "learning_rate": 3.511764437544036e-06, + "loss": 0.5162, + "step": 1200 + }, + { + "epoch": 3.7642913077525453, + "grad_norm": 1.1754904046598893, + "learning_rate": 3.4951210446218385e-06, + "loss": 0.5529, + "step": 1201 + }, + { + "epoch": 3.7674236491777604, + "grad_norm": 1.153840764724347, + "learning_rate": 3.478508828747471e-06, + "loss": 0.4987, + "step": 1202 + }, + { + "epoch": 3.7705559906029755, + "grad_norm": 1.1333187974311192, + "learning_rate": 3.4619278695411495e-06, + "loss": 0.5005, + "step": 1203 + }, + { + "epoch": 3.773688332028191, + "grad_norm": 1.1455993988251625, + "learning_rate": 3.445378246473279e-06, + "loss": 0.5115, + "step": 1204 + }, + { + "epoch": 3.7768206734534067, + "grad_norm": 1.1749646759832453, + "learning_rate": 3.4288600388640714e-06, + "loss": 0.5241, + "step": 1205 + }, + { + "epoch": 3.779953014878622, + "grad_norm": 1.159353327088895, + "learning_rate": 3.4123733258831727e-06, + "loss": 0.5221, + "step": 1206 + }, + { + "epoch": 3.783085356303837, + "grad_norm": 1.1770098803771174, + "learning_rate": 3.3959181865492798e-06, + "loss": 0.536, + "step": 1207 + }, + { + "epoch": 3.7862176977290525, + "grad_norm": 1.173049503757697, + "learning_rate": 3.3794946997297527e-06, + "loss": 0.5423, + "step": 1208 + }, + { + "epoch": 3.789350039154268, + "grad_norm": 1.2771010631863746, + "learning_rate": 3.3631029441402497e-06, + "loss": 0.5587, + "step": 1209 + }, + { + "epoch": 3.792482380579483, + "grad_norm": 1.193386115742984, + "learning_rate": 3.3467429983443477e-06, + "loss": 0.5277, + "step": 1210 + }, + { + "epoch": 3.7956147220046983, + "grad_norm": 1.1915512034829279, + "learning_rate": 3.330414940753164e-06, + "loss": 0.5583, + "step": 1211 + }, + { + "epoch": 3.798747063429914, + "grad_norm": 1.1896725640667438, + "learning_rate": 3.3141188496249755e-06, + "loss": 0.5141, + "step": 1212 + }, + { + "epoch": 3.8018794048551294, + "grad_norm": 1.1687306108915543, + "learning_rate": 3.2978548030648504e-06, + "loss": 0.5443, + "step": 1213 + }, + { + "epoch": 3.8050117462803446, + "grad_norm": 1.1857286795537076, + "learning_rate": 3.2816228790242723e-06, + "loss": 0.5472, + "step": 1214 + }, + { + "epoch": 3.8081440877055597, + "grad_norm": 1.138346768843636, + "learning_rate": 3.2654231553007665e-06, + "loss": 0.549, + "step": 1215 + }, + { + "epoch": 3.8112764291307752, + "grad_norm": 1.1444752892169308, + "learning_rate": 3.249255709537521e-06, + "loss": 0.522, + "step": 1216 + }, + { + "epoch": 3.814408770555991, + "grad_norm": 1.18144246751696, + "learning_rate": 3.233120619223021e-06, + "loss": 0.5303, + "step": 1217 + }, + { + "epoch": 3.817541111981206, + "grad_norm": 1.2096091122117285, + "learning_rate": 3.2170179616906818e-06, + "loss": 0.5207, + "step": 1218 + }, + { + "epoch": 3.820673453406421, + "grad_norm": 1.1530748784567288, + "learning_rate": 3.200947814118467e-06, + "loss": 0.5376, + "step": 1219 + }, + { + "epoch": 3.8238057948316366, + "grad_norm": 1.1215171003312445, + "learning_rate": 3.1849102535285283e-06, + "loss": 0.5082, + "step": 1220 + }, + { + "epoch": 3.826938136256852, + "grad_norm": 1.1703733058989874, + "learning_rate": 3.1689053567868278e-06, + "loss": 0.5337, + "step": 1221 + }, + { + "epoch": 3.8300704776820673, + "grad_norm": 1.130262675345898, + "learning_rate": 3.1529332006027813e-06, + "loss": 0.5404, + "step": 1222 + }, + { + "epoch": 3.8332028191072824, + "grad_norm": 1.1669511760006823, + "learning_rate": 3.1369938615288707e-06, + "loss": 0.5311, + "step": 1223 + }, + { + "epoch": 3.836335160532498, + "grad_norm": 1.1656001584283648, + "learning_rate": 3.1210874159603044e-06, + "loss": 0.5422, + "step": 1224 + }, + { + "epoch": 3.8394675019577136, + "grad_norm": 1.1747394216229974, + "learning_rate": 3.1052139401346216e-06, + "loss": 0.5132, + "step": 1225 + }, + { + "epoch": 3.8425998433829287, + "grad_norm": 1.1495579361244552, + "learning_rate": 3.089373510131354e-06, + "loss": 0.493, + "step": 1226 + }, + { + "epoch": 3.845732184808144, + "grad_norm": 1.180793820104102, + "learning_rate": 3.0735662018716437e-06, + "loss": 0.5291, + "step": 1227 + }, + { + "epoch": 3.8488645262333594, + "grad_norm": 1.124911552092485, + "learning_rate": 3.057792091117889e-06, + "loss": 0.5189, + "step": 1228 + }, + { + "epoch": 3.851996867658575, + "grad_norm": 1.120771756001303, + "learning_rate": 3.042051253473366e-06, + "loss": 0.4857, + "step": 1229 + }, + { + "epoch": 3.85512920908379, + "grad_norm": 1.1265839282142127, + "learning_rate": 3.026343764381887e-06, + "loss": 0.5188, + "step": 1230 + }, + { + "epoch": 3.858261550509005, + "grad_norm": 1.153903476616413, + "learning_rate": 3.0106696991274278e-06, + "loss": 0.518, + "step": 1231 + }, + { + "epoch": 3.861393891934221, + "grad_norm": 1.1653054593485324, + "learning_rate": 2.995029132833768e-06, + "loss": 0.5413, + "step": 1232 + }, + { + "epoch": 3.8645262333594363, + "grad_norm": 1.1219932990648924, + "learning_rate": 2.9794221404641233e-06, + "loss": 0.5584, + "step": 1233 + }, + { + "epoch": 3.8676585747846515, + "grad_norm": 1.1764250161654142, + "learning_rate": 2.9638487968208028e-06, + "loss": 0.5485, + "step": 1234 + }, + { + "epoch": 3.870790916209867, + "grad_norm": 1.1756895255385884, + "learning_rate": 2.9483091765448426e-06, + "loss": 0.5223, + "step": 1235 + }, + { + "epoch": 3.873923257635082, + "grad_norm": 1.1493491848055564, + "learning_rate": 2.9328033541156363e-06, + "loss": 0.5137, + "step": 1236 + }, + { + "epoch": 3.8770555990602977, + "grad_norm": 1.1435017702746313, + "learning_rate": 2.9173314038505986e-06, + "loss": 0.5404, + "step": 1237 + }, + { + "epoch": 3.880187940485513, + "grad_norm": 1.1617452479406913, + "learning_rate": 2.901893399904797e-06, + "loss": 0.5315, + "step": 1238 + }, + { + "epoch": 3.8833202819107284, + "grad_norm": 1.2030832962067164, + "learning_rate": 2.886489416270598e-06, + "loss": 0.5385, + "step": 1239 + }, + { + "epoch": 3.8864526233359435, + "grad_norm": 1.1328735216805788, + "learning_rate": 2.871119526777315e-06, + "loss": 0.5199, + "step": 1240 + }, + { + "epoch": 3.889584964761159, + "grad_norm": 1.1928234546236907, + "learning_rate": 2.855783805090846e-06, + "loss": 0.5258, + "step": 1241 + }, + { + "epoch": 3.8927173061863742, + "grad_norm": 1.208100129151956, + "learning_rate": 2.8404823247133373e-06, + "loss": 0.5168, + "step": 1242 + }, + { + "epoch": 3.89584964761159, + "grad_norm": 1.1414115708828445, + "learning_rate": 2.825215158982809e-06, + "loss": 0.5429, + "step": 1243 + }, + { + "epoch": 3.898981989036805, + "grad_norm": 1.184617005441351, + "learning_rate": 2.8099823810728255e-06, + "loss": 0.5604, + "step": 1244 + }, + { + "epoch": 3.9021143304620205, + "grad_norm": 1.1378204967249268, + "learning_rate": 2.7947840639921308e-06, + "loss": 0.528, + "step": 1245 + }, + { + "epoch": 3.9052466718872356, + "grad_norm": 1.1876757504170918, + "learning_rate": 2.779620280584303e-06, + "loss": 0.5226, + "step": 1246 + }, + { + "epoch": 3.908379013312451, + "grad_norm": 1.1809603854780806, + "learning_rate": 2.764491103527406e-06, + "loss": 0.5266, + "step": 1247 + }, + { + "epoch": 3.9115113547376663, + "grad_norm": 1.1873512461852518, + "learning_rate": 2.7493966053336397e-06, + "loss": 0.5414, + "step": 1248 + }, + { + "epoch": 3.914643696162882, + "grad_norm": 1.1504208913330314, + "learning_rate": 2.734336858348987e-06, + "loss": 0.5283, + "step": 1249 + }, + { + "epoch": 3.917776037588097, + "grad_norm": 1.143048240337088, + "learning_rate": 2.7193119347528842e-06, + "loss": 0.5337, + "step": 1250 + }, + { + "epoch": 3.9209083790133126, + "grad_norm": 1.1689363408230222, + "learning_rate": 2.7043219065578497e-06, + "loss": 0.557, + "step": 1251 + }, + { + "epoch": 3.9240407204385277, + "grad_norm": 1.1555945926450477, + "learning_rate": 2.6893668456091627e-06, + "loss": 0.5256, + "step": 1252 + }, + { + "epoch": 3.9271730618637433, + "grad_norm": 1.158032059169674, + "learning_rate": 2.674446823584507e-06, + "loss": 0.5011, + "step": 1253 + }, + { + "epoch": 3.9303054032889584, + "grad_norm": 1.163791390523606, + "learning_rate": 2.6595619119936267e-06, + "loss": 0.5324, + "step": 1254 + }, + { + "epoch": 3.933437744714174, + "grad_norm": 1.1473109388161766, + "learning_rate": 2.644712182177992e-06, + "loss": 0.5454, + "step": 1255 + }, + { + "epoch": 3.936570086139389, + "grad_norm": 1.1721160935749964, + "learning_rate": 2.62989770531044e-06, + "loss": 0.5144, + "step": 1256 + }, + { + "epoch": 3.9397024275646046, + "grad_norm": 1.2026978621361013, + "learning_rate": 2.6151185523948577e-06, + "loss": 0.5587, + "step": 1257 + }, + { + "epoch": 3.9428347689898198, + "grad_norm": 1.161111467446485, + "learning_rate": 2.6003747942658243e-06, + "loss": 0.5172, + "step": 1258 + }, + { + "epoch": 3.9459671104150353, + "grad_norm": 1.1895161788905066, + "learning_rate": 2.5856665015882686e-06, + "loss": 0.5647, + "step": 1259 + }, + { + "epoch": 3.9490994518402505, + "grad_norm": 1.1667391962309854, + "learning_rate": 2.570993744857151e-06, + "loss": 0.539, + "step": 1260 + }, + { + "epoch": 3.952231793265466, + "grad_norm": 1.1990408248065403, + "learning_rate": 2.5563565943971027e-06, + "loss": 0.5308, + "step": 1261 + }, + { + "epoch": 3.955364134690681, + "grad_norm": 1.177576874471793, + "learning_rate": 2.541755120362108e-06, + "loss": 0.5302, + "step": 1262 + }, + { + "epoch": 3.9584964761158967, + "grad_norm": 1.1329715387380235, + "learning_rate": 2.5271893927351464e-06, + "loss": 0.529, + "step": 1263 + }, + { + "epoch": 3.961628817541112, + "grad_norm": 1.20780770381308, + "learning_rate": 2.5126594813278792e-06, + "loss": 0.5298, + "step": 1264 + }, + { + "epoch": 3.9647611589663274, + "grad_norm": 1.2074087314834707, + "learning_rate": 2.4981654557803026e-06, + "loss": 0.5553, + "step": 1265 + }, + { + "epoch": 3.9678935003915425, + "grad_norm": 1.176718451434302, + "learning_rate": 2.4837073855604186e-06, + "loss": 0.5376, + "step": 1266 + }, + { + "epoch": 3.971025841816758, + "grad_norm": 1.1582592387697839, + "learning_rate": 2.469285339963892e-06, + "loss": 0.5016, + "step": 1267 + }, + { + "epoch": 3.9741581832419732, + "grad_norm": 1.126175949855287, + "learning_rate": 2.4548993881137327e-06, + "loss": 0.5164, + "step": 1268 + }, + { + "epoch": 3.977290524667189, + "grad_norm": 1.182910505553956, + "learning_rate": 2.4405495989599627e-06, + "loss": 0.5606, + "step": 1269 + }, + { + "epoch": 3.980422866092404, + "grad_norm": 1.1661378582361217, + "learning_rate": 2.4262360412792663e-06, + "loss": 0.552, + "step": 1270 + }, + { + "epoch": 3.9835552075176195, + "grad_norm": 1.1721964398123115, + "learning_rate": 2.4119587836746885e-06, + "loss": 0.5481, + "step": 1271 + }, + { + "epoch": 3.9866875489428346, + "grad_norm": 1.1861569431208108, + "learning_rate": 2.3977178945752877e-06, + "loss": 0.5361, + "step": 1272 + }, + { + "epoch": 3.98981989036805, + "grad_norm": 1.142933675959546, + "learning_rate": 2.383513442235812e-06, + "loss": 0.5216, + "step": 1273 + }, + { + "epoch": 3.9929522317932653, + "grad_norm": 1.173762534357803, + "learning_rate": 2.3693454947363782e-06, + "loss": 0.5545, + "step": 1274 + }, + { + "epoch": 3.996084573218481, + "grad_norm": 1.1831215530299448, + "learning_rate": 2.35521411998213e-06, + "loss": 0.5296, + "step": 1275 + }, + { + "epoch": 3.999216914643696, + "grad_norm": 1.1124833568408279, + "learning_rate": 2.341119385702929e-06, + "loss": 0.5065, + "step": 1276 + }, + { + "epoch": 4.0, + "grad_norm": 1.1124833568408279, + "learning_rate": 2.3270613594530235e-06, + "loss": 0.5165, + "step": 1277 + }, + { + "epoch": 4.003132341425215, + "grad_norm": 2.6425185615357405, + "learning_rate": 2.3130401086107247e-06, + "loss": 0.4391, + "step": 1278 + }, + { + "epoch": 4.006264682850431, + "grad_norm": 1.8968989656450992, + "learning_rate": 2.2990557003780833e-06, + "loss": 0.4202, + "step": 1279 + }, + { + "epoch": 4.009397024275646, + "grad_norm": 1.8142320294338747, + "learning_rate": 2.2851082017805704e-06, + "loss": 0.4009, + "step": 1280 + }, + { + "epoch": 4.012529365700861, + "grad_norm": 1.7165806912562718, + "learning_rate": 2.27119767966675e-06, + "loss": 0.405, + "step": 1281 + }, + { + "epoch": 4.0156617071260765, + "grad_norm": 1.5924208377321079, + "learning_rate": 2.2573242007079697e-06, + "loss": 0.421, + "step": 1282 + }, + { + "epoch": 4.0187940485512925, + "grad_norm": 1.5038694947047038, + "learning_rate": 2.2434878313980235e-06, + "loss": 0.4002, + "step": 1283 + }, + { + "epoch": 4.021926389976508, + "grad_norm": 1.340573966307308, + "learning_rate": 2.2296886380528494e-06, + "loss": 0.4272, + "step": 1284 + }, + { + "epoch": 4.025058731401723, + "grad_norm": 1.2987422109641342, + "learning_rate": 2.215926686810206e-06, + "loss": 0.4349, + "step": 1285 + }, + { + "epoch": 4.028191072826938, + "grad_norm": 1.252719539838883, + "learning_rate": 2.202202043629356e-06, + "loss": 0.3898, + "step": 1286 + }, + { + "epoch": 4.031323414252154, + "grad_norm": 1.327752575517584, + "learning_rate": 2.1885147742907465e-06, + "loss": 0.4175, + "step": 1287 + }, + { + "epoch": 4.034455755677369, + "grad_norm": 1.4402839116286432, + "learning_rate": 2.1748649443957004e-06, + "loss": 0.4137, + "step": 1288 + }, + { + "epoch": 4.037588097102584, + "grad_norm": 1.5655168444654703, + "learning_rate": 2.1612526193660975e-06, + "loss": 0.4051, + "step": 1289 + }, + { + "epoch": 4.040720438527799, + "grad_norm": 1.6161366724042316, + "learning_rate": 2.1476778644440553e-06, + "loss": 0.4028, + "step": 1290 + }, + { + "epoch": 4.043852779953015, + "grad_norm": 1.5596875540342134, + "learning_rate": 2.134140744691633e-06, + "loss": 0.3997, + "step": 1291 + }, + { + "epoch": 4.04698512137823, + "grad_norm": 1.4460862400789365, + "learning_rate": 2.1206413249904967e-06, + "loss": 0.4125, + "step": 1292 + }, + { + "epoch": 4.0501174628034455, + "grad_norm": 1.3882745890795896, + "learning_rate": 2.1071796700416334e-06, + "loss": 0.3972, + "step": 1293 + }, + { + "epoch": 4.053249804228661, + "grad_norm": 1.4012869387398328, + "learning_rate": 2.093755844365021e-06, + "loss": 0.4155, + "step": 1294 + }, + { + "epoch": 4.056382145653877, + "grad_norm": 1.3519015902940699, + "learning_rate": 2.0803699122993293e-06, + "loss": 0.4178, + "step": 1295 + }, + { + "epoch": 4.059514487079092, + "grad_norm": 1.3602338471577913, + "learning_rate": 2.067021938001611e-06, + "loss": 0.4193, + "step": 1296 + }, + { + "epoch": 4.062646828504307, + "grad_norm": 1.2707959489599356, + "learning_rate": 2.0537119854469845e-06, + "loss": 0.4235, + "step": 1297 + }, + { + "epoch": 4.065779169929522, + "grad_norm": 1.2304315457883275, + "learning_rate": 2.0404401184283408e-06, + "loss": 0.4281, + "step": 1298 + }, + { + "epoch": 4.068911511354738, + "grad_norm": 1.3126596083493967, + "learning_rate": 2.0272064005560354e-06, + "loss": 0.4116, + "step": 1299 + }, + { + "epoch": 4.072043852779953, + "grad_norm": 1.2164141170274532, + "learning_rate": 2.01401089525757e-06, + "loss": 0.4128, + "step": 1300 + }, + { + "epoch": 4.075176194205168, + "grad_norm": 1.2157628287632507, + "learning_rate": 2.000853665777305e-06, + "loss": 0.3966, + "step": 1301 + }, + { + "epoch": 4.078308535630383, + "grad_norm": 1.1768433581291058, + "learning_rate": 1.987734775176151e-06, + "loss": 0.3773, + "step": 1302 + }, + { + "epoch": 4.081440877055599, + "grad_norm": 1.195358771674673, + "learning_rate": 1.974654286331258e-06, + "loss": 0.4021, + "step": 1303 + }, + { + "epoch": 4.084573218480815, + "grad_norm": 1.2360866223981186, + "learning_rate": 1.961612261935727e-06, + "loss": 0.3914, + "step": 1304 + }, + { + "epoch": 4.08770555990603, + "grad_norm": 1.2186686913113904, + "learning_rate": 1.9486087644983056e-06, + "loss": 0.3896, + "step": 1305 + }, + { + "epoch": 4.090837901331245, + "grad_norm": 1.2346935215427521, + "learning_rate": 1.9356438563430822e-06, + "loss": 0.4038, + "step": 1306 + }, + { + "epoch": 4.093970242756461, + "grad_norm": 1.198362834738886, + "learning_rate": 1.9227175996091952e-06, + "loss": 0.4009, + "step": 1307 + }, + { + "epoch": 4.097102584181676, + "grad_norm": 1.1668330606150041, + "learning_rate": 1.9098300562505266e-06, + "loss": 0.3942, + "step": 1308 + }, + { + "epoch": 4.100234925606891, + "grad_norm": 1.2063732731647043, + "learning_rate": 1.8969812880354167e-06, + "loss": 0.394, + "step": 1309 + }, + { + "epoch": 4.103367267032106, + "grad_norm": 1.2113132171130907, + "learning_rate": 1.8841713565463548e-06, + "loss": 0.3945, + "step": 1310 + }, + { + "epoch": 4.106499608457322, + "grad_norm": 1.2247544654973348, + "learning_rate": 1.871400323179695e-06, + "loss": 0.4096, + "step": 1311 + }, + { + "epoch": 4.109631949882537, + "grad_norm": 1.2881131176691998, + "learning_rate": 1.858668249145359e-06, + "loss": 0.4071, + "step": 1312 + }, + { + "epoch": 4.1127642913077525, + "grad_norm": 1.2380244490730665, + "learning_rate": 1.8459751954665372e-06, + "loss": 0.3887, + "step": 1313 + }, + { + "epoch": 4.115896632732968, + "grad_norm": 1.2457742852610907, + "learning_rate": 1.8333212229794039e-06, + "loss": 0.3938, + "step": 1314 + }, + { + "epoch": 4.119028974158184, + "grad_norm": 1.299735360583674, + "learning_rate": 1.820706392332824e-06, + "loss": 0.4278, + "step": 1315 + }, + { + "epoch": 4.122161315583399, + "grad_norm": 1.22006539661709, + "learning_rate": 1.8081307639880519e-06, + "loss": 0.3969, + "step": 1316 + }, + { + "epoch": 4.125293657008614, + "grad_norm": 1.2043704461002585, + "learning_rate": 1.7955943982184608e-06, + "loss": 0.3772, + "step": 1317 + }, + { + "epoch": 4.128425998433829, + "grad_norm": 1.1954512153531711, + "learning_rate": 1.7830973551092346e-06, + "loss": 0.4232, + "step": 1318 + }, + { + "epoch": 4.131558339859045, + "grad_norm": 1.2198506026583384, + "learning_rate": 1.7706396945570947e-06, + "loss": 0.3996, + "step": 1319 + }, + { + "epoch": 4.13469068128426, + "grad_norm": 1.2218183019573885, + "learning_rate": 1.7582214762700057e-06, + "loss": 0.428, + "step": 1320 + }, + { + "epoch": 4.137823022709475, + "grad_norm": 1.1976267387815713, + "learning_rate": 1.7458427597668893e-06, + "loss": 0.4062, + "step": 1321 + }, + { + "epoch": 4.14095536413469, + "grad_norm": 1.1697967121996906, + "learning_rate": 1.7335036043773402e-06, + "loss": 0.3889, + "step": 1322 + }, + { + "epoch": 4.144087705559906, + "grad_norm": 1.1924932065029525, + "learning_rate": 1.7212040692413468e-06, + "loss": 0.4245, + "step": 1323 + }, + { + "epoch": 4.1472200469851215, + "grad_norm": 1.2147406012062079, + "learning_rate": 1.7089442133089928e-06, + "loss": 0.3927, + "step": 1324 + }, + { + "epoch": 4.150352388410337, + "grad_norm": 1.1788944142485256, + "learning_rate": 1.6967240953401954e-06, + "loss": 0.3862, + "step": 1325 + }, + { + "epoch": 4.153484729835552, + "grad_norm": 1.1947326262463431, + "learning_rate": 1.6845437739044012e-06, + "loss": 0.3828, + "step": 1326 + }, + { + "epoch": 4.156617071260768, + "grad_norm": 1.1173824303336672, + "learning_rate": 1.672403307380328e-06, + "loss": 0.3977, + "step": 1327 + }, + { + "epoch": 4.159749412685983, + "grad_norm": 1.1753324405533268, + "learning_rate": 1.6603027539556682e-06, + "loss": 0.3866, + "step": 1328 + }, + { + "epoch": 4.162881754111198, + "grad_norm": 1.1729189872697827, + "learning_rate": 1.6482421716268215e-06, + "loss": 0.421, + "step": 1329 + }, + { + "epoch": 4.166014095536413, + "grad_norm": 1.2345796879070534, + "learning_rate": 1.6362216181986002e-06, + "loss": 0.3713, + "step": 1330 + }, + { + "epoch": 4.169146436961629, + "grad_norm": 1.155602810779086, + "learning_rate": 1.6242411512839728e-06, + "loss": 0.387, + "step": 1331 + }, + { + "epoch": 4.172278778386844, + "grad_norm": 1.1762864409104852, + "learning_rate": 1.6123008283037778e-06, + "loss": 0.3991, + "step": 1332 + }, + { + "epoch": 4.175411119812059, + "grad_norm": 1.2221930726179149, + "learning_rate": 1.6004007064864468e-06, + "loss": 0.3585, + "step": 1333 + }, + { + "epoch": 4.1785434612372745, + "grad_norm": 1.1528755586082111, + "learning_rate": 1.5885408428677296e-06, + "loss": 0.4074, + "step": 1334 + }, + { + "epoch": 4.1816758026624905, + "grad_norm": 1.1990119644346569, + "learning_rate": 1.5767212942904275e-06, + "loss": 0.4252, + "step": 1335 + }, + { + "epoch": 4.184808144087706, + "grad_norm": 1.190374191471458, + "learning_rate": 1.564942117404119e-06, + "loss": 0.4137, + "step": 1336 + }, + { + "epoch": 4.187940485512921, + "grad_norm": 1.2490135566571798, + "learning_rate": 1.5532033686648785e-06, + "loss": 0.3961, + "step": 1337 + }, + { + "epoch": 4.191072826938136, + "grad_norm": 1.211936868558367, + "learning_rate": 1.5415051043350194e-06, + "loss": 0.3968, + "step": 1338 + }, + { + "epoch": 4.194205168363352, + "grad_norm": 1.2268059684742754, + "learning_rate": 1.5298473804828195e-06, + "loss": 0.3992, + "step": 1339 + }, + { + "epoch": 4.197337509788567, + "grad_norm": 1.2449496248919925, + "learning_rate": 1.518230252982248e-06, + "loss": 0.3926, + "step": 1340 + }, + { + "epoch": 4.200469851213782, + "grad_norm": 1.192327150884329, + "learning_rate": 1.5066537775127044e-06, + "loss": 0.3974, + "step": 1341 + }, + { + "epoch": 4.203602192638997, + "grad_norm": 1.235268675446175, + "learning_rate": 1.4951180095587426e-06, + "loss": 0.3963, + "step": 1342 + }, + { + "epoch": 4.206734534064213, + "grad_norm": 1.2044322613986833, + "learning_rate": 1.4836230044098164e-06, + "loss": 0.4214, + "step": 1343 + }, + { + "epoch": 4.209866875489428, + "grad_norm": 1.233078464789091, + "learning_rate": 1.4721688171600047e-06, + "loss": 0.3956, + "step": 1344 + }, + { + "epoch": 4.2129992169146435, + "grad_norm": 1.2257391290015276, + "learning_rate": 1.4607555027077524e-06, + "loss": 0.424, + "step": 1345 + }, + { + "epoch": 4.216131558339859, + "grad_norm": 1.239469422660483, + "learning_rate": 1.4493831157556094e-06, + "loss": 0.3933, + "step": 1346 + }, + { + "epoch": 4.219263899765075, + "grad_norm": 1.1537827041020894, + "learning_rate": 1.4380517108099623e-06, + "loss": 0.37, + "step": 1347 + }, + { + "epoch": 4.22239624119029, + "grad_norm": 1.2110428750506566, + "learning_rate": 1.426761342180777e-06, + "loss": 0.4159, + "step": 1348 + }, + { + "epoch": 4.225528582615505, + "grad_norm": 1.2034706739426686, + "learning_rate": 1.4155120639813392e-06, + "loss": 0.4007, + "step": 1349 + }, + { + "epoch": 4.22866092404072, + "grad_norm": 1.2096667555894145, + "learning_rate": 1.4043039301279904e-06, + "loss": 0.4168, + "step": 1350 + }, + { + "epoch": 4.231793265465936, + "grad_norm": 1.2136387061332223, + "learning_rate": 1.393136994339872e-06, + "loss": 0.3891, + "step": 1351 + }, + { + "epoch": 4.234925606891151, + "grad_norm": 1.2279471606708432, + "learning_rate": 1.3820113101386712e-06, + "loss": 0.398, + "step": 1352 + }, + { + "epoch": 4.238057948316366, + "grad_norm": 1.1712817094482313, + "learning_rate": 1.3709269308483619e-06, + "loss": 0.3974, + "step": 1353 + }, + { + "epoch": 4.241190289741581, + "grad_norm": 1.2244805625115447, + "learning_rate": 1.359883909594948e-06, + "loss": 0.4101, + "step": 1354 + }, + { + "epoch": 4.244322631166797, + "grad_norm": 1.2055456362307584, + "learning_rate": 1.348882299306209e-06, + "loss": 0.4189, + "step": 1355 + }, + { + "epoch": 4.2474549725920125, + "grad_norm": 1.2472551970779924, + "learning_rate": 1.3379221527114484e-06, + "loss": 0.3974, + "step": 1356 + }, + { + "epoch": 4.250587314017228, + "grad_norm": 1.1928935135771526, + "learning_rate": 1.327003522341237e-06, + "loss": 0.4061, + "step": 1357 + }, + { + "epoch": 4.253719655442444, + "grad_norm": 1.2297720660775902, + "learning_rate": 1.3161264605271662e-06, + "loss": 0.4037, + "step": 1358 + }, + { + "epoch": 4.256851996867659, + "grad_norm": 1.2100253932220972, + "learning_rate": 1.3052910194015922e-06, + "loss": 0.3932, + "step": 1359 + }, + { + "epoch": 4.259984338292874, + "grad_norm": 1.189046936366124, + "learning_rate": 1.2944972508973908e-06, + "loss": 0.3878, + "step": 1360 + }, + { + "epoch": 4.263116679718089, + "grad_norm": 1.1920673345461743, + "learning_rate": 1.2837452067477062e-06, + "loss": 0.3915, + "step": 1361 + }, + { + "epoch": 4.266249021143304, + "grad_norm": 1.2118917776405405, + "learning_rate": 1.273034938485702e-06, + "loss": 0.3939, + "step": 1362 + }, + { + "epoch": 4.26938136256852, + "grad_norm": 1.2126398847084519, + "learning_rate": 1.2623664974443195e-06, + "loss": 0.3988, + "step": 1363 + }, + { + "epoch": 4.272513703993735, + "grad_norm": 1.1928692981682214, + "learning_rate": 1.2517399347560165e-06, + "loss": 0.3966, + "step": 1364 + }, + { + "epoch": 4.27564604541895, + "grad_norm": 1.1977505153553263, + "learning_rate": 1.2411553013525457e-06, + "loss": 0.4051, + "step": 1365 + }, + { + "epoch": 4.2787783868441664, + "grad_norm": 1.222427847128266, + "learning_rate": 1.230612647964694e-06, + "loss": 0.4103, + "step": 1366 + }, + { + "epoch": 4.281910728269382, + "grad_norm": 1.200297992267844, + "learning_rate": 1.2201120251220378e-06, + "loss": 0.4016, + "step": 1367 + }, + { + "epoch": 4.285043069694597, + "grad_norm": 1.216728301148234, + "learning_rate": 1.209653483152714e-06, + "loss": 0.4014, + "step": 1368 + }, + { + "epoch": 4.288175411119812, + "grad_norm": 1.224540090293941, + "learning_rate": 1.1992370721831693e-06, + "loss": 0.3974, + "step": 1369 + }, + { + "epoch": 4.291307752545027, + "grad_norm": 1.1950041179139188, + "learning_rate": 1.1888628421379222e-06, + "loss": 0.4002, + "step": 1370 + }, + { + "epoch": 4.294440093970243, + "grad_norm": 1.190070377224994, + "learning_rate": 1.1785308427393182e-06, + "loss": 0.3964, + "step": 1371 + }, + { + "epoch": 4.297572435395458, + "grad_norm": 1.2008002010734806, + "learning_rate": 1.1682411235073032e-06, + "loss": 0.4076, + "step": 1372 + }, + { + "epoch": 4.300704776820673, + "grad_norm": 1.2132858157877429, + "learning_rate": 1.1579937337591774e-06, + "loss": 0.4015, + "step": 1373 + }, + { + "epoch": 4.303837118245889, + "grad_norm": 1.1826916321444987, + "learning_rate": 1.1477887226093642e-06, + "loss": 0.4171, + "step": 1374 + }, + { + "epoch": 4.306969459671104, + "grad_norm": 1.2089546478448403, + "learning_rate": 1.1376261389691633e-06, + "loss": 0.4146, + "step": 1375 + }, + { + "epoch": 4.3101018010963195, + "grad_norm": 1.206358523522593, + "learning_rate": 1.1275060315465346e-06, + "loss": 0.3872, + "step": 1376 + }, + { + "epoch": 4.313234142521535, + "grad_norm": 1.179621274726758, + "learning_rate": 1.1174284488458509e-06, + "loss": 0.4172, + "step": 1377 + }, + { + "epoch": 4.316366483946751, + "grad_norm": 1.268118607500806, + "learning_rate": 1.1073934391676666e-06, + "loss": 0.4213, + "step": 1378 + }, + { + "epoch": 4.319498825371966, + "grad_norm": 1.2431330503496503, + "learning_rate": 1.097401050608493e-06, + "loss": 0.3729, + "step": 1379 + }, + { + "epoch": 4.322631166797181, + "grad_norm": 1.2057856092166312, + "learning_rate": 1.0874513310605628e-06, + "loss": 0.388, + "step": 1380 + }, + { + "epoch": 4.325763508222396, + "grad_norm": 1.1984630219610473, + "learning_rate": 1.0775443282116017e-06, + "loss": 0.3919, + "step": 1381 + }, + { + "epoch": 4.328895849647612, + "grad_norm": 1.1721834023779714, + "learning_rate": 1.0676800895446016e-06, + "loss": 0.4033, + "step": 1382 + }, + { + "epoch": 4.332028191072827, + "grad_norm": 1.2483769893731758, + "learning_rate": 1.0578586623375863e-06, + "loss": 0.3826, + "step": 1383 + }, + { + "epoch": 4.335160532498042, + "grad_norm": 1.250417832052342, + "learning_rate": 1.048080093663395e-06, + "loss": 0.4275, + "step": 1384 + }, + { + "epoch": 4.338292873923257, + "grad_norm": 1.2565058613082094, + "learning_rate": 1.0383444303894453e-06, + "loss": 0.3947, + "step": 1385 + }, + { + "epoch": 4.341425215348473, + "grad_norm": 1.2252255070005802, + "learning_rate": 1.0286517191775215e-06, + "loss": 0.4061, + "step": 1386 + }, + { + "epoch": 4.3445575567736885, + "grad_norm": 1.235428710435152, + "learning_rate": 1.0190020064835404e-06, + "loss": 0.4089, + "step": 1387 + }, + { + "epoch": 4.347689898198904, + "grad_norm": 1.2195624521585995, + "learning_rate": 1.009395338557333e-06, + "loss": 0.3892, + "step": 1388 + }, + { + "epoch": 4.350822239624119, + "grad_norm": 1.2421525738212085, + "learning_rate": 9.998317614424246e-07, + "loss": 0.4151, + "step": 1389 + }, + { + "epoch": 4.353954581049335, + "grad_norm": 1.2269216736656088, + "learning_rate": 9.903113209758098e-07, + "loss": 0.4203, + "step": 1390 + }, + { + "epoch": 4.35708692247455, + "grad_norm": 1.2627277845413256, + "learning_rate": 9.80834062787731e-07, + "loss": 0.4115, + "step": 1391 + }, + { + "epoch": 4.360219263899765, + "grad_norm": 1.2175493590777189, + "learning_rate": 9.714000323014705e-07, + "loss": 0.3879, + "step": 1392 + }, + { + "epoch": 4.36335160532498, + "grad_norm": 1.1697012465299366, + "learning_rate": 9.62009274733121e-07, + "loss": 0.3977, + "step": 1393 + }, + { + "epoch": 4.366483946750196, + "grad_norm": 1.217611380413171, + "learning_rate": 9.526618350913752e-07, + "loss": 0.3774, + "step": 1394 + }, + { + "epoch": 4.369616288175411, + "grad_norm": 1.1843080929354863, + "learning_rate": 9.43357758177309e-07, + "loss": 0.39, + "step": 1395 + }, + { + "epoch": 4.372748629600626, + "grad_norm": 1.1973637151971002, + "learning_rate": 9.34097088584166e-07, + "loss": 0.4149, + "step": 1396 + }, + { + "epoch": 4.3758809710258415, + "grad_norm": 1.1880735619823894, + "learning_rate": 9.248798706971462e-07, + "loss": 0.3826, + "step": 1397 + }, + { + "epoch": 4.3790133124510575, + "grad_norm": 1.2411864789666913, + "learning_rate": 9.15706148693184e-07, + "loss": 0.4115, + "step": 1398 + }, + { + "epoch": 4.382145653876273, + "grad_norm": 1.2197612576036618, + "learning_rate": 9.065759665407514e-07, + "loss": 0.4113, + "step": 1399 + }, + { + "epoch": 4.385277995301488, + "grad_norm": 1.255965107907724, + "learning_rate": 8.974893679996388e-07, + "loss": 0.3768, + "step": 1400 + }, + { + "epoch": 4.388410336726703, + "grad_norm": 1.1777066026687946, + "learning_rate": 8.884463966207391e-07, + "loss": 0.3983, + "step": 1401 + }, + { + "epoch": 4.391542678151919, + "grad_norm": 1.1768660233012396, + "learning_rate": 8.794470957458523e-07, + "loss": 0.403, + "step": 1402 + }, + { + "epoch": 4.394675019577134, + "grad_norm": 1.2598498922043773, + "learning_rate": 8.704915085074728e-07, + "loss": 0.4133, + "step": 1403 + }, + { + "epoch": 4.397807361002349, + "grad_norm": 1.1732711385614463, + "learning_rate": 8.615796778285735e-07, + "loss": 0.3724, + "step": 1404 + }, + { + "epoch": 4.400939702427564, + "grad_norm": 1.167025579090733, + "learning_rate": 8.527116464224128e-07, + "loss": 0.4062, + "step": 1405 + }, + { + "epoch": 4.40407204385278, + "grad_norm": 1.2023162019645415, + "learning_rate": 8.438874567923261e-07, + "loss": 0.3917, + "step": 1406 + }, + { + "epoch": 4.407204385277995, + "grad_norm": 1.1810461949196573, + "learning_rate": 8.351071512315167e-07, + "loss": 0.4052, + "step": 1407 + }, + { + "epoch": 4.4103367267032105, + "grad_norm": 1.202253524025422, + "learning_rate": 8.263707718228641e-07, + "loss": 0.3979, + "step": 1408 + }, + { + "epoch": 4.413469068128426, + "grad_norm": 1.1576984628276288, + "learning_rate": 8.17678360438704e-07, + "loss": 0.401, + "step": 1409 + }, + { + "epoch": 4.416601409553642, + "grad_norm": 1.2291103726343342, + "learning_rate": 8.090299587406514e-07, + "loss": 0.4177, + "step": 1410 + }, + { + "epoch": 4.419733750978857, + "grad_norm": 1.2466018705955504, + "learning_rate": 8.004256081793782e-07, + "loss": 0.4247, + "step": 1411 + }, + { + "epoch": 4.422866092404072, + "grad_norm": 1.24245531665182, + "learning_rate": 7.918653499944318e-07, + "loss": 0.3814, + "step": 1412 + }, + { + "epoch": 4.425998433829287, + "grad_norm": 1.2207227340577, + "learning_rate": 7.833492252140284e-07, + "loss": 0.4179, + "step": 1413 + }, + { + "epoch": 4.429130775254503, + "grad_norm": 1.2374658186692722, + "learning_rate": 7.748772746548572e-07, + "loss": 0.438, + "step": 1414 + }, + { + "epoch": 4.432263116679718, + "grad_norm": 1.2445922732186314, + "learning_rate": 7.664495389218884e-07, + "loss": 0.4164, + "step": 1415 + }, + { + "epoch": 4.435395458104933, + "grad_norm": 1.2048318810786247, + "learning_rate": 7.580660584081755e-07, + "loss": 0.4114, + "step": 1416 + }, + { + "epoch": 4.438527799530148, + "grad_norm": 1.2092315947423604, + "learning_rate": 7.497268732946605e-07, + "loss": 0.3972, + "step": 1417 + }, + { + "epoch": 4.441660140955364, + "grad_norm": 1.1684399668507264, + "learning_rate": 7.414320235499839e-07, + "loss": 0.4007, + "step": 1418 + }, + { + "epoch": 4.44479248238058, + "grad_norm": 1.2018031522146513, + "learning_rate": 7.331815489302941e-07, + "loss": 0.3926, + "step": 1419 + }, + { + "epoch": 4.447924823805795, + "grad_norm": 1.2377323452479039, + "learning_rate": 7.249754889790539e-07, + "loss": 0.4024, + "step": 1420 + }, + { + "epoch": 4.45105716523101, + "grad_norm": 1.2372044961613387, + "learning_rate": 7.168138830268534e-07, + "loss": 0.4153, + "step": 1421 + }, + { + "epoch": 4.454189506656226, + "grad_norm": 1.2367482818132942, + "learning_rate": 7.086967701912195e-07, + "loss": 0.4145, + "step": 1422 + }, + { + "epoch": 4.457321848081441, + "grad_norm": 1.1720988594585238, + "learning_rate": 7.006241893764298e-07, + "loss": 0.3933, + "step": 1423 + }, + { + "epoch": 4.460454189506656, + "grad_norm": 1.1906519908299524, + "learning_rate": 6.925961792733271e-07, + "loss": 0.4213, + "step": 1424 + }, + { + "epoch": 4.463586530931871, + "grad_norm": 1.1861914037615011, + "learning_rate": 6.846127783591294e-07, + "loss": 0.374, + "step": 1425 + }, + { + "epoch": 4.466718872357087, + "grad_norm": 1.2151391090817059, + "learning_rate": 6.76674024897247e-07, + "loss": 0.3958, + "step": 1426 + }, + { + "epoch": 4.469851213782302, + "grad_norm": 1.1906210035368663, + "learning_rate": 6.687799569371079e-07, + "loss": 0.3997, + "step": 1427 + }, + { + "epoch": 4.4729835552075174, + "grad_norm": 1.2698252501752594, + "learning_rate": 6.609306123139614e-07, + "loss": 0.4152, + "step": 1428 + }, + { + "epoch": 4.476115896632733, + "grad_norm": 1.239086528624892, + "learning_rate": 6.531260286487073e-07, + "loss": 0.4244, + "step": 1429 + }, + { + "epoch": 4.479248238057949, + "grad_norm": 1.239389095459042, + "learning_rate": 6.453662433477137e-07, + "loss": 0.3967, + "step": 1430 + }, + { + "epoch": 4.482380579483164, + "grad_norm": 1.2078463523567677, + "learning_rate": 6.37651293602628e-07, + "loss": 0.3998, + "step": 1431 + }, + { + "epoch": 4.485512920908379, + "grad_norm": 1.2405207380196077, + "learning_rate": 6.299812163902152e-07, + "loss": 0.4079, + "step": 1432 + }, + { + "epoch": 4.488645262333594, + "grad_norm": 1.1965967124161982, + "learning_rate": 6.22356048472168e-07, + "loss": 0.3947, + "step": 1433 + }, + { + "epoch": 4.49177760375881, + "grad_norm": 1.2365962276051872, + "learning_rate": 6.147758263949322e-07, + "loss": 0.4145, + "step": 1434 + }, + { + "epoch": 4.494909945184025, + "grad_norm": 1.2320724240611538, + "learning_rate": 6.072405864895403e-07, + "loss": 0.3989, + "step": 1435 + }, + { + "epoch": 4.49804228660924, + "grad_norm": 1.2278614714663911, + "learning_rate": 5.997503648714254e-07, + "loss": 0.4087, + "step": 1436 + }, + { + "epoch": 4.501174628034455, + "grad_norm": 1.2459676825726136, + "learning_rate": 5.923051974402594e-07, + "loss": 0.4076, + "step": 1437 + }, + { + "epoch": 4.504306969459671, + "grad_norm": 1.2458668198276721, + "learning_rate": 5.849051198797662e-07, + "loss": 0.4248, + "step": 1438 + }, + { + "epoch": 4.5074393108848865, + "grad_norm": 1.209767689172734, + "learning_rate": 5.775501676575702e-07, + "loss": 0.4014, + "step": 1439 + }, + { + "epoch": 4.510571652310102, + "grad_norm": 1.2285591384985166, + "learning_rate": 5.702403760250086e-07, + "loss": 0.4147, + "step": 1440 + }, + { + "epoch": 4.513703993735318, + "grad_norm": 1.1718999655815265, + "learning_rate": 5.629757800169732e-07, + "loss": 0.4044, + "step": 1441 + }, + { + "epoch": 4.516836335160533, + "grad_norm": 1.217216752100384, + "learning_rate": 5.557564144517369e-07, + "loss": 0.4091, + "step": 1442 + }, + { + "epoch": 4.519968676585748, + "grad_norm": 1.1838737229775769, + "learning_rate": 5.485823139307911e-07, + "loss": 0.4034, + "step": 1443 + }, + { + "epoch": 4.523101018010963, + "grad_norm": 1.1673948755376424, + "learning_rate": 5.414535128386766e-07, + "loss": 0.3661, + "step": 1444 + }, + { + "epoch": 4.526233359436178, + "grad_norm": 1.1757487458448712, + "learning_rate": 5.343700453428169e-07, + "loss": 0.3971, + "step": 1445 + }, + { + "epoch": 4.529365700861394, + "grad_norm": 1.2191061942291306, + "learning_rate": 5.273319453933634e-07, + "loss": 0.3885, + "step": 1446 + }, + { + "epoch": 4.532498042286609, + "grad_norm": 1.1814168865664434, + "learning_rate": 5.203392467230229e-07, + "loss": 0.3888, + "step": 1447 + }, + { + "epoch": 4.535630383711824, + "grad_norm": 1.180338891284529, + "learning_rate": 5.133919828468992e-07, + "loss": 0.4031, + "step": 1448 + }, + { + "epoch": 4.53876272513704, + "grad_norm": 1.2002998136423375, + "learning_rate": 5.064901870623362e-07, + "loss": 0.4124, + "step": 1449 + }, + { + "epoch": 4.5418950665622555, + "grad_norm": 1.225998648068942, + "learning_rate": 4.996338924487509e-07, + "loss": 0.4079, + "step": 1450 + }, + { + "epoch": 4.545027407987471, + "grad_norm": 1.232921807498376, + "learning_rate": 4.928231318674836e-07, + "loss": 0.4288, + "step": 1451 + }, + { + "epoch": 4.548159749412686, + "grad_norm": 1.2108344873262966, + "learning_rate": 4.860579379616303e-07, + "loss": 0.3752, + "step": 1452 + }, + { + "epoch": 4.551292090837901, + "grad_norm": 1.1982395456743722, + "learning_rate": 4.793383431558962e-07, + "loss": 0.3956, + "step": 1453 + }, + { + "epoch": 4.554424432263117, + "grad_norm": 1.2245255923948302, + "learning_rate": 4.7266437965643455e-07, + "loss": 0.39, + "step": 1454 + }, + { + "epoch": 4.557556773688332, + "grad_norm": 1.2237018326288942, + "learning_rate": 4.660360794506946e-07, + "loss": 0.4064, + "step": 1455 + }, + { + "epoch": 4.560689115113547, + "grad_norm": 1.203953237163164, + "learning_rate": 4.594534743072654e-07, + "loss": 0.4169, + "step": 1456 + }, + { + "epoch": 4.563821456538763, + "grad_norm": 1.2087498003259758, + "learning_rate": 4.5291659577572753e-07, + "loss": 0.3832, + "step": 1457 + }, + { + "epoch": 4.566953797963978, + "grad_norm": 1.1942569748553682, + "learning_rate": 4.4642547518649824e-07, + "loss": 0.3776, + "step": 1458 + }, + { + "epoch": 4.570086139389193, + "grad_norm": 1.1879804731198786, + "learning_rate": 4.399801436506812e-07, + "loss": 0.3873, + "step": 1459 + }, + { + "epoch": 4.5732184808144085, + "grad_norm": 1.2110866417839938, + "learning_rate": 4.335806320599234e-07, + "loss": 0.399, + "step": 1460 + }, + { + "epoch": 4.576350822239624, + "grad_norm": 1.2296410077584767, + "learning_rate": 4.272269710862587e-07, + "loss": 0.3992, + "step": 1461 + }, + { + "epoch": 4.57948316366484, + "grad_norm": 1.2328738318463752, + "learning_rate": 4.209191911819688e-07, + "loss": 0.4196, + "step": 1462 + }, + { + "epoch": 4.582615505090055, + "grad_norm": 1.1895802067659933, + "learning_rate": 4.1465732257943037e-07, + "loss": 0.4496, + "step": 1463 + }, + { + "epoch": 4.58574784651527, + "grad_norm": 1.2283635780345825, + "learning_rate": 4.084413952909738e-07, + "loss": 0.4111, + "step": 1464 + }, + { + "epoch": 4.588880187940486, + "grad_norm": 1.2164839835142356, + "learning_rate": 4.022714391087379e-07, + "loss": 0.381, + "step": 1465 + }, + { + "epoch": 4.592012529365701, + "grad_norm": 1.207591372814288, + "learning_rate": 3.9614748360452984e-07, + "loss": 0.4174, + "step": 1466 + }, + { + "epoch": 4.595144870790916, + "grad_norm": 1.2361007397271178, + "learning_rate": 3.9006955812967983e-07, + "loss": 0.4093, + "step": 1467 + }, + { + "epoch": 4.598277212216131, + "grad_norm": 1.2148169973114509, + "learning_rate": 3.840376918149025e-07, + "loss": 0.4096, + "step": 1468 + }, + { + "epoch": 4.601409553641347, + "grad_norm": 1.262330409419549, + "learning_rate": 3.780519135701566e-07, + "loss": 0.4104, + "step": 1469 + }, + { + "epoch": 4.604541895066562, + "grad_norm": 1.2330283137163904, + "learning_rate": 3.7211225208450777e-07, + "loss": 0.4192, + "step": 1470 + }, + { + "epoch": 4.6076742364917775, + "grad_norm": 1.2671239020786587, + "learning_rate": 3.6621873582599055e-07, + "loss": 0.4054, + "step": 1471 + }, + { + "epoch": 4.610806577916993, + "grad_norm": 1.2333107036927184, + "learning_rate": 3.603713930414676e-07, + "loss": 0.4063, + "step": 1472 + }, + { + "epoch": 4.613938919342209, + "grad_norm": 1.202856883096194, + "learning_rate": 3.5457025175650084e-07, + "loss": 0.4148, + "step": 1473 + }, + { + "epoch": 4.617071260767424, + "grad_norm": 1.1747823598797158, + "learning_rate": 3.488153397752159e-07, + "loss": 0.3891, + "step": 1474 + }, + { + "epoch": 4.620203602192639, + "grad_norm": 1.1992072519105093, + "learning_rate": 3.431066846801634e-07, + "loss": 0.4267, + "step": 1475 + }, + { + "epoch": 4.623335943617854, + "grad_norm": 1.273775818849566, + "learning_rate": 3.374443138321937e-07, + "loss": 0.3985, + "step": 1476 + }, + { + "epoch": 4.62646828504307, + "grad_norm": 1.2299835011094196, + "learning_rate": 3.3182825437031994e-07, + "loss": 0.4089, + "step": 1477 + }, + { + "epoch": 4.629600626468285, + "grad_norm": 1.2034138009185376, + "learning_rate": 3.262585332115964e-07, + "loss": 0.4043, + "step": 1478 + }, + { + "epoch": 4.6327329678935, + "grad_norm": 1.209424726509648, + "learning_rate": 3.207351770509759e-07, + "loss": 0.4147, + "step": 1479 + }, + { + "epoch": 4.635865309318715, + "grad_norm": 1.1736004823028015, + "learning_rate": 3.152582123611958e-07, + "loss": 0.3687, + "step": 1480 + }, + { + "epoch": 4.6389976507439314, + "grad_norm": 1.2024178154556568, + "learning_rate": 3.098276653926413e-07, + "loss": 0.4049, + "step": 1481 + }, + { + "epoch": 4.642129992169147, + "grad_norm": 1.235968122606252, + "learning_rate": 3.0444356217322893e-07, + "loss": 0.4122, + "step": 1482 + }, + { + "epoch": 4.645262333594362, + "grad_norm": 1.275106262365568, + "learning_rate": 2.9910592850826983e-07, + "loss": 0.4086, + "step": 1483 + }, + { + "epoch": 4.648394675019577, + "grad_norm": 1.2510666555833756, + "learning_rate": 2.9381478998035563e-07, + "loss": 0.4055, + "step": 1484 + }, + { + "epoch": 4.651527016444793, + "grad_norm": 1.2274467661857713, + "learning_rate": 2.8857017194923174e-07, + "loss": 0.4034, + "step": 1485 + }, + { + "epoch": 4.654659357870008, + "grad_norm": 1.1853349236972364, + "learning_rate": 2.833720995516798e-07, + "loss": 0.4195, + "step": 1486 + }, + { + "epoch": 4.657791699295223, + "grad_norm": 1.1854238027175328, + "learning_rate": 2.7822059770138966e-07, + "loss": 0.4074, + "step": 1487 + }, + { + "epoch": 4.660924040720438, + "grad_norm": 1.2436312130461868, + "learning_rate": 2.731156910888477e-07, + "loss": 0.4033, + "step": 1488 + }, + { + "epoch": 4.664056382145654, + "grad_norm": 1.2428796817271779, + "learning_rate": 2.6805740418121453e-07, + "loss": 0.4191, + "step": 1489 + }, + { + "epoch": 4.667188723570869, + "grad_norm": 1.2427783356951405, + "learning_rate": 2.6304576122221035e-07, + "loss": 0.3755, + "step": 1490 + }, + { + "epoch": 4.6703210649960845, + "grad_norm": 1.1881132036518127, + "learning_rate": 2.580807862319912e-07, + "loss": 0.3891, + "step": 1491 + }, + { + "epoch": 4.6734534064213, + "grad_norm": 1.2442335114182685, + "learning_rate": 2.5316250300704726e-07, + "loss": 0.4154, + "step": 1492 + }, + { + "epoch": 4.676585747846516, + "grad_norm": 1.2053182135162017, + "learning_rate": 2.482909351200735e-07, + "loss": 0.3979, + "step": 1493 + }, + { + "epoch": 4.679718089271731, + "grad_norm": 1.1770238326901394, + "learning_rate": 2.434661059198695e-07, + "loss": 0.3948, + "step": 1494 + }, + { + "epoch": 4.682850430696946, + "grad_norm": 1.218619191692209, + "learning_rate": 2.386880385312218e-07, + "loss": 0.371, + "step": 1495 + }, + { + "epoch": 4.685982772122161, + "grad_norm": 1.1637502072533377, + "learning_rate": 2.339567558547906e-07, + "loss": 0.3917, + "step": 1496 + }, + { + "epoch": 4.689115113547377, + "grad_norm": 1.1735101557117311, + "learning_rate": 2.292722805670078e-07, + "loss": 0.4077, + "step": 1497 + }, + { + "epoch": 4.692247454972592, + "grad_norm": 1.2417399018426394, + "learning_rate": 2.2463463511995908e-07, + "loss": 0.4048, + "step": 1498 + }, + { + "epoch": 4.695379796397807, + "grad_norm": 1.1728907713133705, + "learning_rate": 2.200438417412809e-07, + "loss": 0.3891, + "step": 1499 + }, + { + "epoch": 4.698512137823022, + "grad_norm": 1.2016608119778363, + "learning_rate": 2.1549992243405814e-07, + "loss": 0.3882, + "step": 1500 + }, + { + "epoch": 4.701644479248238, + "grad_norm": 1.206612395808125, + "learning_rate": 2.1100289897670877e-07, + "loss": 0.4196, + "step": 1501 + }, + { + "epoch": 4.7047768206734535, + "grad_norm": 1.2493616742742244, + "learning_rate": 2.0655279292288832e-07, + "loss": 0.3967, + "step": 1502 + }, + { + "epoch": 4.707909162098669, + "grad_norm": 1.2226247159036543, + "learning_rate": 2.021496256013833e-07, + "loss": 0.3989, + "step": 1503 + }, + { + "epoch": 4.711041503523884, + "grad_norm": 1.1794232427283857, + "learning_rate": 1.9779341811600795e-07, + "loss": 0.3988, + "step": 1504 + }, + { + "epoch": 4.7141738449491, + "grad_norm": 1.2359058049155338, + "learning_rate": 1.9348419134550322e-07, + "loss": 0.4213, + "step": 1505 + }, + { + "epoch": 4.717306186374315, + "grad_norm": 1.2239539416305099, + "learning_rate": 1.8922196594344023e-07, + "loss": 0.3922, + "step": 1506 + }, + { + "epoch": 4.72043852779953, + "grad_norm": 1.2127862487844647, + "learning_rate": 1.8500676233811466e-07, + "loss": 0.3771, + "step": 1507 + }, + { + "epoch": 4.723570869224745, + "grad_norm": 1.198843093205888, + "learning_rate": 1.8083860073245806e-07, + "loss": 0.4181, + "step": 1508 + }, + { + "epoch": 4.726703210649961, + "grad_norm": 1.2252062247390472, + "learning_rate": 1.7671750110392903e-07, + "loss": 0.403, + "step": 1509 + }, + { + "epoch": 4.729835552075176, + "grad_norm": 1.23819671302605, + "learning_rate": 1.7264348320442992e-07, + "loss": 0.4092, + "step": 1510 + }, + { + "epoch": 4.732967893500391, + "grad_norm": 1.2113129084206005, + "learning_rate": 1.6861656656020464e-07, + "loss": 0.3774, + "step": 1511 + }, + { + "epoch": 4.7361002349256065, + "grad_norm": 1.1845710084452867, + "learning_rate": 1.6463677047174553e-07, + "loss": 0.3841, + "step": 1512 + }, + { + "epoch": 4.7392325763508225, + "grad_norm": 1.1746238547350345, + "learning_rate": 1.6070411401370335e-07, + "loss": 0.3939, + "step": 1513 + }, + { + "epoch": 4.742364917776038, + "grad_norm": 1.212119626132927, + "learning_rate": 1.568186160347951e-07, + "loss": 0.3838, + "step": 1514 + }, + { + "epoch": 4.745497259201253, + "grad_norm": 1.2178905084917646, + "learning_rate": 1.5298029515771195e-07, + "loss": 0.4132, + "step": 1515 + }, + { + "epoch": 4.748629600626469, + "grad_norm": 1.2150279012900946, + "learning_rate": 1.4918916977903264e-07, + "loss": 0.3986, + "step": 1516 + }, + { + "epoch": 4.751761942051684, + "grad_norm": 1.2367095696943584, + "learning_rate": 1.4544525806913234e-07, + "loss": 0.3979, + "step": 1517 + }, + { + "epoch": 4.754894283476899, + "grad_norm": 1.2377643398696656, + "learning_rate": 1.4174857797209951e-07, + "loss": 0.4018, + "step": 1518 + }, + { + "epoch": 4.758026624902114, + "grad_norm": 1.2167876153278898, + "learning_rate": 1.3809914720564478e-07, + "loss": 0.3906, + "step": 1519 + }, + { + "epoch": 4.761158966327329, + "grad_norm": 1.2446471369617926, + "learning_rate": 1.344969832610199e-07, + "loss": 0.3756, + "step": 1520 + }, + { + "epoch": 4.764291307752545, + "grad_norm": 1.1893090822752832, + "learning_rate": 1.3094210340293456e-07, + "loss": 0.4165, + "step": 1521 + }, + { + "epoch": 4.76742364917776, + "grad_norm": 1.2320387940829778, + "learning_rate": 1.274345246694708e-07, + "loss": 0.3913, + "step": 1522 + }, + { + "epoch": 4.7705559906029755, + "grad_norm": 1.216546020584182, + "learning_rate": 1.2397426387200096e-07, + "loss": 0.3959, + "step": 1523 + }, + { + "epoch": 4.7736883320281915, + "grad_norm": 1.2176186473887705, + "learning_rate": 1.205613375951098e-07, + "loss": 0.4247, + "step": 1524 + }, + { + "epoch": 4.776820673453407, + "grad_norm": 1.2283180997381093, + "learning_rate": 1.1719576219651585e-07, + "loss": 0.418, + "step": 1525 + }, + { + "epoch": 4.779953014878622, + "grad_norm": 1.214550392378903, + "learning_rate": 1.1387755380698584e-07, + "loss": 0.3819, + "step": 1526 + }, + { + "epoch": 4.783085356303837, + "grad_norm": 1.2324975421906008, + "learning_rate": 1.1060672833026697e-07, + "loss": 0.4278, + "step": 1527 + }, + { + "epoch": 4.786217697729052, + "grad_norm": 1.238934876307768, + "learning_rate": 1.0738330144300479e-07, + "loss": 0.425, + "step": 1528 + }, + { + "epoch": 4.789350039154268, + "grad_norm": 1.181886288925798, + "learning_rate": 1.0420728859466988e-07, + "loss": 0.4209, + "step": 1529 + }, + { + "epoch": 4.792482380579483, + "grad_norm": 1.1925253947318342, + "learning_rate": 1.0107870500748351e-07, + "loss": 0.3773, + "step": 1530 + }, + { + "epoch": 4.795614722004698, + "grad_norm": 1.2509864343113049, + "learning_rate": 9.799756567634433e-08, + "loss": 0.3853, + "step": 1531 + }, + { + "epoch": 4.798747063429914, + "grad_norm": 1.184682423863237, + "learning_rate": 9.496388536875623e-08, + "loss": 0.417, + "step": 1532 + }, + { + "epoch": 4.801879404855129, + "grad_norm": 1.2246944998945095, + "learning_rate": 9.197767862475837e-08, + "loss": 0.3874, + "step": 1533 + }, + { + "epoch": 4.8050117462803446, + "grad_norm": 1.214520286942335, + "learning_rate": 8.903895975685528e-08, + "loss": 0.4037, + "step": 1534 + }, + { + "epoch": 4.80814408770556, + "grad_norm": 1.2224353069074514, + "learning_rate": 8.614774284994797e-08, + "loss": 0.4099, + "step": 1535 + }, + { + "epoch": 4.811276429130775, + "grad_norm": 1.2068932961719712, + "learning_rate": 8.33040417612685e-08, + "loss": 0.3891, + "step": 1536 + }, + { + "epoch": 4.814408770555991, + "grad_norm": 1.1952441796807236, + "learning_rate": 8.05078701203088e-08, + "loss": 0.3899, + "step": 1537 + }, + { + "epoch": 4.817541111981206, + "grad_norm": 1.226741893844113, + "learning_rate": 7.7759241328762e-08, + "loss": 0.4111, + "step": 1538 + }, + { + "epoch": 4.820673453406421, + "grad_norm": 1.1808940728419945, + "learning_rate": 7.505816856045012e-08, + "loss": 0.3943, + "step": 1539 + }, + { + "epoch": 4.823805794831637, + "grad_norm": 1.203115331813816, + "learning_rate": 7.24046647612675e-08, + "loss": 0.4127, + "step": 1540 + }, + { + "epoch": 4.826938136256852, + "grad_norm": 1.1865631093267626, + "learning_rate": 6.979874264911756e-08, + "loss": 0.4058, + "step": 1541 + }, + { + "epoch": 4.830070477682067, + "grad_norm": 1.185030215141837, + "learning_rate": 6.724041471384835e-08, + "loss": 0.4124, + "step": 1542 + }, + { + "epoch": 4.8332028191072824, + "grad_norm": 1.204462154369856, + "learning_rate": 6.472969321719702e-08, + "loss": 0.4326, + "step": 1543 + }, + { + "epoch": 4.836335160532498, + "grad_norm": 1.2203256822089201, + "learning_rate": 6.226659019272997e-08, + "loss": 0.3917, + "step": 1544 + }, + { + "epoch": 4.839467501957714, + "grad_norm": 1.2180754919299253, + "learning_rate": 5.985111744578165e-08, + "loss": 0.4013, + "step": 1545 + }, + { + "epoch": 4.842599843382929, + "grad_norm": 1.18445065771052, + "learning_rate": 5.748328655340141e-08, + "loss": 0.3912, + "step": 1546 + }, + { + "epoch": 4.845732184808144, + "grad_norm": 1.1965989755685764, + "learning_rate": 5.516310886429899e-08, + "loss": 0.4041, + "step": 1547 + }, + { + "epoch": 4.84886452623336, + "grad_norm": 1.193838779871222, + "learning_rate": 5.2890595498787944e-08, + "loss": 0.4006, + "step": 1548 + }, + { + "epoch": 4.851996867658575, + "grad_norm": 1.1686189895653558, + "learning_rate": 5.0665757348732355e-08, + "loss": 0.4082, + "step": 1549 + }, + { + "epoch": 4.85512920908379, + "grad_norm": 1.2058648478611114, + "learning_rate": 4.848860507749353e-08, + "loss": 0.3881, + "step": 1550 + }, + { + "epoch": 4.858261550509005, + "grad_norm": 1.1637318481244103, + "learning_rate": 4.635914911988448e-08, + "loss": 0.3942, + "step": 1551 + }, + { + "epoch": 4.86139389193422, + "grad_norm": 1.2165645886224812, + "learning_rate": 4.427739968211331e-08, + "loss": 0.4035, + "step": 1552 + }, + { + "epoch": 4.864526233359436, + "grad_norm": 1.2040433534371484, + "learning_rate": 4.2243366741735457e-08, + "loss": 0.3868, + "step": 1553 + }, + { + "epoch": 4.8676585747846515, + "grad_norm": 1.2369435812923857, + "learning_rate": 4.025706004760932e-08, + "loss": 0.3988, + "step": 1554 + }, + { + "epoch": 4.870790916209867, + "grad_norm": 1.2232422405149723, + "learning_rate": 3.831848911984959e-08, + "loss": 0.3968, + "step": 1555 + }, + { + "epoch": 4.873923257635083, + "grad_norm": 1.1711459811080631, + "learning_rate": 3.642766324977509e-08, + "loss": 0.3923, + "step": 1556 + }, + { + "epoch": 4.877055599060298, + "grad_norm": 1.2174185734188516, + "learning_rate": 3.4584591499872146e-08, + "loss": 0.4029, + "step": 1557 + }, + { + "epoch": 4.880187940485513, + "grad_norm": 1.1614173175371763, + "learning_rate": 3.278928270374459e-08, + "loss": 0.3949, + "step": 1558 + }, + { + "epoch": 4.883320281910728, + "grad_norm": 1.2280723648931007, + "learning_rate": 3.104174546607941e-08, + "loss": 0.4199, + "step": 1559 + }, + { + "epoch": 4.886452623335944, + "grad_norm": 1.209148320604221, + "learning_rate": 2.9341988162595593e-08, + "loss": 0.4192, + "step": 1560 + }, + { + "epoch": 4.889584964761159, + "grad_norm": 1.2065698817517443, + "learning_rate": 2.7690018940011996e-08, + "loss": 0.3655, + "step": 1561 + }, + { + "epoch": 4.892717306186374, + "grad_norm": 1.1832085868374411, + "learning_rate": 2.6085845716004034e-08, + "loss": 0.4102, + "step": 1562 + }, + { + "epoch": 4.895849647611589, + "grad_norm": 1.223809874203652, + "learning_rate": 2.452947617916701e-08, + "loss": 0.3826, + "step": 1563 + }, + { + "epoch": 4.898981989036805, + "grad_norm": 1.1564398514888772, + "learning_rate": 2.3020917788981744e-08, + "loss": 0.399, + "step": 1564 + }, + { + "epoch": 4.9021143304620205, + "grad_norm": 1.2322802342710437, + "learning_rate": 2.156017777577346e-08, + "loss": 0.4385, + "step": 1565 + }, + { + "epoch": 4.905246671887236, + "grad_norm": 1.2327395751465036, + "learning_rate": 2.0147263140680717e-08, + "loss": 0.3948, + "step": 1566 + }, + { + "epoch": 4.908379013312451, + "grad_norm": 1.1755915458959958, + "learning_rate": 1.8782180655622096e-08, + "loss": 0.3909, + "step": 1567 + }, + { + "epoch": 4.911511354737667, + "grad_norm": 1.2040808443639759, + "learning_rate": 1.7464936863261785e-08, + "loss": 0.4109, + "step": 1568 + }, + { + "epoch": 4.914643696162882, + "grad_norm": 1.2313492049462424, + "learning_rate": 1.61955380769796e-08, + "loss": 0.3865, + "step": 1569 + }, + { + "epoch": 4.917776037588097, + "grad_norm": 1.2358379850928518, + "learning_rate": 1.497399038084102e-08, + "loss": 0.4126, + "step": 1570 + }, + { + "epoch": 4.920908379013312, + "grad_norm": 1.2159888247443027, + "learning_rate": 1.3800299629568303e-08, + "loss": 0.3979, + "step": 1571 + }, + { + "epoch": 4.924040720438528, + "grad_norm": 1.2314356631041403, + "learning_rate": 1.2674471448508308e-08, + "loss": 0.3954, + "step": 1572 + }, + { + "epoch": 4.927173061863743, + "grad_norm": 1.1763805345232639, + "learning_rate": 1.1596511233610275e-08, + "loss": 0.4174, + "step": 1573 + }, + { + "epoch": 4.930305403288958, + "grad_norm": 1.21950132315641, + "learning_rate": 1.0566424151401412e-08, + "loss": 0.3712, + "step": 1574 + }, + { + "epoch": 4.9334377447141735, + "grad_norm": 1.1857749532551431, + "learning_rate": 9.584215138953579e-09, + "loss": 0.4018, + "step": 1575 + }, + { + "epoch": 4.9365700861393895, + "grad_norm": 1.1885549206002513, + "learning_rate": 8.649888903869973e-09, + "loss": 0.4014, + "step": 1576 + }, + { + "epoch": 4.939702427564605, + "grad_norm": 1.2601185286531782, + "learning_rate": 7.763449924256262e-09, + "loss": 0.3956, + "step": 1577 + }, + { + "epoch": 4.94283476898982, + "grad_norm": 1.1994949171986244, + "learning_rate": 6.924902448699478e-09, + "loss": 0.4229, + "step": 1578 + }, + { + "epoch": 4.945967110415035, + "grad_norm": 1.231920206414556, + "learning_rate": 6.134250496249161e-09, + "loss": 0.4123, + "step": 1579 + }, + { + "epoch": 4.949099451840251, + "grad_norm": 1.231501336996204, + "learning_rate": 5.391497856399585e-09, + "loss": 0.4019, + "step": 1580 + }, + { + "epoch": 4.952231793265466, + "grad_norm": 1.2025254247538497, + "learning_rate": 4.696648089068667e-09, + "loss": 0.4047, + "step": 1581 + }, + { + "epoch": 4.955364134690681, + "grad_norm": 1.2462699073996386, + "learning_rate": 4.0497045245813105e-09, + "loss": 0.3905, + "step": 1582 + }, + { + "epoch": 4.958496476115896, + "grad_norm": 1.2150582015348728, + "learning_rate": 3.450670263654976e-09, + "loss": 0.4034, + "step": 1583 + }, + { + "epoch": 4.961628817541112, + "grad_norm": 1.1657154816919872, + "learning_rate": 2.899548177384137e-09, + "loss": 0.3983, + "step": 1584 + }, + { + "epoch": 4.964761158966327, + "grad_norm": 1.201657869580201, + "learning_rate": 2.396340907225847e-09, + "loss": 0.3985, + "step": 1585 + }, + { + "epoch": 4.9678935003915425, + "grad_norm": 1.1936954506770183, + "learning_rate": 1.9410508649875258e-09, + "loss": 0.4038, + "step": 1586 + }, + { + "epoch": 4.971025841816758, + "grad_norm": 1.1919200016703124, + "learning_rate": 1.5336802328147492e-09, + "loss": 0.3716, + "step": 1587 + }, + { + "epoch": 4.974158183241974, + "grad_norm": 1.2304919494528042, + "learning_rate": 1.1742309631845861e-09, + "loss": 0.4267, + "step": 1588 + }, + { + "epoch": 4.977290524667189, + "grad_norm": 1.2084987865361176, + "learning_rate": 8.627047788911658e-10, + "loss": 0.443, + "step": 1589 + }, + { + "epoch": 4.980422866092404, + "grad_norm": 1.2473534930975694, + "learning_rate": 5.991031730367968e-10, + "loss": 0.4292, + "step": 1590 + }, + { + "epoch": 4.983555207517619, + "grad_norm": 1.2137391828259563, + "learning_rate": 3.8342740903307697e-10, + "loss": 0.4086, + "step": 1591 + }, + { + "epoch": 4.986687548942835, + "grad_norm": 1.239277828293045, + "learning_rate": 2.1567852058423932e-10, + "loss": 0.3963, + "step": 1592 + }, + { + "epoch": 4.98981989036805, + "grad_norm": 1.198137612800272, + "learning_rate": 9.585731168937351e-11, + "loss": 0.3968, + "step": 1593 + }, + { + "epoch": 4.992952231793265, + "grad_norm": 1.2102944840840637, + "learning_rate": 2.396435663687413e-11, + "loss": 0.3788, + "step": 1594 + }, + { + "epoch": 4.99608457321848, + "grad_norm": 1.2188363908623323, + "learning_rate": 0.0, + "loss": 0.3945, + "step": 1595 + }, + { + "epoch": 4.99608457321848, + "step": 1595, + "total_flos": 1.3245374097522688e+18, + "train_loss": 0.0, + "train_runtime": 1.8638, + "train_samples_per_second": 54818.89, + "train_steps_per_second": 855.791 + } + ], + "logging_steps": 1, + "max_steps": 1595, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.3245374097522688e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}