| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.99608457321848, |
| "eval_steps": 500, |
| "global_step": 1595, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0031323414252153485, |
| "grad_norm": 11.050841614106629, |
| "learning_rate": 1.2500000000000002e-07, |
| "loss": 2.1157, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.006264682850430697, |
| "grad_norm": 10.770240996509447, |
| "learning_rate": 2.5000000000000004e-07, |
| "loss": 2.1788, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.009397024275646046, |
| "grad_norm": 10.853179484768518, |
| "learning_rate": 3.75e-07, |
| "loss": 2.109, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.012529365700861394, |
| "grad_norm": 10.639536499246885, |
| "learning_rate": 5.000000000000001e-07, |
| "loss": 2.1164, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.015661707126076743, |
| "grad_norm": 10.748209885589805, |
| "learning_rate": 6.25e-07, |
| "loss": 2.1148, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.018794048551292093, |
| "grad_norm": 10.463406145556146, |
| "learning_rate": 7.5e-07, |
| "loss": 2.1268, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.02192638997650744, |
| "grad_norm": 10.729571035762252, |
| "learning_rate": 8.75e-07, |
| "loss": 2.1396, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.025058731401722788, |
| "grad_norm": 9.834267944614176, |
| "learning_rate": 1.0000000000000002e-06, |
| "loss": 2.1216, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.028191072826938137, |
| "grad_norm": 9.6285022315209, |
| "learning_rate": 1.125e-06, |
| "loss": 2.0811, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.031323414252153486, |
| "grad_norm": 8.068778598644403, |
| "learning_rate": 1.25e-06, |
| "loss": 2.0954, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.03445575567736883, |
| "grad_norm": 7.900246512788569, |
| "learning_rate": 1.3750000000000002e-06, |
| "loss": 2.093, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.037588097102584185, |
| "grad_norm": 7.672686278535105, |
| "learning_rate": 1.5e-06, |
| "loss": 2.035, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.04072043852779953, |
| "grad_norm": 4.005341384803414, |
| "learning_rate": 1.6250000000000001e-06, |
| "loss": 1.9538, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.04385277995301488, |
| "grad_norm": 4.033643898372103, |
| "learning_rate": 1.75e-06, |
| "loss": 1.9371, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.04698512137823023, |
| "grad_norm": 3.9058441402584543, |
| "learning_rate": 1.8750000000000003e-06, |
| "loss": 1.9222, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.050117462803445575, |
| "grad_norm": 3.93510615716937, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 1.9425, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.05324980422866092, |
| "grad_norm": 3.9972887378242508, |
| "learning_rate": 2.125e-06, |
| "loss": 1.871, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.056382145653876274, |
| "grad_norm": 4.637696367766985, |
| "learning_rate": 2.25e-06, |
| "loss": 1.704, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.05951448707909162, |
| "grad_norm": 4.895075769171093, |
| "learning_rate": 2.375e-06, |
| "loss": 1.7081, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.06264682850430697, |
| "grad_norm": 4.887182700750148, |
| "learning_rate": 2.5e-06, |
| "loss": 1.6922, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.06577916992952232, |
| "grad_norm": 4.387829116707637, |
| "learning_rate": 2.6250000000000003e-06, |
| "loss": 1.6223, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.06891151135473766, |
| "grad_norm": 3.9332482151095056, |
| "learning_rate": 2.7500000000000004e-06, |
| "loss": 1.6429, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.07204385277995301, |
| "grad_norm": 3.621520125738942, |
| "learning_rate": 2.875e-06, |
| "loss": 1.5692, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.07517619420516837, |
| "grad_norm": 4.869727787056477, |
| "learning_rate": 3e-06, |
| "loss": 1.4277, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.07830853563038372, |
| "grad_norm": 2.760787151921623, |
| "learning_rate": 3.125e-06, |
| "loss": 1.343, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.08144087705559906, |
| "grad_norm": 2.158543950171073, |
| "learning_rate": 3.2500000000000002e-06, |
| "loss": 1.3304, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.08457321848081441, |
| "grad_norm": 1.8629759121807041, |
| "learning_rate": 3.3750000000000003e-06, |
| "loss": 1.3279, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.08770555990602975, |
| "grad_norm": 1.4660115041698683, |
| "learning_rate": 3.5e-06, |
| "loss": 1.3604, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.0908379013312451, |
| "grad_norm": 1.2902107389850677, |
| "learning_rate": 3.625e-06, |
| "loss": 1.2715, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.09397024275646046, |
| "grad_norm": 1.201092304355642, |
| "learning_rate": 3.7500000000000005e-06, |
| "loss": 1.3227, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0971025841816758, |
| "grad_norm": 1.1572726408932534, |
| "learning_rate": 3.875e-06, |
| "loss": 1.3139, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.10023492560689115, |
| "grad_norm": 1.0860599652926863, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 1.277, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.1033672670321065, |
| "grad_norm": 1.1601900235559528, |
| "learning_rate": 4.125e-06, |
| "loss": 1.3422, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.10649960845732184, |
| "grad_norm": 1.0702922916707374, |
| "learning_rate": 4.25e-06, |
| "loss": 1.2635, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.1096319498825372, |
| "grad_norm": 1.1224186551004272, |
| "learning_rate": 4.3750000000000005e-06, |
| "loss": 1.3686, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.11276429130775255, |
| "grad_norm": 1.0316309615389836, |
| "learning_rate": 4.5e-06, |
| "loss": 1.2546, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.1158966327329679, |
| "grad_norm": 1.0350470961145155, |
| "learning_rate": 4.625000000000001e-06, |
| "loss": 1.2617, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.11902897415818324, |
| "grad_norm": 0.9499273751228338, |
| "learning_rate": 4.75e-06, |
| "loss": 1.2218, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.12216131558339859, |
| "grad_norm": 0.9633898650879267, |
| "learning_rate": 4.875e-06, |
| "loss": 1.2513, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.12529365700861395, |
| "grad_norm": 0.9820874085749698, |
| "learning_rate": 5e-06, |
| "loss": 1.2438, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.1284259984338293, |
| "grad_norm": 0.9949392511683506, |
| "learning_rate": 5.125e-06, |
| "loss": 1.2127, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.13155833985904464, |
| "grad_norm": 0.963927255316921, |
| "learning_rate": 5.2500000000000006e-06, |
| "loss": 1.1986, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.13469068128425998, |
| "grad_norm": 0.9265554217258913, |
| "learning_rate": 5.375e-06, |
| "loss": 1.2378, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.13782302270947533, |
| "grad_norm": 0.895459875699844, |
| "learning_rate": 5.500000000000001e-06, |
| "loss": 1.2106, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.14095536413469067, |
| "grad_norm": 0.8697327947413821, |
| "learning_rate": 5.625e-06, |
| "loss": 1.1968, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.14408770555990602, |
| "grad_norm": 0.8663662696743084, |
| "learning_rate": 5.75e-06, |
| "loss": 1.1794, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.14722004698512137, |
| "grad_norm": 0.925686203130037, |
| "learning_rate": 5.8750000000000005e-06, |
| "loss": 1.2847, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.15035238841033674, |
| "grad_norm": 0.9075811490967844, |
| "learning_rate": 6e-06, |
| "loss": 1.2584, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.1534847298355521, |
| "grad_norm": 0.8432956780425832, |
| "learning_rate": 6.125000000000001e-06, |
| "loss": 1.1543, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.15661707126076743, |
| "grad_norm": 0.8713896231009577, |
| "learning_rate": 6.25e-06, |
| "loss": 1.2038, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.15974941268598278, |
| "grad_norm": 0.7897618524028729, |
| "learning_rate": 6.375e-06, |
| "loss": 1.1681, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.16288175411119812, |
| "grad_norm": 0.8144877727194363, |
| "learning_rate": 6.5000000000000004e-06, |
| "loss": 1.1969, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.16601409553641347, |
| "grad_norm": 0.8337995472721724, |
| "learning_rate": 6.625e-06, |
| "loss": 1.1631, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.16914643696162882, |
| "grad_norm": 0.8286193073403529, |
| "learning_rate": 6.750000000000001e-06, |
| "loss": 1.231, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.17227877838684416, |
| "grad_norm": 0.8272909936450883, |
| "learning_rate": 6.875e-06, |
| "loss": 1.1908, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.1754111198120595, |
| "grad_norm": 0.8159834359002948, |
| "learning_rate": 7e-06, |
| "loss": 1.1909, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.17854346123727485, |
| "grad_norm": 0.8327400175086643, |
| "learning_rate": 7.125e-06, |
| "loss": 1.2279, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.1816758026624902, |
| "grad_norm": 0.8345916890457833, |
| "learning_rate": 7.25e-06, |
| "loss": 1.2322, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.18480814408770557, |
| "grad_norm": 0.8009822359933073, |
| "learning_rate": 7.375000000000001e-06, |
| "loss": 1.1961, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.18794048551292092, |
| "grad_norm": 0.8174246707753029, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 1.159, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.19107282693813626, |
| "grad_norm": 0.8166829885787309, |
| "learning_rate": 7.625e-06, |
| "loss": 1.2641, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.1942051683633516, |
| "grad_norm": 0.8047242090989581, |
| "learning_rate": 7.75e-06, |
| "loss": 1.1689, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.19733750978856696, |
| "grad_norm": 0.7730776885197735, |
| "learning_rate": 7.875e-06, |
| "loss": 1.1269, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.2004698512137823, |
| "grad_norm": 0.7779957947469543, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 1.1762, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.20360219263899765, |
| "grad_norm": 0.8000417296639418, |
| "learning_rate": 8.125000000000001e-06, |
| "loss": 1.2025, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.206734534064213, |
| "grad_norm": 0.8222150496696322, |
| "learning_rate": 8.25e-06, |
| "loss": 1.2, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.20986687548942834, |
| "grad_norm": 0.7895026739737903, |
| "learning_rate": 8.375e-06, |
| "loss": 1.1623, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.21299921691464369, |
| "grad_norm": 0.7784224076989291, |
| "learning_rate": 8.5e-06, |
| "loss": 1.1554, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.21613155833985903, |
| "grad_norm": 0.798211358388717, |
| "learning_rate": 8.625000000000001e-06, |
| "loss": 1.1389, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.2192638997650744, |
| "grad_norm": 0.8027997517397614, |
| "learning_rate": 8.750000000000001e-06, |
| "loss": 1.1578, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.22239624119028975, |
| "grad_norm": 0.7884275020306974, |
| "learning_rate": 8.875e-06, |
| "loss": 1.1476, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.2255285826155051, |
| "grad_norm": 0.8162843570408501, |
| "learning_rate": 9e-06, |
| "loss": 1.197, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.22866092404072044, |
| "grad_norm": 0.8195467948994536, |
| "learning_rate": 9.125e-06, |
| "loss": 1.1463, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.2317932654659358, |
| "grad_norm": 0.785823621627204, |
| "learning_rate": 9.250000000000001e-06, |
| "loss": 1.13, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.23492560689115113, |
| "grad_norm": 0.7858506830531828, |
| "learning_rate": 9.375000000000001e-06, |
| "loss": 1.1392, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.23805794831636648, |
| "grad_norm": 0.8126411047255567, |
| "learning_rate": 9.5e-06, |
| "loss": 1.174, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.24119028974158183, |
| "grad_norm": 0.8134240118545614, |
| "learning_rate": 9.625e-06, |
| "loss": 1.149, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.24432263116679717, |
| "grad_norm": 2.0900195366866186, |
| "learning_rate": 9.75e-06, |
| "loss": 1.1669, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.24745497259201252, |
| "grad_norm": 0.8049430353655749, |
| "learning_rate": 9.875000000000001e-06, |
| "loss": 1.1632, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.2505873140172279, |
| "grad_norm": 0.8154348013283664, |
| "learning_rate": 1e-05, |
| "loss": 1.1752, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.25371965544244324, |
| "grad_norm": 0.7766532231446668, |
| "learning_rate": 1.0125e-05, |
| "loss": 1.1562, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.2568519968676586, |
| "grad_norm": 0.805338167968417, |
| "learning_rate": 1.025e-05, |
| "loss": 1.1875, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.25998433829287393, |
| "grad_norm": 0.7634601102161442, |
| "learning_rate": 1.0375000000000001e-05, |
| "loss": 1.1216, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.2631166797180893, |
| "grad_norm": 0.7954986672515693, |
| "learning_rate": 1.0500000000000001e-05, |
| "loss": 1.1681, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.2662490211433046, |
| "grad_norm": 0.7666715336890474, |
| "learning_rate": 1.0625e-05, |
| "loss": 1.1383, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.26938136256851997, |
| "grad_norm": 0.7841383740610983, |
| "learning_rate": 1.075e-05, |
| "loss": 1.1956, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.2725137039937353, |
| "grad_norm": 0.7593684758700783, |
| "learning_rate": 1.0875e-05, |
| "loss": 1.1199, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.27564604541895066, |
| "grad_norm": 0.7913739392990039, |
| "learning_rate": 1.1000000000000001e-05, |
| "loss": 1.1665, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.278778386844166, |
| "grad_norm": 0.8205958812550826, |
| "learning_rate": 1.1125000000000001e-05, |
| "loss": 1.1756, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.28191072826938135, |
| "grad_norm": 0.7967279563962926, |
| "learning_rate": 1.125e-05, |
| "loss": 1.1559, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.2850430696945967, |
| "grad_norm": 0.7673412238486342, |
| "learning_rate": 1.1375e-05, |
| "loss": 1.1744, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.28817541111981204, |
| "grad_norm": 0.805175551828789, |
| "learning_rate": 1.15e-05, |
| "loss": 1.1629, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.2913077525450274, |
| "grad_norm": 0.7505564301768121, |
| "learning_rate": 1.1625000000000001e-05, |
| "loss": 1.1384, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.29444009397024273, |
| "grad_norm": 0.8039055214719127, |
| "learning_rate": 1.1750000000000001e-05, |
| "loss": 1.146, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.2975724353954581, |
| "grad_norm": 0.7825064856728625, |
| "learning_rate": 1.1875e-05, |
| "loss": 1.1644, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.3007047768206735, |
| "grad_norm": 0.7935646900032081, |
| "learning_rate": 1.2e-05, |
| "loss": 1.1469, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.3038371182458888, |
| "grad_norm": 0.8024905927017434, |
| "learning_rate": 1.2125e-05, |
| "loss": 1.1402, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.3069694596711042, |
| "grad_norm": 0.7830126081823029, |
| "learning_rate": 1.2250000000000001e-05, |
| "loss": 1.1733, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.3101018010963195, |
| "grad_norm": 0.7997702404780472, |
| "learning_rate": 1.2375000000000001e-05, |
| "loss": 1.1991, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.31323414252153486, |
| "grad_norm": 0.7557112262293625, |
| "learning_rate": 1.25e-05, |
| "loss": 1.1463, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.3163664839467502, |
| "grad_norm": 0.7794955385488898, |
| "learning_rate": 1.2625e-05, |
| "loss": 1.192, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.31949882537196556, |
| "grad_norm": 0.8098258584079975, |
| "learning_rate": 1.275e-05, |
| "loss": 1.213, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.3226311667971809, |
| "grad_norm": 0.7578592334982148, |
| "learning_rate": 1.2875000000000001e-05, |
| "loss": 1.1538, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.32576350822239625, |
| "grad_norm": 0.7801947192543522, |
| "learning_rate": 1.3000000000000001e-05, |
| "loss": 1.1362, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.3288958496476116, |
| "grad_norm": 0.803547746608655, |
| "learning_rate": 1.3125e-05, |
| "loss": 1.1337, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.33202819107282694, |
| "grad_norm": 0.749613358405691, |
| "learning_rate": 1.325e-05, |
| "loss": 1.1038, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.3351605324980423, |
| "grad_norm": 0.8427226009611682, |
| "learning_rate": 1.3375e-05, |
| "loss": 1.1611, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.33829287392325763, |
| "grad_norm": 0.8147379441469031, |
| "learning_rate": 1.3500000000000001e-05, |
| "loss": 1.158, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.341425215348473, |
| "grad_norm": 0.7724455205425795, |
| "learning_rate": 1.3625e-05, |
| "loss": 1.1123, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.3445575567736883, |
| "grad_norm": 0.7752938561359628, |
| "learning_rate": 1.375e-05, |
| "loss": 1.1516, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.34768989819890367, |
| "grad_norm": 0.7714492124885363, |
| "learning_rate": 1.3875e-05, |
| "loss": 1.1251, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.350822239624119, |
| "grad_norm": 0.7951581981638454, |
| "learning_rate": 1.4e-05, |
| "loss": 1.1362, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.35395458104933436, |
| "grad_norm": 0.7654142695855861, |
| "learning_rate": 1.4125000000000003e-05, |
| "loss": 1.1494, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.3570869224745497, |
| "grad_norm": 0.7809679002598582, |
| "learning_rate": 1.425e-05, |
| "loss": 1.1533, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.36021926389976505, |
| "grad_norm": 0.8081986701754312, |
| "learning_rate": 1.4375e-05, |
| "loss": 1.1654, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.3633516053249804, |
| "grad_norm": 0.8085004016002348, |
| "learning_rate": 1.45e-05, |
| "loss": 1.1343, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.36648394675019574, |
| "grad_norm": 0.8329554845008073, |
| "learning_rate": 1.4625e-05, |
| "loss": 1.1569, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.36961628817541115, |
| "grad_norm": 0.783780585930527, |
| "learning_rate": 1.4750000000000003e-05, |
| "loss": 1.125, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.3727486296006265, |
| "grad_norm": 0.7912033183545758, |
| "learning_rate": 1.4875000000000002e-05, |
| "loss": 1.1073, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.37588097102584184, |
| "grad_norm": 0.821268337393993, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 1.1386, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.3790133124510572, |
| "grad_norm": 0.832261258255235, |
| "learning_rate": 1.5125e-05, |
| "loss": 1.1683, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.38214565387627253, |
| "grad_norm": 0.8094345539354894, |
| "learning_rate": 1.525e-05, |
| "loss": 1.1508, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.3852779953014879, |
| "grad_norm": 0.8008883805537554, |
| "learning_rate": 1.5375e-05, |
| "loss": 1.1111, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.3884103367267032, |
| "grad_norm": 0.7915578589687696, |
| "learning_rate": 1.55e-05, |
| "loss": 1.1479, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.39154267815191857, |
| "grad_norm": 0.8327731000166747, |
| "learning_rate": 1.5625e-05, |
| "loss": 1.1777, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.3946750195771339, |
| "grad_norm": 0.8189737080837218, |
| "learning_rate": 1.575e-05, |
| "loss": 1.1558, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.39780736100234926, |
| "grad_norm": 0.7783531314158876, |
| "learning_rate": 1.5875e-05, |
| "loss": 1.1095, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.4009397024275646, |
| "grad_norm": 0.8543936473595438, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 1.1427, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.40407204385277995, |
| "grad_norm": 0.8243727971086049, |
| "learning_rate": 1.6125000000000002e-05, |
| "loss": 1.1636, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.4072043852779953, |
| "grad_norm": 0.7919846838112398, |
| "learning_rate": 1.6250000000000002e-05, |
| "loss": 1.1753, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.41033672670321064, |
| "grad_norm": 0.7953800117937856, |
| "learning_rate": 1.6375e-05, |
| "loss": 1.1121, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.413469068128426, |
| "grad_norm": 0.7941547055326952, |
| "learning_rate": 1.65e-05, |
| "loss": 1.1759, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.41660140955364133, |
| "grad_norm": 0.751465418050413, |
| "learning_rate": 1.6625e-05, |
| "loss": 1.1131, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.4197337509788567, |
| "grad_norm": 0.782469109387437, |
| "learning_rate": 1.675e-05, |
| "loss": 1.1319, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.422866092404072, |
| "grad_norm": 0.7537743415586728, |
| "learning_rate": 1.6875e-05, |
| "loss": 1.123, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.42599843382928737, |
| "grad_norm": 0.8266848129941319, |
| "learning_rate": 1.7e-05, |
| "loss": 1.2021, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.4291307752545027, |
| "grad_norm": 0.7907070577194092, |
| "learning_rate": 1.7125e-05, |
| "loss": 1.1311, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.43226311667971806, |
| "grad_norm": 0.770956600368574, |
| "learning_rate": 1.7250000000000003e-05, |
| "loss": 1.1406, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.43539545810493346, |
| "grad_norm": 0.7942699632538319, |
| "learning_rate": 1.7375000000000002e-05, |
| "loss": 1.121, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.4385277995301488, |
| "grad_norm": 0.7669673284070234, |
| "learning_rate": 1.7500000000000002e-05, |
| "loss": 1.1206, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.44166014095536416, |
| "grad_norm": 0.7910094531836759, |
| "learning_rate": 1.7625e-05, |
| "loss": 1.1069, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.4447924823805795, |
| "grad_norm": 0.7711860139608495, |
| "learning_rate": 1.775e-05, |
| "loss": 1.1422, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.44792482380579485, |
| "grad_norm": 0.7957935969277278, |
| "learning_rate": 1.7875e-05, |
| "loss": 1.1059, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.4510571652310102, |
| "grad_norm": 0.7632895910705135, |
| "learning_rate": 1.8e-05, |
| "loss": 1.1327, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.45418950665622554, |
| "grad_norm": 0.7972939663996078, |
| "learning_rate": 1.8125e-05, |
| "loss": 1.1898, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.4573218480814409, |
| "grad_norm": 0.768609653869839, |
| "learning_rate": 1.825e-05, |
| "loss": 1.1361, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.46045418950665623, |
| "grad_norm": 0.74095309811706, |
| "learning_rate": 1.8375e-05, |
| "loss": 1.0726, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.4635865309318716, |
| "grad_norm": 0.7937353734822679, |
| "learning_rate": 1.8500000000000002e-05, |
| "loss": 1.1412, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.4667188723570869, |
| "grad_norm": 0.8231685857372454, |
| "learning_rate": 1.8625000000000002e-05, |
| "loss": 1.1928, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.46985121378230227, |
| "grad_norm": 0.7582039516040869, |
| "learning_rate": 1.8750000000000002e-05, |
| "loss": 1.1157, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.4729835552075176, |
| "grad_norm": 0.7945667336329015, |
| "learning_rate": 1.8875e-05, |
| "loss": 1.1339, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.47611589663273296, |
| "grad_norm": 0.7511917754873464, |
| "learning_rate": 1.9e-05, |
| "loss": 1.115, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.4792482380579483, |
| "grad_norm": 0.7567481788127869, |
| "learning_rate": 1.9125000000000004e-05, |
| "loss": 1.1438, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.48238057948316365, |
| "grad_norm": 0.8035923511844805, |
| "learning_rate": 1.925e-05, |
| "loss": 1.1488, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.485512920908379, |
| "grad_norm": 0.7832113472335267, |
| "learning_rate": 1.9375e-05, |
| "loss": 1.1473, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.48864526233359434, |
| "grad_norm": 0.7835793180590945, |
| "learning_rate": 1.95e-05, |
| "loss": 1.2008, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.4917776037588097, |
| "grad_norm": 0.8058256022750308, |
| "learning_rate": 1.9625e-05, |
| "loss": 1.148, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.49490994518402504, |
| "grad_norm": 0.7847452922181013, |
| "learning_rate": 1.9750000000000002e-05, |
| "loss": 1.1196, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.4980422866092404, |
| "grad_norm": 0.7812512360657365, |
| "learning_rate": 1.9875000000000002e-05, |
| "loss": 1.1419, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.5011746280344558, |
| "grad_norm": 0.8824042145521714, |
| "learning_rate": 2e-05, |
| "loss": 1.0868, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.5043069694596711, |
| "grad_norm": 0.7567570291946419, |
| "learning_rate": 1.9999976035643366e-05, |
| "loss": 1.1777, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.5074393108848865, |
| "grad_norm": 0.7952722559164349, |
| "learning_rate": 1.9999904142688313e-05, |
| "loss": 1.1345, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.5105716523101018, |
| "grad_norm": 0.7743964880190665, |
| "learning_rate": 1.9999784321479416e-05, |
| "loss": 1.1041, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.5137039937353172, |
| "grad_norm": 0.8498468306677794, |
| "learning_rate": 1.999961657259097e-05, |
| "loss": 1.1567, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.5168363351605325, |
| "grad_norm": 0.7947824846601378, |
| "learning_rate": 1.9999400896826965e-05, |
| "loss": 1.1733, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.5199686765857479, |
| "grad_norm": 0.7487074186272767, |
| "learning_rate": 1.999913729522111e-05, |
| "loss": 1.161, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.5231010180109632, |
| "grad_norm": 0.8037325407031123, |
| "learning_rate": 1.999882576903682e-05, |
| "loss": 1.1646, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.5262333594361785, |
| "grad_norm": 0.7799604031117069, |
| "learning_rate": 1.9998466319767188e-05, |
| "loss": 1.153, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.5293657008613939, |
| "grad_norm": 0.8012827006886666, |
| "learning_rate": 1.9998058949135016e-05, |
| "loss": 1.2302, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.5324980422866092, |
| "grad_norm": 0.7752213646522256, |
| "learning_rate": 1.9997603659092773e-05, |
| "loss": 1.1444, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.5356303837118246, |
| "grad_norm": 0.7880510647431511, |
| "learning_rate": 1.9997100451822616e-05, |
| "loss": 1.1522, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.5387627251370399, |
| "grad_norm": 0.802804888286548, |
| "learning_rate": 1.9996549329736345e-05, |
| "loss": 1.1926, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.5418950665622553, |
| "grad_norm": 0.7425430954083608, |
| "learning_rate": 1.9995950295475422e-05, |
| "loss": 1.1118, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.5450274079874706, |
| "grad_norm": 0.794380669252506, |
| "learning_rate": 1.9995303351910934e-05, |
| "loss": 1.1748, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.548159749412686, |
| "grad_norm": 0.7533240888688524, |
| "learning_rate": 1.9994608502143603e-05, |
| "loss": 1.1243, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.5512920908379013, |
| "grad_norm": 0.7646481885397147, |
| "learning_rate": 1.9993865749503754e-05, |
| "loss": 1.1415, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.5544244322631167, |
| "grad_norm": 0.7915751390688428, |
| "learning_rate": 1.9993075097551302e-05, |
| "loss": 1.1005, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.557556773688332, |
| "grad_norm": 0.7921308350778985, |
| "learning_rate": 1.9992236550075746e-05, |
| "loss": 1.1215, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.5606891151135474, |
| "grad_norm": 0.7469409103210944, |
| "learning_rate": 1.9991350111096133e-05, |
| "loss": 1.1023, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.5638214565387627, |
| "grad_norm": 0.7382348346776821, |
| "learning_rate": 1.999041578486105e-05, |
| "loss": 1.1578, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.566953797963978, |
| "grad_norm": 0.7926745472354245, |
| "learning_rate": 1.99894335758486e-05, |
| "loss": 1.1692, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.5700861393891934, |
| "grad_norm": 0.7569926636991706, |
| "learning_rate": 1.998840348876639e-05, |
| "loss": 1.1385, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.5732184808144087, |
| "grad_norm": 0.7458272413839812, |
| "learning_rate": 1.9987325528551492e-05, |
| "loss": 1.1332, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.5763508222396241, |
| "grad_norm": 0.7842596742794642, |
| "learning_rate": 1.9986199700370435e-05, |
| "loss": 1.1599, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.5794831636648394, |
| "grad_norm": 0.7560805393380029, |
| "learning_rate": 1.998502600961916e-05, |
| "loss": 1.0842, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.5826155050900548, |
| "grad_norm": 0.7498681844673027, |
| "learning_rate": 1.9983804461923022e-05, |
| "loss": 1.1539, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.5857478465152701, |
| "grad_norm": 0.7698674252021521, |
| "learning_rate": 1.998253506313674e-05, |
| "loss": 1.1244, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.5888801879404855, |
| "grad_norm": 0.7668778266672499, |
| "learning_rate": 1.998121781934438e-05, |
| "loss": 1.1268, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.5920125293657008, |
| "grad_norm": 0.7473755501240149, |
| "learning_rate": 1.997985273685932e-05, |
| "loss": 1.0964, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.5951448707909162, |
| "grad_norm": 0.7456163683549248, |
| "learning_rate": 1.9978439822224228e-05, |
| "loss": 1.1468, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.5982772122161315, |
| "grad_norm": 0.7586612489506767, |
| "learning_rate": 1.997697908221102e-05, |
| "loss": 1.1148, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.601409553641347, |
| "grad_norm": 0.7526265106209291, |
| "learning_rate": 1.9975470523820835e-05, |
| "loss": 1.1596, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.6045418950665623, |
| "grad_norm": 0.7515061295922136, |
| "learning_rate": 1.9973914154284e-05, |
| "loss": 1.132, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.6076742364917777, |
| "grad_norm": 0.7628083735034853, |
| "learning_rate": 1.997230998105999e-05, |
| "loss": 1.0992, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.610806577916993, |
| "grad_norm": 0.7372667581302207, |
| "learning_rate": 1.9970658011837404e-05, |
| "loss": 1.1549, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.6139389193422083, |
| "grad_norm": 0.7570962352925633, |
| "learning_rate": 1.9968958254533922e-05, |
| "loss": 1.1437, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.6170712607674237, |
| "grad_norm": 0.7760513349389108, |
| "learning_rate": 1.9967210717296256e-05, |
| "loss": 1.1143, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.620203602192639, |
| "grad_norm": 0.7710793602174225, |
| "learning_rate": 1.996541540850013e-05, |
| "loss": 1.1611, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.6233359436178544, |
| "grad_norm": 0.7497240836215765, |
| "learning_rate": 1.9963572336750227e-05, |
| "loss": 1.0894, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.6264682850430697, |
| "grad_norm": 0.769485721788448, |
| "learning_rate": 1.996168151088015e-05, |
| "loss": 1.1326, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.6296006264682851, |
| "grad_norm": 0.7914744857793413, |
| "learning_rate": 1.9959742939952393e-05, |
| "loss": 1.1642, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.6327329678935004, |
| "grad_norm": 0.7749252132709005, |
| "learning_rate": 1.9957756633258264e-05, |
| "loss": 1.1097, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.6358653093187158, |
| "grad_norm": 0.7248448375095949, |
| "learning_rate": 1.9955722600317887e-05, |
| "loss": 1.1149, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.6389976507439311, |
| "grad_norm": 0.7893182002778344, |
| "learning_rate": 1.9953640850880118e-05, |
| "loss": 1.1343, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.6421299921691465, |
| "grad_norm": 0.752880032513866, |
| "learning_rate": 1.995151139492251e-05, |
| "loss": 1.1393, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.6452623335943618, |
| "grad_norm": 0.7715420215410178, |
| "learning_rate": 1.994933424265127e-05, |
| "loss": 1.1498, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.6483946750195771, |
| "grad_norm": 0.7359564724287257, |
| "learning_rate": 1.9947109404501216e-05, |
| "loss": 1.1758, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.6515270164447925, |
| "grad_norm": 0.7587433580271375, |
| "learning_rate": 1.99448368911357e-05, |
| "loss": 1.1346, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.6546593578700078, |
| "grad_norm": 0.7720483403367886, |
| "learning_rate": 1.99425167134466e-05, |
| "loss": 1.1527, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.6577916992952232, |
| "grad_norm": 0.7504516082730889, |
| "learning_rate": 1.9940148882554223e-05, |
| "loss": 1.0908, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.6609240407204385, |
| "grad_norm": 0.7583232566276451, |
| "learning_rate": 1.9937733409807273e-05, |
| "loss": 1.1387, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.6640563821456539, |
| "grad_norm": 0.7275565863134074, |
| "learning_rate": 1.9935270306782804e-05, |
| "loss": 1.1061, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.6671887235708692, |
| "grad_norm": 0.7308212360196517, |
| "learning_rate": 1.9932759585286154e-05, |
| "loss": 1.14, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.6703210649960846, |
| "grad_norm": 0.7395935935887239, |
| "learning_rate": 1.9930201257350883e-05, |
| "loss": 1.1356, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.6734534064212999, |
| "grad_norm": 0.7318764377322893, |
| "learning_rate": 1.9927595335238736e-05, |
| "loss": 1.1468, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.6765857478465153, |
| "grad_norm": 0.716132066894738, |
| "learning_rate": 1.992494183143955e-05, |
| "loss": 1.1329, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.6797180892717306, |
| "grad_norm": 0.783963003986274, |
| "learning_rate": 1.992224075867124e-05, |
| "loss": 1.134, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.682850430696946, |
| "grad_norm": 0.7765840844001187, |
| "learning_rate": 1.991949212987969e-05, |
| "loss": 1.1525, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.6859827721221613, |
| "grad_norm": 0.7664402025464453, |
| "learning_rate": 1.9916695958238736e-05, |
| "loss": 1.1363, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.6891151135473766, |
| "grad_norm": 0.7619234106278223, |
| "learning_rate": 1.9913852257150053e-05, |
| "loss": 1.1043, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.692247454972592, |
| "grad_norm": 0.7503302852552577, |
| "learning_rate": 1.9910961040243148e-05, |
| "loss": 1.1503, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.6953797963978073, |
| "grad_norm": 0.7386647741445785, |
| "learning_rate": 1.9908022321375244e-05, |
| "loss": 1.126, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.6985121378230227, |
| "grad_norm": 0.772006501836167, |
| "learning_rate": 1.9905036114631247e-05, |
| "loss": 1.1324, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.701644479248238, |
| "grad_norm": 0.7498396218310553, |
| "learning_rate": 1.9902002434323658e-05, |
| "loss": 1.13, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.7047768206734534, |
| "grad_norm": 0.7499406276159518, |
| "learning_rate": 1.9898921294992516e-05, |
| "loss": 1.136, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.7079091620986687, |
| "grad_norm": 0.8050655298488321, |
| "learning_rate": 1.9895792711405332e-05, |
| "loss": 1.1725, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.7110415035238841, |
| "grad_norm": 0.785084787806491, |
| "learning_rate": 1.9892616698556996e-05, |
| "loss": 1.1521, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.7141738449490994, |
| "grad_norm": 0.7548865828004917, |
| "learning_rate": 1.9889393271669735e-05, |
| "loss": 1.1182, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.7173061863743148, |
| "grad_norm": 0.7609216107863833, |
| "learning_rate": 1.9886122446193015e-05, |
| "loss": 1.1345, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.7204385277995301, |
| "grad_norm": 0.7547249839650533, |
| "learning_rate": 1.9882804237803487e-05, |
| "loss": 1.1427, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.7235708692247454, |
| "grad_norm": 0.7632257000147761, |
| "learning_rate": 1.9879438662404893e-05, |
| "loss": 1.1235, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.7267032106499608, |
| "grad_norm": 0.7568637017446223, |
| "learning_rate": 1.9876025736128e-05, |
| "loss": 1.1301, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.7298355520751761, |
| "grad_norm": 0.7420518311600437, |
| "learning_rate": 1.987256547533053e-05, |
| "loss": 1.1342, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.7329678935003915, |
| "grad_norm": 0.7646439659344628, |
| "learning_rate": 1.9869057896597068e-05, |
| "loss": 1.1352, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.7361002349256069, |
| "grad_norm": 0.7312821097164371, |
| "learning_rate": 1.9865503016738983e-05, |
| "loss": 1.0321, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.7392325763508223, |
| "grad_norm": 0.7546957399103286, |
| "learning_rate": 1.9861900852794357e-05, |
| "loss": 1.1377, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.7423649177760376, |
| "grad_norm": 0.7194994232421758, |
| "learning_rate": 1.9858251422027903e-05, |
| "loss": 1.1237, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.745497259201253, |
| "grad_norm": 0.7785862815076194, |
| "learning_rate": 1.985455474193087e-05, |
| "loss": 1.1609, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.7486296006264683, |
| "grad_norm": 0.740903933948647, |
| "learning_rate": 1.985081083022097e-05, |
| "loss": 1.0945, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.7517619420516837, |
| "grad_norm": 0.7291780203060889, |
| "learning_rate": 1.984701970484229e-05, |
| "loss": 1.0856, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.754894283476899, |
| "grad_norm": 0.7587452855951344, |
| "learning_rate": 1.9843181383965207e-05, |
| "loss": 1.1665, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.7580266249021144, |
| "grad_norm": 0.7407242126385574, |
| "learning_rate": 1.98392958859863e-05, |
| "loss": 1.1194, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.7611589663273297, |
| "grad_norm": 0.7518927137010917, |
| "learning_rate": 1.9835363229528255e-05, |
| "loss": 1.1213, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.7642913077525451, |
| "grad_norm": 0.749707981917893, |
| "learning_rate": 1.9831383433439798e-05, |
| "loss": 1.1007, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.7674236491777604, |
| "grad_norm": 0.7695157385802572, |
| "learning_rate": 1.982735651679557e-05, |
| "loss": 1.1203, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.7705559906029757, |
| "grad_norm": 0.7438784077225071, |
| "learning_rate": 1.982328249889607e-05, |
| "loss": 1.135, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.7736883320281911, |
| "grad_norm": 0.7415443518523247, |
| "learning_rate": 1.9819161399267545e-05, |
| "loss": 1.076, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.7768206734534064, |
| "grad_norm": 0.7482623409973128, |
| "learning_rate": 1.9814993237661887e-05, |
| "loss": 1.1561, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.7799530148786218, |
| "grad_norm": 0.7460111329694793, |
| "learning_rate": 1.9810778034056562e-05, |
| "loss": 1.1132, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.7830853563038371, |
| "grad_norm": 0.7288330170842255, |
| "learning_rate": 1.98065158086545e-05, |
| "loss": 1.0535, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.7862176977290525, |
| "grad_norm": 0.7314199802562354, |
| "learning_rate": 1.9802206581883992e-05, |
| "loss": 1.1316, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.7893500391542678, |
| "grad_norm": 0.7804910252834903, |
| "learning_rate": 1.9797850374398617e-05, |
| "loss": 1.1219, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.7924823805794832, |
| "grad_norm": 0.7298549749672437, |
| "learning_rate": 1.9793447207077114e-05, |
| "loss": 1.1124, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.7956147220046985, |
| "grad_norm": 0.7551562547765204, |
| "learning_rate": 1.9788997101023294e-05, |
| "loss": 1.1458, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.7987470634299139, |
| "grad_norm": 0.7641456778799793, |
| "learning_rate": 1.9784500077565944e-05, |
| "loss": 1.1743, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.8018794048551292, |
| "grad_norm": 0.7350148397112547, |
| "learning_rate": 1.977995615825872e-05, |
| "loss": 1.137, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.8050117462803446, |
| "grad_norm": 0.742949716719734, |
| "learning_rate": 1.9775365364880043e-05, |
| "loss": 1.1294, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.8081440877055599, |
| "grad_norm": 0.7336184274603342, |
| "learning_rate": 1.9770727719432994e-05, |
| "loss": 1.13, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.8112764291307752, |
| "grad_norm": 0.7150049627370941, |
| "learning_rate": 1.976604324414521e-05, |
| "loss": 1.079, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.8144087705559906, |
| "grad_norm": 0.729004714169207, |
| "learning_rate": 1.9761311961468782e-05, |
| "loss": 1.0903, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.8175411119812059, |
| "grad_norm": 0.7432345451419001, |
| "learning_rate": 1.9756533894080132e-05, |
| "loss": 1.1718, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.8206734534064213, |
| "grad_norm": 0.7485136400635154, |
| "learning_rate": 1.975170906487993e-05, |
| "loss": 1.123, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.8238057948316366, |
| "grad_norm": 0.7792494517105039, |
| "learning_rate": 1.9746837496992953e-05, |
| "loss": 1.1018, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.826938136256852, |
| "grad_norm": 0.7534542131696997, |
| "learning_rate": 1.974191921376801e-05, |
| "loss": 1.1376, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.8300704776820673, |
| "grad_norm": 0.7486965183472643, |
| "learning_rate": 1.9736954238777793e-05, |
| "loss": 1.1112, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.8332028191072827, |
| "grad_norm": 0.7675798633230227, |
| "learning_rate": 1.9731942595818786e-05, |
| "loss": 1.1506, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.836335160532498, |
| "grad_norm": 0.7361146624834083, |
| "learning_rate": 1.9726884308911153e-05, |
| "loss": 1.1072, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.8394675019577134, |
| "grad_norm": 0.7546626067336956, |
| "learning_rate": 1.9721779402298614e-05, |
| "loss": 1.105, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.8425998433829287, |
| "grad_norm": 0.7592163829713124, |
| "learning_rate": 1.9716627900448323e-05, |
| "loss": 1.1045, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.845732184808144, |
| "grad_norm": 0.7263452707909194, |
| "learning_rate": 1.971142982805077e-05, |
| "loss": 1.1131, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.8488645262333594, |
| "grad_norm": 0.7354286459707577, |
| "learning_rate": 1.9706185210019648e-05, |
| "loss": 1.1676, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.8519968676585747, |
| "grad_norm": 0.7423168880868267, |
| "learning_rate": 1.9700894071491736e-05, |
| "loss": 1.1529, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.8551292090837901, |
| "grad_norm": 0.7524433622381994, |
| "learning_rate": 1.969555643782677e-05, |
| "loss": 1.11, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.8582615505090054, |
| "grad_norm": 0.7553141589784657, |
| "learning_rate": 1.969017233460736e-05, |
| "loss": 1.1777, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.8613938919342208, |
| "grad_norm": 0.7431732388608101, |
| "learning_rate": 1.9684741787638808e-05, |
| "loss": 1.1588, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.8645262333594361, |
| "grad_norm": 0.7563523934795757, |
| "learning_rate": 1.9679264822949027e-05, |
| "loss": 1.1277, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.8676585747846516, |
| "grad_norm": 0.743913029928575, |
| "learning_rate": 1.967374146678841e-05, |
| "loss": 1.0578, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.8707909162098669, |
| "grad_norm": 0.76235114176879, |
| "learning_rate": 1.966817174562968e-05, |
| "loss": 1.1283, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.8739232576350823, |
| "grad_norm": 0.7537117820426102, |
| "learning_rate": 1.9662555686167808e-05, |
| "loss": 1.1312, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.8770555990602976, |
| "grad_norm": 0.7425818024035084, |
| "learning_rate": 1.9656893315319838e-05, |
| "loss": 1.1127, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.880187940485513, |
| "grad_norm": 0.7902486001191769, |
| "learning_rate": 1.9651184660224786e-05, |
| "loss": 1.1369, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.8833202819107283, |
| "grad_norm": 0.7580680784170756, |
| "learning_rate": 1.96454297482435e-05, |
| "loss": 1.0847, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.8864526233359437, |
| "grad_norm": 0.7813627798317231, |
| "learning_rate": 1.9639628606958535e-05, |
| "loss": 1.0851, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.889584964761159, |
| "grad_norm": 0.7613328270635236, |
| "learning_rate": 1.963378126417401e-05, |
| "loss": 1.077, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.8927173061863743, |
| "grad_norm": 0.7773885553361843, |
| "learning_rate": 1.9627887747915496e-05, |
| "loss": 1.0678, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.8958496476115897, |
| "grad_norm": 0.7650287423748288, |
| "learning_rate": 1.9621948086429847e-05, |
| "loss": 1.1636, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.898981989036805, |
| "grad_norm": 0.7817877274905041, |
| "learning_rate": 1.96159623081851e-05, |
| "loss": 1.1262, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.9021143304620204, |
| "grad_norm": 0.7527999272006599, |
| "learning_rate": 1.960993044187032e-05, |
| "loss": 1.1082, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.9052466718872357, |
| "grad_norm": 0.7350684686021072, |
| "learning_rate": 1.9603852516395472e-05, |
| "loss": 1.1019, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.9083790133124511, |
| "grad_norm": 0.7222753898804246, |
| "learning_rate": 1.9597728560891266e-05, |
| "loss": 1.1204, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.9115113547376664, |
| "grad_norm": 0.7894532542241856, |
| "learning_rate": 1.9591558604709028e-05, |
| "loss": 1.1405, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.9146436961628818, |
| "grad_norm": 0.7634432547529683, |
| "learning_rate": 1.958534267742057e-05, |
| "loss": 1.1368, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.9177760375880971, |
| "grad_norm": 0.763149271096417, |
| "learning_rate": 1.9579080808818035e-05, |
| "loss": 1.122, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.9209083790133125, |
| "grad_norm": 0.7380685951966892, |
| "learning_rate": 1.9572773028913744e-05, |
| "loss": 1.1044, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.9240407204385278, |
| "grad_norm": 0.7455823416370709, |
| "learning_rate": 1.956641936794008e-05, |
| "loss": 1.1205, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.9271730618637432, |
| "grad_norm": 0.7364688923639685, |
| "learning_rate": 1.956001985634932e-05, |
| "loss": 1.1512, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.9303054032889585, |
| "grad_norm": 0.7134786654834753, |
| "learning_rate": 1.9553574524813507e-05, |
| "loss": 1.1056, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.9334377447141738, |
| "grad_norm": 0.7170844865094715, |
| "learning_rate": 1.9547083404224273e-05, |
| "loss": 1.1325, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.9365700861393892, |
| "grad_norm": 0.6823076058596949, |
| "learning_rate": 1.9540546525692736e-05, |
| "loss": 1.0178, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.9397024275646045, |
| "grad_norm": 0.7416982640172877, |
| "learning_rate": 1.9533963920549307e-05, |
| "loss": 1.1293, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.9428347689898199, |
| "grad_norm": 0.6949473720891305, |
| "learning_rate": 1.9527335620343567e-05, |
| "loss": 1.1152, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.9459671104150352, |
| "grad_norm": 0.7300442646550352, |
| "learning_rate": 1.9520661656844107e-05, |
| "loss": 1.1324, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.9490994518402506, |
| "grad_norm": 0.7453171429773002, |
| "learning_rate": 1.9513942062038375e-05, |
| "loss": 1.1833, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.9522317932654659, |
| "grad_norm": 0.7387019005056833, |
| "learning_rate": 1.9507176868132517e-05, |
| "loss": 1.101, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.9553641346906813, |
| "grad_norm": 0.7116175190537836, |
| "learning_rate": 1.950036610755125e-05, |
| "loss": 1.1204, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.9584964761158966, |
| "grad_norm": 0.7466335049949244, |
| "learning_rate": 1.9493509812937666e-05, |
| "loss": 1.1376, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.961628817541112, |
| "grad_norm": 0.7380075174756867, |
| "learning_rate": 1.94866080171531e-05, |
| "loss": 1.1225, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.9647611589663273, |
| "grad_norm": 0.747788910002534, |
| "learning_rate": 1.9479660753276978e-05, |
| "loss": 1.1651, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.9678935003915426, |
| "grad_norm": 0.7277939144026558, |
| "learning_rate": 1.947266805460664e-05, |
| "loss": 1.1005, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.971025841816758, |
| "grad_norm": 0.7594540196531739, |
| "learning_rate": 1.9465629954657185e-05, |
| "loss": 1.131, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.9741581832419733, |
| "grad_norm": 0.7685734409142018, |
| "learning_rate": 1.9458546487161326e-05, |
| "loss": 1.1645, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.9772905246671887, |
| "grad_norm": 0.7221986487220532, |
| "learning_rate": 1.945141768606921e-05, |
| "loss": 1.0698, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.980422866092404, |
| "grad_norm": 0.7136839321857378, |
| "learning_rate": 1.9444243585548266e-05, |
| "loss": 1.1085, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.9835552075176194, |
| "grad_norm": 0.7257358526029265, |
| "learning_rate": 1.943702421998303e-05, |
| "loss": 1.088, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.9866875489428347, |
| "grad_norm": 0.7283314263257034, |
| "learning_rate": 1.9429759623974992e-05, |
| "loss": 1.1275, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.9898198903680501, |
| "grad_norm": 0.7150867610674063, |
| "learning_rate": 1.942244983234243e-05, |
| "loss": 1.1191, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.9929522317932654, |
| "grad_norm": 0.7256772289345883, |
| "learning_rate": 1.9415094880120236e-05, |
| "loss": 1.1267, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.9960845732184808, |
| "grad_norm": 0.7160733246362201, |
| "learning_rate": 1.940769480255974e-05, |
| "loss": 1.0924, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.9992169146436961, |
| "grad_norm": 0.6968857896809572, |
| "learning_rate": 1.9400249635128575e-05, |
| "loss": 1.0343, |
| "step": 319 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.6968857896809572, |
| "learning_rate": 1.939275941351046e-05, |
| "loss": 1.1068, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.0031323414252153, |
| "grad_norm": 1.6156520613544338, |
| "learning_rate": 1.9385224173605072e-05, |
| "loss": 1.0187, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.0062646828504307, |
| "grad_norm": 0.8615572642140844, |
| "learning_rate": 1.9377643951527834e-05, |
| "loss": 0.9501, |
| "step": 322 |
| }, |
| { |
| "epoch": 1.009397024275646, |
| "grad_norm": 0.8236605914660184, |
| "learning_rate": 1.937001878360979e-05, |
| "loss": 0.9672, |
| "step": 323 |
| }, |
| { |
| "epoch": 1.0125293657008614, |
| "grad_norm": 0.81342630340967, |
| "learning_rate": 1.9362348706397374e-05, |
| "loss": 0.9559, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.0156617071260767, |
| "grad_norm": 0.8581319434910718, |
| "learning_rate": 1.935463375665229e-05, |
| "loss": 0.9808, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.018794048551292, |
| "grad_norm": 0.8129587194790275, |
| "learning_rate": 1.934687397135129e-05, |
| "loss": 0.9512, |
| "step": 326 |
| }, |
| { |
| "epoch": 1.0219263899765074, |
| "grad_norm": 0.8327468346046575, |
| "learning_rate": 1.933906938768604e-05, |
| "loss": 0.9776, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.0250587314017228, |
| "grad_norm": 0.8873213443616459, |
| "learning_rate": 1.9331220043062894e-05, |
| "loss": 0.9568, |
| "step": 328 |
| }, |
| { |
| "epoch": 1.0281910728269381, |
| "grad_norm": 0.8763873438765251, |
| "learning_rate": 1.9323325975102757e-05, |
| "loss": 0.9854, |
| "step": 329 |
| }, |
| { |
| "epoch": 1.0313234142521535, |
| "grad_norm": 0.8737645235450247, |
| "learning_rate": 1.9315387221640874e-05, |
| "loss": 0.9766, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.0344557556773688, |
| "grad_norm": 0.8226677263540446, |
| "learning_rate": 1.9307403820726675e-05, |
| "loss": 0.9439, |
| "step": 331 |
| }, |
| { |
| "epoch": 1.0375880971025842, |
| "grad_norm": 0.8561637749159361, |
| "learning_rate": 1.929937581062357e-05, |
| "loss": 0.962, |
| "step": 332 |
| }, |
| { |
| "epoch": 1.0407204385277995, |
| "grad_norm": 0.8096315231114, |
| "learning_rate": 1.929130322980878e-05, |
| "loss": 0.9454, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.0438527799530148, |
| "grad_norm": 0.8289922628570354, |
| "learning_rate": 1.9283186116973146e-05, |
| "loss": 0.9844, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.0469851213782302, |
| "grad_norm": 0.7901830667535128, |
| "learning_rate": 1.927502451102095e-05, |
| "loss": 0.9493, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.0501174628034455, |
| "grad_norm": 0.7761466037885246, |
| "learning_rate": 1.926681845106971e-05, |
| "loss": 0.9465, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.0532498042286609, |
| "grad_norm": 0.8141355109509392, |
| "learning_rate": 1.925856797645002e-05, |
| "loss": 0.95, |
| "step": 337 |
| }, |
| { |
| "epoch": 1.0563821456538762, |
| "grad_norm": 0.8074412075844187, |
| "learning_rate": 1.925027312670534e-05, |
| "loss": 0.9436, |
| "step": 338 |
| }, |
| { |
| "epoch": 1.0595144870790916, |
| "grad_norm": 0.8272827849228525, |
| "learning_rate": 1.9241933941591826e-05, |
| "loss": 0.9703, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.062646828504307, |
| "grad_norm": 0.7914982733889313, |
| "learning_rate": 1.9233550461078114e-05, |
| "loss": 0.9512, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.0657791699295223, |
| "grad_norm": 0.7994475682723842, |
| "learning_rate": 1.9225122725345145e-05, |
| "loss": 0.9456, |
| "step": 341 |
| }, |
| { |
| "epoch": 1.0689115113547376, |
| "grad_norm": 0.8465580424768789, |
| "learning_rate": 1.9216650774785975e-05, |
| "loss": 0.9432, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.072043852779953, |
| "grad_norm": 0.8466406131696012, |
| "learning_rate": 1.920813465000557e-05, |
| "loss": 0.9697, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.0751761942051683, |
| "grad_norm": 0.8033646805720518, |
| "learning_rate": 1.9199574391820627e-05, |
| "loss": 0.9541, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.0783085356303836, |
| "grad_norm": 0.8224052298742338, |
| "learning_rate": 1.9190970041259352e-05, |
| "loss": 1.0003, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.081440877055599, |
| "grad_norm": 0.7997901412752633, |
| "learning_rate": 1.91823216395613e-05, |
| "loss": 0.9903, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.0845732184808143, |
| "grad_norm": 0.8140939297756549, |
| "learning_rate": 1.917362922817714e-05, |
| "loss": 0.9549, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.0877055599060297, |
| "grad_norm": 0.7824285972233995, |
| "learning_rate": 1.9164892848768485e-05, |
| "loss": 0.9593, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.090837901331245, |
| "grad_norm": 0.7836791598591205, |
| "learning_rate": 1.9156112543207674e-05, |
| "loss": 0.9582, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.0939702427564604, |
| "grad_norm": 0.7911753479916747, |
| "learning_rate": 1.914728835357759e-05, |
| "loss": 0.9348, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.0971025841816757, |
| "grad_norm": 0.7782991822882203, |
| "learning_rate": 1.913842032217143e-05, |
| "loss": 0.9631, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.100234925606891, |
| "grad_norm": 0.8014360461302352, |
| "learning_rate": 1.912950849149253e-05, |
| "loss": 0.9351, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.1033672670321064, |
| "grad_norm": 0.8312030203793829, |
| "learning_rate": 1.912055290425415e-05, |
| "loss": 0.9668, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.1064996084573218, |
| "grad_norm": 0.7891597416989514, |
| "learning_rate": 1.911155360337926e-05, |
| "loss": 0.9432, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.109631949882537, |
| "grad_norm": 0.7878291420752258, |
| "learning_rate": 1.9102510632000365e-05, |
| "loss": 0.9521, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.1127642913077525, |
| "grad_norm": 0.8120607285651226, |
| "learning_rate": 1.909342403345925e-05, |
| "loss": 0.9425, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.1158966327329678, |
| "grad_norm": 0.7856758097045018, |
| "learning_rate": 1.9084293851306815e-05, |
| "loss": 0.9542, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.1190289741581831, |
| "grad_norm": 0.7684160480129255, |
| "learning_rate": 1.9075120129302857e-05, |
| "loss": 0.9454, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.1221613155833985, |
| "grad_norm": 0.8316731012269218, |
| "learning_rate": 1.9065902911415836e-05, |
| "loss": 0.9919, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.1252936570086138, |
| "grad_norm": 0.822006226986475, |
| "learning_rate": 1.905664224182269e-05, |
| "loss": 0.9751, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.1284259984338294, |
| "grad_norm": 0.7696798368031216, |
| "learning_rate": 1.9047338164908628e-05, |
| "loss": 0.9717, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.1315583398590445, |
| "grad_norm": 0.8046009384769924, |
| "learning_rate": 1.903799072526688e-05, |
| "loss": 0.8908, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.13469068128426, |
| "grad_norm": 0.7979550376850155, |
| "learning_rate": 1.9028599967698533e-05, |
| "loss": 0.9904, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.1378230227094752, |
| "grad_norm": 0.8107407192621531, |
| "learning_rate": 1.9019165937212273e-05, |
| "loss": 0.9528, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.1409553641346908, |
| "grad_norm": 0.7767024196508194, |
| "learning_rate": 1.900968867902419e-05, |
| "loss": 0.9159, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.144087705559906, |
| "grad_norm": 0.7927579094268248, |
| "learning_rate": 1.9000168238557574e-05, |
| "loss": 0.9252, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.1472200469851215, |
| "grad_norm": 0.8295313527055139, |
| "learning_rate": 1.899060466144267e-05, |
| "loss": 0.941, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.1503523884103368, |
| "grad_norm": 0.7746260340598791, |
| "learning_rate": 1.8980997993516462e-05, |
| "loss": 0.9237, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.1534847298355522, |
| "grad_norm": 0.7971878626092886, |
| "learning_rate": 1.897134828082248e-05, |
| "loss": 0.9387, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.1566170712607675, |
| "grad_norm": 0.807852831490928, |
| "learning_rate": 1.8961655569610557e-05, |
| "loss": 0.9958, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.1597494126859829, |
| "grad_norm": 0.8147989237592862, |
| "learning_rate": 1.895191990633661e-05, |
| "loss": 0.953, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.1628817541111982, |
| "grad_norm": 0.7959883702024442, |
| "learning_rate": 1.8942141337662415e-05, |
| "loss": 0.9419, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.1660140955364136, |
| "grad_norm": 0.7895917093808684, |
| "learning_rate": 1.89323199104554e-05, |
| "loss": 1.0, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.169146436961629, |
| "grad_norm": 0.7788706927325837, |
| "learning_rate": 1.89224556717884e-05, |
| "loss": 0.9641, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.1722787783868442, |
| "grad_norm": 0.8275735256549743, |
| "learning_rate": 1.891254866893944e-05, |
| "loss": 0.9596, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.1754111198120596, |
| "grad_norm": 0.7901905557676397, |
| "learning_rate": 1.8902598949391507e-05, |
| "loss": 0.9881, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.178543461237275, |
| "grad_norm": 0.7938940344571249, |
| "learning_rate": 1.8892606560832335e-05, |
| "loss": 0.9056, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.1816758026624903, |
| "grad_norm": 0.7774862380646365, |
| "learning_rate": 1.8882571551154152e-05, |
| "loss": 0.9408, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.1848081440877056, |
| "grad_norm": 0.8020328348235701, |
| "learning_rate": 1.8872493968453467e-05, |
| "loss": 0.9251, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.187940485512921, |
| "grad_norm": 0.806279534293864, |
| "learning_rate": 1.8862373861030838e-05, |
| "loss": 0.948, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.1910728269381363, |
| "grad_norm": 0.8025985732944699, |
| "learning_rate": 1.885221127739064e-05, |
| "loss": 0.9588, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.1942051683633517, |
| "grad_norm": 0.7681611064314322, |
| "learning_rate": 1.8842006266240826e-05, |
| "loss": 0.9286, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.197337509788567, |
| "grad_norm": 0.7830003805481669, |
| "learning_rate": 1.88317588764927e-05, |
| "loss": 0.9641, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.2004698512137824, |
| "grad_norm": 0.7682889170068867, |
| "learning_rate": 1.8821469157260687e-05, |
| "loss": 0.9293, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.2036021926389977, |
| "grad_norm": 0.8006318522209638, |
| "learning_rate": 1.8811137157862084e-05, |
| "loss": 0.9283, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.206734534064213, |
| "grad_norm": 0.8064409365222602, |
| "learning_rate": 1.8800762927816835e-05, |
| "loss": 0.9974, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.2098668754894284, |
| "grad_norm": 0.7771414813466695, |
| "learning_rate": 1.8790346516847288e-05, |
| "loss": 0.9414, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.2129992169146437, |
| "grad_norm": 0.8516980588228736, |
| "learning_rate": 1.8779887974877965e-05, |
| "loss": 0.9708, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.216131558339859, |
| "grad_norm": 0.8233081545979598, |
| "learning_rate": 1.876938735203531e-05, |
| "loss": 0.9856, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.2192638997650744, |
| "grad_norm": 0.8148185121214713, |
| "learning_rate": 1.8758844698647457e-05, |
| "loss": 0.9441, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.2223962411902898, |
| "grad_norm": 0.8121759062847519, |
| "learning_rate": 1.8748260065243985e-05, |
| "loss": 0.9865, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.2255285826155051, |
| "grad_norm": 0.8027197651573186, |
| "learning_rate": 1.8737633502555685e-05, |
| "loss": 0.9809, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.2286609240407205, |
| "grad_norm": 0.8058527578911673, |
| "learning_rate": 1.8726965061514297e-05, |
| "loss": 0.9611, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.2317932654659358, |
| "grad_norm": 0.7952817742989116, |
| "learning_rate": 1.8716254793252294e-05, |
| "loss": 0.9786, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.2349256068911512, |
| "grad_norm": 0.7839895990296804, |
| "learning_rate": 1.870550274910261e-05, |
| "loss": 0.9601, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.2380579483163665, |
| "grad_norm": 0.8549762083059709, |
| "learning_rate": 1.869470898059841e-05, |
| "loss": 1.0219, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.2411902897415819, |
| "grad_norm": 0.788953768471007, |
| "learning_rate": 1.868387353947284e-05, |
| "loss": 1.0143, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.2443226311667972, |
| "grad_norm": 0.8217786523602275, |
| "learning_rate": 1.8672996477658767e-05, |
| "loss": 0.9523, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.2474549725920125, |
| "grad_norm": 0.7704666004345789, |
| "learning_rate": 1.8662077847288554e-05, |
| "loss": 0.9207, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.250587314017228, |
| "grad_norm": 0.7525516825767109, |
| "learning_rate": 1.8651117700693794e-05, |
| "loss": 0.9477, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.2537196554424432, |
| "grad_norm": 0.8253629375930066, |
| "learning_rate": 1.8640116090405055e-05, |
| "loss": 0.9731, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.2568519968676586, |
| "grad_norm": 0.7915558764489208, |
| "learning_rate": 1.8629073069151638e-05, |
| "loss": 0.9199, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.259984338292874, |
| "grad_norm": 0.7782487189596384, |
| "learning_rate": 1.8617988689861333e-05, |
| "loss": 0.9199, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.2631166797180893, |
| "grad_norm": 0.7980723005204162, |
| "learning_rate": 1.8606863005660133e-05, |
| "loss": 0.9212, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.2662490211433046, |
| "grad_norm": 0.8048484955000166, |
| "learning_rate": 1.8595696069872013e-05, |
| "loss": 0.9581, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.26938136256852, |
| "grad_norm": 0.8032453880110384, |
| "learning_rate": 1.8584487936018663e-05, |
| "loss": 0.9821, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.2725137039937353, |
| "grad_norm": 0.8005507805303662, |
| "learning_rate": 1.8573238657819225e-05, |
| "loss": 0.9411, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.2756460454189507, |
| "grad_norm": 0.7916521575263914, |
| "learning_rate": 1.8561948289190035e-05, |
| "loss": 0.9357, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.278778386844166, |
| "grad_norm": 0.7979831477850623, |
| "learning_rate": 1.855061688424439e-05, |
| "loss": 0.9679, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.2819107282693813, |
| "grad_norm": 0.7843000585547435, |
| "learning_rate": 1.8539244497292248e-05, |
| "loss": 0.9636, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.2850430696945967, |
| "grad_norm": 0.784278644818112, |
| "learning_rate": 1.8527831182839997e-05, |
| "loss": 1.0007, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.288175411119812, |
| "grad_norm": 0.7946831238148018, |
| "learning_rate": 1.8516376995590185e-05, |
| "loss": 0.9813, |
| "step": 412 |
| }, |
| { |
| "epoch": 1.2913077525450274, |
| "grad_norm": 0.8109210495548596, |
| "learning_rate": 1.850488199044126e-05, |
| "loss": 0.9349, |
| "step": 413 |
| }, |
| { |
| "epoch": 1.2944400939702427, |
| "grad_norm": 0.7934617847842895, |
| "learning_rate": 1.84933462224873e-05, |
| "loss": 0.9416, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.297572435395458, |
| "grad_norm": 0.7902841843838354, |
| "learning_rate": 1.848176974701775e-05, |
| "loss": 0.9346, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.3007047768206734, |
| "grad_norm": 0.779580351091375, |
| "learning_rate": 1.847015261951718e-05, |
| "loss": 0.9739, |
| "step": 416 |
| }, |
| { |
| "epoch": 1.3038371182458888, |
| "grad_norm": 0.7785529627916624, |
| "learning_rate": 1.845849489566498e-05, |
| "loss": 0.9583, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.3069694596711041, |
| "grad_norm": 0.8436224516793075, |
| "learning_rate": 1.8446796631335123e-05, |
| "loss": 0.9681, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.3101018010963195, |
| "grad_norm": 0.8155304400749016, |
| "learning_rate": 1.8435057882595885e-05, |
| "loss": 0.9455, |
| "step": 419 |
| }, |
| { |
| "epoch": 1.3132341425215348, |
| "grad_norm": 0.8154776896881121, |
| "learning_rate": 1.8423278705709573e-05, |
| "loss": 0.9319, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.3163664839467502, |
| "grad_norm": 0.7875316068008983, |
| "learning_rate": 1.841145915713227e-05, |
| "loss": 0.9839, |
| "step": 421 |
| }, |
| { |
| "epoch": 1.3194988253719655, |
| "grad_norm": 0.7878642203922693, |
| "learning_rate": 1.8399599293513555e-05, |
| "loss": 0.9611, |
| "step": 422 |
| }, |
| { |
| "epoch": 1.3226311667971808, |
| "grad_norm": 0.8272412655909572, |
| "learning_rate": 1.8387699171696222e-05, |
| "loss": 0.9676, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.3257635082223962, |
| "grad_norm": 0.8036668057730373, |
| "learning_rate": 1.837575884871603e-05, |
| "loss": 0.9585, |
| "step": 424 |
| }, |
| { |
| "epoch": 1.3288958496476115, |
| "grad_norm": 0.7745104838154117, |
| "learning_rate": 1.8363778381801404e-05, |
| "loss": 0.9586, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.3320281910728269, |
| "grad_norm": 0.8044310075899959, |
| "learning_rate": 1.8351757828373183e-05, |
| "loss": 0.9773, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.3351605324980422, |
| "grad_norm": 0.7816652761458432, |
| "learning_rate": 1.8339697246044333e-05, |
| "loss": 0.9483, |
| "step": 427 |
| }, |
| { |
| "epoch": 1.3382928739232576, |
| "grad_norm": 0.805576362415665, |
| "learning_rate": 1.8327596692619674e-05, |
| "loss": 1.0151, |
| "step": 428 |
| }, |
| { |
| "epoch": 1.341425215348473, |
| "grad_norm": 0.7789246864462362, |
| "learning_rate": 1.83154562260956e-05, |
| "loss": 0.9395, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.3445575567736883, |
| "grad_norm": 0.7843379086528441, |
| "learning_rate": 1.8303275904659807e-05, |
| "loss": 0.9632, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.3476898981989036, |
| "grad_norm": 0.7738505946833321, |
| "learning_rate": 1.829105578669101e-05, |
| "loss": 0.898, |
| "step": 431 |
| }, |
| { |
| "epoch": 1.350822239624119, |
| "grad_norm": 0.773409651313221, |
| "learning_rate": 1.8278795930758656e-05, |
| "loss": 0.925, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.3539545810493343, |
| "grad_norm": 0.775638063691069, |
| "learning_rate": 1.826649639562266e-05, |
| "loss": 0.9296, |
| "step": 433 |
| }, |
| { |
| "epoch": 1.3570869224745497, |
| "grad_norm": 0.8255837346883753, |
| "learning_rate": 1.8254157240233114e-05, |
| "loss": 0.9578, |
| "step": 434 |
| }, |
| { |
| "epoch": 1.360219263899765, |
| "grad_norm": 0.8225702778393079, |
| "learning_rate": 1.8241778523729997e-05, |
| "loss": 0.9938, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.3633516053249803, |
| "grad_norm": 0.8024681843740595, |
| "learning_rate": 1.8229360305442906e-05, |
| "loss": 0.9898, |
| "step": 436 |
| }, |
| { |
| "epoch": 1.3664839467501957, |
| "grad_norm": 0.7659268418287953, |
| "learning_rate": 1.8216902644890768e-05, |
| "loss": 0.9582, |
| "step": 437 |
| }, |
| { |
| "epoch": 1.3696162881754113, |
| "grad_norm": 0.7979578032515822, |
| "learning_rate": 1.8204405601781542e-05, |
| "loss": 0.9539, |
| "step": 438 |
| }, |
| { |
| "epoch": 1.3727486296006264, |
| "grad_norm": 0.7840678597712154, |
| "learning_rate": 1.819186923601195e-05, |
| "loss": 0.9662, |
| "step": 439 |
| }, |
| { |
| "epoch": 1.375880971025842, |
| "grad_norm": 0.7951677564440199, |
| "learning_rate": 1.8179293607667177e-05, |
| "loss": 0.963, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.379013312451057, |
| "grad_norm": 0.7824747662278654, |
| "learning_rate": 1.8166678777020595e-05, |
| "loss": 0.9646, |
| "step": 441 |
| }, |
| { |
| "epoch": 1.3821456538762726, |
| "grad_norm": 0.7889899716482371, |
| "learning_rate": 1.8154024804533464e-05, |
| "loss": 0.9949, |
| "step": 442 |
| }, |
| { |
| "epoch": 1.3852779953014878, |
| "grad_norm": 0.8294307839008276, |
| "learning_rate": 1.8141331750854647e-05, |
| "loss": 0.9745, |
| "step": 443 |
| }, |
| { |
| "epoch": 1.3884103367267033, |
| "grad_norm": 0.8026550217216318, |
| "learning_rate": 1.8128599676820306e-05, |
| "loss": 0.9605, |
| "step": 444 |
| }, |
| { |
| "epoch": 1.3915426781519185, |
| "grad_norm": 0.8043494358409277, |
| "learning_rate": 1.8115828643453647e-05, |
| "loss": 0.9419, |
| "step": 445 |
| }, |
| { |
| "epoch": 1.394675019577134, |
| "grad_norm": 0.7904914993084787, |
| "learning_rate": 1.8103018711964588e-05, |
| "loss": 0.9636, |
| "step": 446 |
| }, |
| { |
| "epoch": 1.3978073610023491, |
| "grad_norm": 0.833878643616376, |
| "learning_rate": 1.8090169943749477e-05, |
| "loss": 0.9799, |
| "step": 447 |
| }, |
| { |
| "epoch": 1.4009397024275647, |
| "grad_norm": 0.8351491373093172, |
| "learning_rate": 1.807728240039081e-05, |
| "loss": 0.954, |
| "step": 448 |
| }, |
| { |
| "epoch": 1.4040720438527798, |
| "grad_norm": 0.8413539861179071, |
| "learning_rate": 1.806435614365692e-05, |
| "loss": 0.9939, |
| "step": 449 |
| }, |
| { |
| "epoch": 1.4072043852779954, |
| "grad_norm": 0.8339283054897446, |
| "learning_rate": 1.8051391235501695e-05, |
| "loss": 0.9699, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.4103367267032105, |
| "grad_norm": 0.8013532736907198, |
| "learning_rate": 1.8038387738064274e-05, |
| "loss": 0.9581, |
| "step": 451 |
| }, |
| { |
| "epoch": 1.413469068128426, |
| "grad_norm": 0.8255816811624092, |
| "learning_rate": 1.8025345713668746e-05, |
| "loss": 1.0, |
| "step": 452 |
| }, |
| { |
| "epoch": 1.4166014095536412, |
| "grad_norm": 0.8242603985956736, |
| "learning_rate": 1.8012265224823853e-05, |
| "loss": 0.9364, |
| "step": 453 |
| }, |
| { |
| "epoch": 1.4197337509788568, |
| "grad_norm": 0.8322278182809092, |
| "learning_rate": 1.79991463342227e-05, |
| "loss": 0.9851, |
| "step": 454 |
| }, |
| { |
| "epoch": 1.422866092404072, |
| "grad_norm": 0.8163638205578437, |
| "learning_rate": 1.7985989104742434e-05, |
| "loss": 0.9499, |
| "step": 455 |
| }, |
| { |
| "epoch": 1.4259984338292875, |
| "grad_norm": 0.8442359156970612, |
| "learning_rate": 1.7972793599443966e-05, |
| "loss": 0.9491, |
| "step": 456 |
| }, |
| { |
| "epoch": 1.4291307752545026, |
| "grad_norm": 0.7893355911335334, |
| "learning_rate": 1.795955988157166e-05, |
| "loss": 0.9816, |
| "step": 457 |
| }, |
| { |
| "epoch": 1.4322631166797182, |
| "grad_norm": 0.806471339752882, |
| "learning_rate": 1.7946288014553017e-05, |
| "loss": 0.9549, |
| "step": 458 |
| }, |
| { |
| "epoch": 1.4353954581049335, |
| "grad_norm": 0.8364165434031879, |
| "learning_rate": 1.7932978061998393e-05, |
| "loss": 0.9836, |
| "step": 459 |
| }, |
| { |
| "epoch": 1.4385277995301489, |
| "grad_norm": 0.7859146658803116, |
| "learning_rate": 1.7919630087700672e-05, |
| "loss": 0.9842, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.4416601409553642, |
| "grad_norm": 0.7690910792708224, |
| "learning_rate": 1.790624415563498e-05, |
| "loss": 0.9396, |
| "step": 461 |
| }, |
| { |
| "epoch": 1.4447924823805796, |
| "grad_norm": 0.8496502770784555, |
| "learning_rate": 1.7892820329958368e-05, |
| "loss": 0.9579, |
| "step": 462 |
| }, |
| { |
| "epoch": 1.447924823805795, |
| "grad_norm": 0.8471130092608972, |
| "learning_rate": 1.7879358675009506e-05, |
| "loss": 0.9973, |
| "step": 463 |
| }, |
| { |
| "epoch": 1.4510571652310102, |
| "grad_norm": 0.784501551931219, |
| "learning_rate": 1.786585925530837e-05, |
| "loss": 0.9641, |
| "step": 464 |
| }, |
| { |
| "epoch": 1.4541895066562256, |
| "grad_norm": 0.836729539160783, |
| "learning_rate": 1.7852322135555946e-05, |
| "loss": 0.9319, |
| "step": 465 |
| }, |
| { |
| "epoch": 1.457321848081441, |
| "grad_norm": 0.8356458620469223, |
| "learning_rate": 1.7838747380633904e-05, |
| "loss": 0.9571, |
| "step": 466 |
| }, |
| { |
| "epoch": 1.4604541895066563, |
| "grad_norm": 0.8216191110677196, |
| "learning_rate": 1.78251350556043e-05, |
| "loss": 0.9553, |
| "step": 467 |
| }, |
| { |
| "epoch": 1.4635865309318716, |
| "grad_norm": 0.7966731084754197, |
| "learning_rate": 1.7811485225709255e-05, |
| "loss": 0.9519, |
| "step": 468 |
| }, |
| { |
| "epoch": 1.466718872357087, |
| "grad_norm": 0.869796047610942, |
| "learning_rate": 1.7797797956370646e-05, |
| "loss": 0.9836, |
| "step": 469 |
| }, |
| { |
| "epoch": 1.4698512137823023, |
| "grad_norm": 0.8527775209514664, |
| "learning_rate": 1.7784073313189796e-05, |
| "loss": 0.9559, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.4729835552075177, |
| "grad_norm": 0.7722201957501883, |
| "learning_rate": 1.7770311361947155e-05, |
| "loss": 0.9424, |
| "step": 471 |
| }, |
| { |
| "epoch": 1.476115896632733, |
| "grad_norm": 0.8231174830960291, |
| "learning_rate": 1.775651216860198e-05, |
| "loss": 1.0208, |
| "step": 472 |
| }, |
| { |
| "epoch": 1.4792482380579484, |
| "grad_norm": 0.8323084233244761, |
| "learning_rate": 1.7742675799292033e-05, |
| "loss": 1.0034, |
| "step": 473 |
| }, |
| { |
| "epoch": 1.4823805794831637, |
| "grad_norm": 0.8153252414548278, |
| "learning_rate": 1.772880232033325e-05, |
| "loss": 0.9844, |
| "step": 474 |
| }, |
| { |
| "epoch": 1.485512920908379, |
| "grad_norm": 0.8189526410041873, |
| "learning_rate": 1.7714891798219432e-05, |
| "loss": 0.9666, |
| "step": 475 |
| }, |
| { |
| "epoch": 1.4886452623335944, |
| "grad_norm": 0.8078003198916532, |
| "learning_rate": 1.770094429962192e-05, |
| "loss": 0.944, |
| "step": 476 |
| }, |
| { |
| "epoch": 1.4917776037588097, |
| "grad_norm": 0.8110193552079895, |
| "learning_rate": 1.768695989138928e-05, |
| "loss": 0.9544, |
| "step": 477 |
| }, |
| { |
| "epoch": 1.494909945184025, |
| "grad_norm": 0.8312565557198122, |
| "learning_rate": 1.767293864054698e-05, |
| "loss": 0.9701, |
| "step": 478 |
| }, |
| { |
| "epoch": 1.4980422866092404, |
| "grad_norm": 0.8061090156479217, |
| "learning_rate": 1.7658880614297075e-05, |
| "loss": 0.9687, |
| "step": 479 |
| }, |
| { |
| "epoch": 1.5011746280344558, |
| "grad_norm": 0.823951800509433, |
| "learning_rate": 1.7644785880017874e-05, |
| "loss": 0.9723, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.5043069694596711, |
| "grad_norm": 0.7965679907922073, |
| "learning_rate": 1.7630654505263628e-05, |
| "loss": 0.9073, |
| "step": 481 |
| }, |
| { |
| "epoch": 1.5074393108848865, |
| "grad_norm": 0.9569281276943931, |
| "learning_rate": 1.7616486557764187e-05, |
| "loss": 0.9808, |
| "step": 482 |
| }, |
| { |
| "epoch": 1.5105716523101018, |
| "grad_norm": 0.8085640633910255, |
| "learning_rate": 1.7602282105424712e-05, |
| "loss": 0.9732, |
| "step": 483 |
| }, |
| { |
| "epoch": 1.5137039937353172, |
| "grad_norm": 0.8532004513694563, |
| "learning_rate": 1.7588041216325314e-05, |
| "loss": 0.9507, |
| "step": 484 |
| }, |
| { |
| "epoch": 1.5168363351605325, |
| "grad_norm": 0.7833644189128819, |
| "learning_rate": 1.7573763958720736e-05, |
| "loss": 0.9411, |
| "step": 485 |
| }, |
| { |
| "epoch": 1.5199686765857479, |
| "grad_norm": 0.8189040727283652, |
| "learning_rate": 1.755945040104004e-05, |
| "loss": 0.9655, |
| "step": 486 |
| }, |
| { |
| "epoch": 1.5231010180109632, |
| "grad_norm": 0.7722121152681558, |
| "learning_rate": 1.754510061188627e-05, |
| "loss": 0.9185, |
| "step": 487 |
| }, |
| { |
| "epoch": 1.5262333594361785, |
| "grad_norm": 0.8037152368297265, |
| "learning_rate": 1.7530714660036112e-05, |
| "loss": 0.9541, |
| "step": 488 |
| }, |
| { |
| "epoch": 1.529365700861394, |
| "grad_norm": 0.8119393404590528, |
| "learning_rate": 1.7516292614439586e-05, |
| "loss": 0.9657, |
| "step": 489 |
| }, |
| { |
| "epoch": 1.5324980422866092, |
| "grad_norm": 0.805216081294044, |
| "learning_rate": 1.7501834544219697e-05, |
| "loss": 0.9448, |
| "step": 490 |
| }, |
| { |
| "epoch": 1.5356303837118246, |
| "grad_norm": 0.7949712346927176, |
| "learning_rate": 1.748734051867212e-05, |
| "loss": 0.9463, |
| "step": 491 |
| }, |
| { |
| "epoch": 1.53876272513704, |
| "grad_norm": 0.77990875986472, |
| "learning_rate": 1.7472810607264853e-05, |
| "loss": 0.9422, |
| "step": 492 |
| }, |
| { |
| "epoch": 1.5418950665622553, |
| "grad_norm": 0.7886283708612627, |
| "learning_rate": 1.7458244879637897e-05, |
| "loss": 0.9487, |
| "step": 493 |
| }, |
| { |
| "epoch": 1.5450274079874706, |
| "grad_norm": 0.7958418765522267, |
| "learning_rate": 1.74436434056029e-05, |
| "loss": 0.9714, |
| "step": 494 |
| }, |
| { |
| "epoch": 1.548159749412686, |
| "grad_norm": 0.8069518461832199, |
| "learning_rate": 1.742900625514285e-05, |
| "loss": 0.9829, |
| "step": 495 |
| }, |
| { |
| "epoch": 1.5512920908379013, |
| "grad_norm": 0.8073763682254401, |
| "learning_rate": 1.7414333498411734e-05, |
| "loss": 0.9569, |
| "step": 496 |
| }, |
| { |
| "epoch": 1.5544244322631167, |
| "grad_norm": 0.8342340049061451, |
| "learning_rate": 1.739962520573418e-05, |
| "loss": 1.0084, |
| "step": 497 |
| }, |
| { |
| "epoch": 1.557556773688332, |
| "grad_norm": 0.8120689308519743, |
| "learning_rate": 1.7384881447605144e-05, |
| "loss": 1.0443, |
| "step": 498 |
| }, |
| { |
| "epoch": 1.5606891151135474, |
| "grad_norm": 0.8010972622074474, |
| "learning_rate": 1.7370102294689563e-05, |
| "loss": 0.959, |
| "step": 499 |
| }, |
| { |
| "epoch": 1.5638214565387627, |
| "grad_norm": 0.7709293012877854, |
| "learning_rate": 1.7355287817822014e-05, |
| "loss": 0.9864, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.566953797963978, |
| "grad_norm": 0.799753436392405, |
| "learning_rate": 1.7340438088006376e-05, |
| "loss": 0.9981, |
| "step": 501 |
| }, |
| { |
| "epoch": 1.5700861393891934, |
| "grad_norm": 0.7779275987751509, |
| "learning_rate": 1.7325553176415496e-05, |
| "loss": 0.9644, |
| "step": 502 |
| }, |
| { |
| "epoch": 1.5732184808144087, |
| "grad_norm": 0.8008846365371122, |
| "learning_rate": 1.731063315439084e-05, |
| "loss": 0.9368, |
| "step": 503 |
| }, |
| { |
| "epoch": 1.576350822239624, |
| "grad_norm": 0.774005109549041, |
| "learning_rate": 1.7295678093442155e-05, |
| "loss": 0.9435, |
| "step": 504 |
| }, |
| { |
| "epoch": 1.5794831636648394, |
| "grad_norm": 0.795159470143028, |
| "learning_rate": 1.728068806524712e-05, |
| "loss": 1.0071, |
| "step": 505 |
| }, |
| { |
| "epoch": 1.5826155050900548, |
| "grad_norm": 0.7963785520950486, |
| "learning_rate": 1.7265663141651016e-05, |
| "loss": 0.9922, |
| "step": 506 |
| }, |
| { |
| "epoch": 1.5857478465152701, |
| "grad_norm": 0.7877046535940556, |
| "learning_rate": 1.7250603394666364e-05, |
| "loss": 0.9514, |
| "step": 507 |
| }, |
| { |
| "epoch": 1.5888801879404855, |
| "grad_norm": 0.7715293903451117, |
| "learning_rate": 1.7235508896472593e-05, |
| "loss": 0.933, |
| "step": 508 |
| }, |
| { |
| "epoch": 1.5920125293657008, |
| "grad_norm": 0.7934117596728014, |
| "learning_rate": 1.72203797194157e-05, |
| "loss": 0.973, |
| "step": 509 |
| }, |
| { |
| "epoch": 1.5951448707909162, |
| "grad_norm": 0.8181710440792368, |
| "learning_rate": 1.720521593600787e-05, |
| "loss": 0.986, |
| "step": 510 |
| }, |
| { |
| "epoch": 1.5982772122161315, |
| "grad_norm": 0.7837650575509727, |
| "learning_rate": 1.7190017618927177e-05, |
| "loss": 0.9612, |
| "step": 511 |
| }, |
| { |
| "epoch": 1.6014095536413469, |
| "grad_norm": 0.8412641440444018, |
| "learning_rate": 1.7174784841017197e-05, |
| "loss": 0.9745, |
| "step": 512 |
| }, |
| { |
| "epoch": 1.6045418950665624, |
| "grad_norm": 0.7881270804166439, |
| "learning_rate": 1.7159517675286667e-05, |
| "loss": 0.9644, |
| "step": 513 |
| }, |
| { |
| "epoch": 1.6076742364917775, |
| "grad_norm": 0.8119820827446707, |
| "learning_rate": 1.7144216194909157e-05, |
| "loss": 0.9886, |
| "step": 514 |
| }, |
| { |
| "epoch": 1.610806577916993, |
| "grad_norm": 0.7998253925780444, |
| "learning_rate": 1.7128880473222688e-05, |
| "loss": 0.9815, |
| "step": 515 |
| }, |
| { |
| "epoch": 1.6139389193422082, |
| "grad_norm": 0.7739056976249491, |
| "learning_rate": 1.7113510583729403e-05, |
| "loss": 0.9766, |
| "step": 516 |
| }, |
| { |
| "epoch": 1.6170712607674238, |
| "grad_norm": 0.8069370636653986, |
| "learning_rate": 1.7098106600095204e-05, |
| "loss": 0.9496, |
| "step": 517 |
| }, |
| { |
| "epoch": 1.620203602192639, |
| "grad_norm": 0.8037770500974222, |
| "learning_rate": 1.7082668596149402e-05, |
| "loss": 0.9646, |
| "step": 518 |
| }, |
| { |
| "epoch": 1.6233359436178545, |
| "grad_norm": 0.7878273006407442, |
| "learning_rate": 1.706719664588437e-05, |
| "loss": 0.9569, |
| "step": 519 |
| }, |
| { |
| "epoch": 1.6264682850430696, |
| "grad_norm": 0.7827154823147211, |
| "learning_rate": 1.7051690823455162e-05, |
| "loss": 0.9405, |
| "step": 520 |
| }, |
| { |
| "epoch": 1.6296006264682852, |
| "grad_norm": 0.8114142283057644, |
| "learning_rate": 1.70361512031792e-05, |
| "loss": 0.9804, |
| "step": 521 |
| }, |
| { |
| "epoch": 1.6327329678935003, |
| "grad_norm": 0.7889387459740889, |
| "learning_rate": 1.702057785953588e-05, |
| "loss": 0.9405, |
| "step": 522 |
| }, |
| { |
| "epoch": 1.6358653093187159, |
| "grad_norm": 0.8311030879338528, |
| "learning_rate": 1.7004970867166238e-05, |
| "loss": 0.9593, |
| "step": 523 |
| }, |
| { |
| "epoch": 1.638997650743931, |
| "grad_norm": 0.7984284134204775, |
| "learning_rate": 1.6989330300872576e-05, |
| "loss": 0.951, |
| "step": 524 |
| }, |
| { |
| "epoch": 1.6421299921691466, |
| "grad_norm": 0.7997161389779351, |
| "learning_rate": 1.6973656235618114e-05, |
| "loss": 0.9379, |
| "step": 525 |
| }, |
| { |
| "epoch": 1.6452623335943617, |
| "grad_norm": 0.7928450659072758, |
| "learning_rate": 1.6957948746526637e-05, |
| "loss": 0.9796, |
| "step": 526 |
| }, |
| { |
| "epoch": 1.6483946750195773, |
| "grad_norm": 0.7864284356410199, |
| "learning_rate": 1.6942207908882115e-05, |
| "loss": 0.9595, |
| "step": 527 |
| }, |
| { |
| "epoch": 1.6515270164447924, |
| "grad_norm": 0.7748245207452046, |
| "learning_rate": 1.692643379812836e-05, |
| "loss": 0.9363, |
| "step": 528 |
| }, |
| { |
| "epoch": 1.654659357870008, |
| "grad_norm": 0.7904035921159812, |
| "learning_rate": 1.691062648986865e-05, |
| "loss": 0.9698, |
| "step": 529 |
| }, |
| { |
| "epoch": 1.657791699295223, |
| "grad_norm": 0.7699380547476845, |
| "learning_rate": 1.6894786059865383e-05, |
| "loss": 0.9229, |
| "step": 530 |
| }, |
| { |
| "epoch": 1.6609240407204386, |
| "grad_norm": 0.756885007762418, |
| "learning_rate": 1.6878912584039698e-05, |
| "loss": 0.9664, |
| "step": 531 |
| }, |
| { |
| "epoch": 1.6640563821456538, |
| "grad_norm": 0.8314070737417165, |
| "learning_rate": 1.686300613847113e-05, |
| "loss": 0.9685, |
| "step": 532 |
| }, |
| { |
| "epoch": 1.6671887235708693, |
| "grad_norm": 0.8339161427130307, |
| "learning_rate": 1.6847066799397224e-05, |
| "loss": 0.9272, |
| "step": 533 |
| }, |
| { |
| "epoch": 1.6703210649960845, |
| "grad_norm": 0.8349954422927233, |
| "learning_rate": 1.683109464321317e-05, |
| "loss": 0.9764, |
| "step": 534 |
| }, |
| { |
| "epoch": 1.6734534064213, |
| "grad_norm": 0.7964469038876641, |
| "learning_rate": 1.6815089746471472e-05, |
| "loss": 1.0066, |
| "step": 535 |
| }, |
| { |
| "epoch": 1.6765857478465152, |
| "grad_norm": 0.7854281643639673, |
| "learning_rate": 1.6799052185881535e-05, |
| "loss": 0.9901, |
| "step": 536 |
| }, |
| { |
| "epoch": 1.6797180892717307, |
| "grad_norm": 0.7974377816831979, |
| "learning_rate": 1.678298203830932e-05, |
| "loss": 0.969, |
| "step": 537 |
| }, |
| { |
| "epoch": 1.6828504306969458, |
| "grad_norm": 0.7734247111191668, |
| "learning_rate": 1.6766879380776983e-05, |
| "loss": 0.9183, |
| "step": 538 |
| }, |
| { |
| "epoch": 1.6859827721221614, |
| "grad_norm": 0.7724972901824376, |
| "learning_rate": 1.6750744290462484e-05, |
| "loss": 0.9186, |
| "step": 539 |
| }, |
| { |
| "epoch": 1.6891151135473765, |
| "grad_norm": 0.7968886343195833, |
| "learning_rate": 1.6734576844699234e-05, |
| "loss": 0.9409, |
| "step": 540 |
| }, |
| { |
| "epoch": 1.692247454972592, |
| "grad_norm": 0.8009596190830679, |
| "learning_rate": 1.671837712097573e-05, |
| "loss": 0.9639, |
| "step": 541 |
| }, |
| { |
| "epoch": 1.6953797963978072, |
| "grad_norm": 0.8367539665803249, |
| "learning_rate": 1.670214519693515e-05, |
| "loss": 0.9655, |
| "step": 542 |
| }, |
| { |
| "epoch": 1.6985121378230228, |
| "grad_norm": 0.8071935125435302, |
| "learning_rate": 1.6685881150375027e-05, |
| "loss": 0.9635, |
| "step": 543 |
| }, |
| { |
| "epoch": 1.701644479248238, |
| "grad_norm": 0.8293405021567984, |
| "learning_rate": 1.6669585059246837e-05, |
| "loss": 0.9681, |
| "step": 544 |
| }, |
| { |
| "epoch": 1.7047768206734535, |
| "grad_norm": 0.8081858624977237, |
| "learning_rate": 1.6653257001655652e-05, |
| "loss": 0.9918, |
| "step": 545 |
| }, |
| { |
| "epoch": 1.7079091620986686, |
| "grad_norm": 0.8057738404404446, |
| "learning_rate": 1.6636897055859754e-05, |
| "loss": 0.9727, |
| "step": 546 |
| }, |
| { |
| "epoch": 1.7110415035238842, |
| "grad_norm": 0.7493081094786209, |
| "learning_rate": 1.6620505300270254e-05, |
| "loss": 0.923, |
| "step": 547 |
| }, |
| { |
| "epoch": 1.7141738449490993, |
| "grad_norm": 0.7989097981932453, |
| "learning_rate": 1.6604081813450724e-05, |
| "loss": 0.9529, |
| "step": 548 |
| }, |
| { |
| "epoch": 1.7173061863743149, |
| "grad_norm": 0.7713974329610848, |
| "learning_rate": 1.6587626674116827e-05, |
| "loss": 0.962, |
| "step": 549 |
| }, |
| { |
| "epoch": 1.72043852779953, |
| "grad_norm": 0.8015438391238155, |
| "learning_rate": 1.657113996113593e-05, |
| "loss": 0.9886, |
| "step": 550 |
| }, |
| { |
| "epoch": 1.7235708692247456, |
| "grad_norm": 0.8065266518656921, |
| "learning_rate": 1.6554621753526724e-05, |
| "loss": 0.9868, |
| "step": 551 |
| }, |
| { |
| "epoch": 1.7267032106499607, |
| "grad_norm": 0.8310150242959837, |
| "learning_rate": 1.6538072130458853e-05, |
| "loss": 0.9989, |
| "step": 552 |
| }, |
| { |
| "epoch": 1.7298355520751763, |
| "grad_norm": 0.7832374226495794, |
| "learning_rate": 1.652149117125253e-05, |
| "loss": 0.9472, |
| "step": 553 |
| }, |
| { |
| "epoch": 1.7329678935003914, |
| "grad_norm": 0.7885410215851165, |
| "learning_rate": 1.6504878955378166e-05, |
| "loss": 0.9987, |
| "step": 554 |
| }, |
| { |
| "epoch": 1.736100234925607, |
| "grad_norm": 0.8206660199757921, |
| "learning_rate": 1.6488235562455964e-05, |
| "loss": 0.9761, |
| "step": 555 |
| }, |
| { |
| "epoch": 1.7392325763508223, |
| "grad_norm": 0.7997233241816074, |
| "learning_rate": 1.6471561072255576e-05, |
| "loss": 0.9768, |
| "step": 556 |
| }, |
| { |
| "epoch": 1.7423649177760376, |
| "grad_norm": 0.7936840487628978, |
| "learning_rate": 1.6454855564695674e-05, |
| "loss": 0.9434, |
| "step": 557 |
| }, |
| { |
| "epoch": 1.745497259201253, |
| "grad_norm": 0.7829379543649524, |
| "learning_rate": 1.643811911984362e-05, |
| "loss": 0.951, |
| "step": 558 |
| }, |
| { |
| "epoch": 1.7486296006264683, |
| "grad_norm": 0.7854616494314223, |
| "learning_rate": 1.6421351817915025e-05, |
| "loss": 0.9417, |
| "step": 559 |
| }, |
| { |
| "epoch": 1.7517619420516837, |
| "grad_norm": 0.7885222818513928, |
| "learning_rate": 1.6404553739273426e-05, |
| "loss": 0.9678, |
| "step": 560 |
| }, |
| { |
| "epoch": 1.754894283476899, |
| "grad_norm": 0.7691482528488672, |
| "learning_rate": 1.6387724964429845e-05, |
| "loss": 0.9235, |
| "step": 561 |
| }, |
| { |
| "epoch": 1.7580266249021144, |
| "grad_norm": 0.7730839050185739, |
| "learning_rate": 1.6370865574042435e-05, |
| "loss": 0.9821, |
| "step": 562 |
| }, |
| { |
| "epoch": 1.7611589663273297, |
| "grad_norm": 0.7879677074806722, |
| "learning_rate": 1.6353975648916086e-05, |
| "loss": 0.9699, |
| "step": 563 |
| }, |
| { |
| "epoch": 1.764291307752545, |
| "grad_norm": 0.7858596514917898, |
| "learning_rate": 1.633705527000203e-05, |
| "loss": 0.994, |
| "step": 564 |
| }, |
| { |
| "epoch": 1.7674236491777604, |
| "grad_norm": 0.8012498378641922, |
| "learning_rate": 1.6320104518397473e-05, |
| "loss": 0.967, |
| "step": 565 |
| }, |
| { |
| "epoch": 1.7705559906029757, |
| "grad_norm": 0.8188666888354701, |
| "learning_rate": 1.6303123475345182e-05, |
| "loss": 0.9785, |
| "step": 566 |
| }, |
| { |
| "epoch": 1.773688332028191, |
| "grad_norm": 0.7981343235293162, |
| "learning_rate": 1.6286112222233113e-05, |
| "loss": 0.975, |
| "step": 567 |
| }, |
| { |
| "epoch": 1.7768206734534064, |
| "grad_norm": 0.7838444823903047, |
| "learning_rate": 1.6269070840594017e-05, |
| "loss": 0.9458, |
| "step": 568 |
| }, |
| { |
| "epoch": 1.7799530148786218, |
| "grad_norm": 0.7846905745122678, |
| "learning_rate": 1.6251999412105036e-05, |
| "loss": 0.9298, |
| "step": 569 |
| }, |
| { |
| "epoch": 1.7830853563038371, |
| "grad_norm": 0.8149058395623431, |
| "learning_rate": 1.6234898018587336e-05, |
| "loss": 0.9436, |
| "step": 570 |
| }, |
| { |
| "epoch": 1.7862176977290525, |
| "grad_norm": 0.7941672166648026, |
| "learning_rate": 1.6217766742005693e-05, |
| "loss": 0.9985, |
| "step": 571 |
| }, |
| { |
| "epoch": 1.7893500391542678, |
| "grad_norm": 0.7710930788046828, |
| "learning_rate": 1.6200605664468113e-05, |
| "loss": 0.97, |
| "step": 572 |
| }, |
| { |
| "epoch": 1.7924823805794832, |
| "grad_norm": 0.7957133349582347, |
| "learning_rate": 1.6183414868225434e-05, |
| "loss": 0.9959, |
| "step": 573 |
| }, |
| { |
| "epoch": 1.7956147220046985, |
| "grad_norm": 0.79291459979055, |
| "learning_rate": 1.616619443567092e-05, |
| "loss": 0.9442, |
| "step": 574 |
| }, |
| { |
| "epoch": 1.7987470634299139, |
| "grad_norm": 0.785169568406983, |
| "learning_rate": 1.6148944449339904e-05, |
| "loss": 0.9639, |
| "step": 575 |
| }, |
| { |
| "epoch": 1.8018794048551292, |
| "grad_norm": 0.7997066561831516, |
| "learning_rate": 1.6131664991909335e-05, |
| "loss": 1.0127, |
| "step": 576 |
| }, |
| { |
| "epoch": 1.8050117462803446, |
| "grad_norm": 0.8141399391865273, |
| "learning_rate": 1.6114356146197445e-05, |
| "loss": 0.9386, |
| "step": 577 |
| }, |
| { |
| "epoch": 1.80814408770556, |
| "grad_norm": 0.8097117540881008, |
| "learning_rate": 1.6097017995163294e-05, |
| "loss": 0.9774, |
| "step": 578 |
| }, |
| { |
| "epoch": 1.8112764291307752, |
| "grad_norm": 0.7730781490152179, |
| "learning_rate": 1.607965062190641e-05, |
| "loss": 0.9453, |
| "step": 579 |
| }, |
| { |
| "epoch": 1.8144087705559906, |
| "grad_norm": 0.7838923521873067, |
| "learning_rate": 1.6062254109666383e-05, |
| "loss": 0.9628, |
| "step": 580 |
| }, |
| { |
| "epoch": 1.817541111981206, |
| "grad_norm": 0.790301862199868, |
| "learning_rate": 1.604482854182245e-05, |
| "loss": 0.9068, |
| "step": 581 |
| }, |
| { |
| "epoch": 1.8206734534064213, |
| "grad_norm": 0.7750930497911785, |
| "learning_rate": 1.6027374001893123e-05, |
| "loss": 0.9927, |
| "step": 582 |
| }, |
| { |
| "epoch": 1.8238057948316366, |
| "grad_norm": 0.7779386940305499, |
| "learning_rate": 1.6009890573535763e-05, |
| "loss": 0.919, |
| "step": 583 |
| }, |
| { |
| "epoch": 1.826938136256852, |
| "grad_norm": 0.7992055706023535, |
| "learning_rate": 1.5992378340546194e-05, |
| "loss": 0.9624, |
| "step": 584 |
| }, |
| { |
| "epoch": 1.8300704776820673, |
| "grad_norm": 0.8153176494752826, |
| "learning_rate": 1.597483738685829e-05, |
| "loss": 1.0051, |
| "step": 585 |
| }, |
| { |
| "epoch": 1.8332028191072827, |
| "grad_norm": 0.7936500253751112, |
| "learning_rate": 1.595726779654359e-05, |
| "loss": 0.9692, |
| "step": 586 |
| }, |
| { |
| "epoch": 1.836335160532498, |
| "grad_norm": 0.8140127238131111, |
| "learning_rate": 1.5939669653810882e-05, |
| "loss": 0.9763, |
| "step": 587 |
| }, |
| { |
| "epoch": 1.8394675019577134, |
| "grad_norm": 0.8084565243659966, |
| "learning_rate": 1.592204304300579e-05, |
| "loss": 0.9402, |
| "step": 588 |
| }, |
| { |
| "epoch": 1.8425998433829287, |
| "grad_norm": 0.7527144155069814, |
| "learning_rate": 1.59043880486104e-05, |
| "loss": 0.9255, |
| "step": 589 |
| }, |
| { |
| "epoch": 1.845732184808144, |
| "grad_norm": 0.7985587310153884, |
| "learning_rate": 1.588670475524283e-05, |
| "loss": 0.9465, |
| "step": 590 |
| }, |
| { |
| "epoch": 1.8488645262333594, |
| "grad_norm": 0.7524132760020612, |
| "learning_rate": 1.5868993247656822e-05, |
| "loss": 0.9865, |
| "step": 591 |
| }, |
| { |
| "epoch": 1.8519968676585747, |
| "grad_norm": 0.7823183770981433, |
| "learning_rate": 1.5851253610741357e-05, |
| "loss": 0.8974, |
| "step": 592 |
| }, |
| { |
| "epoch": 1.85512920908379, |
| "grad_norm": 0.769457387825749, |
| "learning_rate": 1.583348592952023e-05, |
| "loss": 0.9776, |
| "step": 593 |
| }, |
| { |
| "epoch": 1.8582615505090054, |
| "grad_norm": 0.7714656018935003, |
| "learning_rate": 1.581569028915166e-05, |
| "loss": 0.9399, |
| "step": 594 |
| }, |
| { |
| "epoch": 1.8613938919342208, |
| "grad_norm": 0.7784139001699675, |
| "learning_rate": 1.5797866774927848e-05, |
| "loss": 0.9699, |
| "step": 595 |
| }, |
| { |
| "epoch": 1.8645262333594361, |
| "grad_norm": 0.7797773255228968, |
| "learning_rate": 1.5780015472274613e-05, |
| "loss": 0.9404, |
| "step": 596 |
| }, |
| { |
| "epoch": 1.8676585747846515, |
| "grad_norm": 0.7758341272190327, |
| "learning_rate": 1.5762136466750947e-05, |
| "loss": 0.9719, |
| "step": 597 |
| }, |
| { |
| "epoch": 1.870790916209867, |
| "grad_norm": 0.7786162669677469, |
| "learning_rate": 1.5744229844048627e-05, |
| "loss": 0.9436, |
| "step": 598 |
| }, |
| { |
| "epoch": 1.8739232576350822, |
| "grad_norm": 0.7543252031375369, |
| "learning_rate": 1.5726295689991787e-05, |
| "loss": 0.9784, |
| "step": 599 |
| }, |
| { |
| "epoch": 1.8770555990602977, |
| "grad_norm": 0.7757844187122288, |
| "learning_rate": 1.570833409053653e-05, |
| "loss": 0.9713, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.8801879404855129, |
| "grad_norm": 0.7653394710213172, |
| "learning_rate": 1.5690345131770474e-05, |
| "loss": 0.8965, |
| "step": 601 |
| }, |
| { |
| "epoch": 1.8833202819107284, |
| "grad_norm": 0.7896033443700016, |
| "learning_rate": 1.5672328899912397e-05, |
| "loss": 1.0193, |
| "step": 602 |
| }, |
| { |
| "epoch": 1.8864526233359435, |
| "grad_norm": 0.7846649604330512, |
| "learning_rate": 1.5654285481311777e-05, |
| "loss": 0.9474, |
| "step": 603 |
| }, |
| { |
| "epoch": 1.8895849647611591, |
| "grad_norm": 0.8135389235047997, |
| "learning_rate": 1.5636214962448396e-05, |
| "loss": 0.9618, |
| "step": 604 |
| }, |
| { |
| "epoch": 1.8927173061863742, |
| "grad_norm": 0.7824047803609824, |
| "learning_rate": 1.5618117429931923e-05, |
| "loss": 0.9408, |
| "step": 605 |
| }, |
| { |
| "epoch": 1.8958496476115898, |
| "grad_norm": 0.8254165694685803, |
| "learning_rate": 1.5599992970501512e-05, |
| "loss": 0.9767, |
| "step": 606 |
| }, |
| { |
| "epoch": 1.898981989036805, |
| "grad_norm": 0.7942279881179067, |
| "learning_rate": 1.5581841671025358e-05, |
| "loss": 0.9657, |
| "step": 607 |
| }, |
| { |
| "epoch": 1.9021143304620205, |
| "grad_norm": 0.7607717329156005, |
| "learning_rate": 1.55636636185003e-05, |
| "loss": 0.9641, |
| "step": 608 |
| }, |
| { |
| "epoch": 1.9052466718872356, |
| "grad_norm": 0.7992715938950239, |
| "learning_rate": 1.5545458900051417e-05, |
| "loss": 0.9466, |
| "step": 609 |
| }, |
| { |
| "epoch": 1.9083790133124512, |
| "grad_norm": 0.7710025331360174, |
| "learning_rate": 1.552722760293157e-05, |
| "loss": 0.9349, |
| "step": 610 |
| }, |
| { |
| "epoch": 1.9115113547376663, |
| "grad_norm": 0.7995999712279581, |
| "learning_rate": 1.5508969814521026e-05, |
| "loss": 0.9784, |
| "step": 611 |
| }, |
| { |
| "epoch": 1.9146436961628819, |
| "grad_norm": 0.7972698916237193, |
| "learning_rate": 1.5490685622327015e-05, |
| "loss": 0.9954, |
| "step": 612 |
| }, |
| { |
| "epoch": 1.917776037588097, |
| "grad_norm": 0.7658223645888151, |
| "learning_rate": 1.547237511398332e-05, |
| "loss": 0.931, |
| "step": 613 |
| }, |
| { |
| "epoch": 1.9209083790133126, |
| "grad_norm": 0.7891075170093855, |
| "learning_rate": 1.5454038377249845e-05, |
| "loss": 0.9559, |
| "step": 614 |
| }, |
| { |
| "epoch": 1.9240407204385277, |
| "grad_norm": 0.8134205306266102, |
| "learning_rate": 1.5435675500012212e-05, |
| "loss": 0.9907, |
| "step": 615 |
| }, |
| { |
| "epoch": 1.9271730618637433, |
| "grad_norm": 0.803815630288296, |
| "learning_rate": 1.5417286570281334e-05, |
| "loss": 0.9834, |
| "step": 616 |
| }, |
| { |
| "epoch": 1.9303054032889584, |
| "grad_norm": 0.771918493512692, |
| "learning_rate": 1.5398871676192978e-05, |
| "loss": 1.0094, |
| "step": 617 |
| }, |
| { |
| "epoch": 1.933437744714174, |
| "grad_norm": 0.793008120980106, |
| "learning_rate": 1.538043090600737e-05, |
| "loss": 0.9266, |
| "step": 618 |
| }, |
| { |
| "epoch": 1.936570086139389, |
| "grad_norm": 0.7778816113808421, |
| "learning_rate": 1.5361964348108736e-05, |
| "loss": 0.9665, |
| "step": 619 |
| }, |
| { |
| "epoch": 1.9397024275646046, |
| "grad_norm": 0.7989827850535249, |
| "learning_rate": 1.5343472091004925e-05, |
| "loss": 0.9524, |
| "step": 620 |
| }, |
| { |
| "epoch": 1.9428347689898198, |
| "grad_norm": 0.813278631018117, |
| "learning_rate": 1.532495422332694e-05, |
| "loss": 1.0166, |
| "step": 621 |
| }, |
| { |
| "epoch": 1.9459671104150353, |
| "grad_norm": 0.7855425028524953, |
| "learning_rate": 1.5306410833828534e-05, |
| "loss": 0.9816, |
| "step": 622 |
| }, |
| { |
| "epoch": 1.9490994518402505, |
| "grad_norm": 0.7943986605440319, |
| "learning_rate": 1.5287842011385798e-05, |
| "loss": 0.9852, |
| "step": 623 |
| }, |
| { |
| "epoch": 1.952231793265466, |
| "grad_norm": 0.8316767537841315, |
| "learning_rate": 1.52692478449967e-05, |
| "loss": 0.9811, |
| "step": 624 |
| }, |
| { |
| "epoch": 1.9553641346906812, |
| "grad_norm": 0.8079701202216203, |
| "learning_rate": 1.5250628423780684e-05, |
| "loss": 0.989, |
| "step": 625 |
| }, |
| { |
| "epoch": 1.9584964761158967, |
| "grad_norm": 0.8001611635685686, |
| "learning_rate": 1.523198383697825e-05, |
| "loss": 0.9859, |
| "step": 626 |
| }, |
| { |
| "epoch": 1.9616288175411118, |
| "grad_norm": 0.7922734841320317, |
| "learning_rate": 1.5213314173950498e-05, |
| "loss": 0.9785, |
| "step": 627 |
| }, |
| { |
| "epoch": 1.9647611589663274, |
| "grad_norm": 0.7833384810882262, |
| "learning_rate": 1.5194619524178718e-05, |
| "loss": 0.9407, |
| "step": 628 |
| }, |
| { |
| "epoch": 1.9678935003915425, |
| "grad_norm": 0.8098242578519478, |
| "learning_rate": 1.5175899977263963e-05, |
| "loss": 0.9224, |
| "step": 629 |
| }, |
| { |
| "epoch": 1.971025841816758, |
| "grad_norm": 0.8182996603065131, |
| "learning_rate": 1.515715562292662e-05, |
| "loss": 0.9622, |
| "step": 630 |
| }, |
| { |
| "epoch": 1.9741581832419732, |
| "grad_norm": 0.7901742990362269, |
| "learning_rate": 1.513838655100596e-05, |
| "loss": 0.9337, |
| "step": 631 |
| }, |
| { |
| "epoch": 1.9772905246671888, |
| "grad_norm": 0.7615944653532644, |
| "learning_rate": 1.5119592851459726e-05, |
| "loss": 0.9414, |
| "step": 632 |
| }, |
| { |
| "epoch": 1.980422866092404, |
| "grad_norm": 0.7946500220579569, |
| "learning_rate": 1.5100774614363708e-05, |
| "loss": 0.9512, |
| "step": 633 |
| }, |
| { |
| "epoch": 1.9835552075176195, |
| "grad_norm": 0.8076792545370913, |
| "learning_rate": 1.5081931929911298e-05, |
| "loss": 0.973, |
| "step": 634 |
| }, |
| { |
| "epoch": 1.9866875489428346, |
| "grad_norm": 0.7861723773682193, |
| "learning_rate": 1.5063064888413048e-05, |
| "loss": 0.9842, |
| "step": 635 |
| }, |
| { |
| "epoch": 1.9898198903680502, |
| "grad_norm": 0.7955732049686428, |
| "learning_rate": 1.5044173580296267e-05, |
| "loss": 0.9272, |
| "step": 636 |
| }, |
| { |
| "epoch": 1.9929522317932653, |
| "grad_norm": 0.8037074259104876, |
| "learning_rate": 1.5025258096104563e-05, |
| "loss": 0.9697, |
| "step": 637 |
| }, |
| { |
| "epoch": 1.9960845732184809, |
| "grad_norm": 0.8103505085995998, |
| "learning_rate": 1.500631852649742e-05, |
| "loss": 0.948, |
| "step": 638 |
| }, |
| { |
| "epoch": 1.999216914643696, |
| "grad_norm": 0.7802553935331468, |
| "learning_rate": 1.4987354962249749e-05, |
| "loss": 0.9517, |
| "step": 639 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.7802553935331468, |
| "learning_rate": 1.4968367494251486e-05, |
| "loss": 0.9289, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.0093970242756463, |
| "grad_norm": 1.3210123581817999, |
| "learning_rate": 1.4949356213507113e-05, |
| "loss": 0.7754, |
| "step": 641 |
| }, |
| { |
| "epoch": 2.0125293657008614, |
| "grad_norm": 1.3238346751721688, |
| "learning_rate": 1.4930321211135258e-05, |
| "loss": 0.7855, |
| "step": 642 |
| }, |
| { |
| "epoch": 2.015661707126077, |
| "grad_norm": 1.1616452207714454, |
| "learning_rate": 1.4911262578368233e-05, |
| "loss": 0.7959, |
| "step": 643 |
| }, |
| { |
| "epoch": 2.018794048551292, |
| "grad_norm": 0.9412974130528684, |
| "learning_rate": 1.4892180406551618e-05, |
| "loss": 0.7173, |
| "step": 644 |
| }, |
| { |
| "epoch": 2.0219263899765076, |
| "grad_norm": 1.3175139050237945, |
| "learning_rate": 1.48730747871438e-05, |
| "loss": 0.7316, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.0250587314017228, |
| "grad_norm": 1.5954314046677502, |
| "learning_rate": 1.4853945811715563e-05, |
| "loss": 0.7542, |
| "step": 646 |
| }, |
| { |
| "epoch": 2.0281910728269383, |
| "grad_norm": 1.37008550444395, |
| "learning_rate": 1.4834793571949617e-05, |
| "loss": 0.7587, |
| "step": 647 |
| }, |
| { |
| "epoch": 2.0313234142521535, |
| "grad_norm": 1.169303068811387, |
| "learning_rate": 1.4815618159640188e-05, |
| "loss": 0.7195, |
| "step": 648 |
| }, |
| { |
| "epoch": 2.034455755677369, |
| "grad_norm": 1.120757117831837, |
| "learning_rate": 1.4796419666692555e-05, |
| "loss": 0.6884, |
| "step": 649 |
| }, |
| { |
| "epoch": 2.037588097102584, |
| "grad_norm": 1.1374034296234556, |
| "learning_rate": 1.477719818512263e-05, |
| "loss": 0.7608, |
| "step": 650 |
| }, |
| { |
| "epoch": 2.0407204385277997, |
| "grad_norm": 1.018857278719166, |
| "learning_rate": 1.47579538070565e-05, |
| "loss": 0.7136, |
| "step": 651 |
| }, |
| { |
| "epoch": 2.043852779953015, |
| "grad_norm": 0.991608010520256, |
| "learning_rate": 1.4738686624729987e-05, |
| "loss": 0.7105, |
| "step": 652 |
| }, |
| { |
| "epoch": 2.0469851213782304, |
| "grad_norm": 1.0388281087329523, |
| "learning_rate": 1.4719396730488225e-05, |
| "loss": 0.7348, |
| "step": 653 |
| }, |
| { |
| "epoch": 2.0501174628034455, |
| "grad_norm": 1.0171773999697262, |
| "learning_rate": 1.4700084216785189e-05, |
| "loss": 0.7212, |
| "step": 654 |
| }, |
| { |
| "epoch": 2.053249804228661, |
| "grad_norm": 0.9712484652548604, |
| "learning_rate": 1.4680749176183275e-05, |
| "loss": 0.7152, |
| "step": 655 |
| }, |
| { |
| "epoch": 2.0563821456538762, |
| "grad_norm": 0.9445446614613634, |
| "learning_rate": 1.4661391701352844e-05, |
| "loss": 0.7144, |
| "step": 656 |
| }, |
| { |
| "epoch": 2.059514487079092, |
| "grad_norm": 0.9945223666447938, |
| "learning_rate": 1.464201188507178e-05, |
| "loss": 0.7119, |
| "step": 657 |
| }, |
| { |
| "epoch": 2.062646828504307, |
| "grad_norm": 0.9677117020420452, |
| "learning_rate": 1.4622609820225052e-05, |
| "loss": 0.7148, |
| "step": 658 |
| }, |
| { |
| "epoch": 2.0657791699295225, |
| "grad_norm": 1.0137018942940528, |
| "learning_rate": 1.4603185599804259e-05, |
| "loss": 0.7435, |
| "step": 659 |
| }, |
| { |
| "epoch": 2.0689115113547376, |
| "grad_norm": 1.02511003538228, |
| "learning_rate": 1.4583739316907188e-05, |
| "loss": 0.6999, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.072043852779953, |
| "grad_norm": 0.9904700091988614, |
| "learning_rate": 1.4564271064737376e-05, |
| "loss": 0.7449, |
| "step": 661 |
| }, |
| { |
| "epoch": 2.0751761942051683, |
| "grad_norm": 0.9279680631763316, |
| "learning_rate": 1.4544780936603644e-05, |
| "loss": 0.7023, |
| "step": 662 |
| }, |
| { |
| "epoch": 2.078308535630384, |
| "grad_norm": 0.9287114111608203, |
| "learning_rate": 1.4525269025919677e-05, |
| "loss": 0.7163, |
| "step": 663 |
| }, |
| { |
| "epoch": 2.081440877055599, |
| "grad_norm": 0.943750195528307, |
| "learning_rate": 1.4505735426203545e-05, |
| "loss": 0.7377, |
| "step": 664 |
| }, |
| { |
| "epoch": 2.0845732184808146, |
| "grad_norm": 0.9053784915303666, |
| "learning_rate": 1.4486180231077278e-05, |
| "loss": 0.7282, |
| "step": 665 |
| }, |
| { |
| "epoch": 2.0877055599060297, |
| "grad_norm": 0.955797509257669, |
| "learning_rate": 1.446660353426641e-05, |
| "loss": 0.7668, |
| "step": 666 |
| }, |
| { |
| "epoch": 2.0908379013312453, |
| "grad_norm": 0.9130054155780589, |
| "learning_rate": 1.4447005429599537e-05, |
| "loss": 0.695, |
| "step": 667 |
| }, |
| { |
| "epoch": 2.0939702427564604, |
| "grad_norm": 0.9516666064010463, |
| "learning_rate": 1.4427386011007847e-05, |
| "loss": 0.7255, |
| "step": 668 |
| }, |
| { |
| "epoch": 2.097102584181676, |
| "grad_norm": 0.9546422449504612, |
| "learning_rate": 1.4407745372524686e-05, |
| "loss": 0.7224, |
| "step": 669 |
| }, |
| { |
| "epoch": 2.100234925606891, |
| "grad_norm": 0.9448475040935651, |
| "learning_rate": 1.4388083608285112e-05, |
| "loss": 0.7222, |
| "step": 670 |
| }, |
| { |
| "epoch": 2.1033672670321066, |
| "grad_norm": 0.9222655058113182, |
| "learning_rate": 1.4368400812525434e-05, |
| "loss": 0.7251, |
| "step": 671 |
| }, |
| { |
| "epoch": 2.1064996084573218, |
| "grad_norm": 0.9777941043274174, |
| "learning_rate": 1.4348697079582752e-05, |
| "loss": 0.7703, |
| "step": 672 |
| }, |
| { |
| "epoch": 2.1096319498825373, |
| "grad_norm": 0.9749998119435571, |
| "learning_rate": 1.4328972503894525e-05, |
| "loss": 0.7574, |
| "step": 673 |
| }, |
| { |
| "epoch": 2.1127642913077525, |
| "grad_norm": 0.9756161103337109, |
| "learning_rate": 1.4309227179998109e-05, |
| "loss": 0.7646, |
| "step": 674 |
| }, |
| { |
| "epoch": 2.115896632732968, |
| "grad_norm": 0.9561100629580807, |
| "learning_rate": 1.4289461202530297e-05, |
| "loss": 0.7504, |
| "step": 675 |
| }, |
| { |
| "epoch": 2.119028974158183, |
| "grad_norm": 0.9458044599682323, |
| "learning_rate": 1.4269674666226879e-05, |
| "loss": 0.7406, |
| "step": 676 |
| }, |
| { |
| "epoch": 2.1221613155833987, |
| "grad_norm": 0.9193297731491735, |
| "learning_rate": 1.4249867665922175e-05, |
| "loss": 0.6953, |
| "step": 677 |
| }, |
| { |
| "epoch": 2.125293657008614, |
| "grad_norm": 0.938231383030172, |
| "learning_rate": 1.4230040296548588e-05, |
| "loss": 0.7742, |
| "step": 678 |
| }, |
| { |
| "epoch": 2.1284259984338294, |
| "grad_norm": 0.9608218431580744, |
| "learning_rate": 1.4210192653136151e-05, |
| "loss": 0.745, |
| "step": 679 |
| }, |
| { |
| "epoch": 2.1315583398590445, |
| "grad_norm": 0.9131562406248528, |
| "learning_rate": 1.4190324830812067e-05, |
| "loss": 0.706, |
| "step": 680 |
| }, |
| { |
| "epoch": 2.13469068128426, |
| "grad_norm": 0.9330499851170364, |
| "learning_rate": 1.4170436924800251e-05, |
| "loss": 0.7198, |
| "step": 681 |
| }, |
| { |
| "epoch": 2.137823022709475, |
| "grad_norm": 0.9744409362566172, |
| "learning_rate": 1.415052903042087e-05, |
| "loss": 0.748, |
| "step": 682 |
| }, |
| { |
| "epoch": 2.140955364134691, |
| "grad_norm": 0.956855447487281, |
| "learning_rate": 1.4130601243089905e-05, |
| "loss": 0.772, |
| "step": 683 |
| }, |
| { |
| "epoch": 2.144087705559906, |
| "grad_norm": 0.9105854308119294, |
| "learning_rate": 1.4110653658318682e-05, |
| "loss": 0.7066, |
| "step": 684 |
| }, |
| { |
| "epoch": 2.1472200469851215, |
| "grad_norm": 0.9437505928196139, |
| "learning_rate": 1.4090686371713403e-05, |
| "loss": 0.7511, |
| "step": 685 |
| }, |
| { |
| "epoch": 2.1503523884103366, |
| "grad_norm": 0.9754488006918184, |
| "learning_rate": 1.4070699478974698e-05, |
| "loss": 0.7273, |
| "step": 686 |
| }, |
| { |
| "epoch": 2.153484729835552, |
| "grad_norm": 0.9660297851499952, |
| "learning_rate": 1.4050693075897181e-05, |
| "loss": 0.7423, |
| "step": 687 |
| }, |
| { |
| "epoch": 2.1566170712607673, |
| "grad_norm": 0.9710350268049464, |
| "learning_rate": 1.4030667258368968e-05, |
| "loss": 0.7325, |
| "step": 688 |
| }, |
| { |
| "epoch": 2.159749412685983, |
| "grad_norm": 0.9168340841543151, |
| "learning_rate": 1.4010622122371219e-05, |
| "loss": 0.7178, |
| "step": 689 |
| }, |
| { |
| "epoch": 2.162881754111198, |
| "grad_norm": 0.967756101002905, |
| "learning_rate": 1.3990557763977694e-05, |
| "loss": 0.7614, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.1660140955364136, |
| "grad_norm": 0.9340155531023159, |
| "learning_rate": 1.3970474279354283e-05, |
| "loss": 0.7106, |
| "step": 691 |
| }, |
| { |
| "epoch": 2.1691464369616287, |
| "grad_norm": 0.9505636046987749, |
| "learning_rate": 1.3950371764758543e-05, |
| "loss": 0.7243, |
| "step": 692 |
| }, |
| { |
| "epoch": 2.1722787783868442, |
| "grad_norm": 0.9197842895474289, |
| "learning_rate": 1.3930250316539237e-05, |
| "loss": 0.7501, |
| "step": 693 |
| }, |
| { |
| "epoch": 2.1754111198120594, |
| "grad_norm": 0.9424997582167158, |
| "learning_rate": 1.3910110031135884e-05, |
| "loss": 0.7472, |
| "step": 694 |
| }, |
| { |
| "epoch": 2.178543461237275, |
| "grad_norm": 0.9606400745011864, |
| "learning_rate": 1.3889951005078271e-05, |
| "loss": 0.7475, |
| "step": 695 |
| }, |
| { |
| "epoch": 2.18167580266249, |
| "grad_norm": 0.9659045083450463, |
| "learning_rate": 1.3869773334986023e-05, |
| "loss": 0.7488, |
| "step": 696 |
| }, |
| { |
| "epoch": 2.1848081440877056, |
| "grad_norm": 0.984473222173499, |
| "learning_rate": 1.3849577117568115e-05, |
| "loss": 0.7502, |
| "step": 697 |
| }, |
| { |
| "epoch": 2.1879404855129208, |
| "grad_norm": 0.9738775239386929, |
| "learning_rate": 1.3829362449622416e-05, |
| "loss": 0.7482, |
| "step": 698 |
| }, |
| { |
| "epoch": 2.1910728269381363, |
| "grad_norm": 0.9315292619814982, |
| "learning_rate": 1.3809129428035229e-05, |
| "loss": 0.7511, |
| "step": 699 |
| }, |
| { |
| "epoch": 2.1942051683633514, |
| "grad_norm": 0.9347672091944302, |
| "learning_rate": 1.3788878149780827e-05, |
| "loss": 0.7436, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.197337509788567, |
| "grad_norm": 0.9359481225994939, |
| "learning_rate": 1.376860871192098e-05, |
| "loss": 0.7509, |
| "step": 701 |
| }, |
| { |
| "epoch": 2.200469851213782, |
| "grad_norm": 0.9219014695571008, |
| "learning_rate": 1.3748321211604488e-05, |
| "loss": 0.7389, |
| "step": 702 |
| }, |
| { |
| "epoch": 2.2036021926389977, |
| "grad_norm": 0.9880848258280983, |
| "learning_rate": 1.3728015746066736e-05, |
| "loss": 0.7475, |
| "step": 703 |
| }, |
| { |
| "epoch": 2.206734534064213, |
| "grad_norm": 0.9408764680190602, |
| "learning_rate": 1.370769241262921e-05, |
| "loss": 0.7402, |
| "step": 704 |
| }, |
| { |
| "epoch": 2.2098668754894284, |
| "grad_norm": 0.9301989809594925, |
| "learning_rate": 1.3687351308699026e-05, |
| "loss": 0.716, |
| "step": 705 |
| }, |
| { |
| "epoch": 2.2129992169146435, |
| "grad_norm": 0.9179501547426313, |
| "learning_rate": 1.3666992531768482e-05, |
| "loss": 0.7059, |
| "step": 706 |
| }, |
| { |
| "epoch": 2.216131558339859, |
| "grad_norm": 0.9666221682985012, |
| "learning_rate": 1.3646616179414573e-05, |
| "loss": 0.7599, |
| "step": 707 |
| }, |
| { |
| "epoch": 2.219263899765074, |
| "grad_norm": 0.9675432440063244, |
| "learning_rate": 1.3626222349298539e-05, |
| "loss": 0.7518, |
| "step": 708 |
| }, |
| { |
| "epoch": 2.22239624119029, |
| "grad_norm": 0.9390656597784647, |
| "learning_rate": 1.360581113916538e-05, |
| "loss": 0.7252, |
| "step": 709 |
| }, |
| { |
| "epoch": 2.225528582615505, |
| "grad_norm": 0.9581962913041326, |
| "learning_rate": 1.3585382646843396e-05, |
| "loss": 0.7605, |
| "step": 710 |
| }, |
| { |
| "epoch": 2.2286609240407205, |
| "grad_norm": 0.9610602302318673, |
| "learning_rate": 1.356493697024373e-05, |
| "loss": 0.7811, |
| "step": 711 |
| }, |
| { |
| "epoch": 2.2317932654659356, |
| "grad_norm": 0.9867556442165821, |
| "learning_rate": 1.3544474207359875e-05, |
| "loss": 0.7349, |
| "step": 712 |
| }, |
| { |
| "epoch": 2.234925606891151, |
| "grad_norm": 0.9425764813872476, |
| "learning_rate": 1.352399445626722e-05, |
| "loss": 0.7226, |
| "step": 713 |
| }, |
| { |
| "epoch": 2.2380579483163663, |
| "grad_norm": 1.0102139516905289, |
| "learning_rate": 1.3503497815122577e-05, |
| "loss": 0.7671, |
| "step": 714 |
| }, |
| { |
| "epoch": 2.241190289741582, |
| "grad_norm": 0.965501835494664, |
| "learning_rate": 1.3482984382163713e-05, |
| "loss": 0.7449, |
| "step": 715 |
| }, |
| { |
| "epoch": 2.244322631166797, |
| "grad_norm": 0.9131072345452055, |
| "learning_rate": 1.3462454255708868e-05, |
| "loss": 0.714, |
| "step": 716 |
| }, |
| { |
| "epoch": 2.2474549725920125, |
| "grad_norm": 0.9498450701506534, |
| "learning_rate": 1.3441907534156298e-05, |
| "loss": 0.7484, |
| "step": 717 |
| }, |
| { |
| "epoch": 2.2505873140172277, |
| "grad_norm": 0.9689310041592919, |
| "learning_rate": 1.34213443159838e-05, |
| "loss": 0.7772, |
| "step": 718 |
| }, |
| { |
| "epoch": 2.2537196554424432, |
| "grad_norm": 0.9414801485286473, |
| "learning_rate": 1.3400764699748223e-05, |
| "loss": 0.7307, |
| "step": 719 |
| }, |
| { |
| "epoch": 2.256851996867659, |
| "grad_norm": 0.918278021033119, |
| "learning_rate": 1.3380168784085028e-05, |
| "loss": 0.7179, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.259984338292874, |
| "grad_norm": 0.9166637511796559, |
| "learning_rate": 1.3359556667707786e-05, |
| "loss": 0.7142, |
| "step": 721 |
| }, |
| { |
| "epoch": 2.263116679718089, |
| "grad_norm": 0.9495008728633978, |
| "learning_rate": 1.3338928449407721e-05, |
| "loss": 0.7138, |
| "step": 722 |
| }, |
| { |
| "epoch": 2.2662490211433046, |
| "grad_norm": 0.9787409627000769, |
| "learning_rate": 1.3318284228053225e-05, |
| "loss": 0.7126, |
| "step": 723 |
| }, |
| { |
| "epoch": 2.26938136256852, |
| "grad_norm": 0.9627840952028726, |
| "learning_rate": 1.3297624102589395e-05, |
| "loss": 0.7219, |
| "step": 724 |
| }, |
| { |
| "epoch": 2.2725137039937353, |
| "grad_norm": 0.9476059282284458, |
| "learning_rate": 1.3276948172037556e-05, |
| "loss": 0.7392, |
| "step": 725 |
| }, |
| { |
| "epoch": 2.2756460454189504, |
| "grad_norm": 0.9906364986347657, |
| "learning_rate": 1.3256256535494783e-05, |
| "loss": 0.7349, |
| "step": 726 |
| }, |
| { |
| "epoch": 2.278778386844166, |
| "grad_norm": 0.9675118908082809, |
| "learning_rate": 1.3235549292133425e-05, |
| "loss": 0.7364, |
| "step": 727 |
| }, |
| { |
| "epoch": 2.2819107282693816, |
| "grad_norm": 0.9429167511984878, |
| "learning_rate": 1.3214826541200638e-05, |
| "loss": 0.7514, |
| "step": 728 |
| }, |
| { |
| "epoch": 2.2850430696945967, |
| "grad_norm": 0.9847232006632042, |
| "learning_rate": 1.3194088382017897e-05, |
| "loss": 0.7606, |
| "step": 729 |
| }, |
| { |
| "epoch": 2.288175411119812, |
| "grad_norm": 0.9563800977188294, |
| "learning_rate": 1.3173334913980535e-05, |
| "loss": 0.7388, |
| "step": 730 |
| }, |
| { |
| "epoch": 2.2913077525450274, |
| "grad_norm": 0.9735616571758612, |
| "learning_rate": 1.3152566236557251e-05, |
| "loss": 0.7312, |
| "step": 731 |
| }, |
| { |
| "epoch": 2.294440093970243, |
| "grad_norm": 0.9639178348678931, |
| "learning_rate": 1.3131782449289646e-05, |
| "loss": 0.7385, |
| "step": 732 |
| }, |
| { |
| "epoch": 2.297572435395458, |
| "grad_norm": 0.9529934639226041, |
| "learning_rate": 1.3110983651791731e-05, |
| "loss": 0.7194, |
| "step": 733 |
| }, |
| { |
| "epoch": 2.3007047768206736, |
| "grad_norm": 0.9678440459923399, |
| "learning_rate": 1.3090169943749475e-05, |
| "loss": 0.7589, |
| "step": 734 |
| }, |
| { |
| "epoch": 2.3038371182458888, |
| "grad_norm": 0.9540843435038134, |
| "learning_rate": 1.3069341424920301e-05, |
| "loss": 0.7811, |
| "step": 735 |
| }, |
| { |
| "epoch": 2.3069694596711043, |
| "grad_norm": 1.0181162510512147, |
| "learning_rate": 1.3048498195132613e-05, |
| "loss": 0.7492, |
| "step": 736 |
| }, |
| { |
| "epoch": 2.3101018010963195, |
| "grad_norm": 0.9572210519063549, |
| "learning_rate": 1.3027640354285333e-05, |
| "loss": 0.788, |
| "step": 737 |
| }, |
| { |
| "epoch": 2.313234142521535, |
| "grad_norm": 0.9538184834756422, |
| "learning_rate": 1.3006768002347414e-05, |
| "loss": 0.7343, |
| "step": 738 |
| }, |
| { |
| "epoch": 2.31636648394675, |
| "grad_norm": 0.9765399045822022, |
| "learning_rate": 1.298588123935735e-05, |
| "loss": 0.7252, |
| "step": 739 |
| }, |
| { |
| "epoch": 2.3194988253719657, |
| "grad_norm": 0.9618045560134357, |
| "learning_rate": 1.2964980165422701e-05, |
| "loss": 0.7373, |
| "step": 740 |
| }, |
| { |
| "epoch": 2.322631166797181, |
| "grad_norm": 0.9565307835074357, |
| "learning_rate": 1.2944064880719634e-05, |
| "loss": 0.7335, |
| "step": 741 |
| }, |
| { |
| "epoch": 2.3257635082223964, |
| "grad_norm": 0.967282888806741, |
| "learning_rate": 1.2923135485492413e-05, |
| "loss": 0.7845, |
| "step": 742 |
| }, |
| { |
| "epoch": 2.3288958496476115, |
| "grad_norm": 0.9686861328866166, |
| "learning_rate": 1.290219208005294e-05, |
| "loss": 0.7312, |
| "step": 743 |
| }, |
| { |
| "epoch": 2.332028191072827, |
| "grad_norm": 0.9726533064758265, |
| "learning_rate": 1.2881234764780257e-05, |
| "loss": 0.7484, |
| "step": 744 |
| }, |
| { |
| "epoch": 2.3351605324980422, |
| "grad_norm": 0.957502804625888, |
| "learning_rate": 1.2860263640120086e-05, |
| "loss": 0.7295, |
| "step": 745 |
| }, |
| { |
| "epoch": 2.338292873923258, |
| "grad_norm": 1.0246439398068528, |
| "learning_rate": 1.2839278806584323e-05, |
| "loss": 0.7583, |
| "step": 746 |
| }, |
| { |
| "epoch": 2.341425215348473, |
| "grad_norm": 0.9881765469071092, |
| "learning_rate": 1.2818280364750577e-05, |
| "loss": 0.7605, |
| "step": 747 |
| }, |
| { |
| "epoch": 2.3445575567736885, |
| "grad_norm": 0.9226451453922399, |
| "learning_rate": 1.2797268415261681e-05, |
| "loss": 0.7449, |
| "step": 748 |
| }, |
| { |
| "epoch": 2.3476898981989036, |
| "grad_norm": 0.9389466675244833, |
| "learning_rate": 1.2776243058825199e-05, |
| "loss": 0.7508, |
| "step": 749 |
| }, |
| { |
| "epoch": 2.350822239624119, |
| "grad_norm": 0.9680649310436178, |
| "learning_rate": 1.2755204396212965e-05, |
| "loss": 0.7514, |
| "step": 750 |
| }, |
| { |
| "epoch": 2.3539545810493343, |
| "grad_norm": 0.9508971737498871, |
| "learning_rate": 1.273415252826058e-05, |
| "loss": 0.7163, |
| "step": 751 |
| }, |
| { |
| "epoch": 2.35708692247455, |
| "grad_norm": 1.0037413765304937, |
| "learning_rate": 1.2713087555866937e-05, |
| "loss": 0.7523, |
| "step": 752 |
| }, |
| { |
| "epoch": 2.360219263899765, |
| "grad_norm": 0.9675958858875301, |
| "learning_rate": 1.2692009579993736e-05, |
| "loss": 0.752, |
| "step": 753 |
| }, |
| { |
| "epoch": 2.3633516053249806, |
| "grad_norm": 0.9426804383890924, |
| "learning_rate": 1.2670918701665007e-05, |
| "loss": 0.6878, |
| "step": 754 |
| }, |
| { |
| "epoch": 2.3664839467501957, |
| "grad_norm": 0.9505901023922874, |
| "learning_rate": 1.264981502196662e-05, |
| "loss": 0.73, |
| "step": 755 |
| }, |
| { |
| "epoch": 2.3696162881754113, |
| "grad_norm": 0.982586546277607, |
| "learning_rate": 1.2628698642045786e-05, |
| "loss": 0.7187, |
| "step": 756 |
| }, |
| { |
| "epoch": 2.3727486296006264, |
| "grad_norm": 1.0016723707151867, |
| "learning_rate": 1.2607569663110603e-05, |
| "loss": 0.7703, |
| "step": 757 |
| }, |
| { |
| "epoch": 2.375880971025842, |
| "grad_norm": 0.9650854481892198, |
| "learning_rate": 1.258642818642955e-05, |
| "loss": 0.722, |
| "step": 758 |
| }, |
| { |
| "epoch": 2.379013312451057, |
| "grad_norm": 0.9838012057082094, |
| "learning_rate": 1.2565274313331001e-05, |
| "loss": 0.7776, |
| "step": 759 |
| }, |
| { |
| "epoch": 2.3821456538762726, |
| "grad_norm": 0.9546713861914818, |
| "learning_rate": 1.2544108145202748e-05, |
| "loss": 0.7277, |
| "step": 760 |
| }, |
| { |
| "epoch": 2.3852779953014878, |
| "grad_norm": 0.9520183658938295, |
| "learning_rate": 1.2522929783491508e-05, |
| "loss": 0.7334, |
| "step": 761 |
| }, |
| { |
| "epoch": 2.3884103367267033, |
| "grad_norm": 0.9540342124185963, |
| "learning_rate": 1.2501739329702453e-05, |
| "loss": 0.7291, |
| "step": 762 |
| }, |
| { |
| "epoch": 2.3915426781519185, |
| "grad_norm": 0.9549925390587612, |
| "learning_rate": 1.2480536885398697e-05, |
| "loss": 0.7329, |
| "step": 763 |
| }, |
| { |
| "epoch": 2.394675019577134, |
| "grad_norm": 0.9728937913697199, |
| "learning_rate": 1.2459322552200824e-05, |
| "loss": 0.7802, |
| "step": 764 |
| }, |
| { |
| "epoch": 2.397807361002349, |
| "grad_norm": 0.9640580751608233, |
| "learning_rate": 1.2438096431786408e-05, |
| "loss": 0.7606, |
| "step": 765 |
| }, |
| { |
| "epoch": 2.4009397024275647, |
| "grad_norm": 0.9304598478726142, |
| "learning_rate": 1.2416858625889507e-05, |
| "loss": 0.7361, |
| "step": 766 |
| }, |
| { |
| "epoch": 2.40407204385278, |
| "grad_norm": 0.9806923344541586, |
| "learning_rate": 1.2395609236300198e-05, |
| "loss": 0.7155, |
| "step": 767 |
| }, |
| { |
| "epoch": 2.4072043852779954, |
| "grad_norm": 0.969342428109374, |
| "learning_rate": 1.2374348364864068e-05, |
| "loss": 0.7559, |
| "step": 768 |
| }, |
| { |
| "epoch": 2.4103367267032105, |
| "grad_norm": 0.9350655072927523, |
| "learning_rate": 1.2353076113481742e-05, |
| "loss": 0.7544, |
| "step": 769 |
| }, |
| { |
| "epoch": 2.413469068128426, |
| "grad_norm": 0.9447261809188683, |
| "learning_rate": 1.2331792584108375e-05, |
| "loss": 0.7191, |
| "step": 770 |
| }, |
| { |
| "epoch": 2.4166014095536412, |
| "grad_norm": 0.9674909744485723, |
| "learning_rate": 1.2310497878753196e-05, |
| "loss": 0.7686, |
| "step": 771 |
| }, |
| { |
| "epoch": 2.419733750978857, |
| "grad_norm": 0.9854774668922955, |
| "learning_rate": 1.2289192099478979e-05, |
| "loss": 0.7595, |
| "step": 772 |
| }, |
| { |
| "epoch": 2.422866092404072, |
| "grad_norm": 0.9841981999228102, |
| "learning_rate": 1.2267875348401585e-05, |
| "loss": 0.7653, |
| "step": 773 |
| }, |
| { |
| "epoch": 2.4259984338292875, |
| "grad_norm": 0.9641773998632913, |
| "learning_rate": 1.224654772768946e-05, |
| "loss": 0.7338, |
| "step": 774 |
| }, |
| { |
| "epoch": 2.4291307752545026, |
| "grad_norm": 0.947511728550051, |
| "learning_rate": 1.2225209339563144e-05, |
| "loss": 0.7583, |
| "step": 775 |
| }, |
| { |
| "epoch": 2.432263116679718, |
| "grad_norm": 0.9870470819669465, |
| "learning_rate": 1.220386028629479e-05, |
| "loss": 0.781, |
| "step": 776 |
| }, |
| { |
| "epoch": 2.4353954581049333, |
| "grad_norm": 0.9540467312167563, |
| "learning_rate": 1.218250067020766e-05, |
| "loss": 0.732, |
| "step": 777 |
| }, |
| { |
| "epoch": 2.438527799530149, |
| "grad_norm": 0.9742439189723893, |
| "learning_rate": 1.2161130593675643e-05, |
| "loss": 0.7654, |
| "step": 778 |
| }, |
| { |
| "epoch": 2.441660140955364, |
| "grad_norm": 0.947883079443643, |
| "learning_rate": 1.2139750159122771e-05, |
| "loss": 0.7291, |
| "step": 779 |
| }, |
| { |
| "epoch": 2.4447924823805796, |
| "grad_norm": 1.0108972783151455, |
| "learning_rate": 1.2118359469022714e-05, |
| "loss": 0.7343, |
| "step": 780 |
| }, |
| { |
| "epoch": 2.4479248238057947, |
| "grad_norm": 0.9909962900496487, |
| "learning_rate": 1.2096958625898294e-05, |
| "loss": 0.75, |
| "step": 781 |
| }, |
| { |
| "epoch": 2.4510571652310102, |
| "grad_norm": 0.979708350323752, |
| "learning_rate": 1.2075547732321005e-05, |
| "loss": 0.7502, |
| "step": 782 |
| }, |
| { |
| "epoch": 2.4541895066562254, |
| "grad_norm": 0.9605132744547762, |
| "learning_rate": 1.2054126890910499e-05, |
| "loss": 0.7519, |
| "step": 783 |
| }, |
| { |
| "epoch": 2.457321848081441, |
| "grad_norm": 0.9692514022395361, |
| "learning_rate": 1.2032696204334115e-05, |
| "loss": 0.7273, |
| "step": 784 |
| }, |
| { |
| "epoch": 2.460454189506656, |
| "grad_norm": 0.9449589861483914, |
| "learning_rate": 1.2011255775306378e-05, |
| "loss": 0.7511, |
| "step": 785 |
| }, |
| { |
| "epoch": 2.4635865309318716, |
| "grad_norm": 0.9471704960405938, |
| "learning_rate": 1.1989805706588502e-05, |
| "loss": 0.7347, |
| "step": 786 |
| }, |
| { |
| "epoch": 2.4667188723570868, |
| "grad_norm": 0.9320311390343631, |
| "learning_rate": 1.1968346100987906e-05, |
| "loss": 0.7082, |
| "step": 787 |
| }, |
| { |
| "epoch": 2.4698512137823023, |
| "grad_norm": 0.9162015711675013, |
| "learning_rate": 1.1946877061357723e-05, |
| "loss": 0.7287, |
| "step": 788 |
| }, |
| { |
| "epoch": 2.4729835552075174, |
| "grad_norm": 0.9676114383404737, |
| "learning_rate": 1.1925398690596293e-05, |
| "loss": 0.753, |
| "step": 789 |
| }, |
| { |
| "epoch": 2.476115896632733, |
| "grad_norm": 0.9877343226967886, |
| "learning_rate": 1.1903911091646684e-05, |
| "loss": 0.7758, |
| "step": 790 |
| }, |
| { |
| "epoch": 2.479248238057948, |
| "grad_norm": 1.023377404363016, |
| "learning_rate": 1.1882414367496196e-05, |
| "loss": 0.7392, |
| "step": 791 |
| }, |
| { |
| "epoch": 2.4823805794831637, |
| "grad_norm": 0.9945895799174997, |
| "learning_rate": 1.186090862117586e-05, |
| "loss": 0.7502, |
| "step": 792 |
| }, |
| { |
| "epoch": 2.485512920908379, |
| "grad_norm": 1.0038795381065821, |
| "learning_rate": 1.183939395575995e-05, |
| "loss": 0.806, |
| "step": 793 |
| }, |
| { |
| "epoch": 2.4886452623335944, |
| "grad_norm": 0.9926041291737507, |
| "learning_rate": 1.181787047436549e-05, |
| "loss": 0.7617, |
| "step": 794 |
| }, |
| { |
| "epoch": 2.4917776037588095, |
| "grad_norm": 0.9747916863077687, |
| "learning_rate": 1.1796338280151756e-05, |
| "loss": 0.7506, |
| "step": 795 |
| }, |
| { |
| "epoch": 2.494909945184025, |
| "grad_norm": 0.928076659108732, |
| "learning_rate": 1.1774797476319785e-05, |
| "loss": 0.7306, |
| "step": 796 |
| }, |
| { |
| "epoch": 2.49804228660924, |
| "grad_norm": 1.0261524870519174, |
| "learning_rate": 1.175324816611188e-05, |
| "loss": 0.7463, |
| "step": 797 |
| }, |
| { |
| "epoch": 2.501174628034456, |
| "grad_norm": 0.9847678627771624, |
| "learning_rate": 1.1731690452811107e-05, |
| "loss": 0.7281, |
| "step": 798 |
| }, |
| { |
| "epoch": 2.5043069694596714, |
| "grad_norm": 0.9643463190600754, |
| "learning_rate": 1.1710124439740816e-05, |
| "loss": 0.76, |
| "step": 799 |
| }, |
| { |
| "epoch": 2.5074393108848865, |
| "grad_norm": 0.969773920531062, |
| "learning_rate": 1.1688550230264128e-05, |
| "loss": 0.7658, |
| "step": 800 |
| }, |
| { |
| "epoch": 2.5105716523101016, |
| "grad_norm": 0.9626642855194483, |
| "learning_rate": 1.1666967927783461e-05, |
| "loss": 0.7478, |
| "step": 801 |
| }, |
| { |
| "epoch": 2.513703993735317, |
| "grad_norm": 0.9884939163865005, |
| "learning_rate": 1.164537763574001e-05, |
| "loss": 0.7153, |
| "step": 802 |
| }, |
| { |
| "epoch": 2.5168363351605327, |
| "grad_norm": 0.9485195981984704, |
| "learning_rate": 1.1623779457613263e-05, |
| "loss": 0.7493, |
| "step": 803 |
| }, |
| { |
| "epoch": 2.519968676585748, |
| "grad_norm": 0.9438738641259264, |
| "learning_rate": 1.160217349692051e-05, |
| "loss": 0.7592, |
| "step": 804 |
| }, |
| { |
| "epoch": 2.523101018010963, |
| "grad_norm": 0.9775231067597289, |
| "learning_rate": 1.1580559857216347e-05, |
| "loss": 0.7669, |
| "step": 805 |
| }, |
| { |
| "epoch": 2.5262333594361785, |
| "grad_norm": 0.9964813304861763, |
| "learning_rate": 1.1558938642092161e-05, |
| "loss": 0.75, |
| "step": 806 |
| }, |
| { |
| "epoch": 2.529365700861394, |
| "grad_norm": 0.9604452115462686, |
| "learning_rate": 1.1537309955175655e-05, |
| "loss": 0.7367, |
| "step": 807 |
| }, |
| { |
| "epoch": 2.5324980422866092, |
| "grad_norm": 0.9243296564169551, |
| "learning_rate": 1.1515673900130344e-05, |
| "loss": 0.7135, |
| "step": 808 |
| }, |
| { |
| "epoch": 2.5356303837118244, |
| "grad_norm": 0.9840647742781437, |
| "learning_rate": 1.149403058065506e-05, |
| "loss": 0.7744, |
| "step": 809 |
| }, |
| { |
| "epoch": 2.53876272513704, |
| "grad_norm": 0.9363389375104039, |
| "learning_rate": 1.1472380100483438e-05, |
| "loss": 0.7333, |
| "step": 810 |
| }, |
| { |
| "epoch": 2.5418950665622555, |
| "grad_norm": 1.005204535054278, |
| "learning_rate": 1.145072256338345e-05, |
| "loss": 0.7535, |
| "step": 811 |
| }, |
| { |
| "epoch": 2.5450274079874706, |
| "grad_norm": 0.9828566960512756, |
| "learning_rate": 1.1429058073156886e-05, |
| "loss": 0.8036, |
| "step": 812 |
| }, |
| { |
| "epoch": 2.5481597494126857, |
| "grad_norm": 0.9671196176463093, |
| "learning_rate": 1.140738673363886e-05, |
| "loss": 0.7781, |
| "step": 813 |
| }, |
| { |
| "epoch": 2.5512920908379013, |
| "grad_norm": 0.9532345512189722, |
| "learning_rate": 1.1385708648697306e-05, |
| "loss": 0.7126, |
| "step": 814 |
| }, |
| { |
| "epoch": 2.554424432263117, |
| "grad_norm": 0.9251640219717523, |
| "learning_rate": 1.1364023922232503e-05, |
| "loss": 0.7222, |
| "step": 815 |
| }, |
| { |
| "epoch": 2.557556773688332, |
| "grad_norm": 0.9869679713832883, |
| "learning_rate": 1.1342332658176556e-05, |
| "loss": 0.7058, |
| "step": 816 |
| }, |
| { |
| "epoch": 2.560689115113547, |
| "grad_norm": 0.9930580206564761, |
| "learning_rate": 1.13206349604929e-05, |
| "loss": 0.7729, |
| "step": 817 |
| }, |
| { |
| "epoch": 2.5638214565387627, |
| "grad_norm": 0.9420081126831853, |
| "learning_rate": 1.1298930933175805e-05, |
| "loss": 0.7284, |
| "step": 818 |
| }, |
| { |
| "epoch": 2.5669537979639783, |
| "grad_norm": 1.0000782701863635, |
| "learning_rate": 1.1277220680249887e-05, |
| "loss": 0.7776, |
| "step": 819 |
| }, |
| { |
| "epoch": 2.5700861393891934, |
| "grad_norm": 0.9515958392786626, |
| "learning_rate": 1.1255504305769589e-05, |
| "loss": 0.7402, |
| "step": 820 |
| }, |
| { |
| "epoch": 2.5732184808144085, |
| "grad_norm": 0.9398121571554576, |
| "learning_rate": 1.1233781913818707e-05, |
| "loss": 0.7255, |
| "step": 821 |
| }, |
| { |
| "epoch": 2.576350822239624, |
| "grad_norm": 0.9937093752648678, |
| "learning_rate": 1.1212053608509862e-05, |
| "loss": 0.7627, |
| "step": 822 |
| }, |
| { |
| "epoch": 2.5794831636648397, |
| "grad_norm": 0.9791666530858968, |
| "learning_rate": 1.119031949398403e-05, |
| "loss": 0.7689, |
| "step": 823 |
| }, |
| { |
| "epoch": 2.5826155050900548, |
| "grad_norm": 0.9496103182684476, |
| "learning_rate": 1.1168579674410026e-05, |
| "loss": 0.7458, |
| "step": 824 |
| }, |
| { |
| "epoch": 2.58574784651527, |
| "grad_norm": 0.9935430896103922, |
| "learning_rate": 1.1146834253984008e-05, |
| "loss": 0.7673, |
| "step": 825 |
| }, |
| { |
| "epoch": 2.5888801879404855, |
| "grad_norm": 0.9901446257978567, |
| "learning_rate": 1.1125083336928974e-05, |
| "loss": 0.767, |
| "step": 826 |
| }, |
| { |
| "epoch": 2.592012529365701, |
| "grad_norm": 0.9886430099783694, |
| "learning_rate": 1.1103327027494272e-05, |
| "loss": 0.7649, |
| "step": 827 |
| }, |
| { |
| "epoch": 2.595144870790916, |
| "grad_norm": 0.957433731138305, |
| "learning_rate": 1.10815654299551e-05, |
| "loss": 0.7496, |
| "step": 828 |
| }, |
| { |
| "epoch": 2.5982772122161313, |
| "grad_norm": 0.9549512909328034, |
| "learning_rate": 1.1059798648611988e-05, |
| "loss": 0.7479, |
| "step": 829 |
| }, |
| { |
| "epoch": 2.601409553641347, |
| "grad_norm": 0.9729410513377862, |
| "learning_rate": 1.103802678779032e-05, |
| "loss": 0.7215, |
| "step": 830 |
| }, |
| { |
| "epoch": 2.6045418950665624, |
| "grad_norm": 0.9739903952440084, |
| "learning_rate": 1.1016249951839824e-05, |
| "loss": 0.7481, |
| "step": 831 |
| }, |
| { |
| "epoch": 2.6076742364917775, |
| "grad_norm": 0.9413669343604363, |
| "learning_rate": 1.0994468245134071e-05, |
| "loss": 0.7567, |
| "step": 832 |
| }, |
| { |
| "epoch": 2.610806577916993, |
| "grad_norm": 0.9742157973187829, |
| "learning_rate": 1.0972681772069981e-05, |
| "loss": 0.7463, |
| "step": 833 |
| }, |
| { |
| "epoch": 2.6139389193422082, |
| "grad_norm": 0.9778627829815928, |
| "learning_rate": 1.0950890637067317e-05, |
| "loss": 0.7633, |
| "step": 834 |
| }, |
| { |
| "epoch": 2.617071260767424, |
| "grad_norm": 0.9429091429491205, |
| "learning_rate": 1.0929094944568182e-05, |
| "loss": 0.7321, |
| "step": 835 |
| }, |
| { |
| "epoch": 2.620203602192639, |
| "grad_norm": 0.9807225580513256, |
| "learning_rate": 1.0907294799036525e-05, |
| "loss": 0.7599, |
| "step": 836 |
| }, |
| { |
| "epoch": 2.6233359436178545, |
| "grad_norm": 0.9466168414336543, |
| "learning_rate": 1.0885490304957642e-05, |
| "loss": 0.7111, |
| "step": 837 |
| }, |
| { |
| "epoch": 2.6264682850430696, |
| "grad_norm": 0.9473666583781692, |
| "learning_rate": 1.0863681566837666e-05, |
| "loss": 0.7333, |
| "step": 838 |
| }, |
| { |
| "epoch": 2.629600626468285, |
| "grad_norm": 0.9524909535219792, |
| "learning_rate": 1.0841868689203072e-05, |
| "loss": 0.7625, |
| "step": 839 |
| }, |
| { |
| "epoch": 2.6327329678935003, |
| "grad_norm": 0.9432869054961515, |
| "learning_rate": 1.0820051776600175e-05, |
| "loss": 0.745, |
| "step": 840 |
| }, |
| { |
| "epoch": 2.635865309318716, |
| "grad_norm": 0.940106657073856, |
| "learning_rate": 1.0798230933594629e-05, |
| "loss": 0.7438, |
| "step": 841 |
| }, |
| { |
| "epoch": 2.638997650743931, |
| "grad_norm": 0.959855959982091, |
| "learning_rate": 1.0776406264770933e-05, |
| "loss": 0.7728, |
| "step": 842 |
| }, |
| { |
| "epoch": 2.6421299921691466, |
| "grad_norm": 1.0045635045124952, |
| "learning_rate": 1.0754577874731911e-05, |
| "loss": 0.774, |
| "step": 843 |
| }, |
| { |
| "epoch": 2.6452623335943617, |
| "grad_norm": 0.9757139770765741, |
| "learning_rate": 1.073274586809823e-05, |
| "loss": 0.7312, |
| "step": 844 |
| }, |
| { |
| "epoch": 2.6483946750195773, |
| "grad_norm": 0.9438533061715757, |
| "learning_rate": 1.0710910349507881e-05, |
| "loss": 0.7237, |
| "step": 845 |
| }, |
| { |
| "epoch": 2.6515270164447924, |
| "grad_norm": 0.9531758260863951, |
| "learning_rate": 1.0689071423615708e-05, |
| "loss": 0.7348, |
| "step": 846 |
| }, |
| { |
| "epoch": 2.654659357870008, |
| "grad_norm": 0.9446644812284712, |
| "learning_rate": 1.0667229195092868e-05, |
| "loss": 0.7602, |
| "step": 847 |
| }, |
| { |
| "epoch": 2.657791699295223, |
| "grad_norm": 0.9782412701275756, |
| "learning_rate": 1.0645383768626345e-05, |
| "loss": 0.7765, |
| "step": 848 |
| }, |
| { |
| "epoch": 2.6609240407204386, |
| "grad_norm": 1.007139575996801, |
| "learning_rate": 1.0623535248918463e-05, |
| "loss": 0.7751, |
| "step": 849 |
| }, |
| { |
| "epoch": 2.6640563821456538, |
| "grad_norm": 0.972578870009894, |
| "learning_rate": 1.0601683740686366e-05, |
| "loss": 0.8046, |
| "step": 850 |
| }, |
| { |
| "epoch": 2.6671887235708693, |
| "grad_norm": 0.9533200080779424, |
| "learning_rate": 1.0579829348661522e-05, |
| "loss": 0.7268, |
| "step": 851 |
| }, |
| { |
| "epoch": 2.6703210649960845, |
| "grad_norm": 0.9569784524332555, |
| "learning_rate": 1.0557972177589214e-05, |
| "loss": 0.7914, |
| "step": 852 |
| }, |
| { |
| "epoch": 2.6734534064213, |
| "grad_norm": 0.9554744612301984, |
| "learning_rate": 1.0536112332228057e-05, |
| "loss": 0.7656, |
| "step": 853 |
| }, |
| { |
| "epoch": 2.676585747846515, |
| "grad_norm": 0.9379633252744265, |
| "learning_rate": 1.0514249917349473e-05, |
| "loss": 0.7174, |
| "step": 854 |
| }, |
| { |
| "epoch": 2.6797180892717307, |
| "grad_norm": 0.9498518801324679, |
| "learning_rate": 1.0492385037737207e-05, |
| "loss": 0.7359, |
| "step": 855 |
| }, |
| { |
| "epoch": 2.682850430696946, |
| "grad_norm": 0.9323391831700777, |
| "learning_rate": 1.0470517798186813e-05, |
| "loss": 0.7516, |
| "step": 856 |
| }, |
| { |
| "epoch": 2.6859827721221614, |
| "grad_norm": 0.977808841738496, |
| "learning_rate": 1.044864830350515e-05, |
| "loss": 0.7588, |
| "step": 857 |
| }, |
| { |
| "epoch": 2.6891151135473765, |
| "grad_norm": 0.9607140331334884, |
| "learning_rate": 1.0426776658509899e-05, |
| "loss": 0.7297, |
| "step": 858 |
| }, |
| { |
| "epoch": 2.692247454972592, |
| "grad_norm": 0.9644638205133018, |
| "learning_rate": 1.040490296802904e-05, |
| "loss": 0.7602, |
| "step": 859 |
| }, |
| { |
| "epoch": 2.6953797963978072, |
| "grad_norm": 0.9541062791921975, |
| "learning_rate": 1.0383027336900356e-05, |
| "loss": 0.747, |
| "step": 860 |
| }, |
| { |
| "epoch": 2.698512137823023, |
| "grad_norm": 0.9713788632902975, |
| "learning_rate": 1.0361149869970929e-05, |
| "loss": 0.7486, |
| "step": 861 |
| }, |
| { |
| "epoch": 2.701644479248238, |
| "grad_norm": 0.9599164742932305, |
| "learning_rate": 1.0339270672096645e-05, |
| "loss": 0.7677, |
| "step": 862 |
| }, |
| { |
| "epoch": 2.7047768206734535, |
| "grad_norm": 0.9851639258139112, |
| "learning_rate": 1.0317389848141685e-05, |
| "loss": 0.735, |
| "step": 863 |
| }, |
| { |
| "epoch": 2.7079091620986686, |
| "grad_norm": 0.9627844834333039, |
| "learning_rate": 1.0295507502978024e-05, |
| "loss": 0.7423, |
| "step": 864 |
| }, |
| { |
| "epoch": 2.711041503523884, |
| "grad_norm": 0.9528844028329699, |
| "learning_rate": 1.0273623741484924e-05, |
| "loss": 0.7384, |
| "step": 865 |
| }, |
| { |
| "epoch": 2.7141738449490993, |
| "grad_norm": 1.0340380298089817, |
| "learning_rate": 1.0251738668548439e-05, |
| "loss": 0.8031, |
| "step": 866 |
| }, |
| { |
| "epoch": 2.717306186374315, |
| "grad_norm": 0.9538815854838193, |
| "learning_rate": 1.0229852389060905e-05, |
| "loss": 0.7339, |
| "step": 867 |
| }, |
| { |
| "epoch": 2.72043852779953, |
| "grad_norm": 0.9723625755825231, |
| "learning_rate": 1.0207965007920448e-05, |
| "loss": 0.7837, |
| "step": 868 |
| }, |
| { |
| "epoch": 2.7235708692247456, |
| "grad_norm": 0.967760456332842, |
| "learning_rate": 1.0186076630030469e-05, |
| "loss": 0.7522, |
| "step": 869 |
| }, |
| { |
| "epoch": 2.7267032106499607, |
| "grad_norm": 0.956046387699816, |
| "learning_rate": 1.0164187360299143e-05, |
| "loss": 0.7055, |
| "step": 870 |
| }, |
| { |
| "epoch": 2.7298355520751763, |
| "grad_norm": 0.9568472870023063, |
| "learning_rate": 1.0142297303638926e-05, |
| "loss": 0.7284, |
| "step": 871 |
| }, |
| { |
| "epoch": 2.7329678935003914, |
| "grad_norm": 0.988027374434136, |
| "learning_rate": 1.0120406564966043e-05, |
| "loss": 0.7714, |
| "step": 872 |
| }, |
| { |
| "epoch": 2.736100234925607, |
| "grad_norm": 1.028366948188634, |
| "learning_rate": 1.0098515249199988e-05, |
| "loss": 0.8163, |
| "step": 873 |
| }, |
| { |
| "epoch": 2.7392325763508225, |
| "grad_norm": 1.4617689871519928, |
| "learning_rate": 1.0076623461263017e-05, |
| "loss": 0.7347, |
| "step": 874 |
| }, |
| { |
| "epoch": 2.7423649177760376, |
| "grad_norm": 0.9546506131293968, |
| "learning_rate": 1.0054731306079657e-05, |
| "loss": 0.7131, |
| "step": 875 |
| }, |
| { |
| "epoch": 2.7454972592012528, |
| "grad_norm": 0.9591903904399064, |
| "learning_rate": 1.0032838888576194e-05, |
| "loss": 0.72, |
| "step": 876 |
| }, |
| { |
| "epoch": 2.7486296006264683, |
| "grad_norm": 0.9704760247984783, |
| "learning_rate": 1.0010946313680156e-05, |
| "loss": 0.7519, |
| "step": 877 |
| }, |
| { |
| "epoch": 2.751761942051684, |
| "grad_norm": 0.9496820302125051, |
| "learning_rate": 9.989053686319849e-06, |
| "loss": 0.7435, |
| "step": 878 |
| }, |
| { |
| "epoch": 2.754894283476899, |
| "grad_norm": 0.9773806934541943, |
| "learning_rate": 9.967161111423813e-06, |
| "loss": 0.7795, |
| "step": 879 |
| }, |
| { |
| "epoch": 2.758026624902114, |
| "grad_norm": 0.9202467047384671, |
| "learning_rate": 9.945268693920346e-06, |
| "loss": 0.7292, |
| "step": 880 |
| }, |
| { |
| "epoch": 2.7611589663273297, |
| "grad_norm": 0.974687628354556, |
| "learning_rate": 9.923376538736985e-06, |
| "loss": 0.7761, |
| "step": 881 |
| }, |
| { |
| "epoch": 2.7642913077525453, |
| "grad_norm": 0.9294513579702289, |
| "learning_rate": 9.901484750800014e-06, |
| "loss": 0.7235, |
| "step": 882 |
| }, |
| { |
| "epoch": 2.7674236491777604, |
| "grad_norm": 0.9850151969699499, |
| "learning_rate": 9.879593435033957e-06, |
| "loss": 0.7845, |
| "step": 883 |
| }, |
| { |
| "epoch": 2.7705559906029755, |
| "grad_norm": 0.929292922655392, |
| "learning_rate": 9.857702696361076e-06, |
| "loss": 0.75, |
| "step": 884 |
| }, |
| { |
| "epoch": 2.773688332028191, |
| "grad_norm": 0.9256442001669153, |
| "learning_rate": 9.835812639700862e-06, |
| "loss": 0.7215, |
| "step": 885 |
| }, |
| { |
| "epoch": 2.7768206734534067, |
| "grad_norm": 0.9481844945658222, |
| "learning_rate": 9.813923369969535e-06, |
| "loss": 0.7381, |
| "step": 886 |
| }, |
| { |
| "epoch": 2.779953014878622, |
| "grad_norm": 0.9481037841878706, |
| "learning_rate": 9.792034992079554e-06, |
| "loss": 0.7556, |
| "step": 887 |
| }, |
| { |
| "epoch": 2.783085356303837, |
| "grad_norm": 0.952957986686458, |
| "learning_rate": 9.770147610939098e-06, |
| "loss": 0.7339, |
| "step": 888 |
| }, |
| { |
| "epoch": 2.7862176977290525, |
| "grad_norm": 0.9549009956792547, |
| "learning_rate": 9.748261331451566e-06, |
| "loss": 0.7476, |
| "step": 889 |
| }, |
| { |
| "epoch": 2.789350039154268, |
| "grad_norm": 0.9782654041320651, |
| "learning_rate": 9.726376258515077e-06, |
| "loss": 0.7421, |
| "step": 890 |
| }, |
| { |
| "epoch": 2.792482380579483, |
| "grad_norm": 0.9900817275321431, |
| "learning_rate": 9.70449249702198e-06, |
| "loss": 0.7703, |
| "step": 891 |
| }, |
| { |
| "epoch": 2.7956147220046983, |
| "grad_norm": 0.9684826883994128, |
| "learning_rate": 9.682610151858317e-06, |
| "loss": 0.7874, |
| "step": 892 |
| }, |
| { |
| "epoch": 2.798747063429914, |
| "grad_norm": 0.9509743876454371, |
| "learning_rate": 9.660729327903359e-06, |
| "loss": 0.7554, |
| "step": 893 |
| }, |
| { |
| "epoch": 2.8018794048551294, |
| "grad_norm": 0.9651752336989294, |
| "learning_rate": 9.638850130029075e-06, |
| "loss": 0.7081, |
| "step": 894 |
| }, |
| { |
| "epoch": 2.8050117462803446, |
| "grad_norm": 0.9947227809634293, |
| "learning_rate": 9.616972663099648e-06, |
| "loss": 0.7533, |
| "step": 895 |
| }, |
| { |
| "epoch": 2.8081440877055597, |
| "grad_norm": 0.9575600590215773, |
| "learning_rate": 9.595097031970963e-06, |
| "loss": 0.7567, |
| "step": 896 |
| }, |
| { |
| "epoch": 2.8112764291307752, |
| "grad_norm": 0.9658030860447785, |
| "learning_rate": 9.5732233414901e-06, |
| "loss": 0.747, |
| "step": 897 |
| }, |
| { |
| "epoch": 2.814408770555991, |
| "grad_norm": 0.995325139613679, |
| "learning_rate": 9.551351696494854e-06, |
| "loss": 0.7438, |
| "step": 898 |
| }, |
| { |
| "epoch": 2.817541111981206, |
| "grad_norm": 0.9712772975214642, |
| "learning_rate": 9.529482201813192e-06, |
| "loss": 0.7293, |
| "step": 899 |
| }, |
| { |
| "epoch": 2.820673453406421, |
| "grad_norm": 0.9546909848683623, |
| "learning_rate": 9.507614962262796e-06, |
| "loss": 0.768, |
| "step": 900 |
| }, |
| { |
| "epoch": 2.8238057948316366, |
| "grad_norm": 0.9329589715311891, |
| "learning_rate": 9.48575008265053e-06, |
| "loss": 0.7341, |
| "step": 901 |
| }, |
| { |
| "epoch": 2.826938136256852, |
| "grad_norm": 0.9357071849591595, |
| "learning_rate": 9.463887667771946e-06, |
| "loss": 0.7195, |
| "step": 902 |
| }, |
| { |
| "epoch": 2.8300704776820673, |
| "grad_norm": 0.933792474330246, |
| "learning_rate": 9.442027822410786e-06, |
| "loss": 0.7563, |
| "step": 903 |
| }, |
| { |
| "epoch": 2.8332028191072824, |
| "grad_norm": 0.9936060316623524, |
| "learning_rate": 9.420170651338482e-06, |
| "loss": 0.7546, |
| "step": 904 |
| }, |
| { |
| "epoch": 2.836335160532498, |
| "grad_norm": 0.9522662188147043, |
| "learning_rate": 9.398316259313636e-06, |
| "loss": 0.744, |
| "step": 905 |
| }, |
| { |
| "epoch": 2.8394675019577136, |
| "grad_norm": 0.9768599306563925, |
| "learning_rate": 9.376464751081542e-06, |
| "loss": 0.7547, |
| "step": 906 |
| }, |
| { |
| "epoch": 2.8425998433829287, |
| "grad_norm": 0.9612851410888434, |
| "learning_rate": 9.354616231373659e-06, |
| "loss": 0.7662, |
| "step": 907 |
| }, |
| { |
| "epoch": 2.845732184808144, |
| "grad_norm": 0.9419299539508802, |
| "learning_rate": 9.332770804907137e-06, |
| "loss": 0.7293, |
| "step": 908 |
| }, |
| { |
| "epoch": 2.8488645262333594, |
| "grad_norm": 0.9437811849610814, |
| "learning_rate": 9.310928576384293e-06, |
| "loss": 0.7258, |
| "step": 909 |
| }, |
| { |
| "epoch": 2.851996867658575, |
| "grad_norm": 0.9714173836845524, |
| "learning_rate": 9.289089650492119e-06, |
| "loss": 0.732, |
| "step": 910 |
| }, |
| { |
| "epoch": 2.85512920908379, |
| "grad_norm": 0.9482886998971121, |
| "learning_rate": 9.267254131901772e-06, |
| "loss": 0.7273, |
| "step": 911 |
| }, |
| { |
| "epoch": 2.858261550509005, |
| "grad_norm": 0.9868925213214547, |
| "learning_rate": 9.245422125268094e-06, |
| "loss": 0.7215, |
| "step": 912 |
| }, |
| { |
| "epoch": 2.861393891934221, |
| "grad_norm": 0.9529681945509453, |
| "learning_rate": 9.223593735229069e-06, |
| "loss": 0.7498, |
| "step": 913 |
| }, |
| { |
| "epoch": 2.8645262333594363, |
| "grad_norm": 0.9383207113099425, |
| "learning_rate": 9.201769066405373e-06, |
| "loss": 0.7302, |
| "step": 914 |
| }, |
| { |
| "epoch": 2.8676585747846515, |
| "grad_norm": 0.9565533325468113, |
| "learning_rate": 9.179948223399828e-06, |
| "loss": 0.7352, |
| "step": 915 |
| }, |
| { |
| "epoch": 2.870790916209867, |
| "grad_norm": 0.973007632374376, |
| "learning_rate": 9.15813131079693e-06, |
| "loss": 0.7646, |
| "step": 916 |
| }, |
| { |
| "epoch": 2.873923257635082, |
| "grad_norm": 0.9966309662857158, |
| "learning_rate": 9.136318433162336e-06, |
| "loss": 0.724, |
| "step": 917 |
| }, |
| { |
| "epoch": 2.8770555990602977, |
| "grad_norm": 0.9853350271329723, |
| "learning_rate": 9.114509695042362e-06, |
| "loss": 0.7615, |
| "step": 918 |
| }, |
| { |
| "epoch": 2.880187940485513, |
| "grad_norm": 0.9504300421132944, |
| "learning_rate": 9.092705200963478e-06, |
| "loss": 0.7663, |
| "step": 919 |
| }, |
| { |
| "epoch": 2.8833202819107284, |
| "grad_norm": 0.9477723894786849, |
| "learning_rate": 9.070905055431822e-06, |
| "loss": 0.7372, |
| "step": 920 |
| }, |
| { |
| "epoch": 2.8864526233359435, |
| "grad_norm": 0.9495506324641605, |
| "learning_rate": 9.049109362932685e-06, |
| "loss": 0.7323, |
| "step": 921 |
| }, |
| { |
| "epoch": 2.889584964761159, |
| "grad_norm": 0.9315436534258359, |
| "learning_rate": 9.02731822793002e-06, |
| "loss": 0.7298, |
| "step": 922 |
| }, |
| { |
| "epoch": 2.8927173061863742, |
| "grad_norm": 0.9633745464994017, |
| "learning_rate": 9.005531754865929e-06, |
| "loss": 0.7329, |
| "step": 923 |
| }, |
| { |
| "epoch": 2.89584964761159, |
| "grad_norm": 0.9520011329637497, |
| "learning_rate": 8.983750048160178e-06, |
| "loss": 0.7461, |
| "step": 924 |
| }, |
| { |
| "epoch": 2.898981989036805, |
| "grad_norm": 0.947202489605334, |
| "learning_rate": 8.961973212209684e-06, |
| "loss": 0.7353, |
| "step": 925 |
| }, |
| { |
| "epoch": 2.9021143304620205, |
| "grad_norm": 0.9564588694025306, |
| "learning_rate": 8.940201351388016e-06, |
| "loss": 0.7499, |
| "step": 926 |
| }, |
| { |
| "epoch": 2.9052466718872356, |
| "grad_norm": 0.9660352117956935, |
| "learning_rate": 8.918434570044904e-06, |
| "loss": 0.774, |
| "step": 927 |
| }, |
| { |
| "epoch": 2.908379013312451, |
| "grad_norm": 0.9516602740353381, |
| "learning_rate": 8.89667297250573e-06, |
| "loss": 0.7196, |
| "step": 928 |
| }, |
| { |
| "epoch": 2.9115113547376663, |
| "grad_norm": 0.9503514513189356, |
| "learning_rate": 8.874916663071027e-06, |
| "loss": 0.7621, |
| "step": 929 |
| }, |
| { |
| "epoch": 2.914643696162882, |
| "grad_norm": 0.9732615080165525, |
| "learning_rate": 8.853165746015997e-06, |
| "loss": 0.7596, |
| "step": 930 |
| }, |
| { |
| "epoch": 2.917776037588097, |
| "grad_norm": 1.010101952673623, |
| "learning_rate": 8.831420325589976e-06, |
| "loss": 0.7654, |
| "step": 931 |
| }, |
| { |
| "epoch": 2.9209083790133126, |
| "grad_norm": 0.9740311574523944, |
| "learning_rate": 8.809680506015973e-06, |
| "loss": 0.7513, |
| "step": 932 |
| }, |
| { |
| "epoch": 2.9240407204385277, |
| "grad_norm": 0.9960714830033841, |
| "learning_rate": 8.787946391490141e-06, |
| "loss": 0.7656, |
| "step": 933 |
| }, |
| { |
| "epoch": 2.9271730618637433, |
| "grad_norm": 0.9638028878471048, |
| "learning_rate": 8.766218086181298e-06, |
| "loss": 0.7776, |
| "step": 934 |
| }, |
| { |
| "epoch": 2.9303054032889584, |
| "grad_norm": 0.9722845022040784, |
| "learning_rate": 8.744495694230413e-06, |
| "loss": 0.7301, |
| "step": 935 |
| }, |
| { |
| "epoch": 2.933437744714174, |
| "grad_norm": 0.9628137063571912, |
| "learning_rate": 8.722779319750116e-06, |
| "loss": 0.7206, |
| "step": 936 |
| }, |
| { |
| "epoch": 2.936570086139389, |
| "grad_norm": 0.9465035800698545, |
| "learning_rate": 8.701069066824195e-06, |
| "loss": 0.722, |
| "step": 937 |
| }, |
| { |
| "epoch": 2.9397024275646046, |
| "grad_norm": 0.9902093031760724, |
| "learning_rate": 8.679365039507104e-06, |
| "loss": 0.7857, |
| "step": 938 |
| }, |
| { |
| "epoch": 2.9428347689898198, |
| "grad_norm": 0.948550028383559, |
| "learning_rate": 8.657667341823449e-06, |
| "loss": 0.7501, |
| "step": 939 |
| }, |
| { |
| "epoch": 2.9459671104150353, |
| "grad_norm": 0.9610537515494351, |
| "learning_rate": 8.6359760777675e-06, |
| "loss": 0.7242, |
| "step": 940 |
| }, |
| { |
| "epoch": 2.9490994518402505, |
| "grad_norm": 0.9553995440705682, |
| "learning_rate": 8.614291351302697e-06, |
| "loss": 0.7344, |
| "step": 941 |
| }, |
| { |
| "epoch": 2.952231793265466, |
| "grad_norm": 0.9521585819489644, |
| "learning_rate": 8.592613266361145e-06, |
| "loss": 0.7399, |
| "step": 942 |
| }, |
| { |
| "epoch": 2.955364134690681, |
| "grad_norm": 0.9675254410966843, |
| "learning_rate": 8.570941926843117e-06, |
| "loss": 0.7474, |
| "step": 943 |
| }, |
| { |
| "epoch": 2.9584964761158967, |
| "grad_norm": 0.949925314091798, |
| "learning_rate": 8.549277436616551e-06, |
| "loss": 0.7398, |
| "step": 944 |
| }, |
| { |
| "epoch": 2.961628817541112, |
| "grad_norm": 0.9743585973394909, |
| "learning_rate": 8.527619899516567e-06, |
| "loss": 0.7281, |
| "step": 945 |
| }, |
| { |
| "epoch": 2.9647611589663274, |
| "grad_norm": 0.9721394272700253, |
| "learning_rate": 8.505969419344946e-06, |
| "loss": 0.7691, |
| "step": 946 |
| }, |
| { |
| "epoch": 2.9678935003915425, |
| "grad_norm": 0.9756436120696973, |
| "learning_rate": 8.484326099869659e-06, |
| "loss": 0.7346, |
| "step": 947 |
| }, |
| { |
| "epoch": 2.971025841816758, |
| "grad_norm": 0.963968460621726, |
| "learning_rate": 8.462690044824347e-06, |
| "loss": 0.7632, |
| "step": 948 |
| }, |
| { |
| "epoch": 2.9741581832419732, |
| "grad_norm": 0.9886664730071104, |
| "learning_rate": 8.441061357907842e-06, |
| "loss": 0.7795, |
| "step": 949 |
| }, |
| { |
| "epoch": 2.977290524667189, |
| "grad_norm": 0.9402625444423229, |
| "learning_rate": 8.419440142783653e-06, |
| "loss": 0.7237, |
| "step": 950 |
| }, |
| { |
| "epoch": 2.980422866092404, |
| "grad_norm": 0.9602243212134389, |
| "learning_rate": 8.397826503079489e-06, |
| "loss": 0.7561, |
| "step": 951 |
| }, |
| { |
| "epoch": 2.9835552075176195, |
| "grad_norm": 0.9617618663543123, |
| "learning_rate": 8.376220542386742e-06, |
| "loss": 0.7799, |
| "step": 952 |
| }, |
| { |
| "epoch": 2.9866875489428346, |
| "grad_norm": 0.9190755212609129, |
| "learning_rate": 8.354622364259995e-06, |
| "loss": 0.7042, |
| "step": 953 |
| }, |
| { |
| "epoch": 2.98981989036805, |
| "grad_norm": 0.9586872985486566, |
| "learning_rate": 8.333032072216542e-06, |
| "loss": 0.7411, |
| "step": 954 |
| }, |
| { |
| "epoch": 2.9929522317932653, |
| "grad_norm": 0.9490934425198838, |
| "learning_rate": 8.311449769735874e-06, |
| "loss": 0.7699, |
| "step": 955 |
| }, |
| { |
| "epoch": 2.996084573218481, |
| "grad_norm": 0.9483587215466358, |
| "learning_rate": 8.289875560259189e-06, |
| "loss": 0.729, |
| "step": 956 |
| }, |
| { |
| "epoch": 2.999216914643696, |
| "grad_norm": 0.9426645402788387, |
| "learning_rate": 8.268309547188894e-06, |
| "loss": 0.7605, |
| "step": 957 |
| }, |
| { |
| "epoch": 3.0031323414252156, |
| "grad_norm": 2.4449431031791447, |
| "learning_rate": 8.246751833888122e-06, |
| "loss": 1.363, |
| "step": 958 |
| }, |
| { |
| "epoch": 3.0062646828504307, |
| "grad_norm": 1.6451995720214123, |
| "learning_rate": 8.225202523680218e-06, |
| "loss": 0.5471, |
| "step": 959 |
| }, |
| { |
| "epoch": 3.0093970242756463, |
| "grad_norm": 1.6727678935052752, |
| "learning_rate": 8.203661719848249e-06, |
| "loss": 0.5595, |
| "step": 960 |
| }, |
| { |
| "epoch": 3.0125293657008614, |
| "grad_norm": 1.5460240913854986, |
| "learning_rate": 8.182129525634513e-06, |
| "loss": 0.5325, |
| "step": 961 |
| }, |
| { |
| "epoch": 3.015661707126077, |
| "grad_norm": 1.2460927908669523, |
| "learning_rate": 8.160606044240052e-06, |
| "loss": 0.5299, |
| "step": 962 |
| }, |
| { |
| "epoch": 3.018794048551292, |
| "grad_norm": 1.2859700150226807, |
| "learning_rate": 8.139091378824144e-06, |
| "loss": 0.526, |
| "step": 963 |
| }, |
| { |
| "epoch": 3.0219263899765076, |
| "grad_norm": 1.7164212978411557, |
| "learning_rate": 8.117585632503805e-06, |
| "loss": 0.5641, |
| "step": 964 |
| }, |
| { |
| "epoch": 3.0250587314017228, |
| "grad_norm": 1.7851156197762952, |
| "learning_rate": 8.096088908353316e-06, |
| "loss": 0.5354, |
| "step": 965 |
| }, |
| { |
| "epoch": 3.0281910728269383, |
| "grad_norm": 1.7956222015211165, |
| "learning_rate": 8.07460130940371e-06, |
| "loss": 0.5468, |
| "step": 966 |
| }, |
| { |
| "epoch": 3.0313234142521535, |
| "grad_norm": 1.7931934712777515, |
| "learning_rate": 8.05312293864228e-06, |
| "loss": 0.5571, |
| "step": 967 |
| }, |
| { |
| "epoch": 3.034455755677369, |
| "grad_norm": 1.6976805325976927, |
| "learning_rate": 8.031653899012098e-06, |
| "loss": 0.5492, |
| "step": 968 |
| }, |
| { |
| "epoch": 3.037588097102584, |
| "grad_norm": 1.596398489281078, |
| "learning_rate": 8.010194293411502e-06, |
| "loss": 0.4879, |
| "step": 969 |
| }, |
| { |
| "epoch": 3.0407204385277997, |
| "grad_norm": 1.4311365997826093, |
| "learning_rate": 7.988744224693625e-06, |
| "loss": 0.5483, |
| "step": 970 |
| }, |
| { |
| "epoch": 3.043852779953015, |
| "grad_norm": 1.3005396695284464, |
| "learning_rate": 7.967303795665887e-06, |
| "loss": 0.5155, |
| "step": 971 |
| }, |
| { |
| "epoch": 3.0469851213782304, |
| "grad_norm": 1.1670816271780478, |
| "learning_rate": 7.945873109089503e-06, |
| "loss": 0.5483, |
| "step": 972 |
| }, |
| { |
| "epoch": 3.0501174628034455, |
| "grad_norm": 1.2526413892546973, |
| "learning_rate": 7.924452267678998e-06, |
| "loss": 0.5208, |
| "step": 973 |
| }, |
| { |
| "epoch": 3.053249804228661, |
| "grad_norm": 1.3258191230420822, |
| "learning_rate": 7.903041374101707e-06, |
| "loss": 0.5597, |
| "step": 974 |
| }, |
| { |
| "epoch": 3.0563821456538762, |
| "grad_norm": 1.3179145976075899, |
| "learning_rate": 7.881640530977287e-06, |
| "loss": 0.5544, |
| "step": 975 |
| }, |
| { |
| "epoch": 3.059514487079092, |
| "grad_norm": 1.2661427204898215, |
| "learning_rate": 7.860249840877232e-06, |
| "loss": 0.5306, |
| "step": 976 |
| }, |
| { |
| "epoch": 3.062646828504307, |
| "grad_norm": 1.20947370272958, |
| "learning_rate": 7.838869406324358e-06, |
| "loss": 0.529, |
| "step": 977 |
| }, |
| { |
| "epoch": 3.0657791699295225, |
| "grad_norm": 1.1968126239778358, |
| "learning_rate": 7.817499329792342e-06, |
| "loss": 0.5685, |
| "step": 978 |
| }, |
| { |
| "epoch": 3.0689115113547376, |
| "grad_norm": 1.1994215893244842, |
| "learning_rate": 7.796139713705214e-06, |
| "loss": 0.5157, |
| "step": 979 |
| }, |
| { |
| "epoch": 3.072043852779953, |
| "grad_norm": 1.231635060309257, |
| "learning_rate": 7.774790660436857e-06, |
| "loss": 0.5195, |
| "step": 980 |
| }, |
| { |
| "epoch": 3.0751761942051683, |
| "grad_norm": 1.2984303455103845, |
| "learning_rate": 7.753452272310544e-06, |
| "loss": 0.5169, |
| "step": 981 |
| }, |
| { |
| "epoch": 3.078308535630384, |
| "grad_norm": 1.2495741983021167, |
| "learning_rate": 7.732124651598418e-06, |
| "loss": 0.5286, |
| "step": 982 |
| }, |
| { |
| "epoch": 3.081440877055599, |
| "grad_norm": 1.2371927309933792, |
| "learning_rate": 7.710807900521023e-06, |
| "loss": 0.5285, |
| "step": 983 |
| }, |
| { |
| "epoch": 3.0845732184808146, |
| "grad_norm": 1.1598033338515672, |
| "learning_rate": 7.689502121246809e-06, |
| "loss": 0.5285, |
| "step": 984 |
| }, |
| { |
| "epoch": 3.0877055599060297, |
| "grad_norm": 1.137936793677818, |
| "learning_rate": 7.668207415891625e-06, |
| "loss": 0.5257, |
| "step": 985 |
| }, |
| { |
| "epoch": 3.0908379013312453, |
| "grad_norm": 1.2251396369541545, |
| "learning_rate": 7.646923886518263e-06, |
| "loss": 0.5563, |
| "step": 986 |
| }, |
| { |
| "epoch": 3.0939702427564604, |
| "grad_norm": 1.1991699852871776, |
| "learning_rate": 7.625651635135933e-06, |
| "loss": 0.527, |
| "step": 987 |
| }, |
| { |
| "epoch": 3.097102584181676, |
| "grad_norm": 1.1439621826805295, |
| "learning_rate": 7.6043907636998026e-06, |
| "loss": 0.5163, |
| "step": 988 |
| }, |
| { |
| "epoch": 3.100234925606891, |
| "grad_norm": 1.123851307419761, |
| "learning_rate": 7.583141374110495e-06, |
| "loss": 0.5206, |
| "step": 989 |
| }, |
| { |
| "epoch": 3.1033672670321066, |
| "grad_norm": 1.1309411775644675, |
| "learning_rate": 7.561903568213595e-06, |
| "loss": 0.5416, |
| "step": 990 |
| }, |
| { |
| "epoch": 3.1064996084573218, |
| "grad_norm": 1.164015600425183, |
| "learning_rate": 7.540677447799176e-06, |
| "loss": 0.5062, |
| "step": 991 |
| }, |
| { |
| "epoch": 3.1096319498825373, |
| "grad_norm": 1.1836169096484468, |
| "learning_rate": 7.519463114601304e-06, |
| "loss": 0.5476, |
| "step": 992 |
| }, |
| { |
| "epoch": 3.1127642913077525, |
| "grad_norm": 1.142816670241574, |
| "learning_rate": 7.4982606702975505e-06, |
| "loss": 0.5342, |
| "step": 993 |
| }, |
| { |
| "epoch": 3.115896632732968, |
| "grad_norm": 1.1503408738938847, |
| "learning_rate": 7.477070216508493e-06, |
| "loss": 0.5646, |
| "step": 994 |
| }, |
| { |
| "epoch": 3.119028974158183, |
| "grad_norm": 1.150741615449422, |
| "learning_rate": 7.455891854797256e-06, |
| "loss": 0.5175, |
| "step": 995 |
| }, |
| { |
| "epoch": 3.1221613155833987, |
| "grad_norm": 1.2034319285665727, |
| "learning_rate": 7.434725686669002e-06, |
| "loss": 0.5765, |
| "step": 996 |
| }, |
| { |
| "epoch": 3.125293657008614, |
| "grad_norm": 1.1915855507050621, |
| "learning_rate": 7.413571813570453e-06, |
| "loss": 0.5323, |
| "step": 997 |
| }, |
| { |
| "epoch": 3.1284259984338294, |
| "grad_norm": 1.2081976282063787, |
| "learning_rate": 7.392430336889398e-06, |
| "loss": 0.5266, |
| "step": 998 |
| }, |
| { |
| "epoch": 3.1315583398590445, |
| "grad_norm": 1.1322504322898508, |
| "learning_rate": 7.371301357954219e-06, |
| "loss": 0.5129, |
| "step": 999 |
| }, |
| { |
| "epoch": 3.13469068128426, |
| "grad_norm": 1.1548883914063983, |
| "learning_rate": 7.350184978033386e-06, |
| "loss": 0.5193, |
| "step": 1000 |
| }, |
| { |
| "epoch": 3.137823022709475, |
| "grad_norm": 1.1526932585853438, |
| "learning_rate": 7.329081298334994e-06, |
| "loss": 0.5102, |
| "step": 1001 |
| }, |
| { |
| "epoch": 3.140955364134691, |
| "grad_norm": 1.1697102123089387, |
| "learning_rate": 7.307990420006266e-06, |
| "loss": 0.5326, |
| "step": 1002 |
| }, |
| { |
| "epoch": 3.144087705559906, |
| "grad_norm": 1.1265238542719669, |
| "learning_rate": 7.286912444133067e-06, |
| "loss": 0.5299, |
| "step": 1003 |
| }, |
| { |
| "epoch": 3.1472200469851215, |
| "grad_norm": 1.1537612326667075, |
| "learning_rate": 7.265847471739422e-06, |
| "loss": 0.5151, |
| "step": 1004 |
| }, |
| { |
| "epoch": 3.1503523884103366, |
| "grad_norm": 1.1284758438997318, |
| "learning_rate": 7.244795603787036e-06, |
| "loss": 0.5215, |
| "step": 1005 |
| }, |
| { |
| "epoch": 3.153484729835552, |
| "grad_norm": 1.1977561671812984, |
| "learning_rate": 7.2237569411748045e-06, |
| "loss": 0.5126, |
| "step": 1006 |
| }, |
| { |
| "epoch": 3.1566170712607673, |
| "grad_norm": 1.1649308019431355, |
| "learning_rate": 7.202731584738323e-06, |
| "loss": 0.5289, |
| "step": 1007 |
| }, |
| { |
| "epoch": 3.159749412685983, |
| "grad_norm": 1.1377687577701208, |
| "learning_rate": 7.1817196352494245e-06, |
| "loss": 0.5191, |
| "step": 1008 |
| }, |
| { |
| "epoch": 3.162881754111198, |
| "grad_norm": 1.1332419084126149, |
| "learning_rate": 7.160721193415681e-06, |
| "loss": 0.4928, |
| "step": 1009 |
| }, |
| { |
| "epoch": 3.1660140955364136, |
| "grad_norm": 1.1255882422146672, |
| "learning_rate": 7.139736359879916e-06, |
| "loss": 0.505, |
| "step": 1010 |
| }, |
| { |
| "epoch": 3.1691464369616287, |
| "grad_norm": 1.1396125393838465, |
| "learning_rate": 7.118765235219742e-06, |
| "loss": 0.5262, |
| "step": 1011 |
| }, |
| { |
| "epoch": 3.1722787783868442, |
| "grad_norm": 1.221304685970849, |
| "learning_rate": 7.0978079199470596e-06, |
| "loss": 0.5129, |
| "step": 1012 |
| }, |
| { |
| "epoch": 3.1754111198120594, |
| "grad_norm": 1.1646706730711807, |
| "learning_rate": 7.0768645145075885e-06, |
| "loss": 0.5509, |
| "step": 1013 |
| }, |
| { |
| "epoch": 3.178543461237275, |
| "grad_norm": 1.1909834940309783, |
| "learning_rate": 7.055935119280369e-06, |
| "loss": 0.5352, |
| "step": 1014 |
| }, |
| { |
| "epoch": 3.18167580266249, |
| "grad_norm": 1.1275894474381307, |
| "learning_rate": 7.035019834577301e-06, |
| "loss": 0.4987, |
| "step": 1015 |
| }, |
| { |
| "epoch": 3.1848081440877056, |
| "grad_norm": 1.1098330354725712, |
| "learning_rate": 7.014118760642654e-06, |
| "loss": 0.5429, |
| "step": 1016 |
| }, |
| { |
| "epoch": 3.1879404855129208, |
| "grad_norm": 1.1380777641842976, |
| "learning_rate": 6.993231997652587e-06, |
| "loss": 0.516, |
| "step": 1017 |
| }, |
| { |
| "epoch": 3.1910728269381363, |
| "grad_norm": 1.1693763120706757, |
| "learning_rate": 6.972359645714666e-06, |
| "loss": 0.5276, |
| "step": 1018 |
| }, |
| { |
| "epoch": 3.1942051683633514, |
| "grad_norm": 1.0981642819607513, |
| "learning_rate": 6.951501804867392e-06, |
| "loss": 0.5042, |
| "step": 1019 |
| }, |
| { |
| "epoch": 3.197337509788567, |
| "grad_norm": 1.1293301389682295, |
| "learning_rate": 6.930658575079705e-06, |
| "loss": 0.5456, |
| "step": 1020 |
| }, |
| { |
| "epoch": 3.200469851213782, |
| "grad_norm": 1.0968445808904248, |
| "learning_rate": 6.909830056250527e-06, |
| "loss": 0.5079, |
| "step": 1021 |
| }, |
| { |
| "epoch": 3.2036021926389977, |
| "grad_norm": 1.1042880780334279, |
| "learning_rate": 6.889016348208272e-06, |
| "loss": 0.5247, |
| "step": 1022 |
| }, |
| { |
| "epoch": 3.206734534064213, |
| "grad_norm": 1.1473219744960321, |
| "learning_rate": 6.868217550710359e-06, |
| "loss": 0.529, |
| "step": 1023 |
| }, |
| { |
| "epoch": 3.2098668754894284, |
| "grad_norm": 1.1820536568504343, |
| "learning_rate": 6.8474337634427505e-06, |
| "loss": 0.5653, |
| "step": 1024 |
| }, |
| { |
| "epoch": 3.2129992169146435, |
| "grad_norm": 1.1788578418720204, |
| "learning_rate": 6.826665086019467e-06, |
| "loss": 0.556, |
| "step": 1025 |
| }, |
| { |
| "epoch": 3.216131558339859, |
| "grad_norm": 1.1275364008712732, |
| "learning_rate": 6.805911617982105e-06, |
| "loss": 0.5211, |
| "step": 1026 |
| }, |
| { |
| "epoch": 3.219263899765074, |
| "grad_norm": 1.1788948899624612, |
| "learning_rate": 6.785173458799365e-06, |
| "loss": 0.5527, |
| "step": 1027 |
| }, |
| { |
| "epoch": 3.22239624119029, |
| "grad_norm": 1.190831430517554, |
| "learning_rate": 6.764450707866577e-06, |
| "loss": 0.5569, |
| "step": 1028 |
| }, |
| { |
| "epoch": 3.225528582615505, |
| "grad_norm": 1.1631264177590235, |
| "learning_rate": 6.743743464505219e-06, |
| "loss": 0.5389, |
| "step": 1029 |
| }, |
| { |
| "epoch": 3.2286609240407205, |
| "grad_norm": 1.141781399383516, |
| "learning_rate": 6.723051827962445e-06, |
| "loss": 0.5309, |
| "step": 1030 |
| }, |
| { |
| "epoch": 3.2317932654659356, |
| "grad_norm": 1.179512415938292, |
| "learning_rate": 6.7023758974106065e-06, |
| "loss": 0.525, |
| "step": 1031 |
| }, |
| { |
| "epoch": 3.234925606891151, |
| "grad_norm": 1.1944956246302587, |
| "learning_rate": 6.681715771946777e-06, |
| "loss": 0.5541, |
| "step": 1032 |
| }, |
| { |
| "epoch": 3.2380579483163663, |
| "grad_norm": 1.1591633344391639, |
| "learning_rate": 6.661071550592284e-06, |
| "loss": 0.5347, |
| "step": 1033 |
| }, |
| { |
| "epoch": 3.241190289741582, |
| "grad_norm": 1.12772338651315, |
| "learning_rate": 6.6404433322922146e-06, |
| "loss": 0.5155, |
| "step": 1034 |
| }, |
| { |
| "epoch": 3.244322631166797, |
| "grad_norm": 1.165589037264499, |
| "learning_rate": 6.619831215914974e-06, |
| "loss": 0.5346, |
| "step": 1035 |
| }, |
| { |
| "epoch": 3.2474549725920125, |
| "grad_norm": 1.103553776475765, |
| "learning_rate": 6.599235300251778e-06, |
| "loss": 0.5066, |
| "step": 1036 |
| }, |
| { |
| "epoch": 3.2505873140172277, |
| "grad_norm": 1.151353363986568, |
| "learning_rate": 6.578655684016204e-06, |
| "loss": 0.5456, |
| "step": 1037 |
| }, |
| { |
| "epoch": 3.2537196554424432, |
| "grad_norm": 1.159828047686325, |
| "learning_rate": 6.558092465843704e-06, |
| "loss": 0.5341, |
| "step": 1038 |
| }, |
| { |
| "epoch": 3.256851996867659, |
| "grad_norm": 1.1724554056458623, |
| "learning_rate": 6.537545744291134e-06, |
| "loss": 0.5598, |
| "step": 1039 |
| }, |
| { |
| "epoch": 3.259984338292874, |
| "grad_norm": 1.1868205375108176, |
| "learning_rate": 6.517015617836292e-06, |
| "loss": 0.5404, |
| "step": 1040 |
| }, |
| { |
| "epoch": 3.263116679718089, |
| "grad_norm": 1.180384610418905, |
| "learning_rate": 6.496502184877426e-06, |
| "loss": 0.5215, |
| "step": 1041 |
| }, |
| { |
| "epoch": 3.2662490211433046, |
| "grad_norm": 1.205146545691501, |
| "learning_rate": 6.476005543732783e-06, |
| "loss": 0.5314, |
| "step": 1042 |
| }, |
| { |
| "epoch": 3.26938136256852, |
| "grad_norm": 1.1244830342386452, |
| "learning_rate": 6.45552579264013e-06, |
| "loss": 0.5061, |
| "step": 1043 |
| }, |
| { |
| "epoch": 3.2725137039937353, |
| "grad_norm": 1.1324469436671378, |
| "learning_rate": 6.4350630297562726e-06, |
| "loss": 0.5411, |
| "step": 1044 |
| }, |
| { |
| "epoch": 3.2756460454189504, |
| "grad_norm": 1.170084873645457, |
| "learning_rate": 6.414617353156605e-06, |
| "loss": 0.5256, |
| "step": 1045 |
| }, |
| { |
| "epoch": 3.278778386844166, |
| "grad_norm": 1.154539238112036, |
| "learning_rate": 6.394188860834626e-06, |
| "loss": 0.5353, |
| "step": 1046 |
| }, |
| { |
| "epoch": 3.2819107282693816, |
| "grad_norm": 1.145934214062239, |
| "learning_rate": 6.373777650701465e-06, |
| "loss": 0.4949, |
| "step": 1047 |
| }, |
| { |
| "epoch": 3.2850430696945967, |
| "grad_norm": 1.1758118112093967, |
| "learning_rate": 6.35338382058543e-06, |
| "loss": 0.5388, |
| "step": 1048 |
| }, |
| { |
| "epoch": 3.288175411119812, |
| "grad_norm": 1.1930083102260012, |
| "learning_rate": 6.333007468231521e-06, |
| "loss": 0.5357, |
| "step": 1049 |
| }, |
| { |
| "epoch": 3.2913077525450274, |
| "grad_norm": 1.2215776268642229, |
| "learning_rate": 6.312648691300975e-06, |
| "loss": 0.5159, |
| "step": 1050 |
| }, |
| { |
| "epoch": 3.294440093970243, |
| "grad_norm": 1.149266078166264, |
| "learning_rate": 6.292307587370794e-06, |
| "loss": 0.5329, |
| "step": 1051 |
| }, |
| { |
| "epoch": 3.297572435395458, |
| "grad_norm": 1.188650823733469, |
| "learning_rate": 6.271984253933266e-06, |
| "loss": 0.5607, |
| "step": 1052 |
| }, |
| { |
| "epoch": 3.3007047768206736, |
| "grad_norm": 1.1743831994156586, |
| "learning_rate": 6.2516787883955165e-06, |
| "loss": 0.5574, |
| "step": 1053 |
| }, |
| { |
| "epoch": 3.3038371182458888, |
| "grad_norm": 1.1970079452073716, |
| "learning_rate": 6.231391288079027e-06, |
| "loss": 0.5173, |
| "step": 1054 |
| }, |
| { |
| "epoch": 3.3069694596711043, |
| "grad_norm": 1.122279685545005, |
| "learning_rate": 6.211121850219175e-06, |
| "loss": 0.5361, |
| "step": 1055 |
| }, |
| { |
| "epoch": 3.3101018010963195, |
| "grad_norm": 1.225293554009653, |
| "learning_rate": 6.1908705719647735e-06, |
| "loss": 0.555, |
| "step": 1056 |
| }, |
| { |
| "epoch": 3.313234142521535, |
| "grad_norm": 1.1655884451182597, |
| "learning_rate": 6.170637550377587e-06, |
| "loss": 0.5321, |
| "step": 1057 |
| }, |
| { |
| "epoch": 3.31636648394675, |
| "grad_norm": 1.1850726593613186, |
| "learning_rate": 6.150422882431888e-06, |
| "loss": 0.5158, |
| "step": 1058 |
| }, |
| { |
| "epoch": 3.3194988253719657, |
| "grad_norm": 1.1741932650006341, |
| "learning_rate": 6.130226665013979e-06, |
| "loss": 0.5278, |
| "step": 1059 |
| }, |
| { |
| "epoch": 3.322631166797181, |
| "grad_norm": 1.1637151007879059, |
| "learning_rate": 6.110048994921735e-06, |
| "loss": 0.5147, |
| "step": 1060 |
| }, |
| { |
| "epoch": 3.3257635082223964, |
| "grad_norm": 1.19810414141727, |
| "learning_rate": 6.089889968864122e-06, |
| "loss": 0.5523, |
| "step": 1061 |
| }, |
| { |
| "epoch": 3.3288958496476115, |
| "grad_norm": 1.1747671263259907, |
| "learning_rate": 6.069749683460765e-06, |
| "loss": 0.5259, |
| "step": 1062 |
| }, |
| { |
| "epoch": 3.332028191072827, |
| "grad_norm": 1.189733533665569, |
| "learning_rate": 6.049628235241459e-06, |
| "loss": 0.5727, |
| "step": 1063 |
| }, |
| { |
| "epoch": 3.3351605324980422, |
| "grad_norm": 1.1907858250963501, |
| "learning_rate": 6.029525720645719e-06, |
| "loss": 0.5581, |
| "step": 1064 |
| }, |
| { |
| "epoch": 3.338292873923258, |
| "grad_norm": 1.1727785316101889, |
| "learning_rate": 6.009442236022307e-06, |
| "loss": 0.5527, |
| "step": 1065 |
| }, |
| { |
| "epoch": 3.341425215348473, |
| "grad_norm": 1.1473257646366246, |
| "learning_rate": 5.989377877628783e-06, |
| "loss": 0.5354, |
| "step": 1066 |
| }, |
| { |
| "epoch": 3.3445575567736885, |
| "grad_norm": 1.1904303579624806, |
| "learning_rate": 5.969332741631037e-06, |
| "loss": 0.5299, |
| "step": 1067 |
| }, |
| { |
| "epoch": 3.3476898981989036, |
| "grad_norm": 1.1467680391867652, |
| "learning_rate": 5.949306924102822e-06, |
| "loss": 0.5347, |
| "step": 1068 |
| }, |
| { |
| "epoch": 3.350822239624119, |
| "grad_norm": 1.164592432138905, |
| "learning_rate": 5.929300521025304e-06, |
| "loss": 0.5295, |
| "step": 1069 |
| }, |
| { |
| "epoch": 3.3539545810493343, |
| "grad_norm": 1.183763669072413, |
| "learning_rate": 5.9093136282866014e-06, |
| "loss": 0.5309, |
| "step": 1070 |
| }, |
| { |
| "epoch": 3.35708692247455, |
| "grad_norm": 1.1341809428622889, |
| "learning_rate": 5.889346341681319e-06, |
| "loss": 0.5057, |
| "step": 1071 |
| }, |
| { |
| "epoch": 3.360219263899765, |
| "grad_norm": 1.2114438092248463, |
| "learning_rate": 5.869398756910095e-06, |
| "loss": 0.5285, |
| "step": 1072 |
| }, |
| { |
| "epoch": 3.3633516053249806, |
| "grad_norm": 1.1899159606732028, |
| "learning_rate": 5.8494709695791364e-06, |
| "loss": 0.515, |
| "step": 1073 |
| }, |
| { |
| "epoch": 3.3664839467501957, |
| "grad_norm": 1.1585091379070802, |
| "learning_rate": 5.829563075199753e-06, |
| "loss": 0.5215, |
| "step": 1074 |
| }, |
| { |
| "epoch": 3.3696162881754113, |
| "grad_norm": 1.1554751170108555, |
| "learning_rate": 5.809675169187936e-06, |
| "loss": 0.5276, |
| "step": 1075 |
| }, |
| { |
| "epoch": 3.3727486296006264, |
| "grad_norm": 1.1531221192759264, |
| "learning_rate": 5.789807346863849e-06, |
| "loss": 0.5343, |
| "step": 1076 |
| }, |
| { |
| "epoch": 3.375880971025842, |
| "grad_norm": 1.2264179028180235, |
| "learning_rate": 5.769959703451413e-06, |
| "loss": 0.5331, |
| "step": 1077 |
| }, |
| { |
| "epoch": 3.379013312451057, |
| "grad_norm": 1.1909513224731456, |
| "learning_rate": 5.750132334077827e-06, |
| "loss": 0.5221, |
| "step": 1078 |
| }, |
| { |
| "epoch": 3.3821456538762726, |
| "grad_norm": 1.1410796381664094, |
| "learning_rate": 5.730325333773122e-06, |
| "loss": 0.5236, |
| "step": 1079 |
| }, |
| { |
| "epoch": 3.3852779953014878, |
| "grad_norm": 1.165393643113191, |
| "learning_rate": 5.710538797469707e-06, |
| "loss": 0.5363, |
| "step": 1080 |
| }, |
| { |
| "epoch": 3.3884103367267033, |
| "grad_norm": 1.109034591095622, |
| "learning_rate": 5.690772820001894e-06, |
| "loss": 0.5129, |
| "step": 1081 |
| }, |
| { |
| "epoch": 3.3915426781519185, |
| "grad_norm": 1.1479442023962463, |
| "learning_rate": 5.671027496105477e-06, |
| "loss": 0.5767, |
| "step": 1082 |
| }, |
| { |
| "epoch": 3.394675019577134, |
| "grad_norm": 1.1983194565401882, |
| "learning_rate": 5.651302920417252e-06, |
| "loss": 0.5233, |
| "step": 1083 |
| }, |
| { |
| "epoch": 3.397807361002349, |
| "grad_norm": 1.1247461153791098, |
| "learning_rate": 5.63159918747457e-06, |
| "loss": 0.531, |
| "step": 1084 |
| }, |
| { |
| "epoch": 3.4009397024275647, |
| "grad_norm": 1.1645414490351773, |
| "learning_rate": 5.611916391714887e-06, |
| "loss": 0.5356, |
| "step": 1085 |
| }, |
| { |
| "epoch": 3.40407204385278, |
| "grad_norm": 1.1955131839630717, |
| "learning_rate": 5.592254627475313e-06, |
| "loss": 0.5181, |
| "step": 1086 |
| }, |
| { |
| "epoch": 3.4072043852779954, |
| "grad_norm": 1.1669277156801512, |
| "learning_rate": 5.572613988992157e-06, |
| "loss": 0.5505, |
| "step": 1087 |
| }, |
| { |
| "epoch": 3.4103367267032105, |
| "grad_norm": 1.1969803225344355, |
| "learning_rate": 5.552994570400465e-06, |
| "loss": 0.5279, |
| "step": 1088 |
| }, |
| { |
| "epoch": 3.413469068128426, |
| "grad_norm": 1.149976867281222, |
| "learning_rate": 5.5333964657335894e-06, |
| "loss": 0.5176, |
| "step": 1089 |
| }, |
| { |
| "epoch": 3.4166014095536412, |
| "grad_norm": 1.1530538912791253, |
| "learning_rate": 5.513819768922723e-06, |
| "loss": 0.529, |
| "step": 1090 |
| }, |
| { |
| "epoch": 3.419733750978857, |
| "grad_norm": 1.1611367505386145, |
| "learning_rate": 5.49426457379646e-06, |
| "loss": 0.5045, |
| "step": 1091 |
| }, |
| { |
| "epoch": 3.422866092404072, |
| "grad_norm": 1.157629907018517, |
| "learning_rate": 5.474730974080327e-06, |
| "loss": 0.5384, |
| "step": 1092 |
| }, |
| { |
| "epoch": 3.4259984338292875, |
| "grad_norm": 1.1436433927685985, |
| "learning_rate": 5.455219063396356e-06, |
| "loss": 0.5164, |
| "step": 1093 |
| }, |
| { |
| "epoch": 3.4291307752545026, |
| "grad_norm": 1.1624946647803467, |
| "learning_rate": 5.4357289352626284e-06, |
| "loss": 0.518, |
| "step": 1094 |
| }, |
| { |
| "epoch": 3.432263116679718, |
| "grad_norm": 1.2059544843365895, |
| "learning_rate": 5.416260683092814e-06, |
| "loss": 0.5413, |
| "step": 1095 |
| }, |
| { |
| "epoch": 3.4353954581049333, |
| "grad_norm": 1.1848557963338198, |
| "learning_rate": 5.3968144001957445e-06, |
| "loss": 0.556, |
| "step": 1096 |
| }, |
| { |
| "epoch": 3.438527799530149, |
| "grad_norm": 1.2267248653375662, |
| "learning_rate": 5.377390179774951e-06, |
| "loss": 0.5735, |
| "step": 1097 |
| }, |
| { |
| "epoch": 3.441660140955364, |
| "grad_norm": 1.219382842483644, |
| "learning_rate": 5.357988114928221e-06, |
| "loss": 0.5224, |
| "step": 1098 |
| }, |
| { |
| "epoch": 3.4447924823805796, |
| "grad_norm": 1.2019925545507635, |
| "learning_rate": 5.338608298647161e-06, |
| "loss": 0.5188, |
| "step": 1099 |
| }, |
| { |
| "epoch": 3.4479248238057947, |
| "grad_norm": 1.2054098999261758, |
| "learning_rate": 5.319250823816732e-06, |
| "loss": 0.5603, |
| "step": 1100 |
| }, |
| { |
| "epoch": 3.4510571652310102, |
| "grad_norm": 1.149167269524483, |
| "learning_rate": 5.299915783214816e-06, |
| "loss": 0.5103, |
| "step": 1101 |
| }, |
| { |
| "epoch": 3.4541895066562254, |
| "grad_norm": 1.190224838500585, |
| "learning_rate": 5.28060326951178e-06, |
| "loss": 0.5445, |
| "step": 1102 |
| }, |
| { |
| "epoch": 3.457321848081441, |
| "grad_norm": 1.145339136200069, |
| "learning_rate": 5.2613133752700145e-06, |
| "loss": 0.5432, |
| "step": 1103 |
| }, |
| { |
| "epoch": 3.460454189506656, |
| "grad_norm": 1.1801121893769682, |
| "learning_rate": 5.242046192943503e-06, |
| "loss": 0.5321, |
| "step": 1104 |
| }, |
| { |
| "epoch": 3.4635865309318716, |
| "grad_norm": 1.196242124148136, |
| "learning_rate": 5.22280181487737e-06, |
| "loss": 0.5097, |
| "step": 1105 |
| }, |
| { |
| "epoch": 3.4667188723570868, |
| "grad_norm": 1.1558254840616595, |
| "learning_rate": 5.2035803333074435e-06, |
| "loss": 0.515, |
| "step": 1106 |
| }, |
| { |
| "epoch": 3.4698512137823023, |
| "grad_norm": 1.1436458411967738, |
| "learning_rate": 5.184381840359815e-06, |
| "loss": 0.5004, |
| "step": 1107 |
| }, |
| { |
| "epoch": 3.4729835552075174, |
| "grad_norm": 1.137825928298088, |
| "learning_rate": 5.1652064280503884e-06, |
| "loss": 0.5164, |
| "step": 1108 |
| }, |
| { |
| "epoch": 3.476115896632733, |
| "grad_norm": 1.1431285887634195, |
| "learning_rate": 5.146054188284443e-06, |
| "loss": 0.5324, |
| "step": 1109 |
| }, |
| { |
| "epoch": 3.479248238057948, |
| "grad_norm": 1.1848905862756827, |
| "learning_rate": 5.126925212856202e-06, |
| "loss": 0.541, |
| "step": 1110 |
| }, |
| { |
| "epoch": 3.4823805794831637, |
| "grad_norm": 1.15107617387345, |
| "learning_rate": 5.107819593448385e-06, |
| "loss": 0.5049, |
| "step": 1111 |
| }, |
| { |
| "epoch": 3.485512920908379, |
| "grad_norm": 1.1324253201763683, |
| "learning_rate": 5.088737421631767e-06, |
| "loss": 0.5263, |
| "step": 1112 |
| }, |
| { |
| "epoch": 3.4886452623335944, |
| "grad_norm": 1.1791937767344005, |
| "learning_rate": 5.069678788864742e-06, |
| "loss": 0.5301, |
| "step": 1113 |
| }, |
| { |
| "epoch": 3.4917776037588095, |
| "grad_norm": 1.177125639354513, |
| "learning_rate": 5.050643786492888e-06, |
| "loss": 0.5299, |
| "step": 1114 |
| }, |
| { |
| "epoch": 3.494909945184025, |
| "grad_norm": 1.12975187321893, |
| "learning_rate": 5.031632505748516e-06, |
| "loss": 0.5191, |
| "step": 1115 |
| }, |
| { |
| "epoch": 3.49804228660924, |
| "grad_norm": 1.1702301585282073, |
| "learning_rate": 5.012645037750254e-06, |
| "loss": 0.4911, |
| "step": 1116 |
| }, |
| { |
| "epoch": 3.501174628034456, |
| "grad_norm": 1.1898809996788224, |
| "learning_rate": 4.993681473502586e-06, |
| "loss": 0.5846, |
| "step": 1117 |
| }, |
| { |
| "epoch": 3.5043069694596714, |
| "grad_norm": 1.1772803798080353, |
| "learning_rate": 4.97474190389544e-06, |
| "loss": 0.5449, |
| "step": 1118 |
| }, |
| { |
| "epoch": 3.5074393108848865, |
| "grad_norm": 1.1458679793457238, |
| "learning_rate": 4.955826419703736e-06, |
| "loss": 0.527, |
| "step": 1119 |
| }, |
| { |
| "epoch": 3.5105716523101016, |
| "grad_norm": 1.197898331863354, |
| "learning_rate": 4.936935111586953e-06, |
| "loss": 0.5279, |
| "step": 1120 |
| }, |
| { |
| "epoch": 3.513703993735317, |
| "grad_norm": 1.1490696601995432, |
| "learning_rate": 4.918068070088708e-06, |
| "loss": 0.5479, |
| "step": 1121 |
| }, |
| { |
| "epoch": 3.5168363351605327, |
| "grad_norm": 1.1338863485542614, |
| "learning_rate": 4.899225385636293e-06, |
| "loss": 0.5143, |
| "step": 1122 |
| }, |
| { |
| "epoch": 3.519968676585748, |
| "grad_norm": 1.1180638940506065, |
| "learning_rate": 4.880407148540275e-06, |
| "loss": 0.5018, |
| "step": 1123 |
| }, |
| { |
| "epoch": 3.523101018010963, |
| "grad_norm": 1.1649197234392008, |
| "learning_rate": 4.8616134489940455e-06, |
| "loss": 0.5086, |
| "step": 1124 |
| }, |
| { |
| "epoch": 3.5262333594361785, |
| "grad_norm": 1.175823733278329, |
| "learning_rate": 4.8428443770733835e-06, |
| "loss": 0.5128, |
| "step": 1125 |
| }, |
| { |
| "epoch": 3.529365700861394, |
| "grad_norm": 1.1338414552899265, |
| "learning_rate": 4.824100022736036e-06, |
| "loss": 0.4854, |
| "step": 1126 |
| }, |
| { |
| "epoch": 3.5324980422866092, |
| "grad_norm": 1.208715108777811, |
| "learning_rate": 4.805380475821287e-06, |
| "loss": 0.562, |
| "step": 1127 |
| }, |
| { |
| "epoch": 3.5356303837118244, |
| "grad_norm": 1.2832853162474576, |
| "learning_rate": 4.786685826049508e-06, |
| "loss": 0.5227, |
| "step": 1128 |
| }, |
| { |
| "epoch": 3.53876272513704, |
| "grad_norm": 1.20463188289643, |
| "learning_rate": 4.768016163021754e-06, |
| "loss": 0.5343, |
| "step": 1129 |
| }, |
| { |
| "epoch": 3.5418950665622555, |
| "grad_norm": 1.1932857110540964, |
| "learning_rate": 4.749371576219317e-06, |
| "loss": 0.4933, |
| "step": 1130 |
| }, |
| { |
| "epoch": 3.5450274079874706, |
| "grad_norm": 1.1785435772368695, |
| "learning_rate": 4.730752155003303e-06, |
| "loss": 0.5262, |
| "step": 1131 |
| }, |
| { |
| "epoch": 3.5481597494126857, |
| "grad_norm": 1.1881352282816962, |
| "learning_rate": 4.7121579886142065e-06, |
| "loss": 0.5564, |
| "step": 1132 |
| }, |
| { |
| "epoch": 3.5512920908379013, |
| "grad_norm": 1.1608651788367013, |
| "learning_rate": 4.693589166171466e-06, |
| "loss": 0.5287, |
| "step": 1133 |
| }, |
| { |
| "epoch": 3.554424432263117, |
| "grad_norm": 1.2094527888449869, |
| "learning_rate": 4.675045776673067e-06, |
| "loss": 0.5463, |
| "step": 1134 |
| }, |
| { |
| "epoch": 3.557556773688332, |
| "grad_norm": 1.1971780880370575, |
| "learning_rate": 4.65652790899508e-06, |
| "loss": 0.5117, |
| "step": 1135 |
| }, |
| { |
| "epoch": 3.560689115113547, |
| "grad_norm": 1.1896379190718653, |
| "learning_rate": 4.638035651891266e-06, |
| "loss": 0.543, |
| "step": 1136 |
| }, |
| { |
| "epoch": 3.5638214565387627, |
| "grad_norm": 1.123251870010005, |
| "learning_rate": 4.619569093992636e-06, |
| "loss": 0.5263, |
| "step": 1137 |
| }, |
| { |
| "epoch": 3.5669537979639783, |
| "grad_norm": 1.1685045536154566, |
| "learning_rate": 4.601128323807023e-06, |
| "loss": 0.5341, |
| "step": 1138 |
| }, |
| { |
| "epoch": 3.5700861393891934, |
| "grad_norm": 1.1479704004071765, |
| "learning_rate": 4.582713429718667e-06, |
| "loss": 0.5503, |
| "step": 1139 |
| }, |
| { |
| "epoch": 3.5732184808144085, |
| "grad_norm": 1.138427168032927, |
| "learning_rate": 4.56432449998779e-06, |
| "loss": 0.492, |
| "step": 1140 |
| }, |
| { |
| "epoch": 3.576350822239624, |
| "grad_norm": 1.2372766348911877, |
| "learning_rate": 4.545961622750158e-06, |
| "loss": 0.5512, |
| "step": 1141 |
| }, |
| { |
| "epoch": 3.5794831636648397, |
| "grad_norm": 1.1774919589788904, |
| "learning_rate": 4.527624886016686e-06, |
| "loss": 0.5398, |
| "step": 1142 |
| }, |
| { |
| "epoch": 3.5826155050900548, |
| "grad_norm": 1.1693438413445578, |
| "learning_rate": 4.509314377672987e-06, |
| "loss": 0.5444, |
| "step": 1143 |
| }, |
| { |
| "epoch": 3.58574784651527, |
| "grad_norm": 1.206456317247258, |
| "learning_rate": 4.491030185478976e-06, |
| "loss": 0.5362, |
| "step": 1144 |
| }, |
| { |
| "epoch": 3.5888801879404855, |
| "grad_norm": 1.1483671630028152, |
| "learning_rate": 4.472772397068431e-06, |
| "loss": 0.5115, |
| "step": 1145 |
| }, |
| { |
| "epoch": 3.592012529365701, |
| "grad_norm": 1.2246520853364462, |
| "learning_rate": 4.454541099948585e-06, |
| "loss": 0.5305, |
| "step": 1146 |
| }, |
| { |
| "epoch": 3.595144870790916, |
| "grad_norm": 1.1831900976113294, |
| "learning_rate": 4.436336381499701e-06, |
| "loss": 0.5413, |
| "step": 1147 |
| }, |
| { |
| "epoch": 3.5982772122161313, |
| "grad_norm": 1.2043230578018518, |
| "learning_rate": 4.418158328974647e-06, |
| "loss": 0.5196, |
| "step": 1148 |
| }, |
| { |
| "epoch": 3.601409553641347, |
| "grad_norm": 1.1773163602914556, |
| "learning_rate": 4.40000702949849e-06, |
| "loss": 0.5001, |
| "step": 1149 |
| }, |
| { |
| "epoch": 3.6045418950665624, |
| "grad_norm": 1.1478622164586656, |
| "learning_rate": 4.381882570068079e-06, |
| "loss": 0.5303, |
| "step": 1150 |
| }, |
| { |
| "epoch": 3.6076742364917775, |
| "grad_norm": 1.191685305824786, |
| "learning_rate": 4.3637850375516085e-06, |
| "loss": 0.5137, |
| "step": 1151 |
| }, |
| { |
| "epoch": 3.610806577916993, |
| "grad_norm": 1.181732648263047, |
| "learning_rate": 4.345714518688226e-06, |
| "loss": 0.5246, |
| "step": 1152 |
| }, |
| { |
| "epoch": 3.6139389193422082, |
| "grad_norm": 1.1923764986284755, |
| "learning_rate": 4.327671100087605e-06, |
| "loss": 0.5327, |
| "step": 1153 |
| }, |
| { |
| "epoch": 3.617071260767424, |
| "grad_norm": 1.20809690015103, |
| "learning_rate": 4.3096548682295304e-06, |
| "loss": 0.5498, |
| "step": 1154 |
| }, |
| { |
| "epoch": 3.620203602192639, |
| "grad_norm": 1.1491784685338666, |
| "learning_rate": 4.291665909463477e-06, |
| "loss": 0.5041, |
| "step": 1155 |
| }, |
| { |
| "epoch": 3.6233359436178545, |
| "grad_norm": 1.1751280305249536, |
| "learning_rate": 4.273704310008213e-06, |
| "loss": 0.5229, |
| "step": 1156 |
| }, |
| { |
| "epoch": 3.6264682850430696, |
| "grad_norm": 1.1175615334495055, |
| "learning_rate": 4.255770155951374e-06, |
| "loss": 0.5215, |
| "step": 1157 |
| }, |
| { |
| "epoch": 3.629600626468285, |
| "grad_norm": 1.155032398758409, |
| "learning_rate": 4.237863533249057e-06, |
| "loss": 0.5344, |
| "step": 1158 |
| }, |
| { |
| "epoch": 3.6327329678935003, |
| "grad_norm": 1.1988336359099494, |
| "learning_rate": 4.21998452772539e-06, |
| "loss": 0.5579, |
| "step": 1159 |
| }, |
| { |
| "epoch": 3.635865309318716, |
| "grad_norm": 1.1698366279218717, |
| "learning_rate": 4.202133225072153e-06, |
| "loss": 0.5288, |
| "step": 1160 |
| }, |
| { |
| "epoch": 3.638997650743931, |
| "grad_norm": 1.1286869148862564, |
| "learning_rate": 4.1843097108483465e-06, |
| "loss": 0.5245, |
| "step": 1161 |
| }, |
| { |
| "epoch": 3.6421299921691466, |
| "grad_norm": 1.1861981743925876, |
| "learning_rate": 4.16651407047977e-06, |
| "loss": 0.5391, |
| "step": 1162 |
| }, |
| { |
| "epoch": 3.6452623335943617, |
| "grad_norm": 1.2269024042403185, |
| "learning_rate": 4.148746389258645e-06, |
| "loss": 0.5417, |
| "step": 1163 |
| }, |
| { |
| "epoch": 3.6483946750195773, |
| "grad_norm": 1.2098949537700534, |
| "learning_rate": 4.1310067523431795e-06, |
| "loss": 0.5187, |
| "step": 1164 |
| }, |
| { |
| "epoch": 3.6515270164447924, |
| "grad_norm": 1.2311237257384957, |
| "learning_rate": 4.113295244757171e-06, |
| "loss": 0.5112, |
| "step": 1165 |
| }, |
| { |
| "epoch": 3.654659357870008, |
| "grad_norm": 1.1381539159403147, |
| "learning_rate": 4.0956119513896005e-06, |
| "loss": 0.5065, |
| "step": 1166 |
| }, |
| { |
| "epoch": 3.657791699295223, |
| "grad_norm": 1.1621884247268668, |
| "learning_rate": 4.077956956994212e-06, |
| "loss": 0.511, |
| "step": 1167 |
| }, |
| { |
| "epoch": 3.6609240407204386, |
| "grad_norm": 1.168551137549956, |
| "learning_rate": 4.060330346189125e-06, |
| "loss": 0.5177, |
| "step": 1168 |
| }, |
| { |
| "epoch": 3.6640563821456538, |
| "grad_norm": 1.210641000943774, |
| "learning_rate": 4.042732203456413e-06, |
| "loss": 0.5197, |
| "step": 1169 |
| }, |
| { |
| "epoch": 3.6671887235708693, |
| "grad_norm": 1.1705201776756187, |
| "learning_rate": 4.0251626131417135e-06, |
| "loss": 0.5467, |
| "step": 1170 |
| }, |
| { |
| "epoch": 3.6703210649960845, |
| "grad_norm": 1.1690543008503043, |
| "learning_rate": 4.00762165945381e-06, |
| "loss": 0.5404, |
| "step": 1171 |
| }, |
| { |
| "epoch": 3.6734534064213, |
| "grad_norm": 1.1598971397306825, |
| "learning_rate": 3.990109426464237e-06, |
| "loss": 0.5178, |
| "step": 1172 |
| }, |
| { |
| "epoch": 3.676585747846515, |
| "grad_norm": 1.1616293085138196, |
| "learning_rate": 3.972625998106876e-06, |
| "loss": 0.5435, |
| "step": 1173 |
| }, |
| { |
| "epoch": 3.6797180892717307, |
| "grad_norm": 1.1407718103960551, |
| "learning_rate": 3.955171458177551e-06, |
| "loss": 0.5406, |
| "step": 1174 |
| }, |
| { |
| "epoch": 3.682850430696946, |
| "grad_norm": 1.1669175281894106, |
| "learning_rate": 3.937745890333623e-06, |
| "loss": 0.5251, |
| "step": 1175 |
| }, |
| { |
| "epoch": 3.6859827721221614, |
| "grad_norm": 1.1257651461718745, |
| "learning_rate": 3.9203493780935935e-06, |
| "loss": 0.5345, |
| "step": 1176 |
| }, |
| { |
| "epoch": 3.6891151135473765, |
| "grad_norm": 1.1540314221528836, |
| "learning_rate": 3.902982004836709e-06, |
| "loss": 0.5351, |
| "step": 1177 |
| }, |
| { |
| "epoch": 3.692247454972592, |
| "grad_norm": 1.1524838421819237, |
| "learning_rate": 3.885643853802557e-06, |
| "loss": 0.5488, |
| "step": 1178 |
| }, |
| { |
| "epoch": 3.6953797963978072, |
| "grad_norm": 1.188466802503505, |
| "learning_rate": 3.868335008090664e-06, |
| "loss": 0.5142, |
| "step": 1179 |
| }, |
| { |
| "epoch": 3.698512137823023, |
| "grad_norm": 1.138912921077403, |
| "learning_rate": 3.851055550660098e-06, |
| "loss": 0.5269, |
| "step": 1180 |
| }, |
| { |
| "epoch": 3.701644479248238, |
| "grad_norm": 1.1870550813670917, |
| "learning_rate": 3.8338055643290815e-06, |
| "loss": 0.527, |
| "step": 1181 |
| }, |
| { |
| "epoch": 3.7047768206734535, |
| "grad_norm": 1.1829270162712773, |
| "learning_rate": 3.8165851317745705e-06, |
| "loss": 0.526, |
| "step": 1182 |
| }, |
| { |
| "epoch": 3.7079091620986686, |
| "grad_norm": 1.2162157673701168, |
| "learning_rate": 3.799394335531891e-06, |
| "loss": 0.5644, |
| "step": 1183 |
| }, |
| { |
| "epoch": 3.711041503523884, |
| "grad_norm": 1.1789615458756362, |
| "learning_rate": 3.782233257994309e-06, |
| "loss": 0.5369, |
| "step": 1184 |
| }, |
| { |
| "epoch": 3.7141738449490993, |
| "grad_norm": 1.1676281601812897, |
| "learning_rate": 3.7651019814126656e-06, |
| "loss": 0.5243, |
| "step": 1185 |
| }, |
| { |
| "epoch": 3.717306186374315, |
| "grad_norm": 1.1571405119696987, |
| "learning_rate": 3.7480005878949645e-06, |
| "loss": 0.5584, |
| "step": 1186 |
| }, |
| { |
| "epoch": 3.72043852779953, |
| "grad_norm": 1.1512138914116103, |
| "learning_rate": 3.730929159405984e-06, |
| "loss": 0.5272, |
| "step": 1187 |
| }, |
| { |
| "epoch": 3.7235708692247456, |
| "grad_norm": 1.1290612876998436, |
| "learning_rate": 3.713887777766888e-06, |
| "loss": 0.5212, |
| "step": 1188 |
| }, |
| { |
| "epoch": 3.7267032106499607, |
| "grad_norm": 1.1615587465761752, |
| "learning_rate": 3.69687652465482e-06, |
| "loss": 0.5271, |
| "step": 1189 |
| }, |
| { |
| "epoch": 3.7298355520751763, |
| "grad_norm": 1.165644217808434, |
| "learning_rate": 3.679895481602529e-06, |
| "loss": 0.5268, |
| "step": 1190 |
| }, |
| { |
| "epoch": 3.7329678935003914, |
| "grad_norm": 1.1460726933346908, |
| "learning_rate": 3.6629447299979747e-06, |
| "loss": 0.5345, |
| "step": 1191 |
| }, |
| { |
| "epoch": 3.736100234925607, |
| "grad_norm": 1.1701380925076978, |
| "learning_rate": 3.64602435108392e-06, |
| "loss": 0.5482, |
| "step": 1192 |
| }, |
| { |
| "epoch": 3.7392325763508225, |
| "grad_norm": 1.1069644314839941, |
| "learning_rate": 3.629134425957567e-06, |
| "loss": 0.4975, |
| "step": 1193 |
| }, |
| { |
| "epoch": 3.7423649177760376, |
| "grad_norm": 1.1742604377155452, |
| "learning_rate": 3.612275035570155e-06, |
| "loss": 0.5382, |
| "step": 1194 |
| }, |
| { |
| "epoch": 3.7454972592012528, |
| "grad_norm": 1.1594884695567513, |
| "learning_rate": 3.595446260726576e-06, |
| "loss": 0.5276, |
| "step": 1195 |
| }, |
| { |
| "epoch": 3.7486296006264683, |
| "grad_norm": 1.2009203643761532, |
| "learning_rate": 3.578648182084975e-06, |
| "loss": 0.5377, |
| "step": 1196 |
| }, |
| { |
| "epoch": 3.751761942051684, |
| "grad_norm": 1.170884036670357, |
| "learning_rate": 3.561880880156384e-06, |
| "loss": 0.5576, |
| "step": 1197 |
| }, |
| { |
| "epoch": 3.754894283476899, |
| "grad_norm": 1.1959487905145205, |
| "learning_rate": 3.5451444353043263e-06, |
| "loss": 0.5373, |
| "step": 1198 |
| }, |
| { |
| "epoch": 3.758026624902114, |
| "grad_norm": 1.1796481257439184, |
| "learning_rate": 3.5284389277444296e-06, |
| "loss": 0.5313, |
| "step": 1199 |
| }, |
| { |
| "epoch": 3.7611589663273297, |
| "grad_norm": 1.2094185606216385, |
| "learning_rate": 3.511764437544036e-06, |
| "loss": 0.5162, |
| "step": 1200 |
| }, |
| { |
| "epoch": 3.7642913077525453, |
| "grad_norm": 1.1754904046598893, |
| "learning_rate": 3.4951210446218385e-06, |
| "loss": 0.5529, |
| "step": 1201 |
| }, |
| { |
| "epoch": 3.7674236491777604, |
| "grad_norm": 1.153840764724347, |
| "learning_rate": 3.478508828747471e-06, |
| "loss": 0.4987, |
| "step": 1202 |
| }, |
| { |
| "epoch": 3.7705559906029755, |
| "grad_norm": 1.1333187974311192, |
| "learning_rate": 3.4619278695411495e-06, |
| "loss": 0.5005, |
| "step": 1203 |
| }, |
| { |
| "epoch": 3.773688332028191, |
| "grad_norm": 1.1455993988251625, |
| "learning_rate": 3.445378246473279e-06, |
| "loss": 0.5115, |
| "step": 1204 |
| }, |
| { |
| "epoch": 3.7768206734534067, |
| "grad_norm": 1.1749646759832453, |
| "learning_rate": 3.4288600388640714e-06, |
| "loss": 0.5241, |
| "step": 1205 |
| }, |
| { |
| "epoch": 3.779953014878622, |
| "grad_norm": 1.159353327088895, |
| "learning_rate": 3.4123733258831727e-06, |
| "loss": 0.5221, |
| "step": 1206 |
| }, |
| { |
| "epoch": 3.783085356303837, |
| "grad_norm": 1.1770098803771174, |
| "learning_rate": 3.3959181865492798e-06, |
| "loss": 0.536, |
| "step": 1207 |
| }, |
| { |
| "epoch": 3.7862176977290525, |
| "grad_norm": 1.173049503757697, |
| "learning_rate": 3.3794946997297527e-06, |
| "loss": 0.5423, |
| "step": 1208 |
| }, |
| { |
| "epoch": 3.789350039154268, |
| "grad_norm": 1.2771010631863746, |
| "learning_rate": 3.3631029441402497e-06, |
| "loss": 0.5587, |
| "step": 1209 |
| }, |
| { |
| "epoch": 3.792482380579483, |
| "grad_norm": 1.193386115742984, |
| "learning_rate": 3.3467429983443477e-06, |
| "loss": 0.5277, |
| "step": 1210 |
| }, |
| { |
| "epoch": 3.7956147220046983, |
| "grad_norm": 1.1915512034829279, |
| "learning_rate": 3.330414940753164e-06, |
| "loss": 0.5583, |
| "step": 1211 |
| }, |
| { |
| "epoch": 3.798747063429914, |
| "grad_norm": 1.1896725640667438, |
| "learning_rate": 3.3141188496249755e-06, |
| "loss": 0.5141, |
| "step": 1212 |
| }, |
| { |
| "epoch": 3.8018794048551294, |
| "grad_norm": 1.1687306108915543, |
| "learning_rate": 3.2978548030648504e-06, |
| "loss": 0.5443, |
| "step": 1213 |
| }, |
| { |
| "epoch": 3.8050117462803446, |
| "grad_norm": 1.1857286795537076, |
| "learning_rate": 3.2816228790242723e-06, |
| "loss": 0.5472, |
| "step": 1214 |
| }, |
| { |
| "epoch": 3.8081440877055597, |
| "grad_norm": 1.138346768843636, |
| "learning_rate": 3.2654231553007665e-06, |
| "loss": 0.549, |
| "step": 1215 |
| }, |
| { |
| "epoch": 3.8112764291307752, |
| "grad_norm": 1.1444752892169308, |
| "learning_rate": 3.249255709537521e-06, |
| "loss": 0.522, |
| "step": 1216 |
| }, |
| { |
| "epoch": 3.814408770555991, |
| "grad_norm": 1.18144246751696, |
| "learning_rate": 3.233120619223021e-06, |
| "loss": 0.5303, |
| "step": 1217 |
| }, |
| { |
| "epoch": 3.817541111981206, |
| "grad_norm": 1.2096091122117285, |
| "learning_rate": 3.2170179616906818e-06, |
| "loss": 0.5207, |
| "step": 1218 |
| }, |
| { |
| "epoch": 3.820673453406421, |
| "grad_norm": 1.1530748784567288, |
| "learning_rate": 3.200947814118467e-06, |
| "loss": 0.5376, |
| "step": 1219 |
| }, |
| { |
| "epoch": 3.8238057948316366, |
| "grad_norm": 1.1215171003312445, |
| "learning_rate": 3.1849102535285283e-06, |
| "loss": 0.5082, |
| "step": 1220 |
| }, |
| { |
| "epoch": 3.826938136256852, |
| "grad_norm": 1.1703733058989874, |
| "learning_rate": 3.1689053567868278e-06, |
| "loss": 0.5337, |
| "step": 1221 |
| }, |
| { |
| "epoch": 3.8300704776820673, |
| "grad_norm": 1.130262675345898, |
| "learning_rate": 3.1529332006027813e-06, |
| "loss": 0.5404, |
| "step": 1222 |
| }, |
| { |
| "epoch": 3.8332028191072824, |
| "grad_norm": 1.1669511760006823, |
| "learning_rate": 3.1369938615288707e-06, |
| "loss": 0.5311, |
| "step": 1223 |
| }, |
| { |
| "epoch": 3.836335160532498, |
| "grad_norm": 1.1656001584283648, |
| "learning_rate": 3.1210874159603044e-06, |
| "loss": 0.5422, |
| "step": 1224 |
| }, |
| { |
| "epoch": 3.8394675019577136, |
| "grad_norm": 1.1747394216229974, |
| "learning_rate": 3.1052139401346216e-06, |
| "loss": 0.5132, |
| "step": 1225 |
| }, |
| { |
| "epoch": 3.8425998433829287, |
| "grad_norm": 1.1495579361244552, |
| "learning_rate": 3.089373510131354e-06, |
| "loss": 0.493, |
| "step": 1226 |
| }, |
| { |
| "epoch": 3.845732184808144, |
| "grad_norm": 1.180793820104102, |
| "learning_rate": 3.0735662018716437e-06, |
| "loss": 0.5291, |
| "step": 1227 |
| }, |
| { |
| "epoch": 3.8488645262333594, |
| "grad_norm": 1.124911552092485, |
| "learning_rate": 3.057792091117889e-06, |
| "loss": 0.5189, |
| "step": 1228 |
| }, |
| { |
| "epoch": 3.851996867658575, |
| "grad_norm": 1.120771756001303, |
| "learning_rate": 3.042051253473366e-06, |
| "loss": 0.4857, |
| "step": 1229 |
| }, |
| { |
| "epoch": 3.85512920908379, |
| "grad_norm": 1.1265839282142127, |
| "learning_rate": 3.026343764381887e-06, |
| "loss": 0.5188, |
| "step": 1230 |
| }, |
| { |
| "epoch": 3.858261550509005, |
| "grad_norm": 1.153903476616413, |
| "learning_rate": 3.0106696991274278e-06, |
| "loss": 0.518, |
| "step": 1231 |
| }, |
| { |
| "epoch": 3.861393891934221, |
| "grad_norm": 1.1653054593485324, |
| "learning_rate": 2.995029132833768e-06, |
| "loss": 0.5413, |
| "step": 1232 |
| }, |
| { |
| "epoch": 3.8645262333594363, |
| "grad_norm": 1.1219932990648924, |
| "learning_rate": 2.9794221404641233e-06, |
| "loss": 0.5584, |
| "step": 1233 |
| }, |
| { |
| "epoch": 3.8676585747846515, |
| "grad_norm": 1.1764250161654142, |
| "learning_rate": 2.9638487968208028e-06, |
| "loss": 0.5485, |
| "step": 1234 |
| }, |
| { |
| "epoch": 3.870790916209867, |
| "grad_norm": 1.1756895255385884, |
| "learning_rate": 2.9483091765448426e-06, |
| "loss": 0.5223, |
| "step": 1235 |
| }, |
| { |
| "epoch": 3.873923257635082, |
| "grad_norm": 1.1493491848055564, |
| "learning_rate": 2.9328033541156363e-06, |
| "loss": 0.5137, |
| "step": 1236 |
| }, |
| { |
| "epoch": 3.8770555990602977, |
| "grad_norm": 1.1435017702746313, |
| "learning_rate": 2.9173314038505986e-06, |
| "loss": 0.5404, |
| "step": 1237 |
| }, |
| { |
| "epoch": 3.880187940485513, |
| "grad_norm": 1.1617452479406913, |
| "learning_rate": 2.901893399904797e-06, |
| "loss": 0.5315, |
| "step": 1238 |
| }, |
| { |
| "epoch": 3.8833202819107284, |
| "grad_norm": 1.2030832962067164, |
| "learning_rate": 2.886489416270598e-06, |
| "loss": 0.5385, |
| "step": 1239 |
| }, |
| { |
| "epoch": 3.8864526233359435, |
| "grad_norm": 1.1328735216805788, |
| "learning_rate": 2.871119526777315e-06, |
| "loss": 0.5199, |
| "step": 1240 |
| }, |
| { |
| "epoch": 3.889584964761159, |
| "grad_norm": 1.1928234546236907, |
| "learning_rate": 2.855783805090846e-06, |
| "loss": 0.5258, |
| "step": 1241 |
| }, |
| { |
| "epoch": 3.8927173061863742, |
| "grad_norm": 1.208100129151956, |
| "learning_rate": 2.8404823247133373e-06, |
| "loss": 0.5168, |
| "step": 1242 |
| }, |
| { |
| "epoch": 3.89584964761159, |
| "grad_norm": 1.1414115708828445, |
| "learning_rate": 2.825215158982809e-06, |
| "loss": 0.5429, |
| "step": 1243 |
| }, |
| { |
| "epoch": 3.898981989036805, |
| "grad_norm": 1.184617005441351, |
| "learning_rate": 2.8099823810728255e-06, |
| "loss": 0.5604, |
| "step": 1244 |
| }, |
| { |
| "epoch": 3.9021143304620205, |
| "grad_norm": 1.1378204967249268, |
| "learning_rate": 2.7947840639921308e-06, |
| "loss": 0.528, |
| "step": 1245 |
| }, |
| { |
| "epoch": 3.9052466718872356, |
| "grad_norm": 1.1876757504170918, |
| "learning_rate": 2.779620280584303e-06, |
| "loss": 0.5226, |
| "step": 1246 |
| }, |
| { |
| "epoch": 3.908379013312451, |
| "grad_norm": 1.1809603854780806, |
| "learning_rate": 2.764491103527406e-06, |
| "loss": 0.5266, |
| "step": 1247 |
| }, |
| { |
| "epoch": 3.9115113547376663, |
| "grad_norm": 1.1873512461852518, |
| "learning_rate": 2.7493966053336397e-06, |
| "loss": 0.5414, |
| "step": 1248 |
| }, |
| { |
| "epoch": 3.914643696162882, |
| "grad_norm": 1.1504208913330314, |
| "learning_rate": 2.734336858348987e-06, |
| "loss": 0.5283, |
| "step": 1249 |
| }, |
| { |
| "epoch": 3.917776037588097, |
| "grad_norm": 1.143048240337088, |
| "learning_rate": 2.7193119347528842e-06, |
| "loss": 0.5337, |
| "step": 1250 |
| }, |
| { |
| "epoch": 3.9209083790133126, |
| "grad_norm": 1.1689363408230222, |
| "learning_rate": 2.7043219065578497e-06, |
| "loss": 0.557, |
| "step": 1251 |
| }, |
| { |
| "epoch": 3.9240407204385277, |
| "grad_norm": 1.1555945926450477, |
| "learning_rate": 2.6893668456091627e-06, |
| "loss": 0.5256, |
| "step": 1252 |
| }, |
| { |
| "epoch": 3.9271730618637433, |
| "grad_norm": 1.158032059169674, |
| "learning_rate": 2.674446823584507e-06, |
| "loss": 0.5011, |
| "step": 1253 |
| }, |
| { |
| "epoch": 3.9303054032889584, |
| "grad_norm": 1.163791390523606, |
| "learning_rate": 2.6595619119936267e-06, |
| "loss": 0.5324, |
| "step": 1254 |
| }, |
| { |
| "epoch": 3.933437744714174, |
| "grad_norm": 1.1473109388161766, |
| "learning_rate": 2.644712182177992e-06, |
| "loss": 0.5454, |
| "step": 1255 |
| }, |
| { |
| "epoch": 3.936570086139389, |
| "grad_norm": 1.1721160935749964, |
| "learning_rate": 2.62989770531044e-06, |
| "loss": 0.5144, |
| "step": 1256 |
| }, |
| { |
| "epoch": 3.9397024275646046, |
| "grad_norm": 1.2026978621361013, |
| "learning_rate": 2.6151185523948577e-06, |
| "loss": 0.5587, |
| "step": 1257 |
| }, |
| { |
| "epoch": 3.9428347689898198, |
| "grad_norm": 1.161111467446485, |
| "learning_rate": 2.6003747942658243e-06, |
| "loss": 0.5172, |
| "step": 1258 |
| }, |
| { |
| "epoch": 3.9459671104150353, |
| "grad_norm": 1.1895161788905066, |
| "learning_rate": 2.5856665015882686e-06, |
| "loss": 0.5647, |
| "step": 1259 |
| }, |
| { |
| "epoch": 3.9490994518402505, |
| "grad_norm": 1.1667391962309854, |
| "learning_rate": 2.570993744857151e-06, |
| "loss": 0.539, |
| "step": 1260 |
| }, |
| { |
| "epoch": 3.952231793265466, |
| "grad_norm": 1.1990408248065403, |
| "learning_rate": 2.5563565943971027e-06, |
| "loss": 0.5308, |
| "step": 1261 |
| }, |
| { |
| "epoch": 3.955364134690681, |
| "grad_norm": 1.177576874471793, |
| "learning_rate": 2.541755120362108e-06, |
| "loss": 0.5302, |
| "step": 1262 |
| }, |
| { |
| "epoch": 3.9584964761158967, |
| "grad_norm": 1.1329715387380235, |
| "learning_rate": 2.5271893927351464e-06, |
| "loss": 0.529, |
| "step": 1263 |
| }, |
| { |
| "epoch": 3.961628817541112, |
| "grad_norm": 1.20780770381308, |
| "learning_rate": 2.5126594813278792e-06, |
| "loss": 0.5298, |
| "step": 1264 |
| }, |
| { |
| "epoch": 3.9647611589663274, |
| "grad_norm": 1.2074087314834707, |
| "learning_rate": 2.4981654557803026e-06, |
| "loss": 0.5553, |
| "step": 1265 |
| }, |
| { |
| "epoch": 3.9678935003915425, |
| "grad_norm": 1.176718451434302, |
| "learning_rate": 2.4837073855604186e-06, |
| "loss": 0.5376, |
| "step": 1266 |
| }, |
| { |
| "epoch": 3.971025841816758, |
| "grad_norm": 1.1582592387697839, |
| "learning_rate": 2.469285339963892e-06, |
| "loss": 0.5016, |
| "step": 1267 |
| }, |
| { |
| "epoch": 3.9741581832419732, |
| "grad_norm": 1.126175949855287, |
| "learning_rate": 2.4548993881137327e-06, |
| "loss": 0.5164, |
| "step": 1268 |
| }, |
| { |
| "epoch": 3.977290524667189, |
| "grad_norm": 1.182910505553956, |
| "learning_rate": 2.4405495989599627e-06, |
| "loss": 0.5606, |
| "step": 1269 |
| }, |
| { |
| "epoch": 3.980422866092404, |
| "grad_norm": 1.1661378582361217, |
| "learning_rate": 2.4262360412792663e-06, |
| "loss": 0.552, |
| "step": 1270 |
| }, |
| { |
| "epoch": 3.9835552075176195, |
| "grad_norm": 1.1721964398123115, |
| "learning_rate": 2.4119587836746885e-06, |
| "loss": 0.5481, |
| "step": 1271 |
| }, |
| { |
| "epoch": 3.9866875489428346, |
| "grad_norm": 1.1861569431208108, |
| "learning_rate": 2.3977178945752877e-06, |
| "loss": 0.5361, |
| "step": 1272 |
| }, |
| { |
| "epoch": 3.98981989036805, |
| "grad_norm": 1.142933675959546, |
| "learning_rate": 2.383513442235812e-06, |
| "loss": 0.5216, |
| "step": 1273 |
| }, |
| { |
| "epoch": 3.9929522317932653, |
| "grad_norm": 1.173762534357803, |
| "learning_rate": 2.3693454947363782e-06, |
| "loss": 0.5545, |
| "step": 1274 |
| }, |
| { |
| "epoch": 3.996084573218481, |
| "grad_norm": 1.1831215530299448, |
| "learning_rate": 2.35521411998213e-06, |
| "loss": 0.5296, |
| "step": 1275 |
| }, |
| { |
| "epoch": 3.999216914643696, |
| "grad_norm": 1.1124833568408279, |
| "learning_rate": 2.341119385702929e-06, |
| "loss": 0.5065, |
| "step": 1276 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 1.1124833568408279, |
| "learning_rate": 2.3270613594530235e-06, |
| "loss": 0.5165, |
| "step": 1277 |
| }, |
| { |
| "epoch": 4.003132341425215, |
| "grad_norm": 2.6425185615357405, |
| "learning_rate": 2.3130401086107247e-06, |
| "loss": 0.4391, |
| "step": 1278 |
| }, |
| { |
| "epoch": 4.006264682850431, |
| "grad_norm": 1.8968989656450992, |
| "learning_rate": 2.2990557003780833e-06, |
| "loss": 0.4202, |
| "step": 1279 |
| }, |
| { |
| "epoch": 4.009397024275646, |
| "grad_norm": 1.8142320294338747, |
| "learning_rate": 2.2851082017805704e-06, |
| "loss": 0.4009, |
| "step": 1280 |
| }, |
| { |
| "epoch": 4.012529365700861, |
| "grad_norm": 1.7165806912562718, |
| "learning_rate": 2.27119767966675e-06, |
| "loss": 0.405, |
| "step": 1281 |
| }, |
| { |
| "epoch": 4.0156617071260765, |
| "grad_norm": 1.5924208377321079, |
| "learning_rate": 2.2573242007079697e-06, |
| "loss": 0.421, |
| "step": 1282 |
| }, |
| { |
| "epoch": 4.0187940485512925, |
| "grad_norm": 1.5038694947047038, |
| "learning_rate": 2.2434878313980235e-06, |
| "loss": 0.4002, |
| "step": 1283 |
| }, |
| { |
| "epoch": 4.021926389976508, |
| "grad_norm": 1.340573966307308, |
| "learning_rate": 2.2296886380528494e-06, |
| "loss": 0.4272, |
| "step": 1284 |
| }, |
| { |
| "epoch": 4.025058731401723, |
| "grad_norm": 1.2987422109641342, |
| "learning_rate": 2.215926686810206e-06, |
| "loss": 0.4349, |
| "step": 1285 |
| }, |
| { |
| "epoch": 4.028191072826938, |
| "grad_norm": 1.252719539838883, |
| "learning_rate": 2.202202043629356e-06, |
| "loss": 0.3898, |
| "step": 1286 |
| }, |
| { |
| "epoch": 4.031323414252154, |
| "grad_norm": 1.327752575517584, |
| "learning_rate": 2.1885147742907465e-06, |
| "loss": 0.4175, |
| "step": 1287 |
| }, |
| { |
| "epoch": 4.034455755677369, |
| "grad_norm": 1.4402839116286432, |
| "learning_rate": 2.1748649443957004e-06, |
| "loss": 0.4137, |
| "step": 1288 |
| }, |
| { |
| "epoch": 4.037588097102584, |
| "grad_norm": 1.5655168444654703, |
| "learning_rate": 2.1612526193660975e-06, |
| "loss": 0.4051, |
| "step": 1289 |
| }, |
| { |
| "epoch": 4.040720438527799, |
| "grad_norm": 1.6161366724042316, |
| "learning_rate": 2.1476778644440553e-06, |
| "loss": 0.4028, |
| "step": 1290 |
| }, |
| { |
| "epoch": 4.043852779953015, |
| "grad_norm": 1.5596875540342134, |
| "learning_rate": 2.134140744691633e-06, |
| "loss": 0.3997, |
| "step": 1291 |
| }, |
| { |
| "epoch": 4.04698512137823, |
| "grad_norm": 1.4460862400789365, |
| "learning_rate": 2.1206413249904967e-06, |
| "loss": 0.4125, |
| "step": 1292 |
| }, |
| { |
| "epoch": 4.0501174628034455, |
| "grad_norm": 1.3882745890795896, |
| "learning_rate": 2.1071796700416334e-06, |
| "loss": 0.3972, |
| "step": 1293 |
| }, |
| { |
| "epoch": 4.053249804228661, |
| "grad_norm": 1.4012869387398328, |
| "learning_rate": 2.093755844365021e-06, |
| "loss": 0.4155, |
| "step": 1294 |
| }, |
| { |
| "epoch": 4.056382145653877, |
| "grad_norm": 1.3519015902940699, |
| "learning_rate": 2.0803699122993293e-06, |
| "loss": 0.4178, |
| "step": 1295 |
| }, |
| { |
| "epoch": 4.059514487079092, |
| "grad_norm": 1.3602338471577913, |
| "learning_rate": 2.067021938001611e-06, |
| "loss": 0.4193, |
| "step": 1296 |
| }, |
| { |
| "epoch": 4.062646828504307, |
| "grad_norm": 1.2707959489599356, |
| "learning_rate": 2.0537119854469845e-06, |
| "loss": 0.4235, |
| "step": 1297 |
| }, |
| { |
| "epoch": 4.065779169929522, |
| "grad_norm": 1.2304315457883275, |
| "learning_rate": 2.0404401184283408e-06, |
| "loss": 0.4281, |
| "step": 1298 |
| }, |
| { |
| "epoch": 4.068911511354738, |
| "grad_norm": 1.3126596083493967, |
| "learning_rate": 2.0272064005560354e-06, |
| "loss": 0.4116, |
| "step": 1299 |
| }, |
| { |
| "epoch": 4.072043852779953, |
| "grad_norm": 1.2164141170274532, |
| "learning_rate": 2.01401089525757e-06, |
| "loss": 0.4128, |
| "step": 1300 |
| }, |
| { |
| "epoch": 4.075176194205168, |
| "grad_norm": 1.2157628287632507, |
| "learning_rate": 2.000853665777305e-06, |
| "loss": 0.3966, |
| "step": 1301 |
| }, |
| { |
| "epoch": 4.078308535630383, |
| "grad_norm": 1.1768433581291058, |
| "learning_rate": 1.987734775176151e-06, |
| "loss": 0.3773, |
| "step": 1302 |
| }, |
| { |
| "epoch": 4.081440877055599, |
| "grad_norm": 1.195358771674673, |
| "learning_rate": 1.974654286331258e-06, |
| "loss": 0.4021, |
| "step": 1303 |
| }, |
| { |
| "epoch": 4.084573218480815, |
| "grad_norm": 1.2360866223981186, |
| "learning_rate": 1.961612261935727e-06, |
| "loss": 0.3914, |
| "step": 1304 |
| }, |
| { |
| "epoch": 4.08770555990603, |
| "grad_norm": 1.2186686913113904, |
| "learning_rate": 1.9486087644983056e-06, |
| "loss": 0.3896, |
| "step": 1305 |
| }, |
| { |
| "epoch": 4.090837901331245, |
| "grad_norm": 1.2346935215427521, |
| "learning_rate": 1.9356438563430822e-06, |
| "loss": 0.4038, |
| "step": 1306 |
| }, |
| { |
| "epoch": 4.093970242756461, |
| "grad_norm": 1.198362834738886, |
| "learning_rate": 1.9227175996091952e-06, |
| "loss": 0.4009, |
| "step": 1307 |
| }, |
| { |
| "epoch": 4.097102584181676, |
| "grad_norm": 1.1668330606150041, |
| "learning_rate": 1.9098300562505266e-06, |
| "loss": 0.3942, |
| "step": 1308 |
| }, |
| { |
| "epoch": 4.100234925606891, |
| "grad_norm": 1.2063732731647043, |
| "learning_rate": 1.8969812880354167e-06, |
| "loss": 0.394, |
| "step": 1309 |
| }, |
| { |
| "epoch": 4.103367267032106, |
| "grad_norm": 1.2113132171130907, |
| "learning_rate": 1.8841713565463548e-06, |
| "loss": 0.3945, |
| "step": 1310 |
| }, |
| { |
| "epoch": 4.106499608457322, |
| "grad_norm": 1.2247544654973348, |
| "learning_rate": 1.871400323179695e-06, |
| "loss": 0.4096, |
| "step": 1311 |
| }, |
| { |
| "epoch": 4.109631949882537, |
| "grad_norm": 1.2881131176691998, |
| "learning_rate": 1.858668249145359e-06, |
| "loss": 0.4071, |
| "step": 1312 |
| }, |
| { |
| "epoch": 4.1127642913077525, |
| "grad_norm": 1.2380244490730665, |
| "learning_rate": 1.8459751954665372e-06, |
| "loss": 0.3887, |
| "step": 1313 |
| }, |
| { |
| "epoch": 4.115896632732968, |
| "grad_norm": 1.2457742852610907, |
| "learning_rate": 1.8333212229794039e-06, |
| "loss": 0.3938, |
| "step": 1314 |
| }, |
| { |
| "epoch": 4.119028974158184, |
| "grad_norm": 1.299735360583674, |
| "learning_rate": 1.820706392332824e-06, |
| "loss": 0.4278, |
| "step": 1315 |
| }, |
| { |
| "epoch": 4.122161315583399, |
| "grad_norm": 1.22006539661709, |
| "learning_rate": 1.8081307639880519e-06, |
| "loss": 0.3969, |
| "step": 1316 |
| }, |
| { |
| "epoch": 4.125293657008614, |
| "grad_norm": 1.2043704461002585, |
| "learning_rate": 1.7955943982184608e-06, |
| "loss": 0.3772, |
| "step": 1317 |
| }, |
| { |
| "epoch": 4.128425998433829, |
| "grad_norm": 1.1954512153531711, |
| "learning_rate": 1.7830973551092346e-06, |
| "loss": 0.4232, |
| "step": 1318 |
| }, |
| { |
| "epoch": 4.131558339859045, |
| "grad_norm": 1.2198506026583384, |
| "learning_rate": 1.7706396945570947e-06, |
| "loss": 0.3996, |
| "step": 1319 |
| }, |
| { |
| "epoch": 4.13469068128426, |
| "grad_norm": 1.2218183019573885, |
| "learning_rate": 1.7582214762700057e-06, |
| "loss": 0.428, |
| "step": 1320 |
| }, |
| { |
| "epoch": 4.137823022709475, |
| "grad_norm": 1.1976267387815713, |
| "learning_rate": 1.7458427597668893e-06, |
| "loss": 0.4062, |
| "step": 1321 |
| }, |
| { |
| "epoch": 4.14095536413469, |
| "grad_norm": 1.1697967121996906, |
| "learning_rate": 1.7335036043773402e-06, |
| "loss": 0.3889, |
| "step": 1322 |
| }, |
| { |
| "epoch": 4.144087705559906, |
| "grad_norm": 1.1924932065029525, |
| "learning_rate": 1.7212040692413468e-06, |
| "loss": 0.4245, |
| "step": 1323 |
| }, |
| { |
| "epoch": 4.1472200469851215, |
| "grad_norm": 1.2147406012062079, |
| "learning_rate": 1.7089442133089928e-06, |
| "loss": 0.3927, |
| "step": 1324 |
| }, |
| { |
| "epoch": 4.150352388410337, |
| "grad_norm": 1.1788944142485256, |
| "learning_rate": 1.6967240953401954e-06, |
| "loss": 0.3862, |
| "step": 1325 |
| }, |
| { |
| "epoch": 4.153484729835552, |
| "grad_norm": 1.1947326262463431, |
| "learning_rate": 1.6845437739044012e-06, |
| "loss": 0.3828, |
| "step": 1326 |
| }, |
| { |
| "epoch": 4.156617071260768, |
| "grad_norm": 1.1173824303336672, |
| "learning_rate": 1.672403307380328e-06, |
| "loss": 0.3977, |
| "step": 1327 |
| }, |
| { |
| "epoch": 4.159749412685983, |
| "grad_norm": 1.1753324405533268, |
| "learning_rate": 1.6603027539556682e-06, |
| "loss": 0.3866, |
| "step": 1328 |
| }, |
| { |
| "epoch": 4.162881754111198, |
| "grad_norm": 1.1729189872697827, |
| "learning_rate": 1.6482421716268215e-06, |
| "loss": 0.421, |
| "step": 1329 |
| }, |
| { |
| "epoch": 4.166014095536413, |
| "grad_norm": 1.2345796879070534, |
| "learning_rate": 1.6362216181986002e-06, |
| "loss": 0.3713, |
| "step": 1330 |
| }, |
| { |
| "epoch": 4.169146436961629, |
| "grad_norm": 1.155602810779086, |
| "learning_rate": 1.6242411512839728e-06, |
| "loss": 0.387, |
| "step": 1331 |
| }, |
| { |
| "epoch": 4.172278778386844, |
| "grad_norm": 1.1762864409104852, |
| "learning_rate": 1.6123008283037778e-06, |
| "loss": 0.3991, |
| "step": 1332 |
| }, |
| { |
| "epoch": 4.175411119812059, |
| "grad_norm": 1.2221930726179149, |
| "learning_rate": 1.6004007064864468e-06, |
| "loss": 0.3585, |
| "step": 1333 |
| }, |
| { |
| "epoch": 4.1785434612372745, |
| "grad_norm": 1.1528755586082111, |
| "learning_rate": 1.5885408428677296e-06, |
| "loss": 0.4074, |
| "step": 1334 |
| }, |
| { |
| "epoch": 4.1816758026624905, |
| "grad_norm": 1.1990119644346569, |
| "learning_rate": 1.5767212942904275e-06, |
| "loss": 0.4252, |
| "step": 1335 |
| }, |
| { |
| "epoch": 4.184808144087706, |
| "grad_norm": 1.190374191471458, |
| "learning_rate": 1.564942117404119e-06, |
| "loss": 0.4137, |
| "step": 1336 |
| }, |
| { |
| "epoch": 4.187940485512921, |
| "grad_norm": 1.2490135566571798, |
| "learning_rate": 1.5532033686648785e-06, |
| "loss": 0.3961, |
| "step": 1337 |
| }, |
| { |
| "epoch": 4.191072826938136, |
| "grad_norm": 1.211936868558367, |
| "learning_rate": 1.5415051043350194e-06, |
| "loss": 0.3968, |
| "step": 1338 |
| }, |
| { |
| "epoch": 4.194205168363352, |
| "grad_norm": 1.2268059684742754, |
| "learning_rate": 1.5298473804828195e-06, |
| "loss": 0.3992, |
| "step": 1339 |
| }, |
| { |
| "epoch": 4.197337509788567, |
| "grad_norm": 1.2449496248919925, |
| "learning_rate": 1.518230252982248e-06, |
| "loss": 0.3926, |
| "step": 1340 |
| }, |
| { |
| "epoch": 4.200469851213782, |
| "grad_norm": 1.192327150884329, |
| "learning_rate": 1.5066537775127044e-06, |
| "loss": 0.3974, |
| "step": 1341 |
| }, |
| { |
| "epoch": 4.203602192638997, |
| "grad_norm": 1.235268675446175, |
| "learning_rate": 1.4951180095587426e-06, |
| "loss": 0.3963, |
| "step": 1342 |
| }, |
| { |
| "epoch": 4.206734534064213, |
| "grad_norm": 1.2044322613986833, |
| "learning_rate": 1.4836230044098164e-06, |
| "loss": 0.4214, |
| "step": 1343 |
| }, |
| { |
| "epoch": 4.209866875489428, |
| "grad_norm": 1.233078464789091, |
| "learning_rate": 1.4721688171600047e-06, |
| "loss": 0.3956, |
| "step": 1344 |
| }, |
| { |
| "epoch": 4.2129992169146435, |
| "grad_norm": 1.2257391290015276, |
| "learning_rate": 1.4607555027077524e-06, |
| "loss": 0.424, |
| "step": 1345 |
| }, |
| { |
| "epoch": 4.216131558339859, |
| "grad_norm": 1.239469422660483, |
| "learning_rate": 1.4493831157556094e-06, |
| "loss": 0.3933, |
| "step": 1346 |
| }, |
| { |
| "epoch": 4.219263899765075, |
| "grad_norm": 1.1537827041020894, |
| "learning_rate": 1.4380517108099623e-06, |
| "loss": 0.37, |
| "step": 1347 |
| }, |
| { |
| "epoch": 4.22239624119029, |
| "grad_norm": 1.2110428750506566, |
| "learning_rate": 1.426761342180777e-06, |
| "loss": 0.4159, |
| "step": 1348 |
| }, |
| { |
| "epoch": 4.225528582615505, |
| "grad_norm": 1.2034706739426686, |
| "learning_rate": 1.4155120639813392e-06, |
| "loss": 0.4007, |
| "step": 1349 |
| }, |
| { |
| "epoch": 4.22866092404072, |
| "grad_norm": 1.2096667555894145, |
| "learning_rate": 1.4043039301279904e-06, |
| "loss": 0.4168, |
| "step": 1350 |
| }, |
| { |
| "epoch": 4.231793265465936, |
| "grad_norm": 1.2136387061332223, |
| "learning_rate": 1.393136994339872e-06, |
| "loss": 0.3891, |
| "step": 1351 |
| }, |
| { |
| "epoch": 4.234925606891151, |
| "grad_norm": 1.2279471606708432, |
| "learning_rate": 1.3820113101386712e-06, |
| "loss": 0.398, |
| "step": 1352 |
| }, |
| { |
| "epoch": 4.238057948316366, |
| "grad_norm": 1.1712817094482313, |
| "learning_rate": 1.3709269308483619e-06, |
| "loss": 0.3974, |
| "step": 1353 |
| }, |
| { |
| "epoch": 4.241190289741581, |
| "grad_norm": 1.2244805625115447, |
| "learning_rate": 1.359883909594948e-06, |
| "loss": 0.4101, |
| "step": 1354 |
| }, |
| { |
| "epoch": 4.244322631166797, |
| "grad_norm": 1.2055456362307584, |
| "learning_rate": 1.348882299306209e-06, |
| "loss": 0.4189, |
| "step": 1355 |
| }, |
| { |
| "epoch": 4.2474549725920125, |
| "grad_norm": 1.2472551970779924, |
| "learning_rate": 1.3379221527114484e-06, |
| "loss": 0.3974, |
| "step": 1356 |
| }, |
| { |
| "epoch": 4.250587314017228, |
| "grad_norm": 1.1928935135771526, |
| "learning_rate": 1.327003522341237e-06, |
| "loss": 0.4061, |
| "step": 1357 |
| }, |
| { |
| "epoch": 4.253719655442444, |
| "grad_norm": 1.2297720660775902, |
| "learning_rate": 1.3161264605271662e-06, |
| "loss": 0.4037, |
| "step": 1358 |
| }, |
| { |
| "epoch": 4.256851996867659, |
| "grad_norm": 1.2100253932220972, |
| "learning_rate": 1.3052910194015922e-06, |
| "loss": 0.3932, |
| "step": 1359 |
| }, |
| { |
| "epoch": 4.259984338292874, |
| "grad_norm": 1.189046936366124, |
| "learning_rate": 1.2944972508973908e-06, |
| "loss": 0.3878, |
| "step": 1360 |
| }, |
| { |
| "epoch": 4.263116679718089, |
| "grad_norm": 1.1920673345461743, |
| "learning_rate": 1.2837452067477062e-06, |
| "loss": 0.3915, |
| "step": 1361 |
| }, |
| { |
| "epoch": 4.266249021143304, |
| "grad_norm": 1.2118917776405405, |
| "learning_rate": 1.273034938485702e-06, |
| "loss": 0.3939, |
| "step": 1362 |
| }, |
| { |
| "epoch": 4.26938136256852, |
| "grad_norm": 1.2126398847084519, |
| "learning_rate": 1.2623664974443195e-06, |
| "loss": 0.3988, |
| "step": 1363 |
| }, |
| { |
| "epoch": 4.272513703993735, |
| "grad_norm": 1.1928692981682214, |
| "learning_rate": 1.2517399347560165e-06, |
| "loss": 0.3966, |
| "step": 1364 |
| }, |
| { |
| "epoch": 4.27564604541895, |
| "grad_norm": 1.1977505153553263, |
| "learning_rate": 1.2411553013525457e-06, |
| "loss": 0.4051, |
| "step": 1365 |
| }, |
| { |
| "epoch": 4.2787783868441664, |
| "grad_norm": 1.222427847128266, |
| "learning_rate": 1.230612647964694e-06, |
| "loss": 0.4103, |
| "step": 1366 |
| }, |
| { |
| "epoch": 4.281910728269382, |
| "grad_norm": 1.200297992267844, |
| "learning_rate": 1.2201120251220378e-06, |
| "loss": 0.4016, |
| "step": 1367 |
| }, |
| { |
| "epoch": 4.285043069694597, |
| "grad_norm": 1.216728301148234, |
| "learning_rate": 1.209653483152714e-06, |
| "loss": 0.4014, |
| "step": 1368 |
| }, |
| { |
| "epoch": 4.288175411119812, |
| "grad_norm": 1.224540090293941, |
| "learning_rate": 1.1992370721831693e-06, |
| "loss": 0.3974, |
| "step": 1369 |
| }, |
| { |
| "epoch": 4.291307752545027, |
| "grad_norm": 1.1950041179139188, |
| "learning_rate": 1.1888628421379222e-06, |
| "loss": 0.4002, |
| "step": 1370 |
| }, |
| { |
| "epoch": 4.294440093970243, |
| "grad_norm": 1.190070377224994, |
| "learning_rate": 1.1785308427393182e-06, |
| "loss": 0.3964, |
| "step": 1371 |
| }, |
| { |
| "epoch": 4.297572435395458, |
| "grad_norm": 1.2008002010734806, |
| "learning_rate": 1.1682411235073032e-06, |
| "loss": 0.4076, |
| "step": 1372 |
| }, |
| { |
| "epoch": 4.300704776820673, |
| "grad_norm": 1.2132858157877429, |
| "learning_rate": 1.1579937337591774e-06, |
| "loss": 0.4015, |
| "step": 1373 |
| }, |
| { |
| "epoch": 4.303837118245889, |
| "grad_norm": 1.1826916321444987, |
| "learning_rate": 1.1477887226093642e-06, |
| "loss": 0.4171, |
| "step": 1374 |
| }, |
| { |
| "epoch": 4.306969459671104, |
| "grad_norm": 1.2089546478448403, |
| "learning_rate": 1.1376261389691633e-06, |
| "loss": 0.4146, |
| "step": 1375 |
| }, |
| { |
| "epoch": 4.3101018010963195, |
| "grad_norm": 1.206358523522593, |
| "learning_rate": 1.1275060315465346e-06, |
| "loss": 0.3872, |
| "step": 1376 |
| }, |
| { |
| "epoch": 4.313234142521535, |
| "grad_norm": 1.179621274726758, |
| "learning_rate": 1.1174284488458509e-06, |
| "loss": 0.4172, |
| "step": 1377 |
| }, |
| { |
| "epoch": 4.316366483946751, |
| "grad_norm": 1.268118607500806, |
| "learning_rate": 1.1073934391676666e-06, |
| "loss": 0.4213, |
| "step": 1378 |
| }, |
| { |
| "epoch": 4.319498825371966, |
| "grad_norm": 1.2431330503496503, |
| "learning_rate": 1.097401050608493e-06, |
| "loss": 0.3729, |
| "step": 1379 |
| }, |
| { |
| "epoch": 4.322631166797181, |
| "grad_norm": 1.2057856092166312, |
| "learning_rate": 1.0874513310605628e-06, |
| "loss": 0.388, |
| "step": 1380 |
| }, |
| { |
| "epoch": 4.325763508222396, |
| "grad_norm": 1.1984630219610473, |
| "learning_rate": 1.0775443282116017e-06, |
| "loss": 0.3919, |
| "step": 1381 |
| }, |
| { |
| "epoch": 4.328895849647612, |
| "grad_norm": 1.1721834023779714, |
| "learning_rate": 1.0676800895446016e-06, |
| "loss": 0.4033, |
| "step": 1382 |
| }, |
| { |
| "epoch": 4.332028191072827, |
| "grad_norm": 1.2483769893731758, |
| "learning_rate": 1.0578586623375863e-06, |
| "loss": 0.3826, |
| "step": 1383 |
| }, |
| { |
| "epoch": 4.335160532498042, |
| "grad_norm": 1.250417832052342, |
| "learning_rate": 1.048080093663395e-06, |
| "loss": 0.4275, |
| "step": 1384 |
| }, |
| { |
| "epoch": 4.338292873923257, |
| "grad_norm": 1.2565058613082094, |
| "learning_rate": 1.0383444303894453e-06, |
| "loss": 0.3947, |
| "step": 1385 |
| }, |
| { |
| "epoch": 4.341425215348473, |
| "grad_norm": 1.2252255070005802, |
| "learning_rate": 1.0286517191775215e-06, |
| "loss": 0.4061, |
| "step": 1386 |
| }, |
| { |
| "epoch": 4.3445575567736885, |
| "grad_norm": 1.235428710435152, |
| "learning_rate": 1.0190020064835404e-06, |
| "loss": 0.4089, |
| "step": 1387 |
| }, |
| { |
| "epoch": 4.347689898198904, |
| "grad_norm": 1.2195624521585995, |
| "learning_rate": 1.009395338557333e-06, |
| "loss": 0.3892, |
| "step": 1388 |
| }, |
| { |
| "epoch": 4.350822239624119, |
| "grad_norm": 1.2421525738212085, |
| "learning_rate": 9.998317614424246e-07, |
| "loss": 0.4151, |
| "step": 1389 |
| }, |
| { |
| "epoch": 4.353954581049335, |
| "grad_norm": 1.2269216736656088, |
| "learning_rate": 9.903113209758098e-07, |
| "loss": 0.4203, |
| "step": 1390 |
| }, |
| { |
| "epoch": 4.35708692247455, |
| "grad_norm": 1.2627277845413256, |
| "learning_rate": 9.80834062787731e-07, |
| "loss": 0.4115, |
| "step": 1391 |
| }, |
| { |
| "epoch": 4.360219263899765, |
| "grad_norm": 1.2175493590777189, |
| "learning_rate": 9.714000323014705e-07, |
| "loss": 0.3879, |
| "step": 1392 |
| }, |
| { |
| "epoch": 4.36335160532498, |
| "grad_norm": 1.1697012465299366, |
| "learning_rate": 9.62009274733121e-07, |
| "loss": 0.3977, |
| "step": 1393 |
| }, |
| { |
| "epoch": 4.366483946750196, |
| "grad_norm": 1.217611380413171, |
| "learning_rate": 9.526618350913752e-07, |
| "loss": 0.3774, |
| "step": 1394 |
| }, |
| { |
| "epoch": 4.369616288175411, |
| "grad_norm": 1.1843080929354863, |
| "learning_rate": 9.43357758177309e-07, |
| "loss": 0.39, |
| "step": 1395 |
| }, |
| { |
| "epoch": 4.372748629600626, |
| "grad_norm": 1.1973637151971002, |
| "learning_rate": 9.34097088584166e-07, |
| "loss": 0.4149, |
| "step": 1396 |
| }, |
| { |
| "epoch": 4.3758809710258415, |
| "grad_norm": 1.1880735619823894, |
| "learning_rate": 9.248798706971462e-07, |
| "loss": 0.3826, |
| "step": 1397 |
| }, |
| { |
| "epoch": 4.3790133124510575, |
| "grad_norm": 1.2411864789666913, |
| "learning_rate": 9.15706148693184e-07, |
| "loss": 0.4115, |
| "step": 1398 |
| }, |
| { |
| "epoch": 4.382145653876273, |
| "grad_norm": 1.2197612576036618, |
| "learning_rate": 9.065759665407514e-07, |
| "loss": 0.4113, |
| "step": 1399 |
| }, |
| { |
| "epoch": 4.385277995301488, |
| "grad_norm": 1.255965107907724, |
| "learning_rate": 8.974893679996388e-07, |
| "loss": 0.3768, |
| "step": 1400 |
| }, |
| { |
| "epoch": 4.388410336726703, |
| "grad_norm": 1.1777066026687946, |
| "learning_rate": 8.884463966207391e-07, |
| "loss": 0.3983, |
| "step": 1401 |
| }, |
| { |
| "epoch": 4.391542678151919, |
| "grad_norm": 1.1768660233012396, |
| "learning_rate": 8.794470957458523e-07, |
| "loss": 0.403, |
| "step": 1402 |
| }, |
| { |
| "epoch": 4.394675019577134, |
| "grad_norm": 1.2598498922043773, |
| "learning_rate": 8.704915085074728e-07, |
| "loss": 0.4133, |
| "step": 1403 |
| }, |
| { |
| "epoch": 4.397807361002349, |
| "grad_norm": 1.1732711385614463, |
| "learning_rate": 8.615796778285735e-07, |
| "loss": 0.3724, |
| "step": 1404 |
| }, |
| { |
| "epoch": 4.400939702427564, |
| "grad_norm": 1.167025579090733, |
| "learning_rate": 8.527116464224128e-07, |
| "loss": 0.4062, |
| "step": 1405 |
| }, |
| { |
| "epoch": 4.40407204385278, |
| "grad_norm": 1.2023162019645415, |
| "learning_rate": 8.438874567923261e-07, |
| "loss": 0.3917, |
| "step": 1406 |
| }, |
| { |
| "epoch": 4.407204385277995, |
| "grad_norm": 1.1810461949196573, |
| "learning_rate": 8.351071512315167e-07, |
| "loss": 0.4052, |
| "step": 1407 |
| }, |
| { |
| "epoch": 4.4103367267032105, |
| "grad_norm": 1.202253524025422, |
| "learning_rate": 8.263707718228641e-07, |
| "loss": 0.3979, |
| "step": 1408 |
| }, |
| { |
| "epoch": 4.413469068128426, |
| "grad_norm": 1.1576984628276288, |
| "learning_rate": 8.17678360438704e-07, |
| "loss": 0.401, |
| "step": 1409 |
| }, |
| { |
| "epoch": 4.416601409553642, |
| "grad_norm": 1.2291103726343342, |
| "learning_rate": 8.090299587406514e-07, |
| "loss": 0.4177, |
| "step": 1410 |
| }, |
| { |
| "epoch": 4.419733750978857, |
| "grad_norm": 1.2466018705955504, |
| "learning_rate": 8.004256081793782e-07, |
| "loss": 0.4247, |
| "step": 1411 |
| }, |
| { |
| "epoch": 4.422866092404072, |
| "grad_norm": 1.24245531665182, |
| "learning_rate": 7.918653499944318e-07, |
| "loss": 0.3814, |
| "step": 1412 |
| }, |
| { |
| "epoch": 4.425998433829287, |
| "grad_norm": 1.2207227340577, |
| "learning_rate": 7.833492252140284e-07, |
| "loss": 0.4179, |
| "step": 1413 |
| }, |
| { |
| "epoch": 4.429130775254503, |
| "grad_norm": 1.2374658186692722, |
| "learning_rate": 7.748772746548572e-07, |
| "loss": 0.438, |
| "step": 1414 |
| }, |
| { |
| "epoch": 4.432263116679718, |
| "grad_norm": 1.2445922732186314, |
| "learning_rate": 7.664495389218884e-07, |
| "loss": 0.4164, |
| "step": 1415 |
| }, |
| { |
| "epoch": 4.435395458104933, |
| "grad_norm": 1.2048318810786247, |
| "learning_rate": 7.580660584081755e-07, |
| "loss": 0.4114, |
| "step": 1416 |
| }, |
| { |
| "epoch": 4.438527799530148, |
| "grad_norm": 1.2092315947423604, |
| "learning_rate": 7.497268732946605e-07, |
| "loss": 0.3972, |
| "step": 1417 |
| }, |
| { |
| "epoch": 4.441660140955364, |
| "grad_norm": 1.1684399668507264, |
| "learning_rate": 7.414320235499839e-07, |
| "loss": 0.4007, |
| "step": 1418 |
| }, |
| { |
| "epoch": 4.44479248238058, |
| "grad_norm": 1.2018031522146513, |
| "learning_rate": 7.331815489302941e-07, |
| "loss": 0.3926, |
| "step": 1419 |
| }, |
| { |
| "epoch": 4.447924823805795, |
| "grad_norm": 1.2377323452479039, |
| "learning_rate": 7.249754889790539e-07, |
| "loss": 0.4024, |
| "step": 1420 |
| }, |
| { |
| "epoch": 4.45105716523101, |
| "grad_norm": 1.2372044961613387, |
| "learning_rate": 7.168138830268534e-07, |
| "loss": 0.4153, |
| "step": 1421 |
| }, |
| { |
| "epoch": 4.454189506656226, |
| "grad_norm": 1.2367482818132942, |
| "learning_rate": 7.086967701912195e-07, |
| "loss": 0.4145, |
| "step": 1422 |
| }, |
| { |
| "epoch": 4.457321848081441, |
| "grad_norm": 1.1720988594585238, |
| "learning_rate": 7.006241893764298e-07, |
| "loss": 0.3933, |
| "step": 1423 |
| }, |
| { |
| "epoch": 4.460454189506656, |
| "grad_norm": 1.1906519908299524, |
| "learning_rate": 6.925961792733271e-07, |
| "loss": 0.4213, |
| "step": 1424 |
| }, |
| { |
| "epoch": 4.463586530931871, |
| "grad_norm": 1.1861914037615011, |
| "learning_rate": 6.846127783591294e-07, |
| "loss": 0.374, |
| "step": 1425 |
| }, |
| { |
| "epoch": 4.466718872357087, |
| "grad_norm": 1.2151391090817059, |
| "learning_rate": 6.76674024897247e-07, |
| "loss": 0.3958, |
| "step": 1426 |
| }, |
| { |
| "epoch": 4.469851213782302, |
| "grad_norm": 1.1906210035368663, |
| "learning_rate": 6.687799569371079e-07, |
| "loss": 0.3997, |
| "step": 1427 |
| }, |
| { |
| "epoch": 4.4729835552075174, |
| "grad_norm": 1.2698252501752594, |
| "learning_rate": 6.609306123139614e-07, |
| "loss": 0.4152, |
| "step": 1428 |
| }, |
| { |
| "epoch": 4.476115896632733, |
| "grad_norm": 1.239086528624892, |
| "learning_rate": 6.531260286487073e-07, |
| "loss": 0.4244, |
| "step": 1429 |
| }, |
| { |
| "epoch": 4.479248238057949, |
| "grad_norm": 1.239389095459042, |
| "learning_rate": 6.453662433477137e-07, |
| "loss": 0.3967, |
| "step": 1430 |
| }, |
| { |
| "epoch": 4.482380579483164, |
| "grad_norm": 1.2078463523567677, |
| "learning_rate": 6.37651293602628e-07, |
| "loss": 0.3998, |
| "step": 1431 |
| }, |
| { |
| "epoch": 4.485512920908379, |
| "grad_norm": 1.2405207380196077, |
| "learning_rate": 6.299812163902152e-07, |
| "loss": 0.4079, |
| "step": 1432 |
| }, |
| { |
| "epoch": 4.488645262333594, |
| "grad_norm": 1.1965967124161982, |
| "learning_rate": 6.22356048472168e-07, |
| "loss": 0.3947, |
| "step": 1433 |
| }, |
| { |
| "epoch": 4.49177760375881, |
| "grad_norm": 1.2365962276051872, |
| "learning_rate": 6.147758263949322e-07, |
| "loss": 0.4145, |
| "step": 1434 |
| }, |
| { |
| "epoch": 4.494909945184025, |
| "grad_norm": 1.2320724240611538, |
| "learning_rate": 6.072405864895403e-07, |
| "loss": 0.3989, |
| "step": 1435 |
| }, |
| { |
| "epoch": 4.49804228660924, |
| "grad_norm": 1.2278614714663911, |
| "learning_rate": 5.997503648714254e-07, |
| "loss": 0.4087, |
| "step": 1436 |
| }, |
| { |
| "epoch": 4.501174628034455, |
| "grad_norm": 1.2459676825726136, |
| "learning_rate": 5.923051974402594e-07, |
| "loss": 0.4076, |
| "step": 1437 |
| }, |
| { |
| "epoch": 4.504306969459671, |
| "grad_norm": 1.2458668198276721, |
| "learning_rate": 5.849051198797662e-07, |
| "loss": 0.4248, |
| "step": 1438 |
| }, |
| { |
| "epoch": 4.5074393108848865, |
| "grad_norm": 1.209767689172734, |
| "learning_rate": 5.775501676575702e-07, |
| "loss": 0.4014, |
| "step": 1439 |
| }, |
| { |
| "epoch": 4.510571652310102, |
| "grad_norm": 1.2285591384985166, |
| "learning_rate": 5.702403760250086e-07, |
| "loss": 0.4147, |
| "step": 1440 |
| }, |
| { |
| "epoch": 4.513703993735318, |
| "grad_norm": 1.1718999655815265, |
| "learning_rate": 5.629757800169732e-07, |
| "loss": 0.4044, |
| "step": 1441 |
| }, |
| { |
| "epoch": 4.516836335160533, |
| "grad_norm": 1.217216752100384, |
| "learning_rate": 5.557564144517369e-07, |
| "loss": 0.4091, |
| "step": 1442 |
| }, |
| { |
| "epoch": 4.519968676585748, |
| "grad_norm": 1.1838737229775769, |
| "learning_rate": 5.485823139307911e-07, |
| "loss": 0.4034, |
| "step": 1443 |
| }, |
| { |
| "epoch": 4.523101018010963, |
| "grad_norm": 1.1673948755376424, |
| "learning_rate": 5.414535128386766e-07, |
| "loss": 0.3661, |
| "step": 1444 |
| }, |
| { |
| "epoch": 4.526233359436178, |
| "grad_norm": 1.1757487458448712, |
| "learning_rate": 5.343700453428169e-07, |
| "loss": 0.3971, |
| "step": 1445 |
| }, |
| { |
| "epoch": 4.529365700861394, |
| "grad_norm": 1.2191061942291306, |
| "learning_rate": 5.273319453933634e-07, |
| "loss": 0.3885, |
| "step": 1446 |
| }, |
| { |
| "epoch": 4.532498042286609, |
| "grad_norm": 1.1814168865664434, |
| "learning_rate": 5.203392467230229e-07, |
| "loss": 0.3888, |
| "step": 1447 |
| }, |
| { |
| "epoch": 4.535630383711824, |
| "grad_norm": 1.180338891284529, |
| "learning_rate": 5.133919828468992e-07, |
| "loss": 0.4031, |
| "step": 1448 |
| }, |
| { |
| "epoch": 4.53876272513704, |
| "grad_norm": 1.2002998136423375, |
| "learning_rate": 5.064901870623362e-07, |
| "loss": 0.4124, |
| "step": 1449 |
| }, |
| { |
| "epoch": 4.5418950665622555, |
| "grad_norm": 1.225998648068942, |
| "learning_rate": 4.996338924487509e-07, |
| "loss": 0.4079, |
| "step": 1450 |
| }, |
| { |
| "epoch": 4.545027407987471, |
| "grad_norm": 1.232921807498376, |
| "learning_rate": 4.928231318674836e-07, |
| "loss": 0.4288, |
| "step": 1451 |
| }, |
| { |
| "epoch": 4.548159749412686, |
| "grad_norm": 1.2108344873262966, |
| "learning_rate": 4.860579379616303e-07, |
| "loss": 0.3752, |
| "step": 1452 |
| }, |
| { |
| "epoch": 4.551292090837901, |
| "grad_norm": 1.1982395456743722, |
| "learning_rate": 4.793383431558962e-07, |
| "loss": 0.3956, |
| "step": 1453 |
| }, |
| { |
| "epoch": 4.554424432263117, |
| "grad_norm": 1.2245255923948302, |
| "learning_rate": 4.7266437965643455e-07, |
| "loss": 0.39, |
| "step": 1454 |
| }, |
| { |
| "epoch": 4.557556773688332, |
| "grad_norm": 1.2237018326288942, |
| "learning_rate": 4.660360794506946e-07, |
| "loss": 0.4064, |
| "step": 1455 |
| }, |
| { |
| "epoch": 4.560689115113547, |
| "grad_norm": 1.203953237163164, |
| "learning_rate": 4.594534743072654e-07, |
| "loss": 0.4169, |
| "step": 1456 |
| }, |
| { |
| "epoch": 4.563821456538763, |
| "grad_norm": 1.2087498003259758, |
| "learning_rate": 4.5291659577572753e-07, |
| "loss": 0.3832, |
| "step": 1457 |
| }, |
| { |
| "epoch": 4.566953797963978, |
| "grad_norm": 1.1942569748553682, |
| "learning_rate": 4.4642547518649824e-07, |
| "loss": 0.3776, |
| "step": 1458 |
| }, |
| { |
| "epoch": 4.570086139389193, |
| "grad_norm": 1.1879804731198786, |
| "learning_rate": 4.399801436506812e-07, |
| "loss": 0.3873, |
| "step": 1459 |
| }, |
| { |
| "epoch": 4.5732184808144085, |
| "grad_norm": 1.2110866417839938, |
| "learning_rate": 4.335806320599234e-07, |
| "loss": 0.399, |
| "step": 1460 |
| }, |
| { |
| "epoch": 4.576350822239624, |
| "grad_norm": 1.2296410077584767, |
| "learning_rate": 4.272269710862587e-07, |
| "loss": 0.3992, |
| "step": 1461 |
| }, |
| { |
| "epoch": 4.57948316366484, |
| "grad_norm": 1.2328738318463752, |
| "learning_rate": 4.209191911819688e-07, |
| "loss": 0.4196, |
| "step": 1462 |
| }, |
| { |
| "epoch": 4.582615505090055, |
| "grad_norm": 1.1895802067659933, |
| "learning_rate": 4.1465732257943037e-07, |
| "loss": 0.4496, |
| "step": 1463 |
| }, |
| { |
| "epoch": 4.58574784651527, |
| "grad_norm": 1.2283635780345825, |
| "learning_rate": 4.084413952909738e-07, |
| "loss": 0.4111, |
| "step": 1464 |
| }, |
| { |
| "epoch": 4.588880187940486, |
| "grad_norm": 1.2164839835142356, |
| "learning_rate": 4.022714391087379e-07, |
| "loss": 0.381, |
| "step": 1465 |
| }, |
| { |
| "epoch": 4.592012529365701, |
| "grad_norm": 1.207591372814288, |
| "learning_rate": 3.9614748360452984e-07, |
| "loss": 0.4174, |
| "step": 1466 |
| }, |
| { |
| "epoch": 4.595144870790916, |
| "grad_norm": 1.2361007397271178, |
| "learning_rate": 3.9006955812967983e-07, |
| "loss": 0.4093, |
| "step": 1467 |
| }, |
| { |
| "epoch": 4.598277212216131, |
| "grad_norm": 1.2148169973114509, |
| "learning_rate": 3.840376918149025e-07, |
| "loss": 0.4096, |
| "step": 1468 |
| }, |
| { |
| "epoch": 4.601409553641347, |
| "grad_norm": 1.262330409419549, |
| "learning_rate": 3.780519135701566e-07, |
| "loss": 0.4104, |
| "step": 1469 |
| }, |
| { |
| "epoch": 4.604541895066562, |
| "grad_norm": 1.2330283137163904, |
| "learning_rate": 3.7211225208450777e-07, |
| "loss": 0.4192, |
| "step": 1470 |
| }, |
| { |
| "epoch": 4.6076742364917775, |
| "grad_norm": 1.2671239020786587, |
| "learning_rate": 3.6621873582599055e-07, |
| "loss": 0.4054, |
| "step": 1471 |
| }, |
| { |
| "epoch": 4.610806577916993, |
| "grad_norm": 1.2333107036927184, |
| "learning_rate": 3.603713930414676e-07, |
| "loss": 0.4063, |
| "step": 1472 |
| }, |
| { |
| "epoch": 4.613938919342209, |
| "grad_norm": 1.202856883096194, |
| "learning_rate": 3.5457025175650084e-07, |
| "loss": 0.4148, |
| "step": 1473 |
| }, |
| { |
| "epoch": 4.617071260767424, |
| "grad_norm": 1.1747823598797158, |
| "learning_rate": 3.488153397752159e-07, |
| "loss": 0.3891, |
| "step": 1474 |
| }, |
| { |
| "epoch": 4.620203602192639, |
| "grad_norm": 1.1992072519105093, |
| "learning_rate": 3.431066846801634e-07, |
| "loss": 0.4267, |
| "step": 1475 |
| }, |
| { |
| "epoch": 4.623335943617854, |
| "grad_norm": 1.273775818849566, |
| "learning_rate": 3.374443138321937e-07, |
| "loss": 0.3985, |
| "step": 1476 |
| }, |
| { |
| "epoch": 4.62646828504307, |
| "grad_norm": 1.2299835011094196, |
| "learning_rate": 3.3182825437031994e-07, |
| "loss": 0.4089, |
| "step": 1477 |
| }, |
| { |
| "epoch": 4.629600626468285, |
| "grad_norm": 1.2034138009185376, |
| "learning_rate": 3.262585332115964e-07, |
| "loss": 0.4043, |
| "step": 1478 |
| }, |
| { |
| "epoch": 4.6327329678935, |
| "grad_norm": 1.209424726509648, |
| "learning_rate": 3.207351770509759e-07, |
| "loss": 0.4147, |
| "step": 1479 |
| }, |
| { |
| "epoch": 4.635865309318715, |
| "grad_norm": 1.1736004823028015, |
| "learning_rate": 3.152582123611958e-07, |
| "loss": 0.3687, |
| "step": 1480 |
| }, |
| { |
| "epoch": 4.6389976507439314, |
| "grad_norm": 1.2024178154556568, |
| "learning_rate": 3.098276653926413e-07, |
| "loss": 0.4049, |
| "step": 1481 |
| }, |
| { |
| "epoch": 4.642129992169147, |
| "grad_norm": 1.235968122606252, |
| "learning_rate": 3.0444356217322893e-07, |
| "loss": 0.4122, |
| "step": 1482 |
| }, |
| { |
| "epoch": 4.645262333594362, |
| "grad_norm": 1.275106262365568, |
| "learning_rate": 2.9910592850826983e-07, |
| "loss": 0.4086, |
| "step": 1483 |
| }, |
| { |
| "epoch": 4.648394675019577, |
| "grad_norm": 1.2510666555833756, |
| "learning_rate": 2.9381478998035563e-07, |
| "loss": 0.4055, |
| "step": 1484 |
| }, |
| { |
| "epoch": 4.651527016444793, |
| "grad_norm": 1.2274467661857713, |
| "learning_rate": 2.8857017194923174e-07, |
| "loss": 0.4034, |
| "step": 1485 |
| }, |
| { |
| "epoch": 4.654659357870008, |
| "grad_norm": 1.1853349236972364, |
| "learning_rate": 2.833720995516798e-07, |
| "loss": 0.4195, |
| "step": 1486 |
| }, |
| { |
| "epoch": 4.657791699295223, |
| "grad_norm": 1.1854238027175328, |
| "learning_rate": 2.7822059770138966e-07, |
| "loss": 0.4074, |
| "step": 1487 |
| }, |
| { |
| "epoch": 4.660924040720438, |
| "grad_norm": 1.2436312130461868, |
| "learning_rate": 2.731156910888477e-07, |
| "loss": 0.4033, |
| "step": 1488 |
| }, |
| { |
| "epoch": 4.664056382145654, |
| "grad_norm": 1.2428796817271779, |
| "learning_rate": 2.6805740418121453e-07, |
| "loss": 0.4191, |
| "step": 1489 |
| }, |
| { |
| "epoch": 4.667188723570869, |
| "grad_norm": 1.2427783356951405, |
| "learning_rate": 2.6304576122221035e-07, |
| "loss": 0.3755, |
| "step": 1490 |
| }, |
| { |
| "epoch": 4.6703210649960845, |
| "grad_norm": 1.1881132036518127, |
| "learning_rate": 2.580807862319912e-07, |
| "loss": 0.3891, |
| "step": 1491 |
| }, |
| { |
| "epoch": 4.6734534064213, |
| "grad_norm": 1.2442335114182685, |
| "learning_rate": 2.5316250300704726e-07, |
| "loss": 0.4154, |
| "step": 1492 |
| }, |
| { |
| "epoch": 4.676585747846516, |
| "grad_norm": 1.2053182135162017, |
| "learning_rate": 2.482909351200735e-07, |
| "loss": 0.3979, |
| "step": 1493 |
| }, |
| { |
| "epoch": 4.679718089271731, |
| "grad_norm": 1.1770238326901394, |
| "learning_rate": 2.434661059198695e-07, |
| "loss": 0.3948, |
| "step": 1494 |
| }, |
| { |
| "epoch": 4.682850430696946, |
| "grad_norm": 1.218619191692209, |
| "learning_rate": 2.386880385312218e-07, |
| "loss": 0.371, |
| "step": 1495 |
| }, |
| { |
| "epoch": 4.685982772122161, |
| "grad_norm": 1.1637502072533377, |
| "learning_rate": 2.339567558547906e-07, |
| "loss": 0.3917, |
| "step": 1496 |
| }, |
| { |
| "epoch": 4.689115113547377, |
| "grad_norm": 1.1735101557117311, |
| "learning_rate": 2.292722805670078e-07, |
| "loss": 0.4077, |
| "step": 1497 |
| }, |
| { |
| "epoch": 4.692247454972592, |
| "grad_norm": 1.2417399018426394, |
| "learning_rate": 2.2463463511995908e-07, |
| "loss": 0.4048, |
| "step": 1498 |
| }, |
| { |
| "epoch": 4.695379796397807, |
| "grad_norm": 1.1728907713133705, |
| "learning_rate": 2.200438417412809e-07, |
| "loss": 0.3891, |
| "step": 1499 |
| }, |
| { |
| "epoch": 4.698512137823022, |
| "grad_norm": 1.2016608119778363, |
| "learning_rate": 2.1549992243405814e-07, |
| "loss": 0.3882, |
| "step": 1500 |
| }, |
| { |
| "epoch": 4.701644479248238, |
| "grad_norm": 1.206612395808125, |
| "learning_rate": 2.1100289897670877e-07, |
| "loss": 0.4196, |
| "step": 1501 |
| }, |
| { |
| "epoch": 4.7047768206734535, |
| "grad_norm": 1.2493616742742244, |
| "learning_rate": 2.0655279292288832e-07, |
| "loss": 0.3967, |
| "step": 1502 |
| }, |
| { |
| "epoch": 4.707909162098669, |
| "grad_norm": 1.2226247159036543, |
| "learning_rate": 2.021496256013833e-07, |
| "loss": 0.3989, |
| "step": 1503 |
| }, |
| { |
| "epoch": 4.711041503523884, |
| "grad_norm": 1.1794232427283857, |
| "learning_rate": 1.9779341811600795e-07, |
| "loss": 0.3988, |
| "step": 1504 |
| }, |
| { |
| "epoch": 4.7141738449491, |
| "grad_norm": 1.2359058049155338, |
| "learning_rate": 1.9348419134550322e-07, |
| "loss": 0.4213, |
| "step": 1505 |
| }, |
| { |
| "epoch": 4.717306186374315, |
| "grad_norm": 1.2239539416305099, |
| "learning_rate": 1.8922196594344023e-07, |
| "loss": 0.3922, |
| "step": 1506 |
| }, |
| { |
| "epoch": 4.72043852779953, |
| "grad_norm": 1.2127862487844647, |
| "learning_rate": 1.8500676233811466e-07, |
| "loss": 0.3771, |
| "step": 1507 |
| }, |
| { |
| "epoch": 4.723570869224745, |
| "grad_norm": 1.198843093205888, |
| "learning_rate": 1.8083860073245806e-07, |
| "loss": 0.4181, |
| "step": 1508 |
| }, |
| { |
| "epoch": 4.726703210649961, |
| "grad_norm": 1.2252062247390472, |
| "learning_rate": 1.7671750110392903e-07, |
| "loss": 0.403, |
| "step": 1509 |
| }, |
| { |
| "epoch": 4.729835552075176, |
| "grad_norm": 1.23819671302605, |
| "learning_rate": 1.7264348320442992e-07, |
| "loss": 0.4092, |
| "step": 1510 |
| }, |
| { |
| "epoch": 4.732967893500391, |
| "grad_norm": 1.2113129084206005, |
| "learning_rate": 1.6861656656020464e-07, |
| "loss": 0.3774, |
| "step": 1511 |
| }, |
| { |
| "epoch": 4.7361002349256065, |
| "grad_norm": 1.1845710084452867, |
| "learning_rate": 1.6463677047174553e-07, |
| "loss": 0.3841, |
| "step": 1512 |
| }, |
| { |
| "epoch": 4.7392325763508225, |
| "grad_norm": 1.1746238547350345, |
| "learning_rate": 1.6070411401370335e-07, |
| "loss": 0.3939, |
| "step": 1513 |
| }, |
| { |
| "epoch": 4.742364917776038, |
| "grad_norm": 1.212119626132927, |
| "learning_rate": 1.568186160347951e-07, |
| "loss": 0.3838, |
| "step": 1514 |
| }, |
| { |
| "epoch": 4.745497259201253, |
| "grad_norm": 1.2178905084917646, |
| "learning_rate": 1.5298029515771195e-07, |
| "loss": 0.4132, |
| "step": 1515 |
| }, |
| { |
| "epoch": 4.748629600626469, |
| "grad_norm": 1.2150279012900946, |
| "learning_rate": 1.4918916977903264e-07, |
| "loss": 0.3986, |
| "step": 1516 |
| }, |
| { |
| "epoch": 4.751761942051684, |
| "grad_norm": 1.2367095696943584, |
| "learning_rate": 1.4544525806913234e-07, |
| "loss": 0.3979, |
| "step": 1517 |
| }, |
| { |
| "epoch": 4.754894283476899, |
| "grad_norm": 1.2377643398696656, |
| "learning_rate": 1.4174857797209951e-07, |
| "loss": 0.4018, |
| "step": 1518 |
| }, |
| { |
| "epoch": 4.758026624902114, |
| "grad_norm": 1.2167876153278898, |
| "learning_rate": 1.3809914720564478e-07, |
| "loss": 0.3906, |
| "step": 1519 |
| }, |
| { |
| "epoch": 4.761158966327329, |
| "grad_norm": 1.2446471369617926, |
| "learning_rate": 1.344969832610199e-07, |
| "loss": 0.3756, |
| "step": 1520 |
| }, |
| { |
| "epoch": 4.764291307752545, |
| "grad_norm": 1.1893090822752832, |
| "learning_rate": 1.3094210340293456e-07, |
| "loss": 0.4165, |
| "step": 1521 |
| }, |
| { |
| "epoch": 4.76742364917776, |
| "grad_norm": 1.2320387940829778, |
| "learning_rate": 1.274345246694708e-07, |
| "loss": 0.3913, |
| "step": 1522 |
| }, |
| { |
| "epoch": 4.7705559906029755, |
| "grad_norm": 1.216546020584182, |
| "learning_rate": 1.2397426387200096e-07, |
| "loss": 0.3959, |
| "step": 1523 |
| }, |
| { |
| "epoch": 4.7736883320281915, |
| "grad_norm": 1.2176186473887705, |
| "learning_rate": 1.205613375951098e-07, |
| "loss": 0.4247, |
| "step": 1524 |
| }, |
| { |
| "epoch": 4.776820673453407, |
| "grad_norm": 1.2283180997381093, |
| "learning_rate": 1.1719576219651585e-07, |
| "loss": 0.418, |
| "step": 1525 |
| }, |
| { |
| "epoch": 4.779953014878622, |
| "grad_norm": 1.214550392378903, |
| "learning_rate": 1.1387755380698584e-07, |
| "loss": 0.3819, |
| "step": 1526 |
| }, |
| { |
| "epoch": 4.783085356303837, |
| "grad_norm": 1.2324975421906008, |
| "learning_rate": 1.1060672833026697e-07, |
| "loss": 0.4278, |
| "step": 1527 |
| }, |
| { |
| "epoch": 4.786217697729052, |
| "grad_norm": 1.238934876307768, |
| "learning_rate": 1.0738330144300479e-07, |
| "loss": 0.425, |
| "step": 1528 |
| }, |
| { |
| "epoch": 4.789350039154268, |
| "grad_norm": 1.181886288925798, |
| "learning_rate": 1.0420728859466988e-07, |
| "loss": 0.4209, |
| "step": 1529 |
| }, |
| { |
| "epoch": 4.792482380579483, |
| "grad_norm": 1.1925253947318342, |
| "learning_rate": 1.0107870500748351e-07, |
| "loss": 0.3773, |
| "step": 1530 |
| }, |
| { |
| "epoch": 4.795614722004698, |
| "grad_norm": 1.2509864343113049, |
| "learning_rate": 9.799756567634433e-08, |
| "loss": 0.3853, |
| "step": 1531 |
| }, |
| { |
| "epoch": 4.798747063429914, |
| "grad_norm": 1.184682423863237, |
| "learning_rate": 9.496388536875623e-08, |
| "loss": 0.417, |
| "step": 1532 |
| }, |
| { |
| "epoch": 4.801879404855129, |
| "grad_norm": 1.2246944998945095, |
| "learning_rate": 9.197767862475837e-08, |
| "loss": 0.3874, |
| "step": 1533 |
| }, |
| { |
| "epoch": 4.8050117462803446, |
| "grad_norm": 1.214520286942335, |
| "learning_rate": 8.903895975685528e-08, |
| "loss": 0.4037, |
| "step": 1534 |
| }, |
| { |
| "epoch": 4.80814408770556, |
| "grad_norm": 1.2224353069074514, |
| "learning_rate": 8.614774284994797e-08, |
| "loss": 0.4099, |
| "step": 1535 |
| }, |
| { |
| "epoch": 4.811276429130775, |
| "grad_norm": 1.2068932961719712, |
| "learning_rate": 8.33040417612685e-08, |
| "loss": 0.3891, |
| "step": 1536 |
| }, |
| { |
| "epoch": 4.814408770555991, |
| "grad_norm": 1.1952441796807236, |
| "learning_rate": 8.05078701203088e-08, |
| "loss": 0.3899, |
| "step": 1537 |
| }, |
| { |
| "epoch": 4.817541111981206, |
| "grad_norm": 1.226741893844113, |
| "learning_rate": 7.7759241328762e-08, |
| "loss": 0.4111, |
| "step": 1538 |
| }, |
| { |
| "epoch": 4.820673453406421, |
| "grad_norm": 1.1808940728419945, |
| "learning_rate": 7.505816856045012e-08, |
| "loss": 0.3943, |
| "step": 1539 |
| }, |
| { |
| "epoch": 4.823805794831637, |
| "grad_norm": 1.203115331813816, |
| "learning_rate": 7.24046647612675e-08, |
| "loss": 0.4127, |
| "step": 1540 |
| }, |
| { |
| "epoch": 4.826938136256852, |
| "grad_norm": 1.1865631093267626, |
| "learning_rate": 6.979874264911756e-08, |
| "loss": 0.4058, |
| "step": 1541 |
| }, |
| { |
| "epoch": 4.830070477682067, |
| "grad_norm": 1.185030215141837, |
| "learning_rate": 6.724041471384835e-08, |
| "loss": 0.4124, |
| "step": 1542 |
| }, |
| { |
| "epoch": 4.8332028191072824, |
| "grad_norm": 1.204462154369856, |
| "learning_rate": 6.472969321719702e-08, |
| "loss": 0.4326, |
| "step": 1543 |
| }, |
| { |
| "epoch": 4.836335160532498, |
| "grad_norm": 1.2203256822089201, |
| "learning_rate": 6.226659019272997e-08, |
| "loss": 0.3917, |
| "step": 1544 |
| }, |
| { |
| "epoch": 4.839467501957714, |
| "grad_norm": 1.2180754919299253, |
| "learning_rate": 5.985111744578165e-08, |
| "loss": 0.4013, |
| "step": 1545 |
| }, |
| { |
| "epoch": 4.842599843382929, |
| "grad_norm": 1.18445065771052, |
| "learning_rate": 5.748328655340141e-08, |
| "loss": 0.3912, |
| "step": 1546 |
| }, |
| { |
| "epoch": 4.845732184808144, |
| "grad_norm": 1.1965989755685764, |
| "learning_rate": 5.516310886429899e-08, |
| "loss": 0.4041, |
| "step": 1547 |
| }, |
| { |
| "epoch": 4.84886452623336, |
| "grad_norm": 1.193838779871222, |
| "learning_rate": 5.2890595498787944e-08, |
| "loss": 0.4006, |
| "step": 1548 |
| }, |
| { |
| "epoch": 4.851996867658575, |
| "grad_norm": 1.1686189895653558, |
| "learning_rate": 5.0665757348732355e-08, |
| "loss": 0.4082, |
| "step": 1549 |
| }, |
| { |
| "epoch": 4.85512920908379, |
| "grad_norm": 1.2058648478611114, |
| "learning_rate": 4.848860507749353e-08, |
| "loss": 0.3881, |
| "step": 1550 |
| }, |
| { |
| "epoch": 4.858261550509005, |
| "grad_norm": 1.1637318481244103, |
| "learning_rate": 4.635914911988448e-08, |
| "loss": 0.3942, |
| "step": 1551 |
| }, |
| { |
| "epoch": 4.86139389193422, |
| "grad_norm": 1.2165645886224812, |
| "learning_rate": 4.427739968211331e-08, |
| "loss": 0.4035, |
| "step": 1552 |
| }, |
| { |
| "epoch": 4.864526233359436, |
| "grad_norm": 1.2040433534371484, |
| "learning_rate": 4.2243366741735457e-08, |
| "loss": 0.3868, |
| "step": 1553 |
| }, |
| { |
| "epoch": 4.8676585747846515, |
| "grad_norm": 1.2369435812923857, |
| "learning_rate": 4.025706004760932e-08, |
| "loss": 0.3988, |
| "step": 1554 |
| }, |
| { |
| "epoch": 4.870790916209867, |
| "grad_norm": 1.2232422405149723, |
| "learning_rate": 3.831848911984959e-08, |
| "loss": 0.3968, |
| "step": 1555 |
| }, |
| { |
| "epoch": 4.873923257635083, |
| "grad_norm": 1.1711459811080631, |
| "learning_rate": 3.642766324977509e-08, |
| "loss": 0.3923, |
| "step": 1556 |
| }, |
| { |
| "epoch": 4.877055599060298, |
| "grad_norm": 1.2174185734188516, |
| "learning_rate": 3.4584591499872146e-08, |
| "loss": 0.4029, |
| "step": 1557 |
| }, |
| { |
| "epoch": 4.880187940485513, |
| "grad_norm": 1.1614173175371763, |
| "learning_rate": 3.278928270374459e-08, |
| "loss": 0.3949, |
| "step": 1558 |
| }, |
| { |
| "epoch": 4.883320281910728, |
| "grad_norm": 1.2280723648931007, |
| "learning_rate": 3.104174546607941e-08, |
| "loss": 0.4199, |
| "step": 1559 |
| }, |
| { |
| "epoch": 4.886452623335944, |
| "grad_norm": 1.209148320604221, |
| "learning_rate": 2.9341988162595593e-08, |
| "loss": 0.4192, |
| "step": 1560 |
| }, |
| { |
| "epoch": 4.889584964761159, |
| "grad_norm": 1.2065698817517443, |
| "learning_rate": 2.7690018940011996e-08, |
| "loss": 0.3655, |
| "step": 1561 |
| }, |
| { |
| "epoch": 4.892717306186374, |
| "grad_norm": 1.1832085868374411, |
| "learning_rate": 2.6085845716004034e-08, |
| "loss": 0.4102, |
| "step": 1562 |
| }, |
| { |
| "epoch": 4.895849647611589, |
| "grad_norm": 1.223809874203652, |
| "learning_rate": 2.452947617916701e-08, |
| "loss": 0.3826, |
| "step": 1563 |
| }, |
| { |
| "epoch": 4.898981989036805, |
| "grad_norm": 1.1564398514888772, |
| "learning_rate": 2.3020917788981744e-08, |
| "loss": 0.399, |
| "step": 1564 |
| }, |
| { |
| "epoch": 4.9021143304620205, |
| "grad_norm": 1.2322802342710437, |
| "learning_rate": 2.156017777577346e-08, |
| "loss": 0.4385, |
| "step": 1565 |
| }, |
| { |
| "epoch": 4.905246671887236, |
| "grad_norm": 1.2327395751465036, |
| "learning_rate": 2.0147263140680717e-08, |
| "loss": 0.3948, |
| "step": 1566 |
| }, |
| { |
| "epoch": 4.908379013312451, |
| "grad_norm": 1.1755915458959958, |
| "learning_rate": 1.8782180655622096e-08, |
| "loss": 0.3909, |
| "step": 1567 |
| }, |
| { |
| "epoch": 4.911511354737667, |
| "grad_norm": 1.2040808443639759, |
| "learning_rate": 1.7464936863261785e-08, |
| "loss": 0.4109, |
| "step": 1568 |
| }, |
| { |
| "epoch": 4.914643696162882, |
| "grad_norm": 1.2313492049462424, |
| "learning_rate": 1.61955380769796e-08, |
| "loss": 0.3865, |
| "step": 1569 |
| }, |
| { |
| "epoch": 4.917776037588097, |
| "grad_norm": 1.2358379850928518, |
| "learning_rate": 1.497399038084102e-08, |
| "loss": 0.4126, |
| "step": 1570 |
| }, |
| { |
| "epoch": 4.920908379013312, |
| "grad_norm": 1.2159888247443027, |
| "learning_rate": 1.3800299629568303e-08, |
| "loss": 0.3979, |
| "step": 1571 |
| }, |
| { |
| "epoch": 4.924040720438528, |
| "grad_norm": 1.2314356631041403, |
| "learning_rate": 1.2674471448508308e-08, |
| "loss": 0.3954, |
| "step": 1572 |
| }, |
| { |
| "epoch": 4.927173061863743, |
| "grad_norm": 1.1763805345232639, |
| "learning_rate": 1.1596511233610275e-08, |
| "loss": 0.4174, |
| "step": 1573 |
| }, |
| { |
| "epoch": 4.930305403288958, |
| "grad_norm": 1.21950132315641, |
| "learning_rate": 1.0566424151401412e-08, |
| "loss": 0.3712, |
| "step": 1574 |
| }, |
| { |
| "epoch": 4.9334377447141735, |
| "grad_norm": 1.1857749532551431, |
| "learning_rate": 9.584215138953579e-09, |
| "loss": 0.4018, |
| "step": 1575 |
| }, |
| { |
| "epoch": 4.9365700861393895, |
| "grad_norm": 1.1885549206002513, |
| "learning_rate": 8.649888903869973e-09, |
| "loss": 0.4014, |
| "step": 1576 |
| }, |
| { |
| "epoch": 4.939702427564605, |
| "grad_norm": 1.2601185286531782, |
| "learning_rate": 7.763449924256262e-09, |
| "loss": 0.3956, |
| "step": 1577 |
| }, |
| { |
| "epoch": 4.94283476898982, |
| "grad_norm": 1.1994949171986244, |
| "learning_rate": 6.924902448699478e-09, |
| "loss": 0.4229, |
| "step": 1578 |
| }, |
| { |
| "epoch": 4.945967110415035, |
| "grad_norm": 1.231920206414556, |
| "learning_rate": 6.134250496249161e-09, |
| "loss": 0.4123, |
| "step": 1579 |
| }, |
| { |
| "epoch": 4.949099451840251, |
| "grad_norm": 1.231501336996204, |
| "learning_rate": 5.391497856399585e-09, |
| "loss": 0.4019, |
| "step": 1580 |
| }, |
| { |
| "epoch": 4.952231793265466, |
| "grad_norm": 1.2025254247538497, |
| "learning_rate": 4.696648089068667e-09, |
| "loss": 0.4047, |
| "step": 1581 |
| }, |
| { |
| "epoch": 4.955364134690681, |
| "grad_norm": 1.2462699073996386, |
| "learning_rate": 4.0497045245813105e-09, |
| "loss": 0.3905, |
| "step": 1582 |
| }, |
| { |
| "epoch": 4.958496476115896, |
| "grad_norm": 1.2150582015348728, |
| "learning_rate": 3.450670263654976e-09, |
| "loss": 0.4034, |
| "step": 1583 |
| }, |
| { |
| "epoch": 4.961628817541112, |
| "grad_norm": 1.1657154816919872, |
| "learning_rate": 2.899548177384137e-09, |
| "loss": 0.3983, |
| "step": 1584 |
| }, |
| { |
| "epoch": 4.964761158966327, |
| "grad_norm": 1.201657869580201, |
| "learning_rate": 2.396340907225847e-09, |
| "loss": 0.3985, |
| "step": 1585 |
| }, |
| { |
| "epoch": 4.9678935003915425, |
| "grad_norm": 1.1936954506770183, |
| "learning_rate": 1.9410508649875258e-09, |
| "loss": 0.4038, |
| "step": 1586 |
| }, |
| { |
| "epoch": 4.971025841816758, |
| "grad_norm": 1.1919200016703124, |
| "learning_rate": 1.5336802328147492e-09, |
| "loss": 0.3716, |
| "step": 1587 |
| }, |
| { |
| "epoch": 4.974158183241974, |
| "grad_norm": 1.2304919494528042, |
| "learning_rate": 1.1742309631845861e-09, |
| "loss": 0.4267, |
| "step": 1588 |
| }, |
| { |
| "epoch": 4.977290524667189, |
| "grad_norm": 1.2084987865361176, |
| "learning_rate": 8.627047788911658e-10, |
| "loss": 0.443, |
| "step": 1589 |
| }, |
| { |
| "epoch": 4.980422866092404, |
| "grad_norm": 1.2473534930975694, |
| "learning_rate": 5.991031730367968e-10, |
| "loss": 0.4292, |
| "step": 1590 |
| }, |
| { |
| "epoch": 4.983555207517619, |
| "grad_norm": 1.2137391828259563, |
| "learning_rate": 3.8342740903307697e-10, |
| "loss": 0.4086, |
| "step": 1591 |
| }, |
| { |
| "epoch": 4.986687548942835, |
| "grad_norm": 1.239277828293045, |
| "learning_rate": 2.1567852058423932e-10, |
| "loss": 0.3963, |
| "step": 1592 |
| }, |
| { |
| "epoch": 4.98981989036805, |
| "grad_norm": 1.198137612800272, |
| "learning_rate": 9.585731168937351e-11, |
| "loss": 0.3968, |
| "step": 1593 |
| }, |
| { |
| "epoch": 4.992952231793265, |
| "grad_norm": 1.2102944840840637, |
| "learning_rate": 2.396435663687413e-11, |
| "loss": 0.3788, |
| "step": 1594 |
| }, |
| { |
| "epoch": 4.99608457321848, |
| "grad_norm": 1.2188363908623323, |
| "learning_rate": 0.0, |
| "loss": 0.3945, |
| "step": 1595 |
| }, |
| { |
| "epoch": 4.99608457321848, |
| "step": 1595, |
| "total_flos": 1.3245374097522688e+18, |
| "train_loss": 0.0, |
| "train_runtime": 1.8638, |
| "train_samples_per_second": 54818.89, |
| "train_steps_per_second": 855.791 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1595, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.3245374097522688e+18, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|