{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 15.100037750094375, "eval_steps": 500, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003775009437523594, "grad_norm": 12.114919662475586, "learning_rate": 4.5e-07, "loss": 4.3735, "step": 10 }, { "epoch": 0.007550018875047188, "grad_norm": 9.25322437286377, "learning_rate": 9.5e-07, "loss": 4.2737, "step": 20 }, { "epoch": 0.011325028312570781, "grad_norm": 11.06075668334961, "learning_rate": 1.45e-06, "loss": 4.3709, "step": 30 }, { "epoch": 0.015100037750094376, "grad_norm": 11.528477668762207, "learning_rate": 1.95e-06, "loss": 4.5857, "step": 40 }, { "epoch": 0.01887504718761797, "grad_norm": 9.225029945373535, "learning_rate": 2.4500000000000003e-06, "loss": 4.1594, "step": 50 }, { "epoch": 0.022650056625141562, "grad_norm": 9.194082260131836, "learning_rate": 2.95e-06, "loss": 3.9774, "step": 60 }, { "epoch": 0.02642506606266516, "grad_norm": 9.647115707397461, "learning_rate": 3.4500000000000004e-06, "loss": 3.7629, "step": 70 }, { "epoch": 0.03020007550018875, "grad_norm": 7.582672595977783, "learning_rate": 3.95e-06, "loss": 3.3272, "step": 80 }, { "epoch": 0.03397508493771234, "grad_norm": 6.46481466293335, "learning_rate": 4.45e-06, "loss": 2.9437, "step": 90 }, { "epoch": 0.03775009437523594, "grad_norm": 4.374029159545898, "learning_rate": 4.950000000000001e-06, "loss": 2.6371, "step": 100 }, { "epoch": 0.041525103812759534, "grad_norm": 2.939470052719116, "learning_rate": 5.45e-06, "loss": 2.2299, "step": 110 }, { "epoch": 0.045300113250283124, "grad_norm": 2.47430682182312, "learning_rate": 5.95e-06, "loss": 2.0141, "step": 120 }, { "epoch": 0.04907512268780672, "grad_norm": 1.9557963609695435, "learning_rate": 6.45e-06, "loss": 1.8435, "step": 130 }, { "epoch": 0.05285013212533032, "grad_norm": 1.204042673110962, "learning_rate": 6.950000000000001e-06, "loss": 1.5451, "step": 140 }, { "epoch": 0.056625141562853906, "grad_norm": 0.9715531468391418, "learning_rate": 7.45e-06, "loss": 1.5027, "step": 150 }, { "epoch": 0.0604001510003775, "grad_norm": 1.261393427848816, "learning_rate": 7.95e-06, "loss": 1.3894, "step": 160 }, { "epoch": 0.0641751604379011, "grad_norm": 1.091807246208191, "learning_rate": 8.45e-06, "loss": 1.3591, "step": 170 }, { "epoch": 0.06795016987542468, "grad_norm": 0.8383681178092957, "learning_rate": 8.95e-06, "loss": 1.2757, "step": 180 }, { "epoch": 0.07172517931294828, "grad_norm": 1.3401917219161987, "learning_rate": 9.450000000000001e-06, "loss": 1.2084, "step": 190 }, { "epoch": 0.07550018875047187, "grad_norm": 0.8086752891540527, "learning_rate": 9.950000000000001e-06, "loss": 1.3253, "step": 200 }, { "epoch": 0.07927519818799547, "grad_norm": 0.9537988901138306, "learning_rate": 1.045e-05, "loss": 1.171, "step": 210 }, { "epoch": 0.08305020762551907, "grad_norm": 0.8162275552749634, "learning_rate": 1.095e-05, "loss": 1.1737, "step": 220 }, { "epoch": 0.08682521706304266, "grad_norm": 1.144768476486206, "learning_rate": 1.145e-05, "loss": 1.1406, "step": 230 }, { "epoch": 0.09060022650056625, "grad_norm": 1.2052839994430542, "learning_rate": 1.195e-05, "loss": 1.0841, "step": 240 }, { "epoch": 0.09437523593808984, "grad_norm": 1.1619216203689575, "learning_rate": 1.2450000000000001e-05, "loss": 1.2028, "step": 250 }, { "epoch": 0.09815024537561344, "grad_norm": 0.971839189529419, "learning_rate": 1.2950000000000001e-05, "loss": 1.0536, "step": 260 }, { "epoch": 0.10192525481313704, "grad_norm": 1.1455223560333252, "learning_rate": 1.3450000000000002e-05, "loss": 1.0539, "step": 270 }, { "epoch": 0.10570026425066063, "grad_norm": 0.9137556552886963, "learning_rate": 1.3950000000000002e-05, "loss": 1.0439, "step": 280 }, { "epoch": 0.10947527368818422, "grad_norm": 1.4055426120758057, "learning_rate": 1.4449999999999999e-05, "loss": 1.147, "step": 290 }, { "epoch": 0.11325028312570781, "grad_norm": 1.3591498136520386, "learning_rate": 1.4950000000000001e-05, "loss": 1.0227, "step": 300 }, { "epoch": 0.11702529256323141, "grad_norm": 1.004621148109436, "learning_rate": 1.545e-05, "loss": 1.0075, "step": 310 }, { "epoch": 0.120800302000755, "grad_norm": 1.2812256813049316, "learning_rate": 1.595e-05, "loss": 1.0799, "step": 320 }, { "epoch": 0.1245753114382786, "grad_norm": 1.5927451848983765, "learning_rate": 1.645e-05, "loss": 0.9892, "step": 330 }, { "epoch": 0.1283503208758022, "grad_norm": 1.2106682062149048, "learning_rate": 1.6950000000000002e-05, "loss": 0.9776, "step": 340 }, { "epoch": 0.1321253303133258, "grad_norm": 1.2619518041610718, "learning_rate": 1.745e-05, "loss": 0.9786, "step": 350 }, { "epoch": 0.13590033975084936, "grad_norm": 1.5045936107635498, "learning_rate": 1.795e-05, "loss": 0.9763, "step": 360 }, { "epoch": 0.13967534918837296, "grad_norm": 1.4551500082015991, "learning_rate": 1.845e-05, "loss": 0.9857, "step": 370 }, { "epoch": 0.14345035862589656, "grad_norm": 1.3843104839324951, "learning_rate": 1.895e-05, "loss": 0.9163, "step": 380 }, { "epoch": 0.14722536806342015, "grad_norm": 1.4003466367721558, "learning_rate": 1.9450000000000002e-05, "loss": 0.9637, "step": 390 }, { "epoch": 0.15100037750094375, "grad_norm": 1.3285315036773682, "learning_rate": 1.995e-05, "loss": 0.9353, "step": 400 }, { "epoch": 0.15477538693846735, "grad_norm": 1.0920555591583252, "learning_rate": 2.045e-05, "loss": 1.0432, "step": 410 }, { "epoch": 0.15855039637599094, "grad_norm": 1.4697812795639038, "learning_rate": 2.095e-05, "loss": 1.0497, "step": 420 }, { "epoch": 0.16232540581351454, "grad_norm": 1.444948434829712, "learning_rate": 2.145e-05, "loss": 0.9317, "step": 430 }, { "epoch": 0.16610041525103814, "grad_norm": 1.8489030599594116, "learning_rate": 2.195e-05, "loss": 0.9296, "step": 440 }, { "epoch": 0.16987542468856173, "grad_norm": 1.27092444896698, "learning_rate": 2.245e-05, "loss": 0.8998, "step": 450 }, { "epoch": 0.17365043412608533, "grad_norm": 1.3666588068008423, "learning_rate": 2.2950000000000002e-05, "loss": 1.0234, "step": 460 }, { "epoch": 0.1774254435636089, "grad_norm": 0.9460456967353821, "learning_rate": 2.345e-05, "loss": 0.9337, "step": 470 }, { "epoch": 0.1812004530011325, "grad_norm": 1.4684996604919434, "learning_rate": 2.395e-05, "loss": 0.9308, "step": 480 }, { "epoch": 0.1849754624386561, "grad_norm": 1.284343957901001, "learning_rate": 2.445e-05, "loss": 0.908, "step": 490 }, { "epoch": 0.1887504718761797, "grad_norm": 0.9511118531227112, "learning_rate": 2.495e-05, "loss": 0.8959, "step": 500 }, { "epoch": 0.19252548131370328, "grad_norm": 1.701767086982727, "learning_rate": 2.5450000000000002e-05, "loss": 0.9207, "step": 510 }, { "epoch": 0.19630049075122688, "grad_norm": 1.341230034828186, "learning_rate": 2.595e-05, "loss": 0.9837, "step": 520 }, { "epoch": 0.20007550018875048, "grad_norm": 1.1268380880355835, "learning_rate": 2.6450000000000003e-05, "loss": 0.8893, "step": 530 }, { "epoch": 0.20385050962627407, "grad_norm": 1.5450079441070557, "learning_rate": 2.6950000000000005e-05, "loss": 0.8875, "step": 540 }, { "epoch": 0.20762551906379767, "grad_norm": 1.827506184577942, "learning_rate": 2.7450000000000003e-05, "loss": 0.9069, "step": 550 }, { "epoch": 0.21140052850132127, "grad_norm": 1.225759506225586, "learning_rate": 2.7950000000000005e-05, "loss": 0.8977, "step": 560 }, { "epoch": 0.21517553793884484, "grad_norm": 1.1317689418792725, "learning_rate": 2.845e-05, "loss": 0.8673, "step": 570 }, { "epoch": 0.21895054737636843, "grad_norm": 1.3549116849899292, "learning_rate": 2.895e-05, "loss": 0.8613, "step": 580 }, { "epoch": 0.22272555681389203, "grad_norm": 1.5646511316299438, "learning_rate": 2.945e-05, "loss": 0.8847, "step": 590 }, { "epoch": 0.22650056625141562, "grad_norm": 1.453460693359375, "learning_rate": 2.995e-05, "loss": 0.8523, "step": 600 }, { "epoch": 0.23027557568893922, "grad_norm": 1.3521499633789062, "learning_rate": 3.045e-05, "loss": 0.8773, "step": 610 }, { "epoch": 0.23405058512646282, "grad_norm": 1.041245698928833, "learning_rate": 3.095e-05, "loss": 0.857, "step": 620 }, { "epoch": 0.23782559456398641, "grad_norm": 1.2298246622085571, "learning_rate": 3.145e-05, "loss": 0.8364, "step": 630 }, { "epoch": 0.24160060400151, "grad_norm": 1.2487760782241821, "learning_rate": 3.1950000000000004e-05, "loss": 0.8697, "step": 640 }, { "epoch": 0.2453756134390336, "grad_norm": 1.3753890991210938, "learning_rate": 3.245e-05, "loss": 0.8579, "step": 650 }, { "epoch": 0.2491506228765572, "grad_norm": 1.354738712310791, "learning_rate": 3.295e-05, "loss": 0.8574, "step": 660 }, { "epoch": 0.2529256323140808, "grad_norm": 1.3853284120559692, "learning_rate": 3.345000000000001e-05, "loss": 0.8346, "step": 670 }, { "epoch": 0.2567006417516044, "grad_norm": 1.4447784423828125, "learning_rate": 3.3950000000000005e-05, "loss": 0.9262, "step": 680 }, { "epoch": 0.26047565118912797, "grad_norm": 1.2130733728408813, "learning_rate": 3.445e-05, "loss": 0.8176, "step": 690 }, { "epoch": 0.2642506606266516, "grad_norm": 1.698705792427063, "learning_rate": 3.495e-05, "loss": 0.8424, "step": 700 }, { "epoch": 0.26802567006417516, "grad_norm": 1.724214792251587, "learning_rate": 3.545e-05, "loss": 0.9077, "step": 710 }, { "epoch": 0.2718006795016987, "grad_norm": 1.296976923942566, "learning_rate": 3.595e-05, "loss": 0.8253, "step": 720 }, { "epoch": 0.27557568893922235, "grad_norm": 1.2953417301177979, "learning_rate": 3.645e-05, "loss": 0.8949, "step": 730 }, { "epoch": 0.2793506983767459, "grad_norm": 1.267232894897461, "learning_rate": 3.6950000000000004e-05, "loss": 0.8476, "step": 740 }, { "epoch": 0.28312570781426954, "grad_norm": 1.76276695728302, "learning_rate": 3.745e-05, "loss": 0.8981, "step": 750 }, { "epoch": 0.2869007172517931, "grad_norm": 1.5421119928359985, "learning_rate": 3.795e-05, "loss": 0.8343, "step": 760 }, { "epoch": 0.29067572668931674, "grad_norm": 1.1343638896942139, "learning_rate": 3.845e-05, "loss": 0.8316, "step": 770 }, { "epoch": 0.2944507361268403, "grad_norm": 0.9002248048782349, "learning_rate": 3.8950000000000005e-05, "loss": 0.8207, "step": 780 }, { "epoch": 0.29822574556436393, "grad_norm": 1.273121953010559, "learning_rate": 3.9450000000000003e-05, "loss": 0.8782, "step": 790 }, { "epoch": 0.3020007550018875, "grad_norm": 1.6263741254806519, "learning_rate": 3.995e-05, "loss": 0.8359, "step": 800 }, { "epoch": 0.3057757644394111, "grad_norm": 1.3746614456176758, "learning_rate": 4.045000000000001e-05, "loss": 0.7946, "step": 810 }, { "epoch": 0.3095507738769347, "grad_norm": 1.2415761947631836, "learning_rate": 4.095e-05, "loss": 0.7973, "step": 820 }, { "epoch": 0.31332578331445826, "grad_norm": 1.372182846069336, "learning_rate": 4.145e-05, "loss": 0.9107, "step": 830 }, { "epoch": 0.3171007927519819, "grad_norm": 1.5173795223236084, "learning_rate": 4.195e-05, "loss": 0.7988, "step": 840 }, { "epoch": 0.32087580218950545, "grad_norm": 1.5358010530471802, "learning_rate": 4.245e-05, "loss": 0.7724, "step": 850 }, { "epoch": 0.3246508116270291, "grad_norm": 0.867175817489624, "learning_rate": 4.295e-05, "loss": 0.7816, "step": 860 }, { "epoch": 0.32842582106455265, "grad_norm": 0.9137406349182129, "learning_rate": 4.345e-05, "loss": 0.7855, "step": 870 }, { "epoch": 0.33220083050207627, "grad_norm": 0.9933797717094421, "learning_rate": 4.3950000000000004e-05, "loss": 0.8015, "step": 880 }, { "epoch": 0.33597583993959984, "grad_norm": 1.1218225955963135, "learning_rate": 4.445e-05, "loss": 0.7944, "step": 890 }, { "epoch": 0.33975084937712347, "grad_norm": 0.9928627610206604, "learning_rate": 4.495e-05, "loss": 0.785, "step": 900 }, { "epoch": 0.34352585881464703, "grad_norm": 1.0522204637527466, "learning_rate": 4.545000000000001e-05, "loss": 0.7819, "step": 910 }, { "epoch": 0.34730086825217066, "grad_norm": 1.162044644355774, "learning_rate": 4.5950000000000006e-05, "loss": 0.7721, "step": 920 }, { "epoch": 0.3510758776896942, "grad_norm": 0.9964906573295593, "learning_rate": 4.6450000000000004e-05, "loss": 0.7716, "step": 930 }, { "epoch": 0.3548508871272178, "grad_norm": 1.3030009269714355, "learning_rate": 4.695e-05, "loss": 0.7756, "step": 940 }, { "epoch": 0.3586258965647414, "grad_norm": 1.413189172744751, "learning_rate": 4.745e-05, "loss": 0.7709, "step": 950 }, { "epoch": 0.362400906002265, "grad_norm": 1.1650564670562744, "learning_rate": 4.795e-05, "loss": 0.7647, "step": 960 }, { "epoch": 0.3661759154397886, "grad_norm": 1.0290861129760742, "learning_rate": 4.845e-05, "loss": 0.7601, "step": 970 }, { "epoch": 0.3699509248773122, "grad_norm": 1.1548311710357666, "learning_rate": 4.8950000000000004e-05, "loss": 0.8407, "step": 980 }, { "epoch": 0.3737259343148358, "grad_norm": 1.037048578262329, "learning_rate": 4.945e-05, "loss": 0.7576, "step": 990 }, { "epoch": 0.3775009437523594, "grad_norm": 1.5112618207931519, "learning_rate": 4.995e-05, "loss": 0.7462, "step": 1000 }, { "epoch": 0.381275953189883, "grad_norm": 1.65474534034729, "learning_rate": 5.045e-05, "loss": 0.7722, "step": 1010 }, { "epoch": 0.38505096262740657, "grad_norm": 1.3295589685440063, "learning_rate": 5.095e-05, "loss": 0.7551, "step": 1020 }, { "epoch": 0.38882597206493014, "grad_norm": 0.9878939390182495, "learning_rate": 5.145e-05, "loss": 0.7522, "step": 1030 }, { "epoch": 0.39260098150245376, "grad_norm": 1.3945156335830688, "learning_rate": 5.1949999999999996e-05, "loss": 0.7827, "step": 1040 }, { "epoch": 0.39637599093997733, "grad_norm": 1.2583463191986084, "learning_rate": 5.245e-05, "loss": 0.7658, "step": 1050 }, { "epoch": 0.40015100037750095, "grad_norm": 1.0991382598876953, "learning_rate": 5.295e-05, "loss": 0.8056, "step": 1060 }, { "epoch": 0.4039260098150245, "grad_norm": 1.0522119998931885, "learning_rate": 5.345e-05, "loss": 0.8069, "step": 1070 }, { "epoch": 0.40770101925254815, "grad_norm": 0.8524134755134583, "learning_rate": 5.3950000000000004e-05, "loss": 0.8186, "step": 1080 }, { "epoch": 0.4114760286900717, "grad_norm": 1.1254100799560547, "learning_rate": 5.445e-05, "loss": 0.7362, "step": 1090 }, { "epoch": 0.41525103812759534, "grad_norm": 0.7959914207458496, "learning_rate": 5.495e-05, "loss": 0.7387, "step": 1100 }, { "epoch": 0.4190260475651189, "grad_norm": 1.2034642696380615, "learning_rate": 5.545e-05, "loss": 0.7458, "step": 1110 }, { "epoch": 0.42280105700264253, "grad_norm": 1.2874923944473267, "learning_rate": 5.5950000000000005e-05, "loss": 0.7442, "step": 1120 }, { "epoch": 0.4265760664401661, "grad_norm": 1.2025848627090454, "learning_rate": 5.645e-05, "loss": 0.7624, "step": 1130 }, { "epoch": 0.43035107587768967, "grad_norm": 1.1338868141174316, "learning_rate": 5.695e-05, "loss": 0.7311, "step": 1140 }, { "epoch": 0.4341260853152133, "grad_norm": 1.1724085807800293, "learning_rate": 5.745e-05, "loss": 0.7383, "step": 1150 }, { "epoch": 0.43790109475273686, "grad_norm": 1.3248220682144165, "learning_rate": 5.7950000000000006e-05, "loss": 0.7379, "step": 1160 }, { "epoch": 0.4416761041902605, "grad_norm": 1.4836875200271606, "learning_rate": 5.8450000000000005e-05, "loss": 0.7369, "step": 1170 }, { "epoch": 0.44545111362778406, "grad_norm": 1.1852290630340576, "learning_rate": 5.895e-05, "loss": 0.8117, "step": 1180 }, { "epoch": 0.4492261230653077, "grad_norm": 2.466386079788208, "learning_rate": 5.945000000000001e-05, "loss": 0.7997, "step": 1190 }, { "epoch": 0.45300113250283125, "grad_norm": 1.0344855785369873, "learning_rate": 5.995000000000001e-05, "loss": 0.7181, "step": 1200 }, { "epoch": 0.4567761419403549, "grad_norm": 1.1200584173202515, "learning_rate": 6.0450000000000006e-05, "loss": 0.7204, "step": 1210 }, { "epoch": 0.46055115137787844, "grad_norm": 1.0157345533370972, "learning_rate": 6.0950000000000004e-05, "loss": 0.7261, "step": 1220 }, { "epoch": 0.464326160815402, "grad_norm": 0.9601739048957825, "learning_rate": 6.145e-05, "loss": 0.7099, "step": 1230 }, { "epoch": 0.46810117025292564, "grad_norm": 0.8642465472221375, "learning_rate": 6.195e-05, "loss": 0.7242, "step": 1240 }, { "epoch": 0.4718761796904492, "grad_norm": 0.9395157098770142, "learning_rate": 6.245000000000001e-05, "loss": 0.716, "step": 1250 }, { "epoch": 0.47565118912797283, "grad_norm": 1.0848698616027832, "learning_rate": 6.295e-05, "loss": 0.714, "step": 1260 }, { "epoch": 0.4794261985654964, "grad_norm": 1.1483030319213867, "learning_rate": 6.345e-05, "loss": 0.7224, "step": 1270 }, { "epoch": 0.48320120800302, "grad_norm": 0.8236299753189087, "learning_rate": 6.395e-05, "loss": 0.7639, "step": 1280 }, { "epoch": 0.4869762174405436, "grad_norm": 0.9821550846099854, "learning_rate": 6.445e-05, "loss": 0.7136, "step": 1290 }, { "epoch": 0.4907512268780672, "grad_norm": 0.94017493724823, "learning_rate": 6.494999999999999e-05, "loss": 0.7056, "step": 1300 }, { "epoch": 0.4945262363155908, "grad_norm": 0.8587389588356018, "learning_rate": 6.545e-05, "loss": 0.7518, "step": 1310 }, { "epoch": 0.4983012457531144, "grad_norm": 0.9293227791786194, "learning_rate": 6.595e-05, "loss": 0.7058, "step": 1320 }, { "epoch": 0.5020762551906379, "grad_norm": 1.033216118812561, "learning_rate": 6.645e-05, "loss": 0.7106, "step": 1330 }, { "epoch": 0.5058512646281615, "grad_norm": 0.8532618880271912, "learning_rate": 6.695e-05, "loss": 0.7153, "step": 1340 }, { "epoch": 0.5096262740656852, "grad_norm": 0.7834755182266235, "learning_rate": 6.745e-05, "loss": 0.705, "step": 1350 }, { "epoch": 0.5134012835032088, "grad_norm": 0.7160763740539551, "learning_rate": 6.795e-05, "loss": 0.6986, "step": 1360 }, { "epoch": 0.5171762929407323, "grad_norm": 0.9718685150146484, "learning_rate": 6.845e-05, "loss": 0.7002, "step": 1370 }, { "epoch": 0.5209513023782559, "grad_norm": 0.9730781316757202, "learning_rate": 6.895000000000001e-05, "loss": 0.7107, "step": 1380 }, { "epoch": 0.5247263118157796, "grad_norm": 1.1289321184158325, "learning_rate": 6.945000000000001e-05, "loss": 0.6985, "step": 1390 }, { "epoch": 0.5285013212533032, "grad_norm": 0.6597205400466919, "learning_rate": 6.995e-05, "loss": 0.7051, "step": 1400 }, { "epoch": 0.5322763306908267, "grad_norm": 1.0954090356826782, "learning_rate": 7.045e-05, "loss": 0.6965, "step": 1410 }, { "epoch": 0.5360513401283503, "grad_norm": 0.9379609823226929, "learning_rate": 7.095e-05, "loss": 0.6937, "step": 1420 }, { "epoch": 0.5398263495658739, "grad_norm": 1.048135757446289, "learning_rate": 7.145e-05, "loss": 0.6957, "step": 1430 }, { "epoch": 0.5436013590033975, "grad_norm": 1.2906863689422607, "learning_rate": 7.195e-05, "loss": 0.6994, "step": 1440 }, { "epoch": 0.5473763684409211, "grad_norm": 1.1085039377212524, "learning_rate": 7.245000000000001e-05, "loss": 0.6932, "step": 1450 }, { "epoch": 0.5511513778784447, "grad_norm": 1.1647526025772095, "learning_rate": 7.295000000000001e-05, "loss": 0.7641, "step": 1460 }, { "epoch": 0.5549263873159683, "grad_norm": 0.9016187191009521, "learning_rate": 7.345000000000001e-05, "loss": 0.7047, "step": 1470 }, { "epoch": 0.5587013967534918, "grad_norm": 0.7149607539176941, "learning_rate": 7.395000000000001e-05, "loss": 0.6924, "step": 1480 }, { "epoch": 0.5624764061910155, "grad_norm": 0.9485162496566772, "learning_rate": 7.445000000000001e-05, "loss": 0.6896, "step": 1490 }, { "epoch": 0.5662514156285391, "grad_norm": 0.9566022157669067, "learning_rate": 7.495e-05, "loss": 0.6893, "step": 1500 }, { "epoch": 0.5700264250660627, "grad_norm": 0.8616087436676025, "learning_rate": 7.545e-05, "loss": 0.7356, "step": 1510 }, { "epoch": 0.5738014345035862, "grad_norm": 0.7664632201194763, "learning_rate": 7.595e-05, "loss": 0.6909, "step": 1520 }, { "epoch": 0.5775764439411099, "grad_norm": 0.673774003982544, "learning_rate": 7.645e-05, "loss": 0.6804, "step": 1530 }, { "epoch": 0.5813514533786335, "grad_norm": 1.0145224332809448, "learning_rate": 7.695e-05, "loss": 0.696, "step": 1540 }, { "epoch": 0.585126462816157, "grad_norm": 0.805166482925415, "learning_rate": 7.745e-05, "loss": 0.7003, "step": 1550 }, { "epoch": 0.5889014722536806, "grad_norm": 0.7893358469009399, "learning_rate": 7.795e-05, "loss": 0.6933, "step": 1560 }, { "epoch": 0.5926764816912042, "grad_norm": 0.9435643553733826, "learning_rate": 7.845e-05, "loss": 0.6926, "step": 1570 }, { "epoch": 0.5964514911287279, "grad_norm": 0.9065937399864197, "learning_rate": 7.895000000000001e-05, "loss": 0.685, "step": 1580 }, { "epoch": 0.6002265005662514, "grad_norm": 1.084681749343872, "learning_rate": 7.945e-05, "loss": 0.7046, "step": 1590 }, { "epoch": 0.604001510003775, "grad_norm": 0.6447507739067078, "learning_rate": 7.995e-05, "loss": 0.7189, "step": 1600 }, { "epoch": 0.6077765194412986, "grad_norm": 0.855850100517273, "learning_rate": 8.045e-05, "loss": 0.6817, "step": 1610 }, { "epoch": 0.6115515288788222, "grad_norm": 1.1896417140960693, "learning_rate": 8.095e-05, "loss": 0.6844, "step": 1620 }, { "epoch": 0.6153265383163458, "grad_norm": 0.8419579267501831, "learning_rate": 8.145e-05, "loss": 0.6771, "step": 1630 }, { "epoch": 0.6191015477538694, "grad_norm": 0.8358510732650757, "learning_rate": 8.195e-05, "loss": 0.6855, "step": 1640 }, { "epoch": 0.622876557191393, "grad_norm": 0.896302342414856, "learning_rate": 8.245e-05, "loss": 0.6955, "step": 1650 }, { "epoch": 0.6266515666289165, "grad_norm": 1.024187684059143, "learning_rate": 8.295000000000001e-05, "loss": 0.6811, "step": 1660 }, { "epoch": 0.6304265760664401, "grad_norm": 0.8998435735702515, "learning_rate": 8.345000000000001e-05, "loss": 0.6807, "step": 1670 }, { "epoch": 0.6342015855039638, "grad_norm": 0.6220159530639648, "learning_rate": 8.395000000000001e-05, "loss": 0.6795, "step": 1680 }, { "epoch": 0.6379765949414874, "grad_norm": 0.9150254130363464, "learning_rate": 8.445e-05, "loss": 0.6695, "step": 1690 }, { "epoch": 0.6417516043790109, "grad_norm": 0.7432071566581726, "learning_rate": 8.495e-05, "loss": 0.6792, "step": 1700 }, { "epoch": 0.6455266138165345, "grad_norm": 0.808611273765564, "learning_rate": 8.545e-05, "loss": 0.6734, "step": 1710 }, { "epoch": 0.6493016232540582, "grad_norm": 0.6958522200584412, "learning_rate": 8.595e-05, "loss": 0.6867, "step": 1720 }, { "epoch": 0.6530766326915818, "grad_norm": 0.7854651808738708, "learning_rate": 8.645000000000001e-05, "loss": 0.6837, "step": 1730 }, { "epoch": 0.6568516421291053, "grad_norm": 0.7910404205322266, "learning_rate": 8.695000000000001e-05, "loss": 0.6746, "step": 1740 }, { "epoch": 0.6606266515666289, "grad_norm": 0.7932810187339783, "learning_rate": 8.745000000000001e-05, "loss": 0.6642, "step": 1750 }, { "epoch": 0.6644016610041525, "grad_norm": 0.7843324542045593, "learning_rate": 8.795e-05, "loss": 0.6641, "step": 1760 }, { "epoch": 0.6681766704416761, "grad_norm": 0.728706955909729, "learning_rate": 8.845e-05, "loss": 0.6745, "step": 1770 }, { "epoch": 0.6719516798791997, "grad_norm": 0.8016153573989868, "learning_rate": 8.895e-05, "loss": 0.6671, "step": 1780 }, { "epoch": 0.6757266893167233, "grad_norm": 0.695740282535553, "learning_rate": 8.945e-05, "loss": 0.6634, "step": 1790 }, { "epoch": 0.6795016987542469, "grad_norm": 0.8751599788665771, "learning_rate": 8.995e-05, "loss": 0.6747, "step": 1800 }, { "epoch": 0.6832767081917704, "grad_norm": 0.46772128343582153, "learning_rate": 9.045e-05, "loss": 0.6721, "step": 1810 }, { "epoch": 0.6870517176292941, "grad_norm": 0.7569857239723206, "learning_rate": 9.095e-05, "loss": 0.6675, "step": 1820 }, { "epoch": 0.6908267270668177, "grad_norm": 0.853040337562561, "learning_rate": 9.145e-05, "loss": 0.6685, "step": 1830 }, { "epoch": 0.6946017365043413, "grad_norm": 0.6927493214607239, "learning_rate": 9.195e-05, "loss": 0.654, "step": 1840 }, { "epoch": 0.6983767459418648, "grad_norm": 0.6520769000053406, "learning_rate": 9.245e-05, "loss": 0.6641, "step": 1850 }, { "epoch": 0.7021517553793885, "grad_norm": 0.9358047246932983, "learning_rate": 9.295000000000001e-05, "loss": 0.6696, "step": 1860 }, { "epoch": 0.7059267648169121, "grad_norm": 0.7453898787498474, "learning_rate": 9.345000000000001e-05, "loss": 0.6549, "step": 1870 }, { "epoch": 0.7097017742544356, "grad_norm": 0.6966333389282227, "learning_rate": 9.395000000000001e-05, "loss": 0.6653, "step": 1880 }, { "epoch": 0.7134767836919592, "grad_norm": 0.7843179702758789, "learning_rate": 9.445e-05, "loss": 0.6715, "step": 1890 }, { "epoch": 0.7172517931294828, "grad_norm": 0.6947644948959351, "learning_rate": 9.495e-05, "loss": 0.6584, "step": 1900 }, { "epoch": 0.7210268025670065, "grad_norm": 0.8173465132713318, "learning_rate": 9.545e-05, "loss": 0.6649, "step": 1910 }, { "epoch": 0.72480181200453, "grad_norm": 0.9682592749595642, "learning_rate": 9.595e-05, "loss": 0.6549, "step": 1920 }, { "epoch": 0.7285768214420536, "grad_norm": 1.0117123126983643, "learning_rate": 9.645000000000001e-05, "loss": 0.6536, "step": 1930 }, { "epoch": 0.7323518308795772, "grad_norm": 0.9777953028678894, "learning_rate": 9.695000000000001e-05, "loss": 0.6578, "step": 1940 }, { "epoch": 0.7361268403171007, "grad_norm": 0.7957656383514404, "learning_rate": 9.745000000000001e-05, "loss": 0.6566, "step": 1950 }, { "epoch": 0.7399018497546244, "grad_norm": 0.7392054796218872, "learning_rate": 9.795000000000001e-05, "loss": 0.66, "step": 1960 }, { "epoch": 0.743676859192148, "grad_norm": 0.811776340007782, "learning_rate": 9.845000000000001e-05, "loss": 0.6535, "step": 1970 }, { "epoch": 0.7474518686296716, "grad_norm": 0.750037670135498, "learning_rate": 9.895e-05, "loss": 0.651, "step": 1980 }, { "epoch": 0.7512268780671951, "grad_norm": 0.6936712861061096, "learning_rate": 9.945e-05, "loss": 0.6594, "step": 1990 }, { "epoch": 0.7550018875047187, "grad_norm": 0.9359191060066223, "learning_rate": 9.995e-05, "loss": 0.6546, "step": 2000 }, { "epoch": 0.7587768969422424, "grad_norm": 0.8320378065109253, "learning_rate": 9.99999861593158e-05, "loss": 0.6427, "step": 2010 }, { "epoch": 0.762551906379766, "grad_norm": 0.6402168869972229, "learning_rate": 9.999993831498517e-05, "loss": 0.6711, "step": 2020 }, { "epoch": 0.7663269158172895, "grad_norm": 0.6975911259651184, "learning_rate": 9.99998562961682e-05, "loss": 0.6551, "step": 2030 }, { "epoch": 0.7701019252548131, "grad_norm": 0.7861037254333496, "learning_rate": 9.99997401029209e-05, "loss": 0.6548, "step": 2040 }, { "epoch": 0.7738769346923368, "grad_norm": 0.7647865414619446, "learning_rate": 9.999958973532271e-05, "loss": 0.6553, "step": 2050 }, { "epoch": 0.7776519441298603, "grad_norm": 1.0054905414581299, "learning_rate": 9.999940519347642e-05, "loss": 0.6523, "step": 2060 }, { "epoch": 0.7814269535673839, "grad_norm": 0.9435162544250488, "learning_rate": 9.999918647750812e-05, "loss": 0.6573, "step": 2070 }, { "epoch": 0.7852019630049075, "grad_norm": 0.7330409288406372, "learning_rate": 9.999893358756736e-05, "loss": 0.6601, "step": 2080 }, { "epoch": 0.7889769724424311, "grad_norm": 1.0650360584259033, "learning_rate": 9.999864652382692e-05, "loss": 0.6481, "step": 2090 }, { "epoch": 0.7927519818799547, "grad_norm": 0.6721388697624207, "learning_rate": 9.999832528648307e-05, "loss": 0.6395, "step": 2100 }, { "epoch": 0.7965269913174783, "grad_norm": 0.8571961522102356, "learning_rate": 9.999796987575532e-05, "loss": 0.6459, "step": 2110 }, { "epoch": 0.8003020007550019, "grad_norm": 0.575950026512146, "learning_rate": 9.999758029188662e-05, "loss": 0.6459, "step": 2120 }, { "epoch": 0.8040770101925255, "grad_norm": 0.5099409818649292, "learning_rate": 9.999715653514324e-05, "loss": 0.6484, "step": 2130 }, { "epoch": 0.807852019630049, "grad_norm": 0.6396839618682861, "learning_rate": 9.999669860581482e-05, "loss": 0.6504, "step": 2140 }, { "epoch": 0.8116270290675727, "grad_norm": 0.6424649357795715, "learning_rate": 9.999620650421433e-05, "loss": 0.6396, "step": 2150 }, { "epoch": 0.8154020385050963, "grad_norm": 0.7561330199241638, "learning_rate": 9.999568023067813e-05, "loss": 0.6481, "step": 2160 }, { "epoch": 0.8191770479426198, "grad_norm": 0.8530335426330566, "learning_rate": 9.999511978556595e-05, "loss": 0.6392, "step": 2170 }, { "epoch": 0.8229520573801434, "grad_norm": 0.8621054887771606, "learning_rate": 9.99945251692608e-05, "loss": 0.652, "step": 2180 }, { "epoch": 0.8267270668176671, "grad_norm": 0.7959499359130859, "learning_rate": 9.999389638216912e-05, "loss": 0.6384, "step": 2190 }, { "epoch": 0.8305020762551907, "grad_norm": 0.6927378177642822, "learning_rate": 9.999323342472068e-05, "loss": 0.6432, "step": 2200 }, { "epoch": 0.8342770856927142, "grad_norm": 0.5878588557243347, "learning_rate": 9.99925362973686e-05, "loss": 0.6464, "step": 2210 }, { "epoch": 0.8380520951302378, "grad_norm": 0.6220201253890991, "learning_rate": 9.999180500058935e-05, "loss": 0.6978, "step": 2220 }, { "epoch": 0.8418271045677614, "grad_norm": 0.6637720465660095, "learning_rate": 9.99910395348828e-05, "loss": 0.6336, "step": 2230 }, { "epoch": 0.8456021140052851, "grad_norm": 0.7113637924194336, "learning_rate": 9.99902399007721e-05, "loss": 0.633, "step": 2240 }, { "epoch": 0.8493771234428086, "grad_norm": 0.7006509900093079, "learning_rate": 9.99894060988038e-05, "loss": 0.634, "step": 2250 }, { "epoch": 0.8531521328803322, "grad_norm": 0.7758764624595642, "learning_rate": 9.998853812954783e-05, "loss": 0.6327, "step": 2260 }, { "epoch": 0.8569271423178558, "grad_norm": 0.7731608748435974, "learning_rate": 9.998763599359739e-05, "loss": 0.6301, "step": 2270 }, { "epoch": 0.8607021517553793, "grad_norm": 0.9600358009338379, "learning_rate": 9.998669969156912e-05, "loss": 0.6473, "step": 2280 }, { "epoch": 0.864477161192903, "grad_norm": 0.6508148312568665, "learning_rate": 9.998572922410294e-05, "loss": 0.6401, "step": 2290 }, { "epoch": 0.8682521706304266, "grad_norm": 0.6029082536697388, "learning_rate": 9.998472459186219e-05, "loss": 0.6355, "step": 2300 }, { "epoch": 0.8720271800679502, "grad_norm": 0.5723931193351746, "learning_rate": 9.99836857955335e-05, "loss": 0.6352, "step": 2310 }, { "epoch": 0.8758021895054737, "grad_norm": 0.5845490097999573, "learning_rate": 9.998261283582688e-05, "loss": 0.6348, "step": 2320 }, { "epoch": 0.8795771989429974, "grad_norm": 0.702069103717804, "learning_rate": 9.998150571347571e-05, "loss": 0.7144, "step": 2330 }, { "epoch": 0.883352208380521, "grad_norm": 0.7790460586547852, "learning_rate": 9.998036442923667e-05, "loss": 0.6357, "step": 2340 }, { "epoch": 0.8871272178180446, "grad_norm": 0.6690578460693359, "learning_rate": 9.997918898388983e-05, "loss": 0.6327, "step": 2350 }, { "epoch": 0.8909022272555681, "grad_norm": 0.7188066244125366, "learning_rate": 9.997797937823859e-05, "loss": 0.626, "step": 2360 }, { "epoch": 0.8946772366930917, "grad_norm": 0.6960968971252441, "learning_rate": 9.997673561310971e-05, "loss": 0.6359, "step": 2370 }, { "epoch": 0.8984522461306154, "grad_norm": 0.5999962091445923, "learning_rate": 9.997545768935331e-05, "loss": 0.642, "step": 2380 }, { "epoch": 0.9022272555681389, "grad_norm": 0.6590844988822937, "learning_rate": 9.997414560784282e-05, "loss": 0.6334, "step": 2390 }, { "epoch": 0.9060022650056625, "grad_norm": 0.5802249908447266, "learning_rate": 9.997279936947502e-05, "loss": 0.6319, "step": 2400 }, { "epoch": 0.9097772744431861, "grad_norm": 0.5891656279563904, "learning_rate": 9.997141897517006e-05, "loss": 0.6261, "step": 2410 }, { "epoch": 0.9135522838807097, "grad_norm": 0.6824517846107483, "learning_rate": 9.997000442587145e-05, "loss": 0.6261, "step": 2420 }, { "epoch": 0.9173272933182333, "grad_norm": 0.7084891200065613, "learning_rate": 9.9968555722546e-05, "loss": 0.6266, "step": 2430 }, { "epoch": 0.9211023027557569, "grad_norm": 0.5537872314453125, "learning_rate": 9.996707286618387e-05, "loss": 0.6271, "step": 2440 }, { "epoch": 0.9248773121932805, "grad_norm": 0.5806447863578796, "learning_rate": 9.996555585779861e-05, "loss": 0.6356, "step": 2450 }, { "epoch": 0.928652321630804, "grad_norm": 0.6291959285736084, "learning_rate": 9.996400469842708e-05, "loss": 0.6344, "step": 2460 }, { "epoch": 0.9324273310683276, "grad_norm": 0.5299859046936035, "learning_rate": 9.996241938912945e-05, "loss": 0.6256, "step": 2470 }, { "epoch": 0.9362023405058513, "grad_norm": 0.5358120203018188, "learning_rate": 9.99607999309893e-05, "loss": 0.6164, "step": 2480 }, { "epoch": 0.9399773499433749, "grad_norm": 0.4257400929927826, "learning_rate": 9.99591463251135e-05, "loss": 0.6238, "step": 2490 }, { "epoch": 0.9437523593808984, "grad_norm": 0.5488512516021729, "learning_rate": 9.995745857263226e-05, "loss": 0.632, "step": 2500 }, { "epoch": 0.947527368818422, "grad_norm": 0.45664718747138977, "learning_rate": 9.995573667469918e-05, "loss": 0.6276, "step": 2510 }, { "epoch": 0.9513023782559457, "grad_norm": 0.7003333568572998, "learning_rate": 9.995398063249111e-05, "loss": 0.6248, "step": 2520 }, { "epoch": 0.9550773876934693, "grad_norm": 0.5362398028373718, "learning_rate": 9.995219044720833e-05, "loss": 0.6277, "step": 2530 }, { "epoch": 0.9588523971309928, "grad_norm": 0.8197181224822998, "learning_rate": 9.99503661200744e-05, "loss": 0.6219, "step": 2540 }, { "epoch": 0.9626274065685164, "grad_norm": 0.671510636806488, "learning_rate": 9.994850765233624e-05, "loss": 0.6262, "step": 2550 }, { "epoch": 0.96640241600604, "grad_norm": 0.6340009570121765, "learning_rate": 9.994661504526406e-05, "loss": 0.6167, "step": 2560 }, { "epoch": 0.9701774254435636, "grad_norm": 0.5801326632499695, "learning_rate": 9.994468830015148e-05, "loss": 0.6319, "step": 2570 }, { "epoch": 0.9739524348810872, "grad_norm": 0.6170343160629272, "learning_rate": 9.99427274183154e-05, "loss": 0.6233, "step": 2580 }, { "epoch": 0.9777274443186108, "grad_norm": 0.5810487866401672, "learning_rate": 9.994073240109606e-05, "loss": 0.6281, "step": 2590 }, { "epoch": 0.9815024537561344, "grad_norm": 0.46137765049934387, "learning_rate": 9.993870324985703e-05, "loss": 0.6144, "step": 2600 }, { "epoch": 0.9852774631936579, "grad_norm": 0.6433082222938538, "learning_rate": 9.993663996598523e-05, "loss": 0.6129, "step": 2610 }, { "epoch": 0.9890524726311816, "grad_norm": 0.6449345350265503, "learning_rate": 9.993454255089089e-05, "loss": 0.6149, "step": 2620 }, { "epoch": 0.9928274820687052, "grad_norm": 0.6302064061164856, "learning_rate": 9.993241100600757e-05, "loss": 0.6192, "step": 2630 }, { "epoch": 0.9966024915062288, "grad_norm": 0.7228666543960571, "learning_rate": 9.993024533279215e-05, "loss": 0.6179, "step": 2640 }, { "epoch": 1.0003775009437523, "grad_norm": 0.5953015685081482, "learning_rate": 9.992804553272487e-05, "loss": 0.6176, "step": 2650 }, { "epoch": 1.0041525103812758, "grad_norm": 0.586175799369812, "learning_rate": 9.992581160730926e-05, "loss": 0.6095, "step": 2660 }, { "epoch": 1.0079275198187996, "grad_norm": 0.6601322293281555, "learning_rate": 9.992354355807218e-05, "loss": 0.608, "step": 2670 }, { "epoch": 1.011702529256323, "grad_norm": 0.7241320013999939, "learning_rate": 9.992124138656385e-05, "loss": 0.6204, "step": 2680 }, { "epoch": 1.0154775386938468, "grad_norm": 0.7130767107009888, "learning_rate": 9.991890509435774e-05, "loss": 0.6191, "step": 2690 }, { "epoch": 1.0192525481313703, "grad_norm": 0.5906509757041931, "learning_rate": 9.991653468305071e-05, "loss": 0.6216, "step": 2700 }, { "epoch": 1.0230275575688939, "grad_norm": 0.4647708535194397, "learning_rate": 9.991413015426292e-05, "loss": 0.6226, "step": 2710 }, { "epoch": 1.0268025670064176, "grad_norm": 0.5289870500564575, "learning_rate": 9.991169150963781e-05, "loss": 0.6204, "step": 2720 }, { "epoch": 1.030577576443941, "grad_norm": 0.5250043272972107, "learning_rate": 9.990921875084221e-05, "loss": 0.6101, "step": 2730 }, { "epoch": 1.0343525858814646, "grad_norm": 0.553863525390625, "learning_rate": 9.990671187956622e-05, "loss": 0.6236, "step": 2740 }, { "epoch": 1.0381275953189883, "grad_norm": 0.6878931522369385, "learning_rate": 9.990417089752324e-05, "loss": 0.6211, "step": 2750 }, { "epoch": 1.0419026047565119, "grad_norm": 0.5037298798561096, "learning_rate": 9.990159580645001e-05, "loss": 0.6128, "step": 2760 }, { "epoch": 1.0456776141940356, "grad_norm": 0.5697045922279358, "learning_rate": 9.98989866081066e-05, "loss": 0.6088, "step": 2770 }, { "epoch": 1.0494526236315591, "grad_norm": 0.6046783328056335, "learning_rate": 9.989634330427636e-05, "loss": 0.624, "step": 2780 }, { "epoch": 1.0532276330690826, "grad_norm": 0.6193393468856812, "learning_rate": 9.989366589676596e-05, "loss": 0.6133, "step": 2790 }, { "epoch": 1.0570026425066064, "grad_norm": 0.6406804323196411, "learning_rate": 9.98909543874054e-05, "loss": 0.6164, "step": 2800 }, { "epoch": 1.0607776519441299, "grad_norm": 0.6087714433670044, "learning_rate": 9.988820877804795e-05, "loss": 0.6113, "step": 2810 }, { "epoch": 1.0645526613816534, "grad_norm": 0.4003846347332001, "learning_rate": 9.988542907057022e-05, "loss": 0.6065, "step": 2820 }, { "epoch": 1.0683276708191771, "grad_norm": 0.5162433385848999, "learning_rate": 9.988261526687212e-05, "loss": 0.6105, "step": 2830 }, { "epoch": 1.0721026802567006, "grad_norm": 0.5106794238090515, "learning_rate": 9.987976736887685e-05, "loss": 0.6105, "step": 2840 }, { "epoch": 1.0758776896942241, "grad_norm": 0.5757883787155151, "learning_rate": 9.987688537853091e-05, "loss": 0.6105, "step": 2850 }, { "epoch": 1.0796526991317479, "grad_norm": 0.5995069742202759, "learning_rate": 9.987396929780413e-05, "loss": 0.6119, "step": 2860 }, { "epoch": 1.0834277085692714, "grad_norm": 0.5145584940910339, "learning_rate": 9.987101912868962e-05, "loss": 0.6108, "step": 2870 }, { "epoch": 1.087202718006795, "grad_norm": 0.418322890996933, "learning_rate": 9.98680348732038e-05, "loss": 0.6076, "step": 2880 }, { "epoch": 1.0909777274443186, "grad_norm": 0.526392936706543, "learning_rate": 9.986501653338636e-05, "loss": 0.61, "step": 2890 }, { "epoch": 1.0947527368818422, "grad_norm": 0.6863415837287903, "learning_rate": 9.986196411130031e-05, "loss": 0.6483, "step": 2900 }, { "epoch": 1.098527746319366, "grad_norm": 0.8329252600669861, "learning_rate": 9.985887760903197e-05, "loss": 0.6162, "step": 2910 }, { "epoch": 1.1023027557568894, "grad_norm": 0.6749866008758545, "learning_rate": 9.985575702869093e-05, "loss": 0.6157, "step": 2920 }, { "epoch": 1.106077765194413, "grad_norm": 0.6485674381256104, "learning_rate": 9.985260237241008e-05, "loss": 0.6161, "step": 2930 }, { "epoch": 1.1098527746319367, "grad_norm": 0.5751580595970154, "learning_rate": 9.984941364234557e-05, "loss": 0.6196, "step": 2940 }, { "epoch": 1.1136277840694602, "grad_norm": 0.539655864238739, "learning_rate": 9.98461908406769e-05, "loss": 0.6143, "step": 2950 }, { "epoch": 1.1174027935069837, "grad_norm": 0.5958253145217896, "learning_rate": 9.98429339696068e-05, "loss": 0.6111, "step": 2960 }, { "epoch": 1.1211778029445074, "grad_norm": 0.6310438513755798, "learning_rate": 9.983964303136133e-05, "loss": 0.6088, "step": 2970 }, { "epoch": 1.124952812382031, "grad_norm": 0.5665124654769897, "learning_rate": 9.983631802818981e-05, "loss": 0.6046, "step": 2980 }, { "epoch": 1.1287278218195547, "grad_norm": 0.4563884437084198, "learning_rate": 9.983295896236484e-05, "loss": 0.5985, "step": 2990 }, { "epoch": 1.1325028312570782, "grad_norm": 0.5045254230499268, "learning_rate": 9.982956583618232e-05, "loss": 0.6024, "step": 3000 }, { "epoch": 1.1362778406946017, "grad_norm": 0.5779048800468445, "learning_rate": 9.982613865196142e-05, "loss": 0.6088, "step": 3010 }, { "epoch": 1.1400528501321254, "grad_norm": 0.5393741726875305, "learning_rate": 9.982267741204458e-05, "loss": 0.6145, "step": 3020 }, { "epoch": 1.143827859569649, "grad_norm": 0.5488253831863403, "learning_rate": 9.981918211879753e-05, "loss": 0.6019, "step": 3030 }, { "epoch": 1.1476028690071725, "grad_norm": 0.5438612103462219, "learning_rate": 9.981565277460927e-05, "loss": 0.6165, "step": 3040 }, { "epoch": 1.1513778784446962, "grad_norm": 0.5116326808929443, "learning_rate": 9.981208938189206e-05, "loss": 0.6062, "step": 3050 }, { "epoch": 1.1551528878822197, "grad_norm": 0.48625659942626953, "learning_rate": 9.980849194308149e-05, "loss": 0.6113, "step": 3060 }, { "epoch": 1.1589278973197432, "grad_norm": 0.5895056128501892, "learning_rate": 9.98048604606363e-05, "loss": 0.6002, "step": 3070 }, { "epoch": 1.162702906757267, "grad_norm": 0.7021406292915344, "learning_rate": 9.980119493703864e-05, "loss": 0.6139, "step": 3080 }, { "epoch": 1.1664779161947905, "grad_norm": 0.7085283398628235, "learning_rate": 9.979749537479383e-05, "loss": 0.6058, "step": 3090 }, { "epoch": 1.170252925632314, "grad_norm": 0.6100856065750122, "learning_rate": 9.979376177643051e-05, "loss": 0.6003, "step": 3100 }, { "epoch": 1.1740279350698377, "grad_norm": 0.6801294684410095, "learning_rate": 9.978999414450052e-05, "loss": 0.6171, "step": 3110 }, { "epoch": 1.1778029445073612, "grad_norm": 0.47147953510284424, "learning_rate": 9.978619248157904e-05, "loss": 0.6057, "step": 3120 }, { "epoch": 1.1815779539448847, "grad_norm": 0.5377033352851868, "learning_rate": 9.978235679026445e-05, "loss": 0.603, "step": 3130 }, { "epoch": 1.1853529633824085, "grad_norm": 0.6004739999771118, "learning_rate": 9.97784870731784e-05, "loss": 0.5997, "step": 3140 }, { "epoch": 1.189127972819932, "grad_norm": 0.42120230197906494, "learning_rate": 9.977458333296584e-05, "loss": 0.6065, "step": 3150 }, { "epoch": 1.1929029822574557, "grad_norm": 0.4445311427116394, "learning_rate": 9.977064557229492e-05, "loss": 0.6049, "step": 3160 }, { "epoch": 1.1966779916949792, "grad_norm": 0.5501314997673035, "learning_rate": 9.976667379385705e-05, "loss": 0.6177, "step": 3170 }, { "epoch": 1.2004530011325028, "grad_norm": 0.6366961598396301, "learning_rate": 9.97626680003669e-05, "loss": 0.6134, "step": 3180 }, { "epoch": 1.2042280105700265, "grad_norm": 0.5312647223472595, "learning_rate": 9.975862819456242e-05, "loss": 0.5979, "step": 3190 }, { "epoch": 1.20800302000755, "grad_norm": 0.5397608876228333, "learning_rate": 9.975455437920477e-05, "loss": 0.5947, "step": 3200 }, { "epoch": 1.2117780294450737, "grad_norm": 0.4828698933124542, "learning_rate": 9.975044655707834e-05, "loss": 0.6012, "step": 3210 }, { "epoch": 1.2155530388825972, "grad_norm": 0.5150812864303589, "learning_rate": 9.974630473099082e-05, "loss": 0.604, "step": 3220 }, { "epoch": 1.2193280483201208, "grad_norm": 0.6184629201889038, "learning_rate": 9.974212890377311e-05, "loss": 0.5917, "step": 3230 }, { "epoch": 1.2231030577576445, "grad_norm": 0.48113495111465454, "learning_rate": 9.973791907827931e-05, "loss": 0.5953, "step": 3240 }, { "epoch": 1.226878067195168, "grad_norm": 0.4420939087867737, "learning_rate": 9.973367525738683e-05, "loss": 0.6002, "step": 3250 }, { "epoch": 1.2306530766326915, "grad_norm": 0.610998272895813, "learning_rate": 9.972939744399627e-05, "loss": 0.5944, "step": 3260 }, { "epoch": 1.2344280860702153, "grad_norm": 0.5040599703788757, "learning_rate": 9.97250856410315e-05, "loss": 0.5988, "step": 3270 }, { "epoch": 1.2382030955077388, "grad_norm": 0.5534839630126953, "learning_rate": 9.972073985143955e-05, "loss": 0.5982, "step": 3280 }, { "epoch": 1.2419781049452623, "grad_norm": 0.5562023520469666, "learning_rate": 9.971636007819074e-05, "loss": 0.6003, "step": 3290 }, { "epoch": 1.245753114382786, "grad_norm": 0.6595808863639832, "learning_rate": 9.971194632427863e-05, "loss": 0.602, "step": 3300 }, { "epoch": 1.2495281238203095, "grad_norm": 0.7071061134338379, "learning_rate": 9.970749859271997e-05, "loss": 0.6124, "step": 3310 }, { "epoch": 1.253303133257833, "grad_norm": 0.7782847285270691, "learning_rate": 9.970301688655473e-05, "loss": 0.5991, "step": 3320 }, { "epoch": 1.2570781426953568, "grad_norm": 0.7165196537971497, "learning_rate": 9.969850120884612e-05, "loss": 0.5957, "step": 3330 }, { "epoch": 1.2608531521328803, "grad_norm": 0.6089136004447937, "learning_rate": 9.969395156268055e-05, "loss": 0.5891, "step": 3340 }, { "epoch": 1.2646281615704038, "grad_norm": 0.586713969707489, "learning_rate": 9.968936795116768e-05, "loss": 0.594, "step": 3350 }, { "epoch": 1.2684031710079275, "grad_norm": 0.6137309074401855, "learning_rate": 9.968475037744036e-05, "loss": 0.6046, "step": 3360 }, { "epoch": 1.272178180445451, "grad_norm": 0.5665665864944458, "learning_rate": 9.968009884465465e-05, "loss": 0.5905, "step": 3370 }, { "epoch": 1.2759531898829746, "grad_norm": 0.5130990743637085, "learning_rate": 9.967541335598984e-05, "loss": 0.5971, "step": 3380 }, { "epoch": 1.2797281993204983, "grad_norm": 0.6189398765563965, "learning_rate": 9.967069391464841e-05, "loss": 0.5988, "step": 3390 }, { "epoch": 1.2835032087580218, "grad_norm": 0.5427692532539368, "learning_rate": 9.966594052385608e-05, "loss": 0.5979, "step": 3400 }, { "epoch": 1.2872782181955456, "grad_norm": 0.5762437582015991, "learning_rate": 9.96611531868617e-05, "loss": 0.5932, "step": 3410 }, { "epoch": 1.291053227633069, "grad_norm": 0.5316555500030518, "learning_rate": 9.96563319069374e-05, "loss": 0.5962, "step": 3420 }, { "epoch": 1.2948282370705928, "grad_norm": 0.52821284532547, "learning_rate": 9.965147668737847e-05, "loss": 0.5899, "step": 3430 }, { "epoch": 1.2986032465081163, "grad_norm": 0.6211134195327759, "learning_rate": 9.96465875315034e-05, "loss": 0.595, "step": 3440 }, { "epoch": 1.3023782559456398, "grad_norm": 0.4804578125476837, "learning_rate": 9.96416644426539e-05, "loss": 0.5943, "step": 3450 }, { "epoch": 1.3061532653831636, "grad_norm": 0.520340621471405, "learning_rate": 9.963670742419485e-05, "loss": 0.585, "step": 3460 }, { "epoch": 1.309928274820687, "grad_norm": 0.5734260678291321, "learning_rate": 9.96317164795143e-05, "loss": 0.5853, "step": 3470 }, { "epoch": 1.3137032842582106, "grad_norm": 0.6706385612487793, "learning_rate": 9.962669161202356e-05, "loss": 0.6087, "step": 3480 }, { "epoch": 1.3174782936957343, "grad_norm": 0.5032151341438293, "learning_rate": 9.962163282515705e-05, "loss": 0.5877, "step": 3490 }, { "epoch": 1.3212533031332578, "grad_norm": 0.5707389116287231, "learning_rate": 9.961654012237241e-05, "loss": 0.5801, "step": 3500 }, { "epoch": 1.3250283125707814, "grad_norm": 0.5376269817352295, "learning_rate": 9.961141350715044e-05, "loss": 0.5817, "step": 3510 }, { "epoch": 1.328803322008305, "grad_norm": 0.4683854579925537, "learning_rate": 9.960625298299516e-05, "loss": 0.5847, "step": 3520 }, { "epoch": 1.3325783314458286, "grad_norm": 0.4962872266769409, "learning_rate": 9.960105855343372e-05, "loss": 0.5854, "step": 3530 }, { "epoch": 1.3363533408833521, "grad_norm": 0.4632238745689392, "learning_rate": 9.959583022201647e-05, "loss": 0.5844, "step": 3540 }, { "epoch": 1.3401283503208758, "grad_norm": 0.49450498819351196, "learning_rate": 9.959056799231692e-05, "loss": 0.5877, "step": 3550 }, { "epoch": 1.3439033597583994, "grad_norm": 0.45359042286872864, "learning_rate": 9.958527186793176e-05, "loss": 0.5798, "step": 3560 }, { "epoch": 1.3476783691959229, "grad_norm": 0.48816585540771484, "learning_rate": 9.957994185248086e-05, "loss": 0.587, "step": 3570 }, { "epoch": 1.3514533786334466, "grad_norm": 0.5069116353988647, "learning_rate": 9.957457794960718e-05, "loss": 0.5972, "step": 3580 }, { "epoch": 1.3552283880709701, "grad_norm": 0.4354189336299896, "learning_rate": 9.956918016297694e-05, "loss": 0.5872, "step": 3590 }, { "epoch": 1.3590033975084936, "grad_norm": 0.5266293883323669, "learning_rate": 9.956374849627948e-05, "loss": 0.5869, "step": 3600 }, { "epoch": 1.3627784069460174, "grad_norm": 0.4389692544937134, "learning_rate": 9.955828295322728e-05, "loss": 0.5792, "step": 3610 }, { "epoch": 1.3665534163835409, "grad_norm": 0.5600664615631104, "learning_rate": 9.955278353755598e-05, "loss": 0.582, "step": 3620 }, { "epoch": 1.3703284258210646, "grad_norm": 0.6405794024467468, "learning_rate": 9.954725025302439e-05, "loss": 0.5898, "step": 3630 }, { "epoch": 1.3741034352585881, "grad_norm": 0.5375349521636963, "learning_rate": 9.954168310341445e-05, "loss": 0.5863, "step": 3640 }, { "epoch": 1.3778784446961119, "grad_norm": 0.6790446043014526, "learning_rate": 9.953608209253126e-05, "loss": 0.5904, "step": 3650 }, { "epoch": 1.3816534541336354, "grad_norm": 0.4370107352733612, "learning_rate": 9.953044722420307e-05, "loss": 0.5874, "step": 3660 }, { "epoch": 1.385428463571159, "grad_norm": 0.524257242679596, "learning_rate": 9.952477850228124e-05, "loss": 0.5879, "step": 3670 }, { "epoch": 1.3892034730086826, "grad_norm": 0.4980759918689728, "learning_rate": 9.95190759306403e-05, "loss": 0.5817, "step": 3680 }, { "epoch": 1.3929784824462061, "grad_norm": 0.5062684416770935, "learning_rate": 9.951333951317789e-05, "loss": 0.5827, "step": 3690 }, { "epoch": 1.3967534918837297, "grad_norm": 0.7371021509170532, "learning_rate": 9.950756925381479e-05, "loss": 0.5829, "step": 3700 }, { "epoch": 1.4005285013212534, "grad_norm": 0.5569843649864197, "learning_rate": 9.950176515649496e-05, "loss": 0.5779, "step": 3710 }, { "epoch": 1.404303510758777, "grad_norm": 0.5766958594322205, "learning_rate": 9.949592722518542e-05, "loss": 0.5861, "step": 3720 }, { "epoch": 1.4080785201963004, "grad_norm": 0.5430606603622437, "learning_rate": 9.949005546387631e-05, "loss": 0.579, "step": 3730 }, { "epoch": 1.4118535296338242, "grad_norm": 0.49899691343307495, "learning_rate": 9.948414987658098e-05, "loss": 0.5759, "step": 3740 }, { "epoch": 1.4156285390713477, "grad_norm": 0.54237300157547, "learning_rate": 9.94782104673358e-05, "loss": 0.5809, "step": 3750 }, { "epoch": 1.4194035485088712, "grad_norm": 0.6004769802093506, "learning_rate": 9.947223724020034e-05, "loss": 0.5743, "step": 3760 }, { "epoch": 1.423178557946395, "grad_norm": 0.5146884322166443, "learning_rate": 9.94662301992572e-05, "loss": 0.5782, "step": 3770 }, { "epoch": 1.4269535673839184, "grad_norm": 0.5596660375595093, "learning_rate": 9.946018934861216e-05, "loss": 0.5787, "step": 3780 }, { "epoch": 1.430728576821442, "grad_norm": 0.5646491050720215, "learning_rate": 9.945411469239409e-05, "loss": 0.5871, "step": 3790 }, { "epoch": 1.4345035862589657, "grad_norm": 0.6031035780906677, "learning_rate": 9.944800623475497e-05, "loss": 0.5735, "step": 3800 }, { "epoch": 1.4382785956964892, "grad_norm": 0.6148284077644348, "learning_rate": 9.944186397986984e-05, "loss": 0.5807, "step": 3810 }, { "epoch": 1.4420536051340127, "grad_norm": 0.541500449180603, "learning_rate": 9.943568793193687e-05, "loss": 0.5806, "step": 3820 }, { "epoch": 1.4458286145715364, "grad_norm": 0.5454701781272888, "learning_rate": 9.942947809517737e-05, "loss": 0.584, "step": 3830 }, { "epoch": 1.44960362400906, "grad_norm": 0.6133937835693359, "learning_rate": 9.942323447383568e-05, "loss": 0.5811, "step": 3840 }, { "epoch": 1.4533786334465837, "grad_norm": 0.5456222295761108, "learning_rate": 9.941695707217925e-05, "loss": 0.5747, "step": 3850 }, { "epoch": 1.4571536428841072, "grad_norm": 0.5243522524833679, "learning_rate": 9.941064589449865e-05, "loss": 0.5792, "step": 3860 }, { "epoch": 1.460928652321631, "grad_norm": 0.5548741221427917, "learning_rate": 9.940430094510748e-05, "loss": 0.5756, "step": 3870 }, { "epoch": 1.4647036617591545, "grad_norm": 0.65179044008255, "learning_rate": 9.939792222834248e-05, "loss": 0.5788, "step": 3880 }, { "epoch": 1.468478671196678, "grad_norm": 0.6072384715080261, "learning_rate": 9.939150974856343e-05, "loss": 0.5779, "step": 3890 }, { "epoch": 1.4722536806342017, "grad_norm": 0.6061842441558838, "learning_rate": 9.93850635101532e-05, "loss": 0.5735, "step": 3900 }, { "epoch": 1.4760286900717252, "grad_norm": 0.5786122679710388, "learning_rate": 9.937858351751775e-05, "loss": 0.5794, "step": 3910 }, { "epoch": 1.4798036995092487, "grad_norm": 0.5088226199150085, "learning_rate": 9.937206977508604e-05, "loss": 0.5737, "step": 3920 }, { "epoch": 1.4835787089467725, "grad_norm": 0.5596620440483093, "learning_rate": 9.936552228731022e-05, "loss": 0.5912, "step": 3930 }, { "epoch": 1.487353718384296, "grad_norm": 0.4660915732383728, "learning_rate": 9.93589410586654e-05, "loss": 0.5735, "step": 3940 }, { "epoch": 1.4911287278218195, "grad_norm": 0.4994395971298218, "learning_rate": 9.93523260936498e-05, "loss": 0.5807, "step": 3950 }, { "epoch": 1.4949037372593432, "grad_norm": 0.4547739624977112, "learning_rate": 9.934567739678467e-05, "loss": 0.5723, "step": 3960 }, { "epoch": 1.4986787466968667, "grad_norm": 0.4519752264022827, "learning_rate": 9.933899497261433e-05, "loss": 0.5749, "step": 3970 }, { "epoch": 1.5024537561343903, "grad_norm": 0.49861350655555725, "learning_rate": 9.933227882570617e-05, "loss": 0.5825, "step": 3980 }, { "epoch": 1.506228765571914, "grad_norm": 0.5241252183914185, "learning_rate": 9.932552896065062e-05, "loss": 0.582, "step": 3990 }, { "epoch": 1.5100037750094375, "grad_norm": 0.6853564977645874, "learning_rate": 9.931874538206114e-05, "loss": 0.5793, "step": 4000 }, { "epoch": 1.513778784446961, "grad_norm": 0.47244563698768616, "learning_rate": 9.931192809457423e-05, "loss": 0.5737, "step": 4010 }, { "epoch": 1.5175537938844847, "grad_norm": 0.5106404423713684, "learning_rate": 9.930507710284944e-05, "loss": 0.5773, "step": 4020 }, { "epoch": 1.5213288033220083, "grad_norm": 0.5653221011161804, "learning_rate": 9.92981924115694e-05, "loss": 0.5729, "step": 4030 }, { "epoch": 1.5251038127595318, "grad_norm": 0.7388667464256287, "learning_rate": 9.929127402543968e-05, "loss": 0.5695, "step": 4040 }, { "epoch": 1.5288788221970555, "grad_norm": 0.55259770154953, "learning_rate": 9.928432194918895e-05, "loss": 0.5715, "step": 4050 }, { "epoch": 1.5326538316345792, "grad_norm": 0.5290758609771729, "learning_rate": 9.927733618756889e-05, "loss": 0.5753, "step": 4060 }, { "epoch": 1.5364288410721025, "grad_norm": 0.48387640714645386, "learning_rate": 9.92703167453542e-05, "loss": 0.5744, "step": 4070 }, { "epoch": 1.5402038505096263, "grad_norm": 0.5005270838737488, "learning_rate": 9.926326362734263e-05, "loss": 0.5767, "step": 4080 }, { "epoch": 1.54397885994715, "grad_norm": 0.5427155494689941, "learning_rate": 9.925617683835489e-05, "loss": 0.5726, "step": 4090 }, { "epoch": 1.5477538693846733, "grad_norm": 0.5388875603675842, "learning_rate": 9.924905638323472e-05, "loss": 0.5673, "step": 4100 }, { "epoch": 1.551528878822197, "grad_norm": 0.48597651720046997, "learning_rate": 9.92419022668489e-05, "loss": 0.5684, "step": 4110 }, { "epoch": 1.5553038882597208, "grad_norm": 0.5301367044448853, "learning_rate": 9.923471449408723e-05, "loss": 0.5752, "step": 4120 }, { "epoch": 1.5590788976972443, "grad_norm": 0.5001996159553528, "learning_rate": 9.922749306986244e-05, "loss": 0.5629, "step": 4130 }, { "epoch": 1.5628539071347678, "grad_norm": 0.5677825212478638, "learning_rate": 9.922023799911033e-05, "loss": 0.5742, "step": 4140 }, { "epoch": 1.5666289165722915, "grad_norm": 0.5715368390083313, "learning_rate": 9.921294928678968e-05, "loss": 0.5751, "step": 4150 }, { "epoch": 1.570403926009815, "grad_norm": 0.632077157497406, "learning_rate": 9.920562693788222e-05, "loss": 0.5702, "step": 4160 }, { "epoch": 1.5741789354473386, "grad_norm": 0.5607465505599976, "learning_rate": 9.919827095739273e-05, "loss": 0.5713, "step": 4170 }, { "epoch": 1.5779539448848623, "grad_norm": 0.5231792330741882, "learning_rate": 9.919088135034898e-05, "loss": 0.5696, "step": 4180 }, { "epoch": 1.5817289543223858, "grad_norm": 0.46094974875450134, "learning_rate": 9.918345812180165e-05, "loss": 0.5599, "step": 4190 }, { "epoch": 1.5855039637599093, "grad_norm": 0.5033226609230042, "learning_rate": 9.917600127682446e-05, "loss": 0.5706, "step": 4200 }, { "epoch": 1.589278973197433, "grad_norm": 0.5279197096824646, "learning_rate": 9.91685108205141e-05, "loss": 0.5674, "step": 4210 }, { "epoch": 1.5930539826349566, "grad_norm": 0.49496957659721375, "learning_rate": 9.916098675799024e-05, "loss": 0.5644, "step": 4220 }, { "epoch": 1.59682899207248, "grad_norm": 0.5188071131706238, "learning_rate": 9.915342909439548e-05, "loss": 0.5699, "step": 4230 }, { "epoch": 1.6006040015100038, "grad_norm": 0.6602573394775391, "learning_rate": 9.914583783489543e-05, "loss": 0.5699, "step": 4240 }, { "epoch": 1.6043790109475273, "grad_norm": 0.44954216480255127, "learning_rate": 9.913821298467863e-05, "loss": 0.5622, "step": 4250 }, { "epoch": 1.6081540203850508, "grad_norm": 0.40007320046424866, "learning_rate": 9.913055454895661e-05, "loss": 0.5649, "step": 4260 }, { "epoch": 1.6119290298225746, "grad_norm": 0.5066043734550476, "learning_rate": 9.912286253296381e-05, "loss": 0.5689, "step": 4270 }, { "epoch": 1.6157040392600983, "grad_norm": 0.5492740273475647, "learning_rate": 9.911513694195768e-05, "loss": 0.5592, "step": 4280 }, { "epoch": 1.6194790486976216, "grad_norm": 0.40257328748703003, "learning_rate": 9.910737778121859e-05, "loss": 0.5671, "step": 4290 }, { "epoch": 1.6232540581351453, "grad_norm": 0.42792394757270813, "learning_rate": 9.909958505604984e-05, "loss": 0.5624, "step": 4300 }, { "epoch": 1.627029067572669, "grad_norm": 0.5710054039955139, "learning_rate": 9.909175877177767e-05, "loss": 0.5697, "step": 4310 }, { "epoch": 1.6308040770101924, "grad_norm": 0.47160592675209045, "learning_rate": 9.908389893375129e-05, "loss": 0.5684, "step": 4320 }, { "epoch": 1.634579086447716, "grad_norm": 0.6177746653556824, "learning_rate": 9.907600554734283e-05, "loss": 0.5724, "step": 4330 }, { "epoch": 1.6383540958852398, "grad_norm": 0.56391841173172, "learning_rate": 9.906807861794734e-05, "loss": 0.5675, "step": 4340 }, { "epoch": 1.6421291053227633, "grad_norm": 0.6061177849769592, "learning_rate": 9.906011815098279e-05, "loss": 0.5607, "step": 4350 }, { "epoch": 1.6459041147602869, "grad_norm": 0.580389678478241, "learning_rate": 9.90521241518901e-05, "loss": 0.563, "step": 4360 }, { "epoch": 1.6496791241978106, "grad_norm": 0.6616781949996948, "learning_rate": 9.904409662613308e-05, "loss": 0.5664, "step": 4370 }, { "epoch": 1.6534541336353341, "grad_norm": 0.5005329847335815, "learning_rate": 9.903603557919849e-05, "loss": 0.5592, "step": 4380 }, { "epoch": 1.6572291430728576, "grad_norm": 0.4905034601688385, "learning_rate": 9.902794101659594e-05, "loss": 0.5678, "step": 4390 }, { "epoch": 1.6610041525103814, "grad_norm": 0.44465407729148865, "learning_rate": 9.901981294385803e-05, "loss": 0.5684, "step": 4400 }, { "epoch": 1.6647791619479049, "grad_norm": 0.5083572864532471, "learning_rate": 9.901165136654018e-05, "loss": 0.561, "step": 4410 }, { "epoch": 1.6685541713854284, "grad_norm": 0.5734196305274963, "learning_rate": 9.900345629022079e-05, "loss": 0.57, "step": 4420 }, { "epoch": 1.6723291808229521, "grad_norm": 0.568246066570282, "learning_rate": 9.89952277205011e-05, "loss": 0.5666, "step": 4430 }, { "epoch": 1.6761041902604756, "grad_norm": 0.5339452624320984, "learning_rate": 9.898696566300527e-05, "loss": 0.5599, "step": 4440 }, { "epoch": 1.6798791996979991, "grad_norm": 0.5378457307815552, "learning_rate": 9.897867012338032e-05, "loss": 0.5605, "step": 4450 }, { "epoch": 1.6836542091355229, "grad_norm": 0.5373885035514832, "learning_rate": 9.897034110729617e-05, "loss": 0.5641, "step": 4460 }, { "epoch": 1.6874292185730464, "grad_norm": 1.103470802307129, "learning_rate": 9.896197862044564e-05, "loss": 0.5773, "step": 4470 }, { "epoch": 1.69120422801057, "grad_norm": 0.6465436220169067, "learning_rate": 9.895358266854442e-05, "loss": 0.5764, "step": 4480 }, { "epoch": 1.6949792374480936, "grad_norm": 0.5247374176979065, "learning_rate": 9.894515325733103e-05, "loss": 0.5593, "step": 4490 }, { "epoch": 1.6987542468856174, "grad_norm": 0.589596688747406, "learning_rate": 9.893669039256693e-05, "loss": 0.5637, "step": 4500 }, { "epoch": 1.7025292563231407, "grad_norm": 0.4953310489654541, "learning_rate": 9.89281940800364e-05, "loss": 0.5622, "step": 4510 }, { "epoch": 1.7063042657606644, "grad_norm": 0.45567750930786133, "learning_rate": 9.891966432554655e-05, "loss": 0.5594, "step": 4520 }, { "epoch": 1.7100792751981881, "grad_norm": 0.4500698149204254, "learning_rate": 9.891110113492745e-05, "loss": 0.5619, "step": 4530 }, { "epoch": 1.7138542846357114, "grad_norm": 0.4617537260055542, "learning_rate": 9.89025045140319e-05, "loss": 0.5623, "step": 4540 }, { "epoch": 1.7176292940732352, "grad_norm": 0.5609983801841736, "learning_rate": 9.889387446873567e-05, "loss": 0.5629, "step": 4550 }, { "epoch": 1.721404303510759, "grad_norm": 0.6847432851791382, "learning_rate": 9.888521100493726e-05, "loss": 0.57, "step": 4560 }, { "epoch": 1.7251793129482824, "grad_norm": 0.50513756275177, "learning_rate": 9.887651412855809e-05, "loss": 0.5586, "step": 4570 }, { "epoch": 1.728954322385806, "grad_norm": 0.45217183232307434, "learning_rate": 9.88677838455424e-05, "loss": 0.5493, "step": 4580 }, { "epoch": 1.7327293318233297, "grad_norm": 0.445190966129303, "learning_rate": 9.885902016185725e-05, "loss": 0.5597, "step": 4590 }, { "epoch": 1.7365043412608532, "grad_norm": 0.49329379200935364, "learning_rate": 9.885022308349252e-05, "loss": 0.5608, "step": 4600 }, { "epoch": 1.7402793506983767, "grad_norm": 0.531954824924469, "learning_rate": 9.884139261646095e-05, "loss": 0.5646, "step": 4610 }, { "epoch": 1.7440543601359004, "grad_norm": 0.5033367276191711, "learning_rate": 9.883252876679807e-05, "loss": 0.564, "step": 4620 }, { "epoch": 1.747829369573424, "grad_norm": 0.8055509328842163, "learning_rate": 9.882363154056225e-05, "loss": 0.5519, "step": 4630 }, { "epoch": 1.7516043790109475, "grad_norm": 0.592330813407898, "learning_rate": 9.881470094383465e-05, "loss": 0.5565, "step": 4640 }, { "epoch": 1.7553793884484712, "grad_norm": 0.605402946472168, "learning_rate": 9.880573698271924e-05, "loss": 0.5737, "step": 4650 }, { "epoch": 1.7591543978859947, "grad_norm": 0.6107518076896667, "learning_rate": 9.879673966334282e-05, "loss": 0.5569, "step": 4660 }, { "epoch": 1.7629294073235182, "grad_norm": 0.5053714513778687, "learning_rate": 9.878770899185496e-05, "loss": 0.5599, "step": 4670 }, { "epoch": 1.766704416761042, "grad_norm": 0.4542829394340515, "learning_rate": 9.877864497442804e-05, "loss": 0.5606, "step": 4680 }, { "epoch": 1.7704794261985655, "grad_norm": 0.6275748610496521, "learning_rate": 9.876954761725723e-05, "loss": 0.5576, "step": 4690 }, { "epoch": 1.774254435636089, "grad_norm": 0.5331715941429138, "learning_rate": 9.876041692656052e-05, "loss": 0.5567, "step": 4700 }, { "epoch": 1.7780294450736127, "grad_norm": 0.49501025676727295, "learning_rate": 9.87512529085786e-05, "loss": 0.5617, "step": 4710 }, { "epoch": 1.7818044545111362, "grad_norm": 0.4898965060710907, "learning_rate": 9.874205556957503e-05, "loss": 0.5482, "step": 4720 }, { "epoch": 1.7855794639486597, "grad_norm": 0.47128167748451233, "learning_rate": 9.873282491583608e-05, "loss": 0.5579, "step": 4730 }, { "epoch": 1.7893544733861835, "grad_norm": 0.4385862946510315, "learning_rate": 9.872356095367084e-05, "loss": 0.5551, "step": 4740 }, { "epoch": 1.7931294828237072, "grad_norm": 0.5858024954795837, "learning_rate": 9.87142636894111e-05, "loss": 0.5474, "step": 4750 }, { "epoch": 1.7969044922612305, "grad_norm": 0.4746013283729553, "learning_rate": 9.870493312941148e-05, "loss": 0.5596, "step": 4760 }, { "epoch": 1.8006795016987542, "grad_norm": 0.5050995945930481, "learning_rate": 9.869556928004933e-05, "loss": 0.5451, "step": 4770 }, { "epoch": 1.804454511136278, "grad_norm": 0.6011149287223816, "learning_rate": 9.868617214772476e-05, "loss": 0.5543, "step": 4780 }, { "epoch": 1.8082295205738015, "grad_norm": 0.5801728367805481, "learning_rate": 9.867674173886059e-05, "loss": 0.549, "step": 4790 }, { "epoch": 1.812004530011325, "grad_norm": 0.5891419053077698, "learning_rate": 9.866727805990245e-05, "loss": 0.5539, "step": 4800 }, { "epoch": 1.8157795394488487, "grad_norm": 0.538972795009613, "learning_rate": 9.865778111731865e-05, "loss": 0.5506, "step": 4810 }, { "epoch": 1.8195545488863722, "grad_norm": 0.5291750431060791, "learning_rate": 9.864825091760028e-05, "loss": 0.5576, "step": 4820 }, { "epoch": 1.8233295583238958, "grad_norm": 0.5145063400268555, "learning_rate": 9.863868746726112e-05, "loss": 0.5523, "step": 4830 }, { "epoch": 1.8271045677614195, "grad_norm": 0.4349626898765564, "learning_rate": 9.862909077283773e-05, "loss": 0.5525, "step": 4840 }, { "epoch": 1.830879577198943, "grad_norm": 0.5479471683502197, "learning_rate": 9.861946084088933e-05, "loss": 0.5528, "step": 4850 }, { "epoch": 1.8346545866364665, "grad_norm": 0.7466859817504883, "learning_rate": 9.860979767799792e-05, "loss": 0.5539, "step": 4860 }, { "epoch": 1.8384295960739903, "grad_norm": 0.5633376836776733, "learning_rate": 9.860010129076813e-05, "loss": 0.5538, "step": 4870 }, { "epoch": 1.8422046055115138, "grad_norm": 0.49453553557395935, "learning_rate": 9.859037168582741e-05, "loss": 0.5486, "step": 4880 }, { "epoch": 1.8459796149490373, "grad_norm": 0.5297577381134033, "learning_rate": 9.85806088698258e-05, "loss": 0.5394, "step": 4890 }, { "epoch": 1.849754624386561, "grad_norm": 0.5536269545555115, "learning_rate": 9.857081284943612e-05, "loss": 0.5494, "step": 4900 }, { "epoch": 1.8535296338240845, "grad_norm": 0.4998897612094879, "learning_rate": 9.856098363135385e-05, "loss": 0.5459, "step": 4910 }, { "epoch": 1.857304643261608, "grad_norm": 0.48842960596084595, "learning_rate": 9.855112122229717e-05, "loss": 0.5523, "step": 4920 }, { "epoch": 1.8610796526991318, "grad_norm": 0.4732976257801056, "learning_rate": 9.854122562900696e-05, "loss": 0.568, "step": 4930 }, { "epoch": 1.8648546621366553, "grad_norm": 0.47209542989730835, "learning_rate": 9.853129685824673e-05, "loss": 0.5526, "step": 4940 }, { "epoch": 1.8686296715741788, "grad_norm": 0.4802228510379791, "learning_rate": 9.852133491680273e-05, "loss": 0.5431, "step": 4950 }, { "epoch": 1.8724046810117025, "grad_norm": 0.507754385471344, "learning_rate": 9.851133981148385e-05, "loss": 0.5486, "step": 4960 }, { "epoch": 1.8761796904492263, "grad_norm": 0.5926216244697571, "learning_rate": 9.850131154912164e-05, "loss": 0.549, "step": 4970 }, { "epoch": 1.8799546998867496, "grad_norm": 0.49967560172080994, "learning_rate": 9.849125013657031e-05, "loss": 0.5459, "step": 4980 }, { "epoch": 1.8837297093242733, "grad_norm": 0.5526815056800842, "learning_rate": 9.848115558070676e-05, "loss": 0.5439, "step": 4990 }, { "epoch": 1.887504718761797, "grad_norm": 0.5583565831184387, "learning_rate": 9.847102788843055e-05, "loss": 0.5418, "step": 5000 }, { "epoch": 1.8912797281993206, "grad_norm": 0.5497617125511169, "learning_rate": 9.84608670666638e-05, "loss": 0.5473, "step": 5010 }, { "epoch": 1.895054737636844, "grad_norm": 0.6256689429283142, "learning_rate": 9.845067312235138e-05, "loss": 0.5485, "step": 5020 }, { "epoch": 1.8988297470743678, "grad_norm": 0.5378864407539368, "learning_rate": 9.844044606246074e-05, "loss": 0.5432, "step": 5030 }, { "epoch": 1.9026047565118913, "grad_norm": 0.5322272777557373, "learning_rate": 9.843018589398199e-05, "loss": 0.5605, "step": 5040 }, { "epoch": 1.9063797659494148, "grad_norm": 0.47538843750953674, "learning_rate": 9.841989262392785e-05, "loss": 0.5452, "step": 5050 }, { "epoch": 1.9101547753869386, "grad_norm": 0.5338093042373657, "learning_rate": 9.840956625933367e-05, "loss": 0.5443, "step": 5060 }, { "epoch": 1.913929784824462, "grad_norm": 0.5679603815078735, "learning_rate": 9.839920680725743e-05, "loss": 0.5463, "step": 5070 }, { "epoch": 1.9177047942619856, "grad_norm": 0.5604372024536133, "learning_rate": 9.83888142747797e-05, "loss": 0.5508, "step": 5080 }, { "epoch": 1.9214798036995093, "grad_norm": 0.6041125655174255, "learning_rate": 9.83783886690037e-05, "loss": 0.5441, "step": 5090 }, { "epoch": 1.9252548131370328, "grad_norm": 0.6924186944961548, "learning_rate": 9.836792999705524e-05, "loss": 0.546, "step": 5100 }, { "epoch": 1.9290298225745564, "grad_norm": 0.4839281737804413, "learning_rate": 9.835743826608271e-05, "loss": 0.5505, "step": 5110 }, { "epoch": 1.93280483201208, "grad_norm": 0.49724170565605164, "learning_rate": 9.834691348325709e-05, "loss": 0.5384, "step": 5120 }, { "epoch": 1.9365798414496036, "grad_norm": 0.5565958619117737, "learning_rate": 9.833635565577199e-05, "loss": 0.5432, "step": 5130 }, { "epoch": 1.9403548508871271, "grad_norm": 0.5241300463676453, "learning_rate": 9.832576479084359e-05, "loss": 0.5466, "step": 5140 }, { "epoch": 1.9441298603246508, "grad_norm": 0.4891330599784851, "learning_rate": 9.831514089571064e-05, "loss": 0.5466, "step": 5150 }, { "epoch": 1.9479048697621744, "grad_norm": 0.582870364189148, "learning_rate": 9.830448397763447e-05, "loss": 0.5394, "step": 5160 }, { "epoch": 1.9516798791996979, "grad_norm": 0.5557212829589844, "learning_rate": 9.8293794043899e-05, "loss": 0.5426, "step": 5170 }, { "epoch": 1.9554548886372216, "grad_norm": 0.5366427302360535, "learning_rate": 9.828307110181066e-05, "loss": 0.5488, "step": 5180 }, { "epoch": 1.9592298980747453, "grad_norm": 0.512102484703064, "learning_rate": 9.827231515869852e-05, "loss": 0.5463, "step": 5190 }, { "epoch": 1.9630049075122686, "grad_norm": 0.4691350758075714, "learning_rate": 9.826152622191413e-05, "loss": 0.5435, "step": 5200 }, { "epoch": 1.9667799169497924, "grad_norm": 0.5275509357452393, "learning_rate": 9.825070429883167e-05, "loss": 0.5343, "step": 5210 }, { "epoch": 1.970554926387316, "grad_norm": 0.4976699948310852, "learning_rate": 9.82398493968478e-05, "loss": 0.5423, "step": 5220 }, { "epoch": 1.9743299358248394, "grad_norm": 0.5303046107292175, "learning_rate": 9.822896152338173e-05, "loss": 0.5428, "step": 5230 }, { "epoch": 1.9781049452623631, "grad_norm": 0.47112399339675903, "learning_rate": 9.821804068587523e-05, "loss": 0.5357, "step": 5240 }, { "epoch": 1.9818799546998869, "grad_norm": 0.4589209258556366, "learning_rate": 9.820708689179259e-05, "loss": 0.5445, "step": 5250 }, { "epoch": 1.9856549641374104, "grad_norm": 0.47271859645843506, "learning_rate": 9.819610014862063e-05, "loss": 0.5339, "step": 5260 }, { "epoch": 1.989429973574934, "grad_norm": 0.45445945858955383, "learning_rate": 9.818508046386868e-05, "loss": 0.5405, "step": 5270 }, { "epoch": 1.9932049830124576, "grad_norm": 0.5319023728370667, "learning_rate": 9.817402784506859e-05, "loss": 0.5421, "step": 5280 }, { "epoch": 1.9969799924499811, "grad_norm": 0.5033361315727234, "learning_rate": 9.816294229977472e-05, "loss": 0.5405, "step": 5290 }, { "epoch": 2.0007550018875047, "grad_norm": 0.5742616057395935, "learning_rate": 9.815182383556394e-05, "loss": 0.5346, "step": 5300 }, { "epoch": 2.0045300113250284, "grad_norm": 0.591316819190979, "learning_rate": 9.81406724600356e-05, "loss": 0.5396, "step": 5310 }, { "epoch": 2.0083050207625517, "grad_norm": 0.4500562846660614, "learning_rate": 9.812948818081161e-05, "loss": 0.5375, "step": 5320 }, { "epoch": 2.0120800302000754, "grad_norm": 0.5774345993995667, "learning_rate": 9.811827100553623e-05, "loss": 0.5345, "step": 5330 }, { "epoch": 2.015855039637599, "grad_norm": 0.627068042755127, "learning_rate": 9.810702094187638e-05, "loss": 0.5432, "step": 5340 }, { "epoch": 2.019630049075123, "grad_norm": 0.5072972178459167, "learning_rate": 9.809573799752135e-05, "loss": 0.5488, "step": 5350 }, { "epoch": 2.023405058512646, "grad_norm": 0.5516794919967651, "learning_rate": 9.808442218018288e-05, "loss": 0.5383, "step": 5360 }, { "epoch": 2.02718006795017, "grad_norm": 0.5118292570114136, "learning_rate": 9.807307349759527e-05, "loss": 0.539, "step": 5370 }, { "epoch": 2.0309550773876937, "grad_norm": 0.5869454145431519, "learning_rate": 9.806169195751525e-05, "loss": 0.5374, "step": 5380 }, { "epoch": 2.034730086825217, "grad_norm": 0.46649426221847534, "learning_rate": 9.805027756772194e-05, "loss": 0.5426, "step": 5390 }, { "epoch": 2.0385050962627407, "grad_norm": 0.5441772937774658, "learning_rate": 9.803883033601702e-05, "loss": 0.5428, "step": 5400 }, { "epoch": 2.0422801057002644, "grad_norm": 0.4856852889060974, "learning_rate": 9.802735027022453e-05, "loss": 0.5407, "step": 5410 }, { "epoch": 2.0460551151377877, "grad_norm": 0.9828983545303345, "learning_rate": 9.8015837378191e-05, "loss": 0.5436, "step": 5420 }, { "epoch": 2.0498301245753114, "grad_norm": 0.5788944959640503, "learning_rate": 9.80042916677854e-05, "loss": 0.54, "step": 5430 }, { "epoch": 2.053605134012835, "grad_norm": 0.5199455618858337, "learning_rate": 9.799271314689908e-05, "loss": 0.5301, "step": 5440 }, { "epoch": 2.0573801434503585, "grad_norm": 0.5322635173797607, "learning_rate": 9.798110182344588e-05, "loss": 0.5441, "step": 5450 }, { "epoch": 2.061155152887882, "grad_norm": 0.4913989305496216, "learning_rate": 9.796945770536204e-05, "loss": 0.5466, "step": 5460 }, { "epoch": 2.064930162325406, "grad_norm": 0.5005329847335815, "learning_rate": 9.795778080060615e-05, "loss": 0.5479, "step": 5470 }, { "epoch": 2.0687051717629292, "grad_norm": 0.45034486055374146, "learning_rate": 9.79460711171593e-05, "loss": 0.5365, "step": 5480 }, { "epoch": 2.072480181200453, "grad_norm": 0.530417799949646, "learning_rate": 9.793432866302497e-05, "loss": 0.5343, "step": 5490 }, { "epoch": 2.0762551906379767, "grad_norm": 0.5080109238624573, "learning_rate": 9.792255344622897e-05, "loss": 0.5389, "step": 5500 }, { "epoch": 2.0800302000755, "grad_norm": 0.5648706555366516, "learning_rate": 9.791074547481957e-05, "loss": 0.5375, "step": 5510 }, { "epoch": 2.0838052095130237, "grad_norm": 0.4704912602901459, "learning_rate": 9.789890475686742e-05, "loss": 0.5314, "step": 5520 }, { "epoch": 2.0875802189505475, "grad_norm": 0.47333037853240967, "learning_rate": 9.788703130046552e-05, "loss": 0.5357, "step": 5530 }, { "epoch": 2.091355228388071, "grad_norm": 0.5688567161560059, "learning_rate": 9.78751251137293e-05, "loss": 0.5406, "step": 5540 }, { "epoch": 2.0951302378255945, "grad_norm": 0.5980736613273621, "learning_rate": 9.786318620479646e-05, "loss": 0.541, "step": 5550 }, { "epoch": 2.0989052472631182, "grad_norm": 0.5263574719429016, "learning_rate": 9.78512145818272e-05, "loss": 0.5337, "step": 5560 }, { "epoch": 2.102680256700642, "grad_norm": 0.6168349981307983, "learning_rate": 9.783921025300398e-05, "loss": 0.534, "step": 5570 }, { "epoch": 2.1064552661381652, "grad_norm": 0.5928676128387451, "learning_rate": 9.782717322653164e-05, "loss": 0.5298, "step": 5580 }, { "epoch": 2.110230275575689, "grad_norm": 0.5435881018638611, "learning_rate": 9.781510351063738e-05, "loss": 0.5395, "step": 5590 }, { "epoch": 2.1140052850132127, "grad_norm": 0.5674425363540649, "learning_rate": 9.780300111357075e-05, "loss": 0.5384, "step": 5600 }, { "epoch": 2.117780294450736, "grad_norm": 0.60903400182724, "learning_rate": 9.779086604360361e-05, "loss": 0.5377, "step": 5610 }, { "epoch": 2.1215553038882597, "grad_norm": 0.6011250615119934, "learning_rate": 9.777869830903019e-05, "loss": 0.5371, "step": 5620 }, { "epoch": 2.1253303133257835, "grad_norm": 0.5067744851112366, "learning_rate": 9.776649791816698e-05, "loss": 0.529, "step": 5630 }, { "epoch": 2.1291053227633068, "grad_norm": 0.5736885070800781, "learning_rate": 9.775426487935285e-05, "loss": 0.5298, "step": 5640 }, { "epoch": 2.1328803322008305, "grad_norm": 0.5602228045463562, "learning_rate": 9.774199920094898e-05, "loss": 0.5312, "step": 5650 }, { "epoch": 2.1366553416383542, "grad_norm": 0.5565608739852905, "learning_rate": 9.772970089133884e-05, "loss": 0.531, "step": 5660 }, { "epoch": 2.1404303510758775, "grad_norm": 0.6767264604568481, "learning_rate": 9.77173699589282e-05, "loss": 0.53, "step": 5670 }, { "epoch": 2.1442053605134013, "grad_norm": 0.5440645813941956, "learning_rate": 9.770500641214513e-05, "loss": 0.5317, "step": 5680 }, { "epoch": 2.147980369950925, "grad_norm": 0.6627269387245178, "learning_rate": 9.769261025944003e-05, "loss": 0.5445, "step": 5690 }, { "epoch": 2.1517553793884483, "grad_norm": 1.5312882661819458, "learning_rate": 9.768018150928552e-05, "loss": 0.5425, "step": 5700 }, { "epoch": 2.155530388825972, "grad_norm": 0.621485710144043, "learning_rate": 9.766772017017654e-05, "loss": 0.5347, "step": 5710 }, { "epoch": 2.1593053982634958, "grad_norm": 0.5858162641525269, "learning_rate": 9.765522625063032e-05, "loss": 0.5268, "step": 5720 }, { "epoch": 2.163080407701019, "grad_norm": 0.4868731200695038, "learning_rate": 9.76426997591863e-05, "loss": 0.5361, "step": 5730 }, { "epoch": 2.166855417138543, "grad_norm": 0.5163987874984741, "learning_rate": 9.763014070440625e-05, "loss": 0.5292, "step": 5740 }, { "epoch": 2.1706304265760665, "grad_norm": 0.5586041808128357, "learning_rate": 9.761754909487415e-05, "loss": 0.5264, "step": 5750 }, { "epoch": 2.17440543601359, "grad_norm": 0.45366203784942627, "learning_rate": 9.760492493919626e-05, "loss": 0.5405, "step": 5760 }, { "epoch": 2.1781804454511136, "grad_norm": 0.5266132354736328, "learning_rate": 9.759226824600105e-05, "loss": 0.5403, "step": 5770 }, { "epoch": 2.1819554548886373, "grad_norm": 0.516366183757782, "learning_rate": 9.757957902393928e-05, "loss": 0.5345, "step": 5780 }, { "epoch": 2.185730464326161, "grad_norm": 0.4906608760356903, "learning_rate": 9.756685728168387e-05, "loss": 0.5437, "step": 5790 }, { "epoch": 2.1895054737636843, "grad_norm": 0.5202317237854004, "learning_rate": 9.755410302793004e-05, "loss": 0.5299, "step": 5800 }, { "epoch": 2.193280483201208, "grad_norm": 0.5337672829627991, "learning_rate": 9.754131627139522e-05, "loss": 0.5359, "step": 5810 }, { "epoch": 2.197055492638732, "grad_norm": 0.5173420906066895, "learning_rate": 9.752849702081901e-05, "loss": 0.5345, "step": 5820 }, { "epoch": 2.200830502076255, "grad_norm": 0.4675118625164032, "learning_rate": 9.751564528496324e-05, "loss": 0.531, "step": 5830 }, { "epoch": 2.204605511513779, "grad_norm": 0.48218834400177, "learning_rate": 9.750276107261197e-05, "loss": 0.5279, "step": 5840 }, { "epoch": 2.2083805209513026, "grad_norm": 0.5200138688087463, "learning_rate": 9.748984439257142e-05, "loss": 0.5289, "step": 5850 }, { "epoch": 2.212155530388826, "grad_norm": 0.6025936603546143, "learning_rate": 9.747689525367005e-05, "loss": 0.526, "step": 5860 }, { "epoch": 2.2159305398263496, "grad_norm": 0.4721621572971344, "learning_rate": 9.746391366475845e-05, "loss": 0.535, "step": 5870 }, { "epoch": 2.2197055492638733, "grad_norm": 0.5410889983177185, "learning_rate": 9.745089963470942e-05, "loss": 0.5316, "step": 5880 }, { "epoch": 2.2234805587013966, "grad_norm": 0.5121446251869202, "learning_rate": 9.743785317241791e-05, "loss": 0.5283, "step": 5890 }, { "epoch": 2.2272555681389203, "grad_norm": 0.48920613527297974, "learning_rate": 9.742477428680108e-05, "loss": 0.5278, "step": 5900 }, { "epoch": 2.231030577576444, "grad_norm": 0.47083306312561035, "learning_rate": 9.741166298679821e-05, "loss": 0.5301, "step": 5910 }, { "epoch": 2.2348055870139674, "grad_norm": 0.45593857765197754, "learning_rate": 9.739851928137076e-05, "loss": 0.5278, "step": 5920 }, { "epoch": 2.238580596451491, "grad_norm": 0.5642737746238708, "learning_rate": 9.738534317950232e-05, "loss": 0.5226, "step": 5930 }, { "epoch": 2.242355605889015, "grad_norm": 0.5394318103790283, "learning_rate": 9.737213469019864e-05, "loss": 0.5278, "step": 5940 }, { "epoch": 2.246130615326538, "grad_norm": 0.5335193872451782, "learning_rate": 9.735889382248757e-05, "loss": 0.5239, "step": 5950 }, { "epoch": 2.249905624764062, "grad_norm": 0.5988762974739075, "learning_rate": 9.734562058541916e-05, "loss": 0.5418, "step": 5960 }, { "epoch": 2.2536806342015856, "grad_norm": 0.5146274566650391, "learning_rate": 9.733231498806552e-05, "loss": 0.5384, "step": 5970 }, { "epoch": 2.2574556436391093, "grad_norm": 0.5223647356033325, "learning_rate": 9.731897703952088e-05, "loss": 0.5321, "step": 5980 }, { "epoch": 2.2612306530766326, "grad_norm": 0.5377148985862732, "learning_rate": 9.730560674890165e-05, "loss": 0.5249, "step": 5990 }, { "epoch": 2.2650056625141564, "grad_norm": 0.6117594838142395, "learning_rate": 9.729220412534628e-05, "loss": 0.524, "step": 6000 }, { "epoch": 2.2687806719516797, "grad_norm": 0.612998366355896, "learning_rate": 9.72787691780153e-05, "loss": 0.5314, "step": 6010 }, { "epoch": 2.2725556813892034, "grad_norm": 0.5014058947563171, "learning_rate": 9.726530191609142e-05, "loss": 0.5267, "step": 6020 }, { "epoch": 2.276330690826727, "grad_norm": 0.5796371698379517, "learning_rate": 9.725180234877937e-05, "loss": 0.5323, "step": 6030 }, { "epoch": 2.280105700264251, "grad_norm": 0.6417374610900879, "learning_rate": 9.723827048530595e-05, "loss": 0.5308, "step": 6040 }, { "epoch": 2.283880709701774, "grad_norm": 0.4562576711177826, "learning_rate": 9.722470633492011e-05, "loss": 0.5298, "step": 6050 }, { "epoch": 2.287655719139298, "grad_norm": 0.5561509728431702, "learning_rate": 9.721110990689278e-05, "loss": 0.5265, "step": 6060 }, { "epoch": 2.2914307285768216, "grad_norm": 0.5673072338104248, "learning_rate": 9.719748121051699e-05, "loss": 0.5195, "step": 6070 }, { "epoch": 2.295205738014345, "grad_norm": 0.5953671336174011, "learning_rate": 9.718382025510785e-05, "loss": 0.5248, "step": 6080 }, { "epoch": 2.2989807474518686, "grad_norm": 0.6203503608703613, "learning_rate": 9.717012705000248e-05, "loss": 0.5264, "step": 6090 }, { "epoch": 2.3027557568893924, "grad_norm": 0.5094435811042786, "learning_rate": 9.715640160456006e-05, "loss": 0.5325, "step": 6100 }, { "epoch": 2.3065307663269157, "grad_norm": 0.5434731245040894, "learning_rate": 9.714264392816181e-05, "loss": 0.5243, "step": 6110 }, { "epoch": 2.3103057757644394, "grad_norm": 0.5409674048423767, "learning_rate": 9.712885403021095e-05, "loss": 0.523, "step": 6120 }, { "epoch": 2.314080785201963, "grad_norm": 0.5891657471656799, "learning_rate": 9.711503192013276e-05, "loss": 0.529, "step": 6130 }, { "epoch": 2.3178557946394864, "grad_norm": 0.6148802042007446, "learning_rate": 9.710117760737454e-05, "loss": 0.5234, "step": 6140 }, { "epoch": 2.32163080407701, "grad_norm": 0.5786886215209961, "learning_rate": 9.708729110140554e-05, "loss": 0.524, "step": 6150 }, { "epoch": 2.325405813514534, "grad_norm": 0.5547001361846924, "learning_rate": 9.707337241171709e-05, "loss": 0.5212, "step": 6160 }, { "epoch": 2.329180822952057, "grad_norm": 0.6747357845306396, "learning_rate": 9.705942154782245e-05, "loss": 0.5274, "step": 6170 }, { "epoch": 2.332955832389581, "grad_norm": 0.4640480875968933, "learning_rate": 9.704543851925696e-05, "loss": 0.5254, "step": 6180 }, { "epoch": 2.3367308418271047, "grad_norm": 0.5004093647003174, "learning_rate": 9.703142333557784e-05, "loss": 0.5208, "step": 6190 }, { "epoch": 2.340505851264628, "grad_norm": 0.5276467204093933, "learning_rate": 9.701737600636436e-05, "loss": 0.5205, "step": 6200 }, { "epoch": 2.3442808607021517, "grad_norm": 0.5766600370407104, "learning_rate": 9.700329654121771e-05, "loss": 0.5145, "step": 6210 }, { "epoch": 2.3480558701396754, "grad_norm": 0.4839555025100708, "learning_rate": 9.698918494976109e-05, "loss": 0.5234, "step": 6220 }, { "epoch": 2.351830879577199, "grad_norm": 0.5838026404380798, "learning_rate": 9.697504124163965e-05, "loss": 0.5232, "step": 6230 }, { "epoch": 2.3556058890147225, "grad_norm": 0.553741991519928, "learning_rate": 9.696086542652045e-05, "loss": 0.5223, "step": 6240 }, { "epoch": 2.359380898452246, "grad_norm": 0.48992642760276794, "learning_rate": 9.694665751409256e-05, "loss": 0.521, "step": 6250 }, { "epoch": 2.3631559078897695, "grad_norm": 0.511161208152771, "learning_rate": 9.693241751406694e-05, "loss": 0.5186, "step": 6260 }, { "epoch": 2.366930917327293, "grad_norm": 0.480486124753952, "learning_rate": 9.69181454361765e-05, "loss": 0.517, "step": 6270 }, { "epoch": 2.370705926764817, "grad_norm": 0.49377861618995667, "learning_rate": 9.690384129017605e-05, "loss": 0.5187, "step": 6280 }, { "epoch": 2.3744809362023407, "grad_norm": 0.5575188398361206, "learning_rate": 9.688950508584237e-05, "loss": 0.5194, "step": 6290 }, { "epoch": 2.378255945639864, "grad_norm": 0.501025378704071, "learning_rate": 9.68751368329741e-05, "loss": 0.519, "step": 6300 }, { "epoch": 2.3820309550773877, "grad_norm": 0.5015434622764587, "learning_rate": 9.68607365413918e-05, "loss": 0.5213, "step": 6310 }, { "epoch": 2.3858059645149114, "grad_norm": 0.4865148663520813, "learning_rate": 9.684630422093797e-05, "loss": 0.5212, "step": 6320 }, { "epoch": 2.3895809739524347, "grad_norm": 0.545136034488678, "learning_rate": 9.683183988147693e-05, "loss": 0.5199, "step": 6330 }, { "epoch": 2.3933559833899585, "grad_norm": 0.4947832226753235, "learning_rate": 9.681734353289491e-05, "loss": 0.5217, "step": 6340 }, { "epoch": 2.397130992827482, "grad_norm": 0.554629921913147, "learning_rate": 9.680281518510006e-05, "loss": 0.5204, "step": 6350 }, { "epoch": 2.4009060022650055, "grad_norm": 0.49678510427474976, "learning_rate": 9.678825484802236e-05, "loss": 0.5212, "step": 6360 }, { "epoch": 2.4046810117025292, "grad_norm": 0.5372890830039978, "learning_rate": 9.677366253161365e-05, "loss": 0.5252, "step": 6370 }, { "epoch": 2.408456021140053, "grad_norm": 0.4868631660938263, "learning_rate": 9.675903824584765e-05, "loss": 0.5309, "step": 6380 }, { "epoch": 2.4122310305775763, "grad_norm": 0.5607171654701233, "learning_rate": 9.674438200071991e-05, "loss": 0.5197, "step": 6390 }, { "epoch": 2.4160060400151, "grad_norm": 0.6623253226280212, "learning_rate": 9.672969380624786e-05, "loss": 0.5247, "step": 6400 }, { "epoch": 2.4197810494526237, "grad_norm": 0.5090137720108032, "learning_rate": 9.671497367247073e-05, "loss": 0.5169, "step": 6410 }, { "epoch": 2.4235560588901475, "grad_norm": 0.4954577684402466, "learning_rate": 9.670022160944959e-05, "loss": 0.5109, "step": 6420 }, { "epoch": 2.4273310683276708, "grad_norm": 0.471021831035614, "learning_rate": 9.668543762726734e-05, "loss": 0.5147, "step": 6430 }, { "epoch": 2.4311060777651945, "grad_norm": 0.5328172445297241, "learning_rate": 9.667062173602868e-05, "loss": 0.5144, "step": 6440 }, { "epoch": 2.434881087202718, "grad_norm": 0.596956193447113, "learning_rate": 9.665577394586013e-05, "loss": 0.5231, "step": 6450 }, { "epoch": 2.4386560966402415, "grad_norm": 0.5546612739562988, "learning_rate": 9.664089426691006e-05, "loss": 0.5233, "step": 6460 }, { "epoch": 2.4424311060777653, "grad_norm": 0.5410290360450745, "learning_rate": 9.662598270934852e-05, "loss": 0.5234, "step": 6470 }, { "epoch": 2.446206115515289, "grad_norm": 0.5477890968322754, "learning_rate": 9.661103928336748e-05, "loss": 0.5153, "step": 6480 }, { "epoch": 2.4499811249528123, "grad_norm": 0.5000752210617065, "learning_rate": 9.659606399918058e-05, "loss": 0.5104, "step": 6490 }, { "epoch": 2.453756134390336, "grad_norm": 0.5385310053825378, "learning_rate": 9.658105686702334e-05, "loss": 0.5143, "step": 6500 }, { "epoch": 2.4575311438278598, "grad_norm": 0.48516371846199036, "learning_rate": 9.656601789715294e-05, "loss": 0.5117, "step": 6510 }, { "epoch": 2.461306153265383, "grad_norm": 0.574826717376709, "learning_rate": 9.655094709984842e-05, "loss": 0.5114, "step": 6520 }, { "epoch": 2.465081162702907, "grad_norm": 0.5119508504867554, "learning_rate": 9.653584448541048e-05, "loss": 0.519, "step": 6530 }, { "epoch": 2.4688561721404305, "grad_norm": 0.5045502781867981, "learning_rate": 9.652071006416166e-05, "loss": 0.5134, "step": 6540 }, { "epoch": 2.472631181577954, "grad_norm": 0.4496314227581024, "learning_rate": 9.650554384644617e-05, "loss": 0.513, "step": 6550 }, { "epoch": 2.4764061910154775, "grad_norm": 0.45445775985717773, "learning_rate": 9.649034584263e-05, "loss": 0.5129, "step": 6560 }, { "epoch": 2.4801812004530013, "grad_norm": 0.5979011058807373, "learning_rate": 9.64751160631008e-05, "loss": 0.5236, "step": 6570 }, { "epoch": 2.4839562098905246, "grad_norm": 0.595025897026062, "learning_rate": 9.645985451826803e-05, "loss": 0.5233, "step": 6580 }, { "epoch": 2.4877312193280483, "grad_norm": 0.5131351351737976, "learning_rate": 9.644456121856275e-05, "loss": 0.5186, "step": 6590 }, { "epoch": 2.491506228765572, "grad_norm": 0.5269580483436584, "learning_rate": 9.642923617443784e-05, "loss": 0.527, "step": 6600 }, { "epoch": 2.4952812382030953, "grad_norm": 0.43365490436553955, "learning_rate": 9.641387939636782e-05, "loss": 0.5164, "step": 6610 }, { "epoch": 2.499056247640619, "grad_norm": 0.5464170575141907, "learning_rate": 9.639849089484888e-05, "loss": 0.5193, "step": 6620 }, { "epoch": 2.502831257078143, "grad_norm": 0.5362778306007385, "learning_rate": 9.638307068039893e-05, "loss": 0.5204, "step": 6630 }, { "epoch": 2.506606266515666, "grad_norm": 0.5616941452026367, "learning_rate": 9.636761876355753e-05, "loss": 0.5144, "step": 6640 }, { "epoch": 2.51038127595319, "grad_norm": 0.4664197862148285, "learning_rate": 9.635213515488593e-05, "loss": 0.5116, "step": 6650 }, { "epoch": 2.5141562853907136, "grad_norm": 0.5451635122299194, "learning_rate": 9.633661986496702e-05, "loss": 0.5167, "step": 6660 }, { "epoch": 2.5179312948282373, "grad_norm": 0.47441262006759644, "learning_rate": 9.632107290440535e-05, "loss": 0.5114, "step": 6670 }, { "epoch": 2.5217063042657606, "grad_norm": 0.5057079195976257, "learning_rate": 9.630549428382715e-05, "loss": 0.518, "step": 6680 }, { "epoch": 2.5254813137032843, "grad_norm": 0.5143377184867859, "learning_rate": 9.628988401388024e-05, "loss": 0.5096, "step": 6690 }, { "epoch": 2.5292563231408076, "grad_norm": 0.485270231962204, "learning_rate": 9.627424210523407e-05, "loss": 0.5115, "step": 6700 }, { "epoch": 2.5330313325783314, "grad_norm": 0.5043366551399231, "learning_rate": 9.62585685685798e-05, "loss": 0.5142, "step": 6710 }, { "epoch": 2.536806342015855, "grad_norm": 0.5317198038101196, "learning_rate": 9.624286341463008e-05, "loss": 0.5193, "step": 6720 }, { "epoch": 2.540581351453379, "grad_norm": 0.4412330985069275, "learning_rate": 9.622712665411927e-05, "loss": 0.5119, "step": 6730 }, { "epoch": 2.544356360890902, "grad_norm": 0.5754643082618713, "learning_rate": 9.621135829780328e-05, "loss": 0.515, "step": 6740 }, { "epoch": 2.548131370328426, "grad_norm": 0.5370513200759888, "learning_rate": 9.619555835645964e-05, "loss": 0.5132, "step": 6750 }, { "epoch": 2.551906379765949, "grad_norm": 0.5530955791473389, "learning_rate": 9.617972684088747e-05, "loss": 0.5167, "step": 6760 }, { "epoch": 2.555681389203473, "grad_norm": 0.5820709466934204, "learning_rate": 9.616386376190745e-05, "loss": 0.5118, "step": 6770 }, { "epoch": 2.5594563986409966, "grad_norm": 0.48913341760635376, "learning_rate": 9.614796913036184e-05, "loss": 0.5152, "step": 6780 }, { "epoch": 2.5632314080785203, "grad_norm": 0.48742911219596863, "learning_rate": 9.61320429571145e-05, "loss": 0.5074, "step": 6790 }, { "epoch": 2.5670064175160436, "grad_norm": 0.935467004776001, "learning_rate": 9.611608525305078e-05, "loss": 0.5215, "step": 6800 }, { "epoch": 2.5707814269535674, "grad_norm": 0.4997836947441101, "learning_rate": 9.610009602907764e-05, "loss": 0.5082, "step": 6810 }, { "epoch": 2.574556436391091, "grad_norm": 0.47966811060905457, "learning_rate": 9.608407529612357e-05, "loss": 0.5044, "step": 6820 }, { "epoch": 2.5783314458286144, "grad_norm": 0.5587261915206909, "learning_rate": 9.606802306513857e-05, "loss": 0.506, "step": 6830 }, { "epoch": 2.582106455266138, "grad_norm": 0.4997664988040924, "learning_rate": 9.605193934709421e-05, "loss": 0.5145, "step": 6840 }, { "epoch": 2.585881464703662, "grad_norm": 0.5600175261497498, "learning_rate": 9.603582415298354e-05, "loss": 0.5111, "step": 6850 }, { "epoch": 2.5896564741411856, "grad_norm": 0.49723485112190247, "learning_rate": 9.601967749382117e-05, "loss": 0.5114, "step": 6860 }, { "epoch": 2.593431483578709, "grad_norm": 0.4990619421005249, "learning_rate": 9.600349938064316e-05, "loss": 0.5086, "step": 6870 }, { "epoch": 2.5972064930162326, "grad_norm": 0.4644363522529602, "learning_rate": 9.598728982450711e-05, "loss": 0.5134, "step": 6880 }, { "epoch": 2.600981502453756, "grad_norm": 0.4740346670150757, "learning_rate": 9.597104883649211e-05, "loss": 0.5176, "step": 6890 }, { "epoch": 2.6047565118912797, "grad_norm": 0.5966189503669739, "learning_rate": 9.59547764276987e-05, "loss": 0.5044, "step": 6900 }, { "epoch": 2.6085315213288034, "grad_norm": 0.5578002333641052, "learning_rate": 9.593847260924892e-05, "loss": 0.5072, "step": 6910 }, { "epoch": 2.612306530766327, "grad_norm": 0.5736666321754456, "learning_rate": 9.59221373922863e-05, "loss": 0.5144, "step": 6920 }, { "epoch": 2.6160815402038504, "grad_norm": 0.6220727562904358, "learning_rate": 9.59057707879758e-05, "loss": 0.5031, "step": 6930 }, { "epoch": 2.619856549641374, "grad_norm": 0.565088152885437, "learning_rate": 9.588937280750382e-05, "loss": 0.5083, "step": 6940 }, { "epoch": 2.6236315590788974, "grad_norm": 0.5720441937446594, "learning_rate": 9.587294346207824e-05, "loss": 0.5303, "step": 6950 }, { "epoch": 2.627406568516421, "grad_norm": 0.6338963508605957, "learning_rate": 9.585648276292836e-05, "loss": 0.5149, "step": 6960 }, { "epoch": 2.631181577953945, "grad_norm": 0.504148542881012, "learning_rate": 9.583999072130492e-05, "loss": 0.5045, "step": 6970 }, { "epoch": 2.6349565873914687, "grad_norm": 0.5418031215667725, "learning_rate": 9.582346734848005e-05, "loss": 0.5128, "step": 6980 }, { "epoch": 2.638731596828992, "grad_norm": 0.5693923234939575, "learning_rate": 9.580691265574735e-05, "loss": 0.5106, "step": 6990 }, { "epoch": 2.6425066062665157, "grad_norm": 0.4889356195926666, "learning_rate": 9.579032665442178e-05, "loss": 0.5053, "step": 7000 }, { "epoch": 2.6462816157040394, "grad_norm": 0.5362417697906494, "learning_rate": 9.577370935583971e-05, "loss": 0.5196, "step": 7010 }, { "epoch": 2.6500566251415627, "grad_norm": 0.572538435459137, "learning_rate": 9.575706077135893e-05, "loss": 0.5022, "step": 7020 }, { "epoch": 2.6538316345790864, "grad_norm": 0.5328758955001831, "learning_rate": 9.574038091235857e-05, "loss": 0.5186, "step": 7030 }, { "epoch": 2.65760664401661, "grad_norm": 0.4795249104499817, "learning_rate": 9.572366979023916e-05, "loss": 0.5051, "step": 7040 }, { "epoch": 2.661381653454134, "grad_norm": 1.024570345878601, "learning_rate": 9.57069274164226e-05, "loss": 0.5087, "step": 7050 }, { "epoch": 2.665156662891657, "grad_norm": 0.641263484954834, "learning_rate": 9.569015380235213e-05, "loss": 0.5094, "step": 7060 }, { "epoch": 2.668931672329181, "grad_norm": 0.5658965706825256, "learning_rate": 9.567334895949238e-05, "loss": 0.5089, "step": 7070 }, { "epoch": 2.6727066817667042, "grad_norm": 0.6503294706344604, "learning_rate": 9.565651289932928e-05, "loss": 0.5071, "step": 7080 }, { "epoch": 2.676481691204228, "grad_norm": 0.5400107502937317, "learning_rate": 9.563964563337015e-05, "loss": 0.5072, "step": 7090 }, { "epoch": 2.6802567006417517, "grad_norm": 0.5809869766235352, "learning_rate": 9.562274717314357e-05, "loss": 0.5055, "step": 7100 }, { "epoch": 2.6840317100792754, "grad_norm": 0.5175840854644775, "learning_rate": 9.56058175301995e-05, "loss": 0.5084, "step": 7110 }, { "epoch": 2.6878067195167987, "grad_norm": 0.510823130607605, "learning_rate": 9.558885671610918e-05, "loss": 0.5142, "step": 7120 }, { "epoch": 2.6915817289543225, "grad_norm": 0.5192266702651978, "learning_rate": 9.557186474246517e-05, "loss": 0.51, "step": 7130 }, { "epoch": 2.6953567383918458, "grad_norm": 0.4871031939983368, "learning_rate": 9.555484162088133e-05, "loss": 0.5029, "step": 7140 }, { "epoch": 2.6991317478293695, "grad_norm": 0.6060387492179871, "learning_rate": 9.553778736299279e-05, "loss": 0.5032, "step": 7150 }, { "epoch": 2.7029067572668932, "grad_norm": 0.5182626247406006, "learning_rate": 9.552070198045599e-05, "loss": 0.5086, "step": 7160 }, { "epoch": 2.706681766704417, "grad_norm": 0.509029746055603, "learning_rate": 9.550358548494863e-05, "loss": 0.5076, "step": 7170 }, { "epoch": 2.7104567761419402, "grad_norm": 0.5284494757652283, "learning_rate": 9.548643788816965e-05, "loss": 0.4986, "step": 7180 }, { "epoch": 2.714231785579464, "grad_norm": 0.5049721598625183, "learning_rate": 9.546925920183929e-05, "loss": 0.5059, "step": 7190 }, { "epoch": 2.7180067950169873, "grad_norm": 0.5326721668243408, "learning_rate": 9.545204943769902e-05, "loss": 0.5078, "step": 7200 }, { "epoch": 2.721781804454511, "grad_norm": 0.5484986305236816, "learning_rate": 9.543480860751155e-05, "loss": 0.5036, "step": 7210 }, { "epoch": 2.7255568138920347, "grad_norm": 0.5126481652259827, "learning_rate": 9.541753672306081e-05, "loss": 0.5037, "step": 7220 }, { "epoch": 2.7293318233295585, "grad_norm": 0.4992707669734955, "learning_rate": 9.540023379615198e-05, "loss": 0.499, "step": 7230 }, { "epoch": 2.7331068327670818, "grad_norm": 0.46070438623428345, "learning_rate": 9.538289983861146e-05, "loss": 0.504, "step": 7240 }, { "epoch": 2.7368818422046055, "grad_norm": 0.6228576898574829, "learning_rate": 9.536553486228683e-05, "loss": 0.4989, "step": 7250 }, { "epoch": 2.7406568516421292, "grad_norm": 0.5347064137458801, "learning_rate": 9.534813887904689e-05, "loss": 0.5057, "step": 7260 }, { "epoch": 2.7444318610796525, "grad_norm": 0.5296137928962708, "learning_rate": 9.533071190078163e-05, "loss": 0.5062, "step": 7270 }, { "epoch": 2.7482068705171763, "grad_norm": 0.521789014339447, "learning_rate": 9.531325393940225e-05, "loss": 0.5067, "step": 7280 }, { "epoch": 2.7519818799547, "grad_norm": 0.5194920897483826, "learning_rate": 9.529576500684105e-05, "loss": 0.4991, "step": 7290 }, { "epoch": 2.7557568893922237, "grad_norm": 0.6369245648384094, "learning_rate": 9.52782451150516e-05, "loss": 0.5752, "step": 7300 }, { "epoch": 2.759531898829747, "grad_norm": 0.6341930031776428, "learning_rate": 9.526069427600857e-05, "loss": 0.5042, "step": 7310 }, { "epoch": 2.7633069082672708, "grad_norm": 0.5371559262275696, "learning_rate": 9.524311250170776e-05, "loss": 0.5055, "step": 7320 }, { "epoch": 2.767081917704794, "grad_norm": 0.6049063205718994, "learning_rate": 9.522549980416619e-05, "loss": 0.5062, "step": 7330 }, { "epoch": 2.770856927142318, "grad_norm": 0.5581164956092834, "learning_rate": 9.520785619542196e-05, "loss": 0.5072, "step": 7340 }, { "epoch": 2.7746319365798415, "grad_norm": 0.5462380051612854, "learning_rate": 9.519018168753428e-05, "loss": 0.5026, "step": 7350 }, { "epoch": 2.7784069460173653, "grad_norm": 0.5068135857582092, "learning_rate": 9.517247629258357e-05, "loss": 0.5122, "step": 7360 }, { "epoch": 2.7821819554548886, "grad_norm": 0.5248288512229919, "learning_rate": 9.515474002267124e-05, "loss": 0.5, "step": 7370 }, { "epoch": 2.7859569648924123, "grad_norm": 0.5446317791938782, "learning_rate": 9.513697288991989e-05, "loss": 0.5096, "step": 7380 }, { "epoch": 2.7897319743299356, "grad_norm": 0.5539740324020386, "learning_rate": 9.51191749064732e-05, "loss": 0.5037, "step": 7390 }, { "epoch": 2.7935069837674593, "grad_norm": 0.48345574736595154, "learning_rate": 9.51013460844959e-05, "loss": 0.4938, "step": 7400 }, { "epoch": 2.797281993204983, "grad_norm": 0.538747251033783, "learning_rate": 9.508348643617382e-05, "loss": 0.501, "step": 7410 }, { "epoch": 2.801057002642507, "grad_norm": 0.4825133681297302, "learning_rate": 9.50655959737139e-05, "loss": 0.4987, "step": 7420 }, { "epoch": 2.80483201208003, "grad_norm": 0.5534849762916565, "learning_rate": 9.504767470934405e-05, "loss": 0.4992, "step": 7430 }, { "epoch": 2.808607021517554, "grad_norm": 0.5716150999069214, "learning_rate": 9.502972265531332e-05, "loss": 0.5085, "step": 7440 }, { "epoch": 2.8123820309550775, "grad_norm": 0.47607553005218506, "learning_rate": 9.501173982389175e-05, "loss": 0.501, "step": 7450 }, { "epoch": 2.816157040392601, "grad_norm": 0.5296433568000793, "learning_rate": 9.499372622737047e-05, "loss": 0.5004, "step": 7460 }, { "epoch": 2.8199320498301246, "grad_norm": 0.5382222533226013, "learning_rate": 9.497568187806156e-05, "loss": 0.5099, "step": 7470 }, { "epoch": 2.8237070592676483, "grad_norm": 0.480925977230072, "learning_rate": 9.495760678829819e-05, "loss": 0.5031, "step": 7480 }, { "epoch": 2.8274820687051716, "grad_norm": 0.5002157092094421, "learning_rate": 9.49395009704345e-05, "loss": 0.5087, "step": 7490 }, { "epoch": 2.8312570781426953, "grad_norm": 0.517042338848114, "learning_rate": 9.492136443684565e-05, "loss": 0.5033, "step": 7500 }, { "epoch": 2.835032087580219, "grad_norm": 0.5893968343734741, "learning_rate": 9.49031971999278e-05, "loss": 0.5025, "step": 7510 }, { "epoch": 2.8388070970177424, "grad_norm": 0.47264131903648376, "learning_rate": 9.488499927209806e-05, "loss": 0.4947, "step": 7520 }, { "epoch": 2.842582106455266, "grad_norm": 0.5126714706420898, "learning_rate": 9.486677066579456e-05, "loss": 0.4967, "step": 7530 }, { "epoch": 2.84635711589279, "grad_norm": 0.5384365916252136, "learning_rate": 9.484851139347639e-05, "loss": 0.4961, "step": 7540 }, { "epoch": 2.8501321253303136, "grad_norm": 0.5498970746994019, "learning_rate": 9.483022146762358e-05, "loss": 0.4979, "step": 7550 }, { "epoch": 2.853907134767837, "grad_norm": 0.5383314490318298, "learning_rate": 9.48119009007371e-05, "loss": 0.4993, "step": 7560 }, { "epoch": 2.8576821442053606, "grad_norm": 0.46135392785072327, "learning_rate": 9.479354970533892e-05, "loss": 0.4989, "step": 7570 }, { "epoch": 2.861457153642884, "grad_norm": 0.5344179272651672, "learning_rate": 9.477516789397189e-05, "loss": 0.4975, "step": 7580 }, { "epoch": 2.8652321630804076, "grad_norm": 0.47662675380706787, "learning_rate": 9.475675547919983e-05, "loss": 0.5002, "step": 7590 }, { "epoch": 2.8690071725179314, "grad_norm": 0.5206663012504578, "learning_rate": 9.47383124736074e-05, "loss": 0.4969, "step": 7600 }, { "epoch": 2.872782181955455, "grad_norm": 0.5600720643997192, "learning_rate": 9.471983888980028e-05, "loss": 0.4965, "step": 7610 }, { "epoch": 2.8765571913929784, "grad_norm": 0.5266720056533813, "learning_rate": 9.470133474040494e-05, "loss": 0.5027, "step": 7620 }, { "epoch": 2.880332200830502, "grad_norm": 2.0006062984466553, "learning_rate": 9.468280003806882e-05, "loss": 0.5084, "step": 7630 }, { "epoch": 2.8841072102680254, "grad_norm": 0.5050598978996277, "learning_rate": 9.46642347954602e-05, "loss": 0.5037, "step": 7640 }, { "epoch": 2.887882219705549, "grad_norm": 0.47348907589912415, "learning_rate": 9.464563902526826e-05, "loss": 0.494, "step": 7650 }, { "epoch": 2.891657229143073, "grad_norm": 0.5870567560195923, "learning_rate": 9.462701274020303e-05, "loss": 0.4916, "step": 7660 }, { "epoch": 2.8954322385805966, "grad_norm": 0.580295979976654, "learning_rate": 9.46083559529954e-05, "loss": 0.4945, "step": 7670 }, { "epoch": 2.89920724801812, "grad_norm": 0.5126973390579224, "learning_rate": 9.458966867639712e-05, "loss": 0.4977, "step": 7680 }, { "epoch": 2.9029822574556436, "grad_norm": 0.5632660984992981, "learning_rate": 9.457095092318074e-05, "loss": 0.5005, "step": 7690 }, { "epoch": 2.9067572668931674, "grad_norm": 0.5285568833351135, "learning_rate": 9.455220270613969e-05, "loss": 0.4945, "step": 7700 }, { "epoch": 2.9105322763306907, "grad_norm": 0.8250537514686584, "learning_rate": 9.45334240380882e-05, "loss": 0.49, "step": 7710 }, { "epoch": 2.9143072857682144, "grad_norm": 0.5369483232498169, "learning_rate": 9.451461493186129e-05, "loss": 0.4984, "step": 7720 }, { "epoch": 2.918082295205738, "grad_norm": 0.49857333302497864, "learning_rate": 9.449577540031482e-05, "loss": 0.4919, "step": 7730 }, { "epoch": 2.921857304643262, "grad_norm": 0.5103145241737366, "learning_rate": 9.447690545632544e-05, "loss": 0.4908, "step": 7740 }, { "epoch": 2.925632314080785, "grad_norm": 0.5372836589813232, "learning_rate": 9.445800511279058e-05, "loss": 0.496, "step": 7750 }, { "epoch": 2.929407323518309, "grad_norm": 0.5468769669532776, "learning_rate": 9.443907438262844e-05, "loss": 0.5104, "step": 7760 }, { "epoch": 2.933182332955832, "grad_norm": 0.6511932015419006, "learning_rate": 9.442011327877798e-05, "loss": 0.4951, "step": 7770 }, { "epoch": 2.936957342393356, "grad_norm": 0.5281775593757629, "learning_rate": 9.440112181419895e-05, "loss": 0.498, "step": 7780 }, { "epoch": 2.9407323518308797, "grad_norm": 0.6145192384719849, "learning_rate": 9.438210000187186e-05, "loss": 0.4998, "step": 7790 }, { "epoch": 2.9445073612684034, "grad_norm": 0.5092405676841736, "learning_rate": 9.43630478547979e-05, "loss": 0.4879, "step": 7800 }, { "epoch": 2.9482823707059267, "grad_norm": 0.5056966543197632, "learning_rate": 9.434396538599906e-05, "loss": 0.49, "step": 7810 }, { "epoch": 2.9520573801434504, "grad_norm": 0.49092942476272583, "learning_rate": 9.4324852608518e-05, "loss": 0.4925, "step": 7820 }, { "epoch": 2.9558323895809737, "grad_norm": 0.48439252376556396, "learning_rate": 9.430570953541816e-05, "loss": 0.4888, "step": 7830 }, { "epoch": 2.9596073990184975, "grad_norm": 0.553947389125824, "learning_rate": 9.428653617978362e-05, "loss": 0.4942, "step": 7840 }, { "epoch": 2.963382408456021, "grad_norm": 0.5610196590423584, "learning_rate": 9.42673325547192e-05, "loss": 0.4906, "step": 7850 }, { "epoch": 2.967157417893545, "grad_norm": 0.6619260907173157, "learning_rate": 9.42480986733504e-05, "loss": 0.4923, "step": 7860 }, { "epoch": 2.970932427331068, "grad_norm": 0.5648078918457031, "learning_rate": 9.422883454882338e-05, "loss": 0.4951, "step": 7870 }, { "epoch": 2.974707436768592, "grad_norm": 0.505842387676239, "learning_rate": 9.4209540194305e-05, "loss": 0.5008, "step": 7880 }, { "epoch": 2.9784824462061152, "grad_norm": 0.5517747402191162, "learning_rate": 9.419021562298278e-05, "loss": 0.4977, "step": 7890 }, { "epoch": 2.982257455643639, "grad_norm": 0.5466206073760986, "learning_rate": 9.417086084806486e-05, "loss": 0.5011, "step": 7900 }, { "epoch": 2.9860324650811627, "grad_norm": 0.5874969363212585, "learning_rate": 9.415147588278005e-05, "loss": 0.4834, "step": 7910 }, { "epoch": 2.9898074745186864, "grad_norm": 0.6627646684646606, "learning_rate": 9.413206074037781e-05, "loss": 0.4895, "step": 7920 }, { "epoch": 2.9935824839562097, "grad_norm": 0.5905627608299255, "learning_rate": 9.411261543412819e-05, "loss": 0.4971, "step": 7930 }, { "epoch": 2.9973574933937335, "grad_norm": 0.5630371570587158, "learning_rate": 9.409313997732188e-05, "loss": 0.4933, "step": 7940 }, { "epoch": 3.001132502831257, "grad_norm": 0.5096926093101501, "learning_rate": 9.407363438327018e-05, "loss": 0.5077, "step": 7950 }, { "epoch": 3.0049075122687805, "grad_norm": 0.545479416847229, "learning_rate": 9.405409866530496e-05, "loss": 0.4925, "step": 7960 }, { "epoch": 3.0086825217063042, "grad_norm": 0.4964505434036255, "learning_rate": 9.40345328367787e-05, "loss": 0.4872, "step": 7970 }, { "epoch": 3.012457531143828, "grad_norm": 0.5932561159133911, "learning_rate": 9.401493691106446e-05, "loss": 0.4889, "step": 7980 }, { "epoch": 3.0162325405813513, "grad_norm": 0.4858354926109314, "learning_rate": 9.39953109015559e-05, "loss": 0.4962, "step": 7990 }, { "epoch": 3.020007550018875, "grad_norm": 0.4681301712989807, "learning_rate": 9.397565482166718e-05, "loss": 0.493, "step": 8000 }, { "epoch": 3.0237825594563987, "grad_norm": 0.5294594168663025, "learning_rate": 9.395596868483305e-05, "loss": 0.4941, "step": 8010 }, { "epoch": 3.027557568893922, "grad_norm": 0.5004336833953857, "learning_rate": 9.393625250450881e-05, "loss": 0.4886, "step": 8020 }, { "epoch": 3.0313325783314458, "grad_norm": 0.5321143269538879, "learning_rate": 9.391650629417028e-05, "loss": 0.497, "step": 8030 }, { "epoch": 3.0351075877689695, "grad_norm": 0.6064821481704712, "learning_rate": 9.38967300673138e-05, "loss": 0.4979, "step": 8040 }, { "epoch": 3.0388825972064932, "grad_norm": 0.5963532328605652, "learning_rate": 9.387692383745624e-05, "loss": 0.4883, "step": 8050 }, { "epoch": 3.0426576066440165, "grad_norm": 0.510930597782135, "learning_rate": 9.385708761813496e-05, "loss": 0.4906, "step": 8060 }, { "epoch": 3.0464326160815403, "grad_norm": 0.627373456954956, "learning_rate": 9.383722142290787e-05, "loss": 0.4881, "step": 8070 }, { "epoch": 3.050207625519064, "grad_norm": 0.5580819249153137, "learning_rate": 9.381732526535328e-05, "loss": 0.4941, "step": 8080 }, { "epoch": 3.0539826349565873, "grad_norm": 0.4533594250679016, "learning_rate": 9.379739915907004e-05, "loss": 0.4924, "step": 8090 }, { "epoch": 3.057757644394111, "grad_norm": 0.47666409611701965, "learning_rate": 9.377744311767746e-05, "loss": 0.4957, "step": 8100 }, { "epoch": 3.0615326538316348, "grad_norm": 0.5064053535461426, "learning_rate": 9.375745715481532e-05, "loss": 0.4978, "step": 8110 }, { "epoch": 3.065307663269158, "grad_norm": 0.5491259694099426, "learning_rate": 9.373744128414382e-05, "loss": 0.4879, "step": 8120 }, { "epoch": 3.069082672706682, "grad_norm": 0.5816760063171387, "learning_rate": 9.371739551934359e-05, "loss": 0.4947, "step": 8130 }, { "epoch": 3.0728576821442055, "grad_norm": 0.63339763879776, "learning_rate": 9.36973198741158e-05, "loss": 0.4921, "step": 8140 }, { "epoch": 3.076632691581729, "grad_norm": 0.48095303773880005, "learning_rate": 9.367721436218189e-05, "loss": 0.5039, "step": 8150 }, { "epoch": 3.0804077010192525, "grad_norm": 1.313503623008728, "learning_rate": 9.365707899728383e-05, "loss": 0.4874, "step": 8160 }, { "epoch": 3.0841827104567763, "grad_norm": 0.5337556004524231, "learning_rate": 9.363691379318391e-05, "loss": 0.4949, "step": 8170 }, { "epoch": 3.0879577198942996, "grad_norm": 0.5201692581176758, "learning_rate": 9.361671876366491e-05, "loss": 0.4903, "step": 8180 }, { "epoch": 3.0917327293318233, "grad_norm": 0.48047223687171936, "learning_rate": 9.35964939225299e-05, "loss": 0.4895, "step": 8190 }, { "epoch": 3.095507738769347, "grad_norm": 0.5520181655883789, "learning_rate": 9.35762392836024e-05, "loss": 0.5046, "step": 8200 }, { "epoch": 3.0992827482068703, "grad_norm": 0.5537580847740173, "learning_rate": 9.355595486072624e-05, "loss": 0.4904, "step": 8210 }, { "epoch": 3.103057757644394, "grad_norm": 0.5307925343513489, "learning_rate": 9.353564066776563e-05, "loss": 0.4851, "step": 8220 }, { "epoch": 3.106832767081918, "grad_norm": 0.5735664367675781, "learning_rate": 9.351529671860516e-05, "loss": 0.4852, "step": 8230 }, { "epoch": 3.110607776519441, "grad_norm": 0.553432047367096, "learning_rate": 9.349492302714969e-05, "loss": 0.5036, "step": 8240 }, { "epoch": 3.114382785956965, "grad_norm": 0.6025165319442749, "learning_rate": 9.347451960732447e-05, "loss": 0.4889, "step": 8250 }, { "epoch": 3.1181577953944886, "grad_norm": 10.570920944213867, "learning_rate": 9.345408647307506e-05, "loss": 0.5668, "step": 8260 }, { "epoch": 3.121932804832012, "grad_norm": 0.6134909391403198, "learning_rate": 9.343362363836726e-05, "loss": 0.488, "step": 8270 }, { "epoch": 3.1257078142695356, "grad_norm": 0.5113944411277771, "learning_rate": 9.341313111718728e-05, "loss": 0.4943, "step": 8280 }, { "epoch": 3.1294828237070593, "grad_norm": 0.586234986782074, "learning_rate": 9.339260892354153e-05, "loss": 0.4888, "step": 8290 }, { "epoch": 3.133257833144583, "grad_norm": 0.5822952389717102, "learning_rate": 9.337205707145676e-05, "loss": 0.5001, "step": 8300 }, { "epoch": 3.1370328425821064, "grad_norm": 0.6024602651596069, "learning_rate": 9.335147557497994e-05, "loss": 0.4862, "step": 8310 }, { "epoch": 3.14080785201963, "grad_norm": 0.5428383946418762, "learning_rate": 9.333086444817835e-05, "loss": 0.4889, "step": 8320 }, { "epoch": 3.144582861457154, "grad_norm": 0.5743405818939209, "learning_rate": 9.331022370513951e-05, "loss": 0.4873, "step": 8330 }, { "epoch": 3.148357870894677, "grad_norm": 0.497543066740036, "learning_rate": 9.328955335997113e-05, "loss": 0.4962, "step": 8340 }, { "epoch": 3.152132880332201, "grad_norm": 0.4729820489883423, "learning_rate": 9.326885342680125e-05, "loss": 0.4873, "step": 8350 }, { "epoch": 3.1559078897697246, "grad_norm": 0.5927625298500061, "learning_rate": 9.324812391977806e-05, "loss": 0.4924, "step": 8360 }, { "epoch": 3.159682899207248, "grad_norm": 0.5837076902389526, "learning_rate": 9.322736485306995e-05, "loss": 0.4886, "step": 8370 }, { "epoch": 3.1634579086447716, "grad_norm": 0.5531527996063232, "learning_rate": 9.32065762408656e-05, "loss": 0.4897, "step": 8380 }, { "epoch": 3.1672329180822953, "grad_norm": 0.5851390361785889, "learning_rate": 9.318575809737377e-05, "loss": 0.4864, "step": 8390 }, { "epoch": 3.1710079275198186, "grad_norm": 0.5682461261749268, "learning_rate": 9.316491043682351e-05, "loss": 0.4849, "step": 8400 }, { "epoch": 3.1747829369573424, "grad_norm": 0.623104989528656, "learning_rate": 9.3144033273464e-05, "loss": 0.5003, "step": 8410 }, { "epoch": 3.178557946394866, "grad_norm": 0.6800477504730225, "learning_rate": 9.312312662156457e-05, "loss": 0.4887, "step": 8420 }, { "epoch": 3.1823329558323894, "grad_norm": 0.5521610975265503, "learning_rate": 9.310219049541471e-05, "loss": 0.5015, "step": 8430 }, { "epoch": 3.186107965269913, "grad_norm": 0.619483232498169, "learning_rate": 9.308122490932409e-05, "loss": 0.4849, "step": 8440 }, { "epoch": 3.189882974707437, "grad_norm": 0.5283095240592957, "learning_rate": 9.306022987762246e-05, "loss": 0.4877, "step": 8450 }, { "epoch": 3.19365798414496, "grad_norm": 0.569065511226654, "learning_rate": 9.303920541465974e-05, "loss": 0.4844, "step": 8460 }, { "epoch": 3.197432993582484, "grad_norm": 0.4644882082939148, "learning_rate": 9.301815153480598e-05, "loss": 0.4875, "step": 8470 }, { "epoch": 3.2012080030200076, "grad_norm": 0.5540872812271118, "learning_rate": 9.299706825245126e-05, "loss": 0.4936, "step": 8480 }, { "epoch": 3.2049830124575314, "grad_norm": 0.5254048705101013, "learning_rate": 9.297595558200583e-05, "loss": 0.4812, "step": 8490 }, { "epoch": 3.2087580218950547, "grad_norm": 0.4865260422229767, "learning_rate": 9.295481353790001e-05, "loss": 0.4874, "step": 8500 }, { "epoch": 3.2125330313325784, "grad_norm": 0.6707580089569092, "learning_rate": 9.293364213458417e-05, "loss": 0.4893, "step": 8510 }, { "epoch": 3.216308040770102, "grad_norm": 0.5525599718093872, "learning_rate": 9.291244138652877e-05, "loss": 0.4847, "step": 8520 }, { "epoch": 3.2200830502076254, "grad_norm": 0.5599222183227539, "learning_rate": 9.289121130822431e-05, "loss": 0.4823, "step": 8530 }, { "epoch": 3.223858059645149, "grad_norm": 0.6073276996612549, "learning_rate": 9.286995191418137e-05, "loss": 0.481, "step": 8540 }, { "epoch": 3.227633069082673, "grad_norm": 0.5141738057136536, "learning_rate": 9.284866321893056e-05, "loss": 0.4807, "step": 8550 }, { "epoch": 3.231408078520196, "grad_norm": 0.5618724822998047, "learning_rate": 9.282734523702244e-05, "loss": 0.4911, "step": 8560 }, { "epoch": 3.23518308795772, "grad_norm": 0.5364367961883545, "learning_rate": 9.280599798302771e-05, "loss": 0.4817, "step": 8570 }, { "epoch": 3.2389580973952437, "grad_norm": 0.5237820744514465, "learning_rate": 9.278462147153699e-05, "loss": 0.4798, "step": 8580 }, { "epoch": 3.242733106832767, "grad_norm": 0.5670002698898315, "learning_rate": 9.276321571716094e-05, "loss": 0.4839, "step": 8590 }, { "epoch": 3.2465081162702907, "grad_norm": 0.5286555290222168, "learning_rate": 9.274178073453017e-05, "loss": 0.4827, "step": 8600 }, { "epoch": 3.2502831257078144, "grad_norm": 0.7837060689926147, "learning_rate": 9.27203165382953e-05, "loss": 0.4848, "step": 8610 }, { "epoch": 3.2540581351453377, "grad_norm": 0.5733458995819092, "learning_rate": 9.26988231431269e-05, "loss": 0.4874, "step": 8620 }, { "epoch": 3.2578331445828614, "grad_norm": 0.5995279550552368, "learning_rate": 9.267730056371551e-05, "loss": 0.4856, "step": 8630 }, { "epoch": 3.261608154020385, "grad_norm": 0.585941731929779, "learning_rate": 9.265574881477161e-05, "loss": 0.495, "step": 8640 }, { "epoch": 3.2653831634579085, "grad_norm": 0.592987596988678, "learning_rate": 9.263416791102561e-05, "loss": 0.4823, "step": 8650 }, { "epoch": 3.269158172895432, "grad_norm": 0.47990015149116516, "learning_rate": 9.261255786722786e-05, "loss": 0.487, "step": 8660 }, { "epoch": 3.272933182332956, "grad_norm": 0.49177876114845276, "learning_rate": 9.259091869814864e-05, "loss": 0.4745, "step": 8670 }, { "epoch": 3.2767081917704797, "grad_norm": 0.5345214009284973, "learning_rate": 9.256925041857807e-05, "loss": 0.4883, "step": 8680 }, { "epoch": 3.280483201208003, "grad_norm": 0.5280712842941284, "learning_rate": 9.254755304332626e-05, "loss": 0.4837, "step": 8690 }, { "epoch": 3.2842582106455267, "grad_norm": 0.5407823920249939, "learning_rate": 9.252582658722313e-05, "loss": 0.4824, "step": 8700 }, { "epoch": 3.28803322008305, "grad_norm": 0.510508120059967, "learning_rate": 9.250407106511853e-05, "loss": 0.4763, "step": 8710 }, { "epoch": 3.2918082295205737, "grad_norm": 0.5562644004821777, "learning_rate": 9.248228649188215e-05, "loss": 0.4756, "step": 8720 }, { "epoch": 3.2955832389580975, "grad_norm": 0.5866154432296753, "learning_rate": 9.246047288240354e-05, "loss": 0.4792, "step": 8730 }, { "epoch": 3.299358248395621, "grad_norm": 0.5227904319763184, "learning_rate": 9.243863025159208e-05, "loss": 0.48, "step": 8740 }, { "epoch": 3.3031332578331445, "grad_norm": 0.6132093667984009, "learning_rate": 9.241675861437704e-05, "loss": 0.4762, "step": 8750 }, { "epoch": 3.3069082672706682, "grad_norm": 0.5366108417510986, "learning_rate": 9.239485798570742e-05, "loss": 0.4813, "step": 8760 }, { "epoch": 3.310683276708192, "grad_norm": 0.5920717716217041, "learning_rate": 9.237292838055215e-05, "loss": 0.4834, "step": 8770 }, { "epoch": 3.3144582861457152, "grad_norm": 0.5619731545448303, "learning_rate": 9.235096981389988e-05, "loss": 0.4846, "step": 8780 }, { "epoch": 3.318233295583239, "grad_norm": 0.45175281167030334, "learning_rate": 9.232898230075909e-05, "loss": 0.4807, "step": 8790 }, { "epoch": 3.3220083050207627, "grad_norm": 0.613134503364563, "learning_rate": 9.230696585615801e-05, "loss": 0.4724, "step": 8800 }, { "epoch": 3.325783314458286, "grad_norm": 0.5235288739204407, "learning_rate": 9.228492049514473e-05, "loss": 0.4767, "step": 8810 }, { "epoch": 3.3295583238958097, "grad_norm": 0.5533393025398254, "learning_rate": 9.2262846232787e-05, "loss": 0.4836, "step": 8820 }, { "epoch": 3.3333333333333335, "grad_norm": 0.5668758749961853, "learning_rate": 9.224074308417238e-05, "loss": 0.485, "step": 8830 }, { "epoch": 3.3371083427708568, "grad_norm": 0.568915069103241, "learning_rate": 9.221861106440818e-05, "loss": 0.4767, "step": 8840 }, { "epoch": 3.3408833522083805, "grad_norm": 0.6423974633216858, "learning_rate": 9.219645018862142e-05, "loss": 0.4766, "step": 8850 }, { "epoch": 3.3446583616459042, "grad_norm": 0.8083536028862, "learning_rate": 9.217426047195882e-05, "loss": 0.4734, "step": 8860 }, { "epoch": 3.3484333710834275, "grad_norm": 0.5543457865715027, "learning_rate": 9.21520419295869e-05, "loss": 0.4806, "step": 8870 }, { "epoch": 3.3522083805209513, "grad_norm": 0.5718511939048767, "learning_rate": 9.212979457669178e-05, "loss": 0.4761, "step": 8880 }, { "epoch": 3.355983389958475, "grad_norm": 0.6094539165496826, "learning_rate": 9.210751842847931e-05, "loss": 0.4789, "step": 8890 }, { "epoch": 3.3597583993959983, "grad_norm": 0.5675216913223267, "learning_rate": 9.208521350017506e-05, "loss": 0.4796, "step": 8900 }, { "epoch": 3.363533408833522, "grad_norm": 0.5290115475654602, "learning_rate": 9.206287980702422e-05, "loss": 0.477, "step": 8910 }, { "epoch": 3.3673084182710458, "grad_norm": 0.5056905150413513, "learning_rate": 9.204051736429166e-05, "loss": 0.4738, "step": 8920 }, { "epoch": 3.3710834277085695, "grad_norm": 0.8471300601959229, "learning_rate": 9.20181261872619e-05, "loss": 0.4825, "step": 8930 }, { "epoch": 3.374858437146093, "grad_norm": 0.516025185585022, "learning_rate": 9.19957062912391e-05, "loss": 0.4708, "step": 8940 }, { "epoch": 3.3786334465836165, "grad_norm": 0.5982151627540588, "learning_rate": 9.197325769154704e-05, "loss": 0.4698, "step": 8950 }, { "epoch": 3.38240845602114, "grad_norm": 0.6391962766647339, "learning_rate": 9.195078040352914e-05, "loss": 0.477, "step": 8960 }, { "epoch": 3.3861834654586636, "grad_norm": 0.6175360679626465, "learning_rate": 9.192827444254839e-05, "loss": 0.4771, "step": 8970 }, { "epoch": 3.3899584748961873, "grad_norm": 0.5909090042114258, "learning_rate": 9.190573982398744e-05, "loss": 0.4789, "step": 8980 }, { "epoch": 3.393733484333711, "grad_norm": 0.6406884789466858, "learning_rate": 9.188317656324845e-05, "loss": 0.4735, "step": 8990 }, { "epoch": 3.3975084937712343, "grad_norm": 0.509844183921814, "learning_rate": 9.186058467575322e-05, "loss": 0.4786, "step": 9000 }, { "epoch": 3.401283503208758, "grad_norm": 0.524660587310791, "learning_rate": 9.183796417694309e-05, "loss": 0.4789, "step": 9010 }, { "epoch": 3.405058512646282, "grad_norm": 0.5184488892555237, "learning_rate": 9.181531508227897e-05, "loss": 0.4794, "step": 9020 }, { "epoch": 3.408833522083805, "grad_norm": 0.4693617522716522, "learning_rate": 9.179263740724125e-05, "loss": 0.483, "step": 9030 }, { "epoch": 3.412608531521329, "grad_norm": 0.5032320618629456, "learning_rate": 9.176993116732996e-05, "loss": 0.4748, "step": 9040 }, { "epoch": 3.4163835409588525, "grad_norm": 0.5809365510940552, "learning_rate": 9.17471963780646e-05, "loss": 0.4671, "step": 9050 }, { "epoch": 3.420158550396376, "grad_norm": 0.547820508480072, "learning_rate": 9.172443305498414e-05, "loss": 0.4729, "step": 9060 }, { "epoch": 3.4239335598338996, "grad_norm": 0.48663732409477234, "learning_rate": 9.170164121364714e-05, "loss": 0.475, "step": 9070 }, { "epoch": 3.4277085692714233, "grad_norm": 0.6125722527503967, "learning_rate": 9.167882086963158e-05, "loss": 0.4802, "step": 9080 }, { "epoch": 3.4314835787089466, "grad_norm": 0.5475627183914185, "learning_rate": 9.165597203853499e-05, "loss": 0.4822, "step": 9090 }, { "epoch": 3.4352585881464703, "grad_norm": 1.4172425270080566, "learning_rate": 9.163309473597428e-05, "loss": 0.4796, "step": 9100 }, { "epoch": 3.439033597583994, "grad_norm": 0.5351138114929199, "learning_rate": 9.161018897758591e-05, "loss": 0.4766, "step": 9110 }, { "epoch": 3.4428086070215174, "grad_norm": 0.6007091999053955, "learning_rate": 9.158725477902572e-05, "loss": 0.4755, "step": 9120 }, { "epoch": 3.446583616459041, "grad_norm": 0.5067334175109863, "learning_rate": 9.156429215596905e-05, "loss": 0.4923, "step": 9130 }, { "epoch": 3.450358625896565, "grad_norm": 0.7098965644836426, "learning_rate": 9.154130112411061e-05, "loss": 0.4741, "step": 9140 }, { "epoch": 3.454133635334088, "grad_norm": 0.555362343788147, "learning_rate": 9.151828169916455e-05, "loss": 0.4741, "step": 9150 }, { "epoch": 3.457908644771612, "grad_norm": 0.5457745790481567, "learning_rate": 9.149523389686445e-05, "loss": 0.4723, "step": 9160 }, { "epoch": 3.4616836542091356, "grad_norm": 0.5627350807189941, "learning_rate": 9.147215773296327e-05, "loss": 0.4732, "step": 9170 }, { "epoch": 3.4654586636466593, "grad_norm": 0.5835473537445068, "learning_rate": 9.144905322323334e-05, "loss": 0.4693, "step": 9180 }, { "epoch": 3.4692336730841826, "grad_norm": 0.5265465974807739, "learning_rate": 9.142592038346635e-05, "loss": 0.477, "step": 9190 }, { "epoch": 3.4730086825217064, "grad_norm": 0.5562682747840881, "learning_rate": 9.140275922947343e-05, "loss": 0.4757, "step": 9200 }, { "epoch": 3.4767836919592296, "grad_norm": 0.6338258981704712, "learning_rate": 9.137956977708496e-05, "loss": 0.4843, "step": 9210 }, { "epoch": 3.4805587013967534, "grad_norm": 0.5517134666442871, "learning_rate": 9.135635204215075e-05, "loss": 0.47, "step": 9220 }, { "epoch": 3.484333710834277, "grad_norm": 0.6941575407981873, "learning_rate": 9.133310604053986e-05, "loss": 0.4773, "step": 9230 }, { "epoch": 3.488108720271801, "grad_norm": 0.532227098941803, "learning_rate": 9.130983178814077e-05, "loss": 0.4677, "step": 9240 }, { "epoch": 3.491883729709324, "grad_norm": 0.5691418051719666, "learning_rate": 9.128652930086116e-05, "loss": 0.476, "step": 9250 }, { "epoch": 3.495658739146848, "grad_norm": 0.5184800028800964, "learning_rate": 9.126319859462808e-05, "loss": 0.4822, "step": 9260 }, { "epoch": 3.4994337485843716, "grad_norm": 0.55794358253479, "learning_rate": 9.123983968538787e-05, "loss": 0.4786, "step": 9270 }, { "epoch": 3.503208758021895, "grad_norm": 0.5698232650756836, "learning_rate": 9.121645258910609e-05, "loss": 0.4716, "step": 9280 }, { "epoch": 3.5069837674594186, "grad_norm": 0.53619784116745, "learning_rate": 9.119303732176762e-05, "loss": 0.4761, "step": 9290 }, { "epoch": 3.5107587768969424, "grad_norm": 0.5619392395019531, "learning_rate": 9.116959389937656e-05, "loss": 0.4741, "step": 9300 }, { "epoch": 3.5145337863344657, "grad_norm": 0.5324605107307434, "learning_rate": 9.114612233795628e-05, "loss": 0.4683, "step": 9310 }, { "epoch": 3.5183087957719894, "grad_norm": 0.5709056258201599, "learning_rate": 9.112262265354937e-05, "loss": 0.4778, "step": 9320 }, { "epoch": 3.522083805209513, "grad_norm": 0.5106092691421509, "learning_rate": 9.109909486221763e-05, "loss": 0.4673, "step": 9330 }, { "epoch": 3.5258588146470364, "grad_norm": 0.5639387369155884, "learning_rate": 9.107553898004208e-05, "loss": 0.4741, "step": 9340 }, { "epoch": 3.52963382408456, "grad_norm": 0.5696660280227661, "learning_rate": 9.105195502312298e-05, "loss": 0.4716, "step": 9350 }, { "epoch": 3.533408833522084, "grad_norm": 0.5708215236663818, "learning_rate": 9.102834300757972e-05, "loss": 0.4703, "step": 9360 }, { "epoch": 3.5371838429596076, "grad_norm": 0.5708835124969482, "learning_rate": 9.100470294955087e-05, "loss": 0.4707, "step": 9370 }, { "epoch": 3.540958852397131, "grad_norm": 0.49455785751342773, "learning_rate": 9.098103486519423e-05, "loss": 0.4693, "step": 9380 }, { "epoch": 3.5447338618346547, "grad_norm": 0.5507346987724304, "learning_rate": 9.095733877068666e-05, "loss": 0.468, "step": 9390 }, { "epoch": 3.548508871272178, "grad_norm": 0.5424711108207703, "learning_rate": 9.093361468222426e-05, "loss": 0.4784, "step": 9400 }, { "epoch": 3.5522838807097017, "grad_norm": 0.5007194876670837, "learning_rate": 9.090986261602221e-05, "loss": 0.4659, "step": 9410 }, { "epoch": 3.5560588901472254, "grad_norm": 0.6124007701873779, "learning_rate": 9.088608258831482e-05, "loss": 0.4768, "step": 9420 }, { "epoch": 3.559833899584749, "grad_norm": 0.5798971056938171, "learning_rate": 9.08622746153555e-05, "loss": 0.4814, "step": 9430 }, { "epoch": 3.5636089090222725, "grad_norm": 0.5349039435386658, "learning_rate": 9.083843871341679e-05, "loss": 0.4676, "step": 9440 }, { "epoch": 3.567383918459796, "grad_norm": 0.557998776435852, "learning_rate": 9.08145748987903e-05, "loss": 0.4758, "step": 9450 }, { "epoch": 3.5711589278973195, "grad_norm": 0.5525006055831909, "learning_rate": 9.079068318778674e-05, "loss": 0.4692, "step": 9460 }, { "epoch": 3.574933937334843, "grad_norm": 0.6234112977981567, "learning_rate": 9.076676359673586e-05, "loss": 0.4658, "step": 9470 }, { "epoch": 3.578708946772367, "grad_norm": 0.6095515489578247, "learning_rate": 9.074281614198645e-05, "loss": 0.4685, "step": 9480 }, { "epoch": 3.5824839562098907, "grad_norm": 0.5489786267280579, "learning_rate": 9.071884083990638e-05, "loss": 0.4732, "step": 9490 }, { "epoch": 3.586258965647414, "grad_norm": 0.5583869814872742, "learning_rate": 9.069483770688259e-05, "loss": 0.4704, "step": 9500 }, { "epoch": 3.5900339750849377, "grad_norm": 0.5421421527862549, "learning_rate": 9.067080675932093e-05, "loss": 0.4735, "step": 9510 }, { "epoch": 3.5938089845224614, "grad_norm": 0.6016372442245483, "learning_rate": 9.064674801364634e-05, "loss": 0.4794, "step": 9520 }, { "epoch": 3.5975839939599847, "grad_norm": 0.5686151385307312, "learning_rate": 9.062266148630279e-05, "loss": 0.4758, "step": 9530 }, { "epoch": 3.6013590033975085, "grad_norm": 0.5975391864776611, "learning_rate": 9.059854719375313e-05, "loss": 0.4748, "step": 9540 }, { "epoch": 3.605134012835032, "grad_norm": 0.5515162944793701, "learning_rate": 9.057440515247932e-05, "loss": 0.4738, "step": 9550 }, { "epoch": 3.608909022272556, "grad_norm": 0.5409767627716064, "learning_rate": 9.055023537898215e-05, "loss": 0.4786, "step": 9560 }, { "epoch": 3.6126840317100792, "grad_norm": 0.4877788722515106, "learning_rate": 9.052603788978149e-05, "loss": 0.4849, "step": 9570 }, { "epoch": 3.616459041147603, "grad_norm": 0.6089911460876465, "learning_rate": 9.050181270141606e-05, "loss": 0.4728, "step": 9580 }, { "epoch": 3.6202340505851263, "grad_norm": 0.64964359998703, "learning_rate": 9.047755983044355e-05, "loss": 0.4684, "step": 9590 }, { "epoch": 3.62400906002265, "grad_norm": 0.5478320121765137, "learning_rate": 9.045327929344058e-05, "loss": 0.4765, "step": 9600 }, { "epoch": 3.6277840694601737, "grad_norm": 0.5902840495109558, "learning_rate": 9.042897110700268e-05, "loss": 0.4705, "step": 9610 }, { "epoch": 3.6315590788976975, "grad_norm": 0.5538387894630432, "learning_rate": 9.040463528774423e-05, "loss": 0.473, "step": 9620 }, { "epoch": 3.6353340883352208, "grad_norm": 0.5468869209289551, "learning_rate": 9.038027185229856e-05, "loss": 0.4745, "step": 9630 }, { "epoch": 3.6391090977727445, "grad_norm": 0.6634562015533447, "learning_rate": 9.035588081731784e-05, "loss": 0.4749, "step": 9640 }, { "epoch": 3.642884107210268, "grad_norm": 0.5892646908760071, "learning_rate": 9.033146219947313e-05, "loss": 0.4674, "step": 9650 }, { "epoch": 3.6466591166477915, "grad_norm": 0.5538533926010132, "learning_rate": 9.030701601545429e-05, "loss": 0.4665, "step": 9660 }, { "epoch": 3.6504341260853153, "grad_norm": 0.5980792045593262, "learning_rate": 9.028254228197009e-05, "loss": 0.4635, "step": 9670 }, { "epoch": 3.654209135522839, "grad_norm": 0.5423790216445923, "learning_rate": 9.025804101574808e-05, "loss": 0.4637, "step": 9680 }, { "epoch": 3.6579841449603623, "grad_norm": 0.5625879168510437, "learning_rate": 9.023351223353463e-05, "loss": 0.4744, "step": 9690 }, { "epoch": 3.661759154397886, "grad_norm": 0.5283533930778503, "learning_rate": 9.020895595209498e-05, "loss": 0.4697, "step": 9700 }, { "epoch": 3.6655341638354098, "grad_norm": 0.569664478302002, "learning_rate": 9.018437218821305e-05, "loss": 0.4591, "step": 9710 }, { "epoch": 3.669309173272933, "grad_norm": 0.5049920678138733, "learning_rate": 9.015976095869166e-05, "loss": 0.4713, "step": 9720 }, { "epoch": 3.673084182710457, "grad_norm": 0.5221184492111206, "learning_rate": 9.013512228035235e-05, "loss": 0.4634, "step": 9730 }, { "epoch": 3.6768591921479805, "grad_norm": 0.6969385743141174, "learning_rate": 9.01104561700354e-05, "loss": 0.4678, "step": 9740 }, { "epoch": 3.680634201585504, "grad_norm": 0.6279971599578857, "learning_rate": 9.008576264459987e-05, "loss": 0.4684, "step": 9750 }, { "epoch": 3.6844092110230275, "grad_norm": 0.7082238793373108, "learning_rate": 9.006104172092357e-05, "loss": 0.4734, "step": 9760 }, { "epoch": 3.6881842204605513, "grad_norm": 0.5783301591873169, "learning_rate": 9.0036293415903e-05, "loss": 0.494, "step": 9770 }, { "epoch": 3.6919592298980746, "grad_norm": 0.5902776122093201, "learning_rate": 9.00115177464534e-05, "loss": 0.5231, "step": 9780 }, { "epoch": 3.6957342393355983, "grad_norm": 0.6361207365989685, "learning_rate": 8.998671472950872e-05, "loss": 0.4692, "step": 9790 }, { "epoch": 3.699509248773122, "grad_norm": 0.5117760300636292, "learning_rate": 8.996188438202156e-05, "loss": 0.4679, "step": 9800 }, { "epoch": 3.7032842582106458, "grad_norm": 0.6711943745613098, "learning_rate": 8.993702672096324e-05, "loss": 0.4669, "step": 9810 }, { "epoch": 3.707059267648169, "grad_norm": 0.5595587491989136, "learning_rate": 8.991214176332377e-05, "loss": 0.4706, "step": 9820 }, { "epoch": 3.710834277085693, "grad_norm": 0.5421229004859924, "learning_rate": 8.988722952611171e-05, "loss": 0.47, "step": 9830 }, { "epoch": 3.714609286523216, "grad_norm": 0.5640469789505005, "learning_rate": 8.986229002635441e-05, "loss": 0.4672, "step": 9840 }, { "epoch": 3.71838429596074, "grad_norm": 0.5714800357818604, "learning_rate": 8.983732328109775e-05, "loss": 0.4649, "step": 9850 }, { "epoch": 3.7221593053982636, "grad_norm": 0.5369188189506531, "learning_rate": 8.981232930740627e-05, "loss": 0.4677, "step": 9860 }, { "epoch": 3.7259343148357873, "grad_norm": 0.8434441089630127, "learning_rate": 8.978730812236313e-05, "loss": 0.4664, "step": 9870 }, { "epoch": 3.7297093242733106, "grad_norm": 0.5156511664390564, "learning_rate": 8.976225974307004e-05, "loss": 0.4658, "step": 9880 }, { "epoch": 3.7334843337108343, "grad_norm": 0.5229775905609131, "learning_rate": 8.973718418664736e-05, "loss": 0.4692, "step": 9890 }, { "epoch": 3.7372593431483576, "grad_norm": 0.5336432456970215, "learning_rate": 8.971208147023397e-05, "loss": 0.4629, "step": 9900 }, { "epoch": 3.7410343525858814, "grad_norm": 0.5237134695053101, "learning_rate": 8.968695161098739e-05, "loss": 0.4709, "step": 9910 }, { "epoch": 3.744809362023405, "grad_norm": 0.5551522970199585, "learning_rate": 8.96617946260836e-05, "loss": 0.4639, "step": 9920 }, { "epoch": 3.748584371460929, "grad_norm": 0.5532174706459045, "learning_rate": 8.963661053271716e-05, "loss": 0.4654, "step": 9930 }, { "epoch": 3.752359380898452, "grad_norm": 0.5492597818374634, "learning_rate": 8.961139934810117e-05, "loss": 0.4636, "step": 9940 }, { "epoch": 3.756134390335976, "grad_norm": 0.7447717785835266, "learning_rate": 8.958616108946725e-05, "loss": 0.4703, "step": 9950 }, { "epoch": 3.7599093997734996, "grad_norm": 0.5668697357177734, "learning_rate": 8.95608957740655e-05, "loss": 0.4647, "step": 9960 }, { "epoch": 3.763684409211023, "grad_norm": 0.6193100214004517, "learning_rate": 8.953560341916452e-05, "loss": 0.4757, "step": 9970 }, { "epoch": 3.7674594186485466, "grad_norm": 0.522419273853302, "learning_rate": 8.95102840420514e-05, "loss": 0.4662, "step": 9980 }, { "epoch": 3.7712344280860703, "grad_norm": 0.5857412219047546, "learning_rate": 8.94849376600317e-05, "loss": 0.4657, "step": 9990 }, { "epoch": 3.775009437523594, "grad_norm": 0.557212769985199, "learning_rate": 8.945956429042943e-05, "loss": 0.4557, "step": 10000 }, { "epoch": 3.7787844469611174, "grad_norm": 0.600222647190094, "learning_rate": 8.943416395058705e-05, "loss": 0.4661, "step": 10010 }, { "epoch": 3.782559456398641, "grad_norm": 0.6195318102836609, "learning_rate": 8.940873665786544e-05, "loss": 0.467, "step": 10020 }, { "epoch": 3.7863344658361644, "grad_norm": 0.8778566122055054, "learning_rate": 8.938328242964394e-05, "loss": 0.4602, "step": 10030 }, { "epoch": 3.790109475273688, "grad_norm": 0.6205072999000549, "learning_rate": 8.935780128332026e-05, "loss": 0.4604, "step": 10040 }, { "epoch": 3.793884484711212, "grad_norm": 0.5959478616714478, "learning_rate": 8.933229323631052e-05, "loss": 0.4582, "step": 10050 }, { "epoch": 3.7976594941487356, "grad_norm": 0.5634093880653381, "learning_rate": 8.930675830604925e-05, "loss": 0.4649, "step": 10060 }, { "epoch": 3.801434503586259, "grad_norm": 0.5344751477241516, "learning_rate": 8.92811965099893e-05, "loss": 0.465, "step": 10070 }, { "epoch": 3.8052095130237826, "grad_norm": 0.5962951183319092, "learning_rate": 8.925560786560194e-05, "loss": 0.4676, "step": 10080 }, { "epoch": 3.808984522461306, "grad_norm": 0.5603546500205994, "learning_rate": 8.922999239037677e-05, "loss": 0.4597, "step": 10090 }, { "epoch": 3.8127595318988297, "grad_norm": 0.5416293144226074, "learning_rate": 8.920435010182171e-05, "loss": 0.4627, "step": 10100 }, { "epoch": 3.8165345413363534, "grad_norm": 0.5338233709335327, "learning_rate": 8.917868101746302e-05, "loss": 0.4745, "step": 10110 }, { "epoch": 3.820309550773877, "grad_norm": 0.6542847752571106, "learning_rate": 8.91529851548453e-05, "loss": 0.4676, "step": 10120 }, { "epoch": 3.8240845602114004, "grad_norm": 0.6008440256118774, "learning_rate": 8.912726253153142e-05, "loss": 0.4685, "step": 10130 }, { "epoch": 3.827859569648924, "grad_norm": 0.6180557012557983, "learning_rate": 8.910151316510255e-05, "loss": 0.4589, "step": 10140 }, { "epoch": 3.8316345790864474, "grad_norm": 0.6951040625572205, "learning_rate": 8.907573707315813e-05, "loss": 0.4624, "step": 10150 }, { "epoch": 3.835409588523971, "grad_norm": 0.5660848021507263, "learning_rate": 8.904993427331588e-05, "loss": 0.4647, "step": 10160 }, { "epoch": 3.839184597961495, "grad_norm": 0.5535010099411011, "learning_rate": 8.902410478321176e-05, "loss": 0.4602, "step": 10170 }, { "epoch": 3.8429596073990187, "grad_norm": 0.5885663628578186, "learning_rate": 8.899824862050002e-05, "loss": 0.4686, "step": 10180 }, { "epoch": 3.846734616836542, "grad_norm": 0.6347377300262451, "learning_rate": 8.897236580285308e-05, "loss": 0.4584, "step": 10190 }, { "epoch": 3.8505096262740657, "grad_norm": 0.6041572093963623, "learning_rate": 8.894645634796159e-05, "loss": 0.4584, "step": 10200 }, { "epoch": 3.8542846357115894, "grad_norm": 0.5319682359695435, "learning_rate": 8.892052027353444e-05, "loss": 0.4608, "step": 10210 }, { "epoch": 3.8580596451491127, "grad_norm": 0.5556275248527527, "learning_rate": 8.889455759729866e-05, "loss": 0.465, "step": 10220 }, { "epoch": 3.8618346545866364, "grad_norm": 0.6249091029167175, "learning_rate": 8.886856833699955e-05, "loss": 0.4611, "step": 10230 }, { "epoch": 3.86560966402416, "grad_norm": 0.5192009806632996, "learning_rate": 8.884255251040046e-05, "loss": 0.4656, "step": 10240 }, { "epoch": 3.869384673461684, "grad_norm": 0.5966506600379944, "learning_rate": 8.8816510135283e-05, "loss": 0.4614, "step": 10250 }, { "epoch": 3.873159682899207, "grad_norm": 0.5522090196609497, "learning_rate": 8.879044122944688e-05, "loss": 0.4583, "step": 10260 }, { "epoch": 3.876934692336731, "grad_norm": 0.8987534046173096, "learning_rate": 8.876434581070996e-05, "loss": 0.4639, "step": 10270 }, { "epoch": 3.8807097017742542, "grad_norm": 0.5576398372650146, "learning_rate": 8.87382238969082e-05, "loss": 0.4597, "step": 10280 }, { "epoch": 3.884484711211778, "grad_norm": 0.5064070820808411, "learning_rate": 8.871207550589568e-05, "loss": 0.4612, "step": 10290 }, { "epoch": 3.8882597206493017, "grad_norm": 0.536137580871582, "learning_rate": 8.868590065554458e-05, "loss": 0.4681, "step": 10300 }, { "epoch": 3.8920347300868254, "grad_norm": 0.5211617350578308, "learning_rate": 8.865969936374519e-05, "loss": 0.458, "step": 10310 }, { "epoch": 3.8958097395243487, "grad_norm": 0.5458037853240967, "learning_rate": 8.863347164840581e-05, "loss": 0.469, "step": 10320 }, { "epoch": 3.8995847489618725, "grad_norm": 0.5937955379486084, "learning_rate": 8.860721752745285e-05, "loss": 0.4657, "step": 10330 }, { "epoch": 3.9033597583993958, "grad_norm": 0.5774049758911133, "learning_rate": 8.858093701883077e-05, "loss": 0.454, "step": 10340 }, { "epoch": 3.9071347678369195, "grad_norm": 0.5954073667526245, "learning_rate": 8.8554630140502e-05, "loss": 0.4606, "step": 10350 }, { "epoch": 3.910909777274443, "grad_norm": 0.568393349647522, "learning_rate": 8.85282969104471e-05, "loss": 0.4559, "step": 10360 }, { "epoch": 3.914684786711967, "grad_norm": 0.7279831767082214, "learning_rate": 8.850193734666456e-05, "loss": 0.462, "step": 10370 }, { "epoch": 3.9184597961494902, "grad_norm": 0.610234260559082, "learning_rate": 8.84755514671709e-05, "loss": 0.4536, "step": 10380 }, { "epoch": 3.922234805587014, "grad_norm": 0.6353870630264282, "learning_rate": 8.84491392900006e-05, "loss": 0.456, "step": 10390 }, { "epoch": 3.9260098150245377, "grad_norm": 0.6264570951461792, "learning_rate": 8.842270083320617e-05, "loss": 0.4607, "step": 10400 }, { "epoch": 3.929784824462061, "grad_norm": 0.5441379547119141, "learning_rate": 8.839623611485801e-05, "loss": 0.4638, "step": 10410 }, { "epoch": 3.9335598338995847, "grad_norm": 0.5710632801055908, "learning_rate": 8.836974515304453e-05, "loss": 0.4552, "step": 10420 }, { "epoch": 3.9373348433371085, "grad_norm": 0.6466676592826843, "learning_rate": 8.834322796587204e-05, "loss": 0.4643, "step": 10430 }, { "epoch": 3.9411098527746318, "grad_norm": 0.5665135383605957, "learning_rate": 8.831668457146478e-05, "loss": 0.4565, "step": 10440 }, { "epoch": 3.9448848622121555, "grad_norm": 0.48944804072380066, "learning_rate": 8.829011498796493e-05, "loss": 0.4624, "step": 10450 }, { "epoch": 3.9486598716496792, "grad_norm": 0.4873522222042084, "learning_rate": 8.826351923353253e-05, "loss": 0.4627, "step": 10460 }, { "epoch": 3.9524348810872025, "grad_norm": 0.7126550674438477, "learning_rate": 8.823689732634555e-05, "loss": 0.4545, "step": 10470 }, { "epoch": 3.9562098905247263, "grad_norm": 0.5510223507881165, "learning_rate": 8.82102492845998e-05, "loss": 0.466, "step": 10480 }, { "epoch": 3.95998489996225, "grad_norm": 0.5616976022720337, "learning_rate": 8.818357512650896e-05, "loss": 0.4593, "step": 10490 }, { "epoch": 3.9637599093997737, "grad_norm": 0.523526668548584, "learning_rate": 8.815687487030458e-05, "loss": 0.4557, "step": 10500 }, { "epoch": 3.967534918837297, "grad_norm": 0.5704669952392578, "learning_rate": 8.8130148534236e-05, "loss": 0.4687, "step": 10510 }, { "epoch": 3.9713099282748208, "grad_norm": 0.6883910894393921, "learning_rate": 8.810339613657047e-05, "loss": 0.4598, "step": 10520 }, { "epoch": 3.975084937712344, "grad_norm": 0.9575225114822388, "learning_rate": 8.807661769559295e-05, "loss": 0.4557, "step": 10530 }, { "epoch": 3.978859947149868, "grad_norm": 0.5738497376441956, "learning_rate": 8.804981322960628e-05, "loss": 0.4612, "step": 10540 }, { "epoch": 3.9826349565873915, "grad_norm": 0.530934751033783, "learning_rate": 8.802298275693106e-05, "loss": 0.4532, "step": 10550 }, { "epoch": 3.9864099660249153, "grad_norm": 0.5506056547164917, "learning_rate": 8.799612629590568e-05, "loss": 0.4542, "step": 10560 }, { "epoch": 3.9901849754624386, "grad_norm": 0.6080657243728638, "learning_rate": 8.796924386488624e-05, "loss": 0.4601, "step": 10570 }, { "epoch": 3.9939599848999623, "grad_norm": 0.5610547661781311, "learning_rate": 8.794233548224666e-05, "loss": 0.4545, "step": 10580 }, { "epoch": 3.9977349943374856, "grad_norm": 0.5351642966270447, "learning_rate": 8.791540116637853e-05, "loss": 0.4598, "step": 10590 }, { "epoch": 4.001510003775009, "grad_norm": 0.5863658785820007, "learning_rate": 8.788844093569124e-05, "loss": 0.4532, "step": 10600 }, { "epoch": 4.005285013212533, "grad_norm": 0.530657172203064, "learning_rate": 8.786145480861184e-05, "loss": 0.4474, "step": 10610 }, { "epoch": 4.009060022650057, "grad_norm": 0.6831036806106567, "learning_rate": 8.783444280358507e-05, "loss": 0.4602, "step": 10620 }, { "epoch": 4.0128350320875805, "grad_norm": 0.579365074634552, "learning_rate": 8.780740493907342e-05, "loss": 0.4698, "step": 10630 }, { "epoch": 4.016610041525103, "grad_norm": 0.5720619559288025, "learning_rate": 8.778034123355698e-05, "loss": 0.4594, "step": 10640 }, { "epoch": 4.020385050962627, "grad_norm": 0.6572504043579102, "learning_rate": 8.775325170553357e-05, "loss": 0.4755, "step": 10650 }, { "epoch": 4.024160060400151, "grad_norm": 0.584198534488678, "learning_rate": 8.77261363735186e-05, "loss": 0.4579, "step": 10660 }, { "epoch": 4.027935069837675, "grad_norm": 0.5127620100975037, "learning_rate": 8.769899525604517e-05, "loss": 0.4559, "step": 10670 }, { "epoch": 4.031710079275198, "grad_norm": 0.5613308548927307, "learning_rate": 8.767182837166397e-05, "loss": 0.4607, "step": 10680 }, { "epoch": 4.035485088712722, "grad_norm": 0.49144065380096436, "learning_rate": 8.764463573894328e-05, "loss": 0.4517, "step": 10690 }, { "epoch": 4.039260098150246, "grad_norm": 0.533999502658844, "learning_rate": 8.761741737646902e-05, "loss": 0.4584, "step": 10700 }, { "epoch": 4.043035107587769, "grad_norm": 0.521372377872467, "learning_rate": 8.759017330284471e-05, "loss": 0.4653, "step": 10710 }, { "epoch": 4.046810117025292, "grad_norm": 0.6814863681793213, "learning_rate": 8.756290353669142e-05, "loss": 0.4665, "step": 10720 }, { "epoch": 4.050585126462816, "grad_norm": 0.5394108891487122, "learning_rate": 8.753560809664774e-05, "loss": 0.4549, "step": 10730 }, { "epoch": 4.05436013590034, "grad_norm": 0.624742329120636, "learning_rate": 8.750828700136986e-05, "loss": 0.4544, "step": 10740 }, { "epoch": 4.058135145337864, "grad_norm": 0.6260218620300293, "learning_rate": 8.74809402695315e-05, "loss": 0.4608, "step": 10750 }, { "epoch": 4.061910154775387, "grad_norm": 0.5796566009521484, "learning_rate": 8.745356791982391e-05, "loss": 0.4623, "step": 10760 }, { "epoch": 4.06568516421291, "grad_norm": 0.5631840825080872, "learning_rate": 8.742616997095578e-05, "loss": 0.4543, "step": 10770 }, { "epoch": 4.069460173650434, "grad_norm": 0.6347076892852783, "learning_rate": 8.739874644165341e-05, "loss": 0.4613, "step": 10780 }, { "epoch": 4.073235183087958, "grad_norm": 0.5708393454551697, "learning_rate": 8.737129735066048e-05, "loss": 0.4609, "step": 10790 }, { "epoch": 4.077010192525481, "grad_norm": 0.5556269288063049, "learning_rate": 8.734382271673821e-05, "loss": 0.4542, "step": 10800 }, { "epoch": 4.080785201963005, "grad_norm": 0.587650716304779, "learning_rate": 8.731632255866525e-05, "loss": 0.4546, "step": 10810 }, { "epoch": 4.084560211400529, "grad_norm": 0.5899844765663147, "learning_rate": 8.728879689523767e-05, "loss": 0.4493, "step": 10820 }, { "epoch": 4.088335220838052, "grad_norm": 0.5399772524833679, "learning_rate": 8.726124574526905e-05, "loss": 0.4597, "step": 10830 }, { "epoch": 4.092110230275575, "grad_norm": 0.534979522228241, "learning_rate": 8.72336691275903e-05, "loss": 0.4519, "step": 10840 }, { "epoch": 4.095885239713099, "grad_norm": 0.5515267848968506, "learning_rate": 8.720606706104979e-05, "loss": 0.453, "step": 10850 }, { "epoch": 4.099660249150623, "grad_norm": 0.5918541550636292, "learning_rate": 8.71784395645133e-05, "loss": 0.4586, "step": 10860 }, { "epoch": 4.103435258588147, "grad_norm": 0.5579758286476135, "learning_rate": 8.715078665686392e-05, "loss": 0.4613, "step": 10870 }, { "epoch": 4.10721026802567, "grad_norm": 0.5700817108154297, "learning_rate": 8.712310835700218e-05, "loss": 0.4591, "step": 10880 }, { "epoch": 4.110985277463194, "grad_norm": 0.5911275148391724, "learning_rate": 8.709540468384591e-05, "loss": 0.4544, "step": 10890 }, { "epoch": 4.114760286900717, "grad_norm": 0.6004135012626648, "learning_rate": 8.706767565633033e-05, "loss": 0.4484, "step": 10900 }, { "epoch": 4.118535296338241, "grad_norm": 0.6015802621841431, "learning_rate": 8.7039921293408e-05, "loss": 0.4672, "step": 10910 }, { "epoch": 4.122310305775764, "grad_norm": 0.6620537638664246, "learning_rate": 8.70121416140487e-05, "loss": 0.4506, "step": 10920 }, { "epoch": 4.126085315213288, "grad_norm": 0.5017173886299133, "learning_rate": 8.698433663723962e-05, "loss": 0.4554, "step": 10930 }, { "epoch": 4.129860324650812, "grad_norm": 0.5419957637786865, "learning_rate": 8.695650638198518e-05, "loss": 0.456, "step": 10940 }, { "epoch": 4.133635334088336, "grad_norm": 0.6066375374794006, "learning_rate": 8.692865086730713e-05, "loss": 0.4519, "step": 10950 }, { "epoch": 4.1374103435258585, "grad_norm": 0.5887925624847412, "learning_rate": 8.69007701122444e-05, "loss": 0.4557, "step": 10960 }, { "epoch": 4.141185352963382, "grad_norm": 0.5485467910766602, "learning_rate": 8.687286413585328e-05, "loss": 0.4606, "step": 10970 }, { "epoch": 4.144960362400906, "grad_norm": 0.5847924947738647, "learning_rate": 8.684493295720719e-05, "loss": 0.4481, "step": 10980 }, { "epoch": 4.14873537183843, "grad_norm": 0.6698928475379944, "learning_rate": 8.681697659539685e-05, "loss": 0.4543, "step": 10990 }, { "epoch": 4.152510381275953, "grad_norm": 0.578162431716919, "learning_rate": 8.678899506953019e-05, "loss": 0.4537, "step": 11000 }, { "epoch": 4.156285390713477, "grad_norm": 0.5858516097068787, "learning_rate": 8.676098839873227e-05, "loss": 0.4531, "step": 11010 }, { "epoch": 4.160060400151, "grad_norm": 0.6587435603141785, "learning_rate": 8.673295660214545e-05, "loss": 0.4573, "step": 11020 }, { "epoch": 4.163835409588524, "grad_norm": 0.5828598141670227, "learning_rate": 8.670489969892914e-05, "loss": 0.4543, "step": 11030 }, { "epoch": 4.1676104190260475, "grad_norm": 0.6038519740104675, "learning_rate": 8.667681770826e-05, "loss": 0.448, "step": 11040 }, { "epoch": 4.171385428463571, "grad_norm": 0.5508007407188416, "learning_rate": 8.66487106493318e-05, "loss": 0.4513, "step": 11050 }, { "epoch": 4.175160437901095, "grad_norm": 0.5896838903427124, "learning_rate": 8.662057854135544e-05, "loss": 0.4543, "step": 11060 }, { "epoch": 4.178935447338619, "grad_norm": 0.5867491960525513, "learning_rate": 8.659242140355897e-05, "loss": 0.449, "step": 11070 }, { "epoch": 4.182710456776142, "grad_norm": 0.6184877157211304, "learning_rate": 8.65642392551875e-05, "loss": 0.4619, "step": 11080 }, { "epoch": 4.186485466213665, "grad_norm": 0.6027196049690247, "learning_rate": 8.65360321155033e-05, "loss": 0.4591, "step": 11090 }, { "epoch": 4.190260475651189, "grad_norm": 0.588248610496521, "learning_rate": 8.650780000378566e-05, "loss": 0.4567, "step": 11100 }, { "epoch": 4.194035485088713, "grad_norm": 0.5748762488365173, "learning_rate": 8.647954293933096e-05, "loss": 0.4568, "step": 11110 }, { "epoch": 4.1978104945262364, "grad_norm": 0.5934643149375916, "learning_rate": 8.645126094145264e-05, "loss": 0.4538, "step": 11120 }, { "epoch": 4.20158550396376, "grad_norm": 0.6221973896026611, "learning_rate": 8.642295402948117e-05, "loss": 0.4602, "step": 11130 }, { "epoch": 4.205360513401284, "grad_norm": 0.5034570097923279, "learning_rate": 8.639462222276409e-05, "loss": 0.4543, "step": 11140 }, { "epoch": 4.209135522838807, "grad_norm": 0.6960034966468811, "learning_rate": 8.636626554066589e-05, "loss": 0.4548, "step": 11150 }, { "epoch": 4.2129105322763305, "grad_norm": 0.5671536326408386, "learning_rate": 8.633788400256811e-05, "loss": 0.4484, "step": 11160 }, { "epoch": 4.216685541713854, "grad_norm": 0.48511791229248047, "learning_rate": 8.630947762786927e-05, "loss": 0.4558, "step": 11170 }, { "epoch": 4.220460551151378, "grad_norm": 0.6003875732421875, "learning_rate": 8.628104643598483e-05, "loss": 0.4512, "step": 11180 }, { "epoch": 4.224235560588902, "grad_norm": 0.544647753238678, "learning_rate": 8.625259044634726e-05, "loss": 0.4552, "step": 11190 }, { "epoch": 4.228010570026425, "grad_norm": 0.5275722146034241, "learning_rate": 8.622410967840597e-05, "loss": 0.4498, "step": 11200 }, { "epoch": 4.231785579463948, "grad_norm": 0.5396122932434082, "learning_rate": 8.619560415162731e-05, "loss": 0.4526, "step": 11210 }, { "epoch": 4.235560588901472, "grad_norm": 0.5510291457176208, "learning_rate": 8.616707388549447e-05, "loss": 0.4465, "step": 11220 }, { "epoch": 4.239335598338996, "grad_norm": 0.6948357820510864, "learning_rate": 8.613851889950771e-05, "loss": 0.4498, "step": 11230 }, { "epoch": 4.2431106077765195, "grad_norm": 0.6219735145568848, "learning_rate": 8.610993921318402e-05, "loss": 0.4567, "step": 11240 }, { "epoch": 4.246885617214043, "grad_norm": 0.5520322322845459, "learning_rate": 8.608133484605738e-05, "loss": 0.4523, "step": 11250 }, { "epoch": 4.250660626651567, "grad_norm": 0.5849624872207642, "learning_rate": 8.605270581767859e-05, "loss": 0.4526, "step": 11260 }, { "epoch": 4.25443563608909, "grad_norm": 0.5052843689918518, "learning_rate": 8.602405214761536e-05, "loss": 0.4525, "step": 11270 }, { "epoch": 4.2582106455266135, "grad_norm": 0.518661618232727, "learning_rate": 8.599537385545215e-05, "loss": 0.4503, "step": 11280 }, { "epoch": 4.261985654964137, "grad_norm": 0.5502333045005798, "learning_rate": 8.596667096079032e-05, "loss": 0.4477, "step": 11290 }, { "epoch": 4.265760664401661, "grad_norm": 0.6355045437812805, "learning_rate": 8.593794348324806e-05, "loss": 0.4492, "step": 11300 }, { "epoch": 4.269535673839185, "grad_norm": 0.5576472878456116, "learning_rate": 8.590919144246028e-05, "loss": 0.4517, "step": 11310 }, { "epoch": 4.2733106832767085, "grad_norm": 0.5880599021911621, "learning_rate": 8.588041485807876e-05, "loss": 0.4478, "step": 11320 }, { "epoch": 4.277085692714232, "grad_norm": 0.5388513207435608, "learning_rate": 8.585161374977202e-05, "loss": 0.4523, "step": 11330 }, { "epoch": 4.280860702151755, "grad_norm": 0.5540359616279602, "learning_rate": 8.582278813722533e-05, "loss": 0.4484, "step": 11340 }, { "epoch": 4.284635711589279, "grad_norm": 0.6409743428230286, "learning_rate": 8.579393804014076e-05, "loss": 0.4489, "step": 11350 }, { "epoch": 4.2884107210268025, "grad_norm": 0.6246625185012817, "learning_rate": 8.576506347823703e-05, "loss": 0.4488, "step": 11360 }, { "epoch": 4.292185730464326, "grad_norm": 0.7034008502960205, "learning_rate": 8.573616447124968e-05, "loss": 0.4492, "step": 11370 }, { "epoch": 4.29596073990185, "grad_norm": 0.6897478103637695, "learning_rate": 8.570724103893086e-05, "loss": 0.4477, "step": 11380 }, { "epoch": 4.299735749339374, "grad_norm": 0.5276645421981812, "learning_rate": 8.567829320104951e-05, "loss": 0.4424, "step": 11390 }, { "epoch": 4.303510758776897, "grad_norm": 0.5786343812942505, "learning_rate": 8.564932097739118e-05, "loss": 0.4503, "step": 11400 }, { "epoch": 4.30728576821442, "grad_norm": 0.5901648998260498, "learning_rate": 8.562032438775811e-05, "loss": 0.4426, "step": 11410 }, { "epoch": 4.311060777651944, "grad_norm": 0.60638427734375, "learning_rate": 8.559130345196921e-05, "loss": 0.4519, "step": 11420 }, { "epoch": 4.314835787089468, "grad_norm": 0.6317540407180786, "learning_rate": 8.556225818986e-05, "loss": 0.4567, "step": 11430 }, { "epoch": 4.3186107965269915, "grad_norm": 0.5992687344551086, "learning_rate": 8.553318862128265e-05, "loss": 0.4547, "step": 11440 }, { "epoch": 4.322385805964515, "grad_norm": 0.6526298522949219, "learning_rate": 8.550409476610593e-05, "loss": 0.4417, "step": 11450 }, { "epoch": 4.326160815402038, "grad_norm": 0.5577118992805481, "learning_rate": 8.547497664421522e-05, "loss": 0.4517, "step": 11460 }, { "epoch": 4.329935824839562, "grad_norm": 0.6398622989654541, "learning_rate": 8.544583427551249e-05, "loss": 0.4474, "step": 11470 }, { "epoch": 4.333710834277086, "grad_norm": 0.5577737092971802, "learning_rate": 8.541666767991628e-05, "loss": 0.4508, "step": 11480 }, { "epoch": 4.337485843714609, "grad_norm": 0.5385574698448181, "learning_rate": 8.538747687736166e-05, "loss": 0.4417, "step": 11490 }, { "epoch": 4.341260853152133, "grad_norm": 0.595480740070343, "learning_rate": 8.535826188780026e-05, "loss": 0.4463, "step": 11500 }, { "epoch": 4.345035862589657, "grad_norm": 0.5585368871688843, "learning_rate": 8.532902273120029e-05, "loss": 0.4421, "step": 11510 }, { "epoch": 4.34881087202718, "grad_norm": 0.6641288995742798, "learning_rate": 8.52997594275464e-05, "loss": 0.444, "step": 11520 }, { "epoch": 4.352585881464703, "grad_norm": 0.5479051470756531, "learning_rate": 8.52704719968398e-05, "loss": 0.4611, "step": 11530 }, { "epoch": 4.356360890902227, "grad_norm": 0.5787330269813538, "learning_rate": 8.524116045909818e-05, "loss": 0.4524, "step": 11540 }, { "epoch": 4.360135900339751, "grad_norm": 0.5680203437805176, "learning_rate": 8.521182483435569e-05, "loss": 0.4497, "step": 11550 }, { "epoch": 4.363910909777275, "grad_norm": 0.6969168186187744, "learning_rate": 8.518246514266295e-05, "loss": 0.4511, "step": 11560 }, { "epoch": 4.367685919214798, "grad_norm": 0.6396414041519165, "learning_rate": 8.515308140408703e-05, "loss": 0.4435, "step": 11570 }, { "epoch": 4.371460928652322, "grad_norm": 0.6805716753005981, "learning_rate": 8.512367363871145e-05, "loss": 0.4603, "step": 11580 }, { "epoch": 4.375235938089845, "grad_norm": 0.712642252445221, "learning_rate": 8.509424186663614e-05, "loss": 0.4488, "step": 11590 }, { "epoch": 4.379010947527369, "grad_norm": 0.6053329110145569, "learning_rate": 8.506478610797743e-05, "loss": 0.4522, "step": 11600 }, { "epoch": 4.382785956964892, "grad_norm": 0.5076550245285034, "learning_rate": 8.503530638286805e-05, "loss": 0.4405, "step": 11610 }, { "epoch": 4.386560966402416, "grad_norm": 0.6560049057006836, "learning_rate": 8.500580271145712e-05, "loss": 0.4475, "step": 11620 }, { "epoch": 4.39033597583994, "grad_norm": 0.5975435972213745, "learning_rate": 8.497627511391014e-05, "loss": 0.4492, "step": 11630 }, { "epoch": 4.394110985277464, "grad_norm": 0.6190789341926575, "learning_rate": 8.494672361040891e-05, "loss": 0.4587, "step": 11640 }, { "epoch": 4.397885994714986, "grad_norm": 0.6516814827919006, "learning_rate": 8.491714822115162e-05, "loss": 0.4436, "step": 11650 }, { "epoch": 4.40166100415251, "grad_norm": 0.5875669121742249, "learning_rate": 8.488754896635277e-05, "loss": 0.445, "step": 11660 }, { "epoch": 4.405436013590034, "grad_norm": 0.49789679050445557, "learning_rate": 8.485792586624317e-05, "loss": 0.4551, "step": 11670 }, { "epoch": 4.409211023027558, "grad_norm": 0.6263242959976196, "learning_rate": 8.482827894106993e-05, "loss": 0.4457, "step": 11680 }, { "epoch": 4.412986032465081, "grad_norm": 0.5868949294090271, "learning_rate": 8.479860821109646e-05, "loss": 0.4527, "step": 11690 }, { "epoch": 4.416761041902605, "grad_norm": 0.6368350386619568, "learning_rate": 8.476891369660239e-05, "loss": 0.4464, "step": 11700 }, { "epoch": 4.420536051340128, "grad_norm": 0.5921139717102051, "learning_rate": 8.473919541788366e-05, "loss": 0.4423, "step": 11710 }, { "epoch": 4.424311060777652, "grad_norm": 0.5540608763694763, "learning_rate": 8.470945339525245e-05, "loss": 0.4407, "step": 11720 }, { "epoch": 4.428086070215175, "grad_norm": 0.5866498351097107, "learning_rate": 8.467968764903713e-05, "loss": 0.4408, "step": 11730 }, { "epoch": 4.431861079652699, "grad_norm": 0.6014940738677979, "learning_rate": 8.46498981995823e-05, "loss": 0.4541, "step": 11740 }, { "epoch": 4.435636089090223, "grad_norm": 0.5633965134620667, "learning_rate": 8.462008506724879e-05, "loss": 0.4428, "step": 11750 }, { "epoch": 4.439411098527747, "grad_norm": 0.5375277400016785, "learning_rate": 8.459024827241359e-05, "loss": 0.4449, "step": 11760 }, { "epoch": 4.4431861079652695, "grad_norm": 0.566908597946167, "learning_rate": 8.456038783546985e-05, "loss": 0.4425, "step": 11770 }, { "epoch": 4.446961117402793, "grad_norm": 0.5475563406944275, "learning_rate": 8.453050377682691e-05, "loss": 0.4469, "step": 11780 }, { "epoch": 4.450736126840317, "grad_norm": 0.5368761420249939, "learning_rate": 8.450059611691026e-05, "loss": 0.441, "step": 11790 }, { "epoch": 4.454511136277841, "grad_norm": 0.561191976070404, "learning_rate": 8.447066487616146e-05, "loss": 0.4444, "step": 11800 }, { "epoch": 4.458286145715364, "grad_norm": 0.6211469173431396, "learning_rate": 8.444071007503826e-05, "loss": 0.443, "step": 11810 }, { "epoch": 4.462061155152888, "grad_norm": 0.6266103982925415, "learning_rate": 8.441073173401449e-05, "loss": 0.4392, "step": 11820 }, { "epoch": 4.465836164590412, "grad_norm": 0.5911449193954468, "learning_rate": 8.438072987358006e-05, "loss": 0.4502, "step": 11830 }, { "epoch": 4.469611174027935, "grad_norm": 0.7175887227058411, "learning_rate": 8.435070451424094e-05, "loss": 0.4447, "step": 11840 }, { "epoch": 4.4733861834654585, "grad_norm": 0.5663455724716187, "learning_rate": 8.432065567651919e-05, "loss": 0.4387, "step": 11850 }, { "epoch": 4.477161192902982, "grad_norm": 0.8334063291549683, "learning_rate": 8.429058338095291e-05, "loss": 0.4453, "step": 11860 }, { "epoch": 4.480936202340506, "grad_norm": 0.5345563888549805, "learning_rate": 8.426048764809624e-05, "loss": 0.446, "step": 11870 }, { "epoch": 4.48471121177803, "grad_norm": 0.5361315608024597, "learning_rate": 8.423036849851932e-05, "loss": 0.4468, "step": 11880 }, { "epoch": 4.488486221215553, "grad_norm": 0.624729573726654, "learning_rate": 8.42002259528083e-05, "loss": 0.4496, "step": 11890 }, { "epoch": 4.492261230653076, "grad_norm": 0.5892227292060852, "learning_rate": 8.417006003156532e-05, "loss": 0.4537, "step": 11900 }, { "epoch": 4.4960362400906, "grad_norm": 0.5822799205780029, "learning_rate": 8.413987075540852e-05, "loss": 0.4537, "step": 11910 }, { "epoch": 4.499811249528124, "grad_norm": 0.656360924243927, "learning_rate": 8.4109658144972e-05, "loss": 0.4475, "step": 11920 }, { "epoch": 4.5035862589656475, "grad_norm": 0.6090147495269775, "learning_rate": 8.407942222090573e-05, "loss": 0.4528, "step": 11930 }, { "epoch": 4.507361268403171, "grad_norm": 0.6417983174324036, "learning_rate": 8.404916300387576e-05, "loss": 0.4479, "step": 11940 }, { "epoch": 4.511136277840695, "grad_norm": 0.6105488538742065, "learning_rate": 8.401888051456391e-05, "loss": 0.4563, "step": 11950 }, { "epoch": 4.514911287278219, "grad_norm": 0.6055814027786255, "learning_rate": 8.398857477366803e-05, "loss": 0.4442, "step": 11960 }, { "epoch": 4.5186862967157415, "grad_norm": 0.6824854612350464, "learning_rate": 8.395824580190178e-05, "loss": 0.4473, "step": 11970 }, { "epoch": 4.522461306153265, "grad_norm": 0.5766848921775818, "learning_rate": 8.392789361999473e-05, "loss": 0.4444, "step": 11980 }, { "epoch": 4.526236315590789, "grad_norm": 0.5585546493530273, "learning_rate": 8.38975182486923e-05, "loss": 0.4419, "step": 11990 }, { "epoch": 4.530011325028313, "grad_norm": 0.5684893727302551, "learning_rate": 8.386711970875581e-05, "loss": 0.4381, "step": 12000 }, { "epoch": 4.5337863344658365, "grad_norm": 0.5678910613059998, "learning_rate": 8.383669802096232e-05, "loss": 0.4394, "step": 12010 }, { "epoch": 4.537561343903359, "grad_norm": 0.5642674565315247, "learning_rate": 8.38062532061048e-05, "loss": 0.4401, "step": 12020 }, { "epoch": 4.541336353340883, "grad_norm": 0.5675539374351501, "learning_rate": 8.3775785284992e-05, "loss": 0.4403, "step": 12030 }, { "epoch": 4.545111362778407, "grad_norm": 0.6496412754058838, "learning_rate": 8.374529427844843e-05, "loss": 0.4428, "step": 12040 }, { "epoch": 4.5488863722159305, "grad_norm": 0.6716994643211365, "learning_rate": 8.371478020731442e-05, "loss": 0.4419, "step": 12050 }, { "epoch": 4.552661381653454, "grad_norm": 0.5777934193611145, "learning_rate": 8.368424309244607e-05, "loss": 0.4414, "step": 12060 }, { "epoch": 4.556436391090978, "grad_norm": 0.6052347421646118, "learning_rate": 8.365368295471517e-05, "loss": 0.4444, "step": 12070 }, { "epoch": 4.560211400528502, "grad_norm": 0.6728206276893616, "learning_rate": 8.362309981500931e-05, "loss": 0.4366, "step": 12080 }, { "epoch": 4.563986409966025, "grad_norm": 0.576066792011261, "learning_rate": 8.359249369423177e-05, "loss": 0.4433, "step": 12090 }, { "epoch": 4.567761419403548, "grad_norm": 0.566498875617981, "learning_rate": 8.356186461330155e-05, "loss": 0.444, "step": 12100 }, { "epoch": 4.571536428841072, "grad_norm": 0.5168943405151367, "learning_rate": 8.353121259315334e-05, "loss": 0.4345, "step": 12110 }, { "epoch": 4.575311438278596, "grad_norm": 0.6019729375839233, "learning_rate": 8.350053765473751e-05, "loss": 0.4342, "step": 12120 }, { "epoch": 4.5790864477161195, "grad_norm": 0.6395158767700195, "learning_rate": 8.346983981902005e-05, "loss": 0.4388, "step": 12130 }, { "epoch": 4.582861457153643, "grad_norm": 0.6709539890289307, "learning_rate": 8.343911910698271e-05, "loss": 0.4448, "step": 12140 }, { "epoch": 4.586636466591166, "grad_norm": 0.6900480389595032, "learning_rate": 8.340837553962278e-05, "loss": 0.4453, "step": 12150 }, { "epoch": 4.59041147602869, "grad_norm": 0.6444860696792603, "learning_rate": 8.337760913795316e-05, "loss": 0.4453, "step": 12160 }, { "epoch": 4.5941864854662136, "grad_norm": 0.5633404850959778, "learning_rate": 8.334681992300244e-05, "loss": 0.4447, "step": 12170 }, { "epoch": 4.597961494903737, "grad_norm": 0.5962954163551331, "learning_rate": 8.331600791581475e-05, "loss": 0.4448, "step": 12180 }, { "epoch": 4.601736504341261, "grad_norm": 0.5366525053977966, "learning_rate": 8.328517313744978e-05, "loss": 0.4433, "step": 12190 }, { "epoch": 4.605511513778785, "grad_norm": 0.5605337619781494, "learning_rate": 8.325431560898286e-05, "loss": 0.4383, "step": 12200 }, { "epoch": 4.6092865232163085, "grad_norm": 0.7788808941841125, "learning_rate": 8.322343535150478e-05, "loss": 0.4409, "step": 12210 }, { "epoch": 4.613061532653831, "grad_norm": 0.6166380643844604, "learning_rate": 8.319253238612191e-05, "loss": 0.4415, "step": 12220 }, { "epoch": 4.616836542091355, "grad_norm": 0.5688962936401367, "learning_rate": 8.316160673395614e-05, "loss": 0.4401, "step": 12230 }, { "epoch": 4.620611551528879, "grad_norm": 0.6172645688056946, "learning_rate": 8.313065841614487e-05, "loss": 0.4439, "step": 12240 }, { "epoch": 4.6243865609664025, "grad_norm": 0.7282325029373169, "learning_rate": 8.309968745384096e-05, "loss": 0.4383, "step": 12250 }, { "epoch": 4.628161570403926, "grad_norm": 0.519595742225647, "learning_rate": 8.306869386821282e-05, "loss": 0.4384, "step": 12260 }, { "epoch": 4.631936579841449, "grad_norm": 0.5448688268661499, "learning_rate": 8.30376776804442e-05, "loss": 0.4418, "step": 12270 }, { "epoch": 4.635711589278973, "grad_norm": 0.5774092674255371, "learning_rate": 8.300663891173443e-05, "loss": 0.4396, "step": 12280 }, { "epoch": 4.639486598716497, "grad_norm": 0.6279200315475464, "learning_rate": 8.297557758329822e-05, "loss": 0.4434, "step": 12290 }, { "epoch": 4.64326160815402, "grad_norm": 0.6159646511077881, "learning_rate": 8.294449371636564e-05, "loss": 0.4406, "step": 12300 }, { "epoch": 4.647036617591544, "grad_norm": 0.6391906142234802, "learning_rate": 8.291338733218226e-05, "loss": 0.4379, "step": 12310 }, { "epoch": 4.650811627029068, "grad_norm": 0.6429669260978699, "learning_rate": 8.2882258452009e-05, "loss": 0.4328, "step": 12320 }, { "epoch": 4.6545866364665915, "grad_norm": 0.5893004536628723, "learning_rate": 8.285110709712214e-05, "loss": 0.4385, "step": 12330 }, { "epoch": 4.658361645904114, "grad_norm": 0.5372092127799988, "learning_rate": 8.281993328881337e-05, "loss": 0.434, "step": 12340 }, { "epoch": 4.662136655341638, "grad_norm": 0.5648928284645081, "learning_rate": 8.278873704838964e-05, "loss": 0.4438, "step": 12350 }, { "epoch": 4.665911664779162, "grad_norm": 0.5447239875793457, "learning_rate": 8.275751839717334e-05, "loss": 0.4314, "step": 12360 }, { "epoch": 4.669686674216686, "grad_norm": 0.5694062113761902, "learning_rate": 8.272627735650208e-05, "loss": 0.443, "step": 12370 }, { "epoch": 4.673461683654209, "grad_norm": 0.6510103940963745, "learning_rate": 8.269501394772884e-05, "loss": 0.4352, "step": 12380 }, { "epoch": 4.677236693091733, "grad_norm": 0.5314978361129761, "learning_rate": 8.266372819222189e-05, "loss": 0.4385, "step": 12390 }, { "epoch": 4.681011702529256, "grad_norm": 0.6842690110206604, "learning_rate": 8.26324201113647e-05, "loss": 0.4378, "step": 12400 }, { "epoch": 4.68478671196678, "grad_norm": 0.6124461889266968, "learning_rate": 8.260108972655606e-05, "loss": 0.4416, "step": 12410 }, { "epoch": 4.688561721404303, "grad_norm": 0.568488359451294, "learning_rate": 8.256973705921e-05, "loss": 0.4378, "step": 12420 }, { "epoch": 4.692336730841827, "grad_norm": 0.5795332193374634, "learning_rate": 8.25383621307558e-05, "loss": 0.4411, "step": 12430 }, { "epoch": 4.696111740279351, "grad_norm": 0.5848597884178162, "learning_rate": 8.25069649626379e-05, "loss": 0.4318, "step": 12440 }, { "epoch": 4.699886749716875, "grad_norm": 0.5752859115600586, "learning_rate": 8.247554557631596e-05, "loss": 0.4373, "step": 12450 }, { "epoch": 4.703661759154398, "grad_norm": 0.5697815418243408, "learning_rate": 8.244410399326483e-05, "loss": 0.4434, "step": 12460 }, { "epoch": 4.707436768591921, "grad_norm": 0.5662826895713806, "learning_rate": 8.241264023497457e-05, "loss": 0.4412, "step": 12470 }, { "epoch": 4.711211778029445, "grad_norm": 0.5779763460159302, "learning_rate": 8.238115432295034e-05, "loss": 0.4347, "step": 12480 }, { "epoch": 4.714986787466969, "grad_norm": 0.5694817304611206, "learning_rate": 8.234964627871247e-05, "loss": 0.44, "step": 12490 }, { "epoch": 4.718761796904492, "grad_norm": 0.6271215677261353, "learning_rate": 8.231811612379639e-05, "loss": 0.4388, "step": 12500 }, { "epoch": 4.722536806342016, "grad_norm": 0.647670567035675, "learning_rate": 8.228656387975268e-05, "loss": 0.4424, "step": 12510 }, { "epoch": 4.726311815779539, "grad_norm": 0.58333820104599, "learning_rate": 8.225498956814702e-05, "loss": 0.4432, "step": 12520 }, { "epoch": 4.730086825217063, "grad_norm": 0.5770929455757141, "learning_rate": 8.222339321056014e-05, "loss": 0.4366, "step": 12530 }, { "epoch": 4.733861834654586, "grad_norm": 0.5523325800895691, "learning_rate": 8.219177482858785e-05, "loss": 0.4332, "step": 12540 }, { "epoch": 4.73763684409211, "grad_norm": 0.8121219873428345, "learning_rate": 8.216013444384099e-05, "loss": 0.4383, "step": 12550 }, { "epoch": 4.741411853529634, "grad_norm": 0.5820472240447998, "learning_rate": 8.21284720779455e-05, "loss": 0.4418, "step": 12560 }, { "epoch": 4.745186862967158, "grad_norm": 0.6363267302513123, "learning_rate": 8.209678775254231e-05, "loss": 0.4436, "step": 12570 }, { "epoch": 4.748961872404681, "grad_norm": 0.5754694938659668, "learning_rate": 8.206508148928733e-05, "loss": 0.448, "step": 12580 }, { "epoch": 4.752736881842204, "grad_norm": 0.6008169054985046, "learning_rate": 8.203335330985151e-05, "loss": 0.4448, "step": 12590 }, { "epoch": 4.756511891279728, "grad_norm": 0.6524280905723572, "learning_rate": 8.200160323592076e-05, "loss": 0.4572, "step": 12600 }, { "epoch": 4.760286900717252, "grad_norm": 0.5874181389808655, "learning_rate": 8.196983128919598e-05, "loss": 0.4404, "step": 12610 }, { "epoch": 4.764061910154775, "grad_norm": 0.8897712826728821, "learning_rate": 8.193803749139295e-05, "loss": 0.4319, "step": 12620 }, { "epoch": 4.767836919592299, "grad_norm": 0.5824182033538818, "learning_rate": 8.190622186424244e-05, "loss": 0.4436, "step": 12630 }, { "epoch": 4.771611929029823, "grad_norm": 0.6077965497970581, "learning_rate": 8.187438442949016e-05, "loss": 0.4422, "step": 12640 }, { "epoch": 4.775386938467346, "grad_norm": 0.7708337903022766, "learning_rate": 8.184252520889668e-05, "loss": 0.4312, "step": 12650 }, { "epoch": 4.7791619479048695, "grad_norm": 0.5734137296676636, "learning_rate": 8.181064422423748e-05, "loss": 0.4382, "step": 12660 }, { "epoch": 4.782936957342393, "grad_norm": 0.5938032865524292, "learning_rate": 8.177874149730289e-05, "loss": 0.4417, "step": 12670 }, { "epoch": 4.786711966779917, "grad_norm": 0.6791886687278748, "learning_rate": 8.174681704989816e-05, "loss": 0.4365, "step": 12680 }, { "epoch": 4.790486976217441, "grad_norm": 0.6398690342903137, "learning_rate": 8.171487090384333e-05, "loss": 0.4407, "step": 12690 }, { "epoch": 4.794261985654964, "grad_norm": 0.6101489067077637, "learning_rate": 8.168290308097328e-05, "loss": 0.4296, "step": 12700 }, { "epoch": 4.798036995092488, "grad_norm": 0.6080114841461182, "learning_rate": 8.165091360313774e-05, "loss": 0.4463, "step": 12710 }, { "epoch": 4.801812004530011, "grad_norm": 0.5545832514762878, "learning_rate": 8.161890249220119e-05, "loss": 0.4354, "step": 12720 }, { "epoch": 4.805587013967535, "grad_norm": 0.6806719899177551, "learning_rate": 8.158686977004295e-05, "loss": 0.4329, "step": 12730 }, { "epoch": 4.8093620234050585, "grad_norm": 0.6922208070755005, "learning_rate": 8.155481545855706e-05, "loss": 0.4361, "step": 12740 }, { "epoch": 4.813137032842582, "grad_norm": 0.6132349371910095, "learning_rate": 8.152273957965233e-05, "loss": 0.4392, "step": 12750 }, { "epoch": 4.816912042280106, "grad_norm": 0.5297583341598511, "learning_rate": 8.149064215525237e-05, "loss": 0.4371, "step": 12760 }, { "epoch": 4.82068705171763, "grad_norm": 0.5206873416900635, "learning_rate": 8.14585232072954e-05, "loss": 0.4358, "step": 12770 }, { "epoch": 4.8244620611551525, "grad_norm": 0.5716885328292847, "learning_rate": 8.142638275773449e-05, "loss": 0.4391, "step": 12780 }, { "epoch": 4.828237070592676, "grad_norm": 0.5915391445159912, "learning_rate": 8.139422082853729e-05, "loss": 0.4369, "step": 12790 }, { "epoch": 4.8320120800302, "grad_norm": 0.5491467118263245, "learning_rate": 8.136203744168618e-05, "loss": 0.4377, "step": 12800 }, { "epoch": 4.835787089467724, "grad_norm": 0.5877269506454468, "learning_rate": 8.132983261917819e-05, "loss": 0.4327, "step": 12810 }, { "epoch": 4.8395620989052475, "grad_norm": 0.8957927227020264, "learning_rate": 8.129760638302504e-05, "loss": 0.4363, "step": 12820 }, { "epoch": 4.843337108342771, "grad_norm": 0.6469703912734985, "learning_rate": 8.126535875525305e-05, "loss": 0.4354, "step": 12830 }, { "epoch": 4.847112117780295, "grad_norm": 0.5585578680038452, "learning_rate": 8.123308975790316e-05, "loss": 0.4417, "step": 12840 }, { "epoch": 4.850887127217818, "grad_norm": 0.858641505241394, "learning_rate": 8.120079941303094e-05, "loss": 0.4365, "step": 12850 }, { "epoch": 4.8546621366553415, "grad_norm": 0.6256632804870605, "learning_rate": 8.116848774270651e-05, "loss": 0.44, "step": 12860 }, { "epoch": 4.858437146092865, "grad_norm": 0.6245275735855103, "learning_rate": 8.113615476901461e-05, "loss": 0.4331, "step": 12870 }, { "epoch": 4.862212155530389, "grad_norm": 0.6392505168914795, "learning_rate": 8.110380051405454e-05, "loss": 0.4316, "step": 12880 }, { "epoch": 4.865987164967913, "grad_norm": 0.6834761500358582, "learning_rate": 8.107142499994009e-05, "loss": 0.4369, "step": 12890 }, { "epoch": 4.869762174405436, "grad_norm": 0.6249173879623413, "learning_rate": 8.103902824879966e-05, "loss": 0.4378, "step": 12900 }, { "epoch": 4.873537183842959, "grad_norm": 0.7045850157737732, "learning_rate": 8.10066102827761e-05, "loss": 0.4355, "step": 12910 }, { "epoch": 4.877312193280483, "grad_norm": 0.6210693120956421, "learning_rate": 8.097417112402676e-05, "loss": 0.4387, "step": 12920 }, { "epoch": 4.881087202718007, "grad_norm": 0.6081416010856628, "learning_rate": 8.094171079472355e-05, "loss": 0.4362, "step": 12930 }, { "epoch": 4.8848622121555305, "grad_norm": 0.651164710521698, "learning_rate": 8.090922931705277e-05, "loss": 0.4349, "step": 12940 }, { "epoch": 4.888637221593054, "grad_norm": 0.6259034276008606, "learning_rate": 8.08767267132152e-05, "loss": 0.4379, "step": 12950 }, { "epoch": 4.892412231030578, "grad_norm": 0.5768272876739502, "learning_rate": 8.084420300542608e-05, "loss": 0.4296, "step": 12960 }, { "epoch": 4.896187240468101, "grad_norm": 0.6301991939544678, "learning_rate": 8.081165821591505e-05, "loss": 0.434, "step": 12970 }, { "epoch": 4.899962249905625, "grad_norm": 0.6101955771446228, "learning_rate": 8.077909236692615e-05, "loss": 0.4301, "step": 12980 }, { "epoch": 4.903737259343148, "grad_norm": 0.6158044338226318, "learning_rate": 8.074650548071787e-05, "loss": 0.4378, "step": 12990 }, { "epoch": 4.907512268780672, "grad_norm": 0.5665403604507446, "learning_rate": 8.071389757956301e-05, "loss": 0.436, "step": 13000 }, { "epoch": 4.911287278218196, "grad_norm": 0.6046426892280579, "learning_rate": 8.068126868574876e-05, "loss": 0.4314, "step": 13010 }, { "epoch": 4.9150622876557195, "grad_norm": 0.5915471315383911, "learning_rate": 8.064861882157668e-05, "loss": 0.435, "step": 13020 }, { "epoch": 4.918837297093242, "grad_norm": 0.6827958226203918, "learning_rate": 8.061594800936263e-05, "loss": 0.4365, "step": 13030 }, { "epoch": 4.922612306530766, "grad_norm": 0.5622991323471069, "learning_rate": 8.058325627143681e-05, "loss": 0.4344, "step": 13040 }, { "epoch": 4.92638731596829, "grad_norm": 0.5483256578445435, "learning_rate": 8.055054363014372e-05, "loss": 0.4383, "step": 13050 }, { "epoch": 4.930162325405814, "grad_norm": 0.6701416969299316, "learning_rate": 8.051781010784211e-05, "loss": 0.4397, "step": 13060 }, { "epoch": 4.933937334843337, "grad_norm": 0.5472255945205688, "learning_rate": 8.048505572690506e-05, "loss": 0.4302, "step": 13070 }, { "epoch": 4.937712344280861, "grad_norm": 0.655945897102356, "learning_rate": 8.045228050971988e-05, "loss": 0.4367, "step": 13080 }, { "epoch": 4.941487353718385, "grad_norm": 0.5672075748443604, "learning_rate": 8.041948447868814e-05, "loss": 0.4292, "step": 13090 }, { "epoch": 4.945262363155908, "grad_norm": 0.7679628133773804, "learning_rate": 8.038666765622558e-05, "loss": 0.4333, "step": 13100 }, { "epoch": 4.949037372593431, "grad_norm": 0.5666245818138123, "learning_rate": 8.03538300647622e-05, "loss": 0.443, "step": 13110 }, { "epoch": 4.952812382030955, "grad_norm": 0.6425677537918091, "learning_rate": 8.03209717267422e-05, "loss": 0.4288, "step": 13120 }, { "epoch": 4.956587391468479, "grad_norm": 0.585444450378418, "learning_rate": 8.028809266462395e-05, "loss": 0.4317, "step": 13130 }, { "epoch": 4.9603624009060026, "grad_norm": 0.6069588661193848, "learning_rate": 8.025519290087994e-05, "loss": 0.4352, "step": 13140 }, { "epoch": 4.964137410343525, "grad_norm": 0.5473458170890808, "learning_rate": 8.022227245799688e-05, "loss": 0.4407, "step": 13150 }, { "epoch": 4.967912419781049, "grad_norm": 0.5702902674674988, "learning_rate": 8.018933135847557e-05, "loss": 0.4293, "step": 13160 }, { "epoch": 4.971687429218573, "grad_norm": 0.6425644755363464, "learning_rate": 8.015636962483096e-05, "loss": 0.4394, "step": 13170 }, { "epoch": 4.975462438656097, "grad_norm": 0.5478054285049438, "learning_rate": 8.012338727959205e-05, "loss": 0.4267, "step": 13180 }, { "epoch": 4.97923744809362, "grad_norm": 0.6049672961235046, "learning_rate": 8.009038434530198e-05, "loss": 0.4381, "step": 13190 }, { "epoch": 4.983012457531144, "grad_norm": 0.5757096409797668, "learning_rate": 8.005736084451796e-05, "loss": 0.4292, "step": 13200 }, { "epoch": 4.986787466968668, "grad_norm": 0.6315653920173645, "learning_rate": 8.002431679981122e-05, "loss": 0.4249, "step": 13210 }, { "epoch": 4.990562476406191, "grad_norm": 0.6188434958457947, "learning_rate": 7.999125223376706e-05, "loss": 0.4312, "step": 13220 }, { "epoch": 4.994337485843714, "grad_norm": 0.5495557188987732, "learning_rate": 7.99581671689848e-05, "loss": 0.4272, "step": 13230 }, { "epoch": 4.998112495281238, "grad_norm": 0.6032848954200745, "learning_rate": 7.992506162807775e-05, "loss": 0.4378, "step": 13240 }, { "epoch": 5.001887504718762, "grad_norm": 0.563420832157135, "learning_rate": 7.989193563367328e-05, "loss": 0.4333, "step": 13250 }, { "epoch": 5.005662514156286, "grad_norm": 0.5446894764900208, "learning_rate": 7.985878920841266e-05, "loss": 0.4252, "step": 13260 }, { "epoch": 5.009437523593809, "grad_norm": 0.6635614633560181, "learning_rate": 7.982562237495117e-05, "loss": 0.4289, "step": 13270 }, { "epoch": 5.013212533031332, "grad_norm": 0.6846327781677246, "learning_rate": 7.979243515595802e-05, "loss": 0.4408, "step": 13280 }, { "epoch": 5.016987542468856, "grad_norm": 0.6627950668334961, "learning_rate": 7.975922757411636e-05, "loss": 0.4443, "step": 13290 }, { "epoch": 5.02076255190638, "grad_norm": 0.5397953987121582, "learning_rate": 7.972599965212329e-05, "loss": 0.4393, "step": 13300 }, { "epoch": 5.024537561343903, "grad_norm": 0.5803020000457764, "learning_rate": 7.969275141268973e-05, "loss": 0.4316, "step": 13310 }, { "epoch": 5.028312570781427, "grad_norm": 0.625580370426178, "learning_rate": 7.96594828785406e-05, "loss": 0.4343, "step": 13320 }, { "epoch": 5.032087580218951, "grad_norm": 0.5824500918388367, "learning_rate": 7.962619407241456e-05, "loss": 0.4314, "step": 13330 }, { "epoch": 5.035862589656474, "grad_norm": 0.6776422262191772, "learning_rate": 7.959288501706424e-05, "loss": 0.444, "step": 13340 }, { "epoch": 5.0396375990939974, "grad_norm": 0.5621914267539978, "learning_rate": 7.955955573525605e-05, "loss": 0.435, "step": 13350 }, { "epoch": 5.043412608531521, "grad_norm": 0.6005207300186157, "learning_rate": 7.952620624977026e-05, "loss": 0.4282, "step": 13360 }, { "epoch": 5.047187617969045, "grad_norm": 0.5816413760185242, "learning_rate": 7.949283658340089e-05, "loss": 0.4396, "step": 13370 }, { "epoch": 5.050962627406569, "grad_norm": 0.6071826219558716, "learning_rate": 7.945944675895585e-05, "loss": 0.4344, "step": 13380 }, { "epoch": 5.054737636844092, "grad_norm": 0.5671345591545105, "learning_rate": 7.942603679925671e-05, "loss": 0.4225, "step": 13390 }, { "epoch": 5.058512646281616, "grad_norm": 0.5134578943252563, "learning_rate": 7.93926067271389e-05, "loss": 0.4412, "step": 13400 }, { "epoch": 5.062287655719139, "grad_norm": 0.5748825669288635, "learning_rate": 7.935915656545155e-05, "loss": 0.4246, "step": 13410 }, { "epoch": 5.066062665156663, "grad_norm": 0.6098251938819885, "learning_rate": 7.932568633705752e-05, "loss": 0.4316, "step": 13420 }, { "epoch": 5.069837674594186, "grad_norm": 0.5781272649765015, "learning_rate": 7.929219606483341e-05, "loss": 0.4375, "step": 13430 }, { "epoch": 5.07361268403171, "grad_norm": 0.5721918940544128, "learning_rate": 7.925868577166948e-05, "loss": 0.4328, "step": 13440 }, { "epoch": 5.077387693469234, "grad_norm": 0.6408784985542297, "learning_rate": 7.922515548046974e-05, "loss": 0.4351, "step": 13450 }, { "epoch": 5.081162702906758, "grad_norm": 0.5710509419441223, "learning_rate": 7.919160521415179e-05, "loss": 0.4305, "step": 13460 }, { "epoch": 5.0849377123442805, "grad_norm": 0.6249564290046692, "learning_rate": 7.915803499564694e-05, "loss": 0.4246, "step": 13470 }, { "epoch": 5.088712721781804, "grad_norm": 0.6229961514472961, "learning_rate": 7.912444484790013e-05, "loss": 0.4307, "step": 13480 }, { "epoch": 5.092487731219328, "grad_norm": 0.5835480093955994, "learning_rate": 7.909083479386987e-05, "loss": 0.4262, "step": 13490 }, { "epoch": 5.096262740656852, "grad_norm": 0.6598435044288635, "learning_rate": 7.905720485652836e-05, "loss": 0.4326, "step": 13500 }, { "epoch": 5.100037750094375, "grad_norm": 0.6232707500457764, "learning_rate": 7.902355505886132e-05, "loss": 0.4312, "step": 13510 }, { "epoch": 5.103812759531899, "grad_norm": 0.5751008987426758, "learning_rate": 7.898988542386805e-05, "loss": 0.4318, "step": 13520 }, { "epoch": 5.107587768969422, "grad_norm": 0.7185839414596558, "learning_rate": 7.895619597456147e-05, "loss": 0.431, "step": 13530 }, { "epoch": 5.111362778406946, "grad_norm": 0.6064713597297668, "learning_rate": 7.892248673396798e-05, "loss": 0.4428, "step": 13540 }, { "epoch": 5.1151377878444695, "grad_norm": 0.5585970282554626, "learning_rate": 7.888875772512754e-05, "loss": 0.4285, "step": 13550 }, { "epoch": 5.118912797281993, "grad_norm": 0.590541660785675, "learning_rate": 7.885500897109359e-05, "loss": 0.4327, "step": 13560 }, { "epoch": 5.122687806719517, "grad_norm": 0.6079216003417969, "learning_rate": 7.882124049493309e-05, "loss": 0.4219, "step": 13570 }, { "epoch": 5.126462816157041, "grad_norm": 0.6529505848884583, "learning_rate": 7.878745231972649e-05, "loss": 0.4195, "step": 13580 }, { "epoch": 5.130237825594564, "grad_norm": 0.6022513508796692, "learning_rate": 7.875364446856766e-05, "loss": 0.4318, "step": 13590 }, { "epoch": 5.134012835032087, "grad_norm": 0.6363440752029419, "learning_rate": 7.871981696456398e-05, "loss": 0.4323, "step": 13600 }, { "epoch": 5.137787844469611, "grad_norm": 0.6193153262138367, "learning_rate": 7.868596983083623e-05, "loss": 0.4285, "step": 13610 }, { "epoch": 5.141562853907135, "grad_norm": 0.7710440158843994, "learning_rate": 7.865210309051858e-05, "loss": 0.4266, "step": 13620 }, { "epoch": 5.1453378633446585, "grad_norm": 0.5872613787651062, "learning_rate": 7.861821676675863e-05, "loss": 0.4347, "step": 13630 }, { "epoch": 5.149112872782182, "grad_norm": 0.536893367767334, "learning_rate": 7.858431088271739e-05, "loss": 0.4304, "step": 13640 }, { "epoch": 5.152887882219706, "grad_norm": 0.5925413966178894, "learning_rate": 7.855038546156918e-05, "loss": 0.4284, "step": 13650 }, { "epoch": 5.156662891657229, "grad_norm": 0.562650203704834, "learning_rate": 7.851644052650173e-05, "loss": 0.4268, "step": 13660 }, { "epoch": 5.1604379010947525, "grad_norm": 0.5925480723381042, "learning_rate": 7.848247610071609e-05, "loss": 0.4251, "step": 13670 }, { "epoch": 5.164212910532276, "grad_norm": 0.6896264553070068, "learning_rate": 7.844849220742658e-05, "loss": 0.43, "step": 13680 }, { "epoch": 5.1679879199698, "grad_norm": 0.643406867980957, "learning_rate": 7.841448886986092e-05, "loss": 0.4298, "step": 13690 }, { "epoch": 5.171762929407324, "grad_norm": 0.5790614485740662, "learning_rate": 7.838046611126004e-05, "loss": 0.4333, "step": 13700 }, { "epoch": 5.1755379388448475, "grad_norm": 0.6230219602584839, "learning_rate": 7.834642395487819e-05, "loss": 0.432, "step": 13710 }, { "epoch": 5.17931294828237, "grad_norm": 0.6495296955108643, "learning_rate": 7.831236242398285e-05, "loss": 0.4336, "step": 13720 }, { "epoch": 5.183087957719894, "grad_norm": 0.5876701474189758, "learning_rate": 7.827828154185477e-05, "loss": 0.4362, "step": 13730 }, { "epoch": 5.186862967157418, "grad_norm": 0.5457055568695068, "learning_rate": 7.82441813317879e-05, "loss": 0.43, "step": 13740 }, { "epoch": 5.1906379765949415, "grad_norm": 0.618160605430603, "learning_rate": 7.821006181708944e-05, "loss": 0.4281, "step": 13750 }, { "epoch": 5.194412986032465, "grad_norm": 0.5848748087882996, "learning_rate": 7.81759230210797e-05, "loss": 0.4397, "step": 13760 }, { "epoch": 5.198187995469989, "grad_norm": 0.5916726589202881, "learning_rate": 7.814176496709227e-05, "loss": 0.4268, "step": 13770 }, { "epoch": 5.201963004907512, "grad_norm": 0.5741629004478455, "learning_rate": 7.810758767847385e-05, "loss": 0.4286, "step": 13780 }, { "epoch": 5.205738014345036, "grad_norm": 0.548261284828186, "learning_rate": 7.807339117858427e-05, "loss": 0.4242, "step": 13790 }, { "epoch": 5.209513023782559, "grad_norm": 0.5949342846870422, "learning_rate": 7.803917549079655e-05, "loss": 0.4291, "step": 13800 }, { "epoch": 5.213288033220083, "grad_norm": 0.611446738243103, "learning_rate": 7.800494063849679e-05, "loss": 0.4322, "step": 13810 }, { "epoch": 5.217063042657607, "grad_norm": 0.5855134129524231, "learning_rate": 7.797068664508416e-05, "loss": 0.4329, "step": 13820 }, { "epoch": 5.2208380520951305, "grad_norm": 0.615392804145813, "learning_rate": 7.793641353397096e-05, "loss": 0.4321, "step": 13830 }, { "epoch": 5.224613061532654, "grad_norm": 0.6607218980789185, "learning_rate": 7.790212132858253e-05, "loss": 0.4314, "step": 13840 }, { "epoch": 5.228388070970177, "grad_norm": 0.6020424365997314, "learning_rate": 7.786781005235728e-05, "loss": 0.4282, "step": 13850 }, { "epoch": 5.232163080407701, "grad_norm": 0.5580222606658936, "learning_rate": 7.783347972874662e-05, "loss": 0.4303, "step": 13860 }, { "epoch": 5.235938089845225, "grad_norm": 0.5938263535499573, "learning_rate": 7.779913038121504e-05, "loss": 0.4318, "step": 13870 }, { "epoch": 5.239713099282748, "grad_norm": 0.5820572376251221, "learning_rate": 7.776476203323997e-05, "loss": 0.4244, "step": 13880 }, { "epoch": 5.243488108720272, "grad_norm": 0.5716726779937744, "learning_rate": 7.773037470831185e-05, "loss": 0.4268, "step": 13890 }, { "epoch": 5.247263118157796, "grad_norm": 0.6699029803276062, "learning_rate": 7.76959684299341e-05, "loss": 0.4448, "step": 13900 }, { "epoch": 5.251038127595319, "grad_norm": 0.6544981002807617, "learning_rate": 7.76615432216231e-05, "loss": 0.4324, "step": 13910 }, { "epoch": 5.254813137032842, "grad_norm": 0.6434966325759888, "learning_rate": 7.762709910690811e-05, "loss": 0.431, "step": 13920 }, { "epoch": 5.258588146470366, "grad_norm": 0.6706520915031433, "learning_rate": 7.759263610933141e-05, "loss": 0.4239, "step": 13930 }, { "epoch": 5.26236315590789, "grad_norm": 0.7404385209083557, "learning_rate": 7.755815425244811e-05, "loss": 0.4211, "step": 13940 }, { "epoch": 5.266138165345414, "grad_norm": 0.750700831413269, "learning_rate": 7.752365355982624e-05, "loss": 0.4297, "step": 13950 }, { "epoch": 5.269913174782937, "grad_norm": 0.6595298051834106, "learning_rate": 7.748913405504668e-05, "loss": 0.4395, "step": 13960 }, { "epoch": 5.27368818422046, "grad_norm": 0.7262568473815918, "learning_rate": 7.745459576170322e-05, "loss": 0.4189, "step": 13970 }, { "epoch": 5.277463193657984, "grad_norm": 0.6093210577964783, "learning_rate": 7.742003870340242e-05, "loss": 0.4403, "step": 13980 }, { "epoch": 5.281238203095508, "grad_norm": 0.6888540387153625, "learning_rate": 7.738546290376373e-05, "loss": 0.4357, "step": 13990 }, { "epoch": 5.285013212533031, "grad_norm": 0.9906258583068848, "learning_rate": 7.735086838641937e-05, "loss": 0.4655, "step": 14000 }, { "epoch": 5.288788221970555, "grad_norm": 0.6373776197433472, "learning_rate": 7.731625517501437e-05, "loss": 0.4296, "step": 14010 }, { "epoch": 5.292563231408079, "grad_norm": 0.5870838761329651, "learning_rate": 7.728162329320655e-05, "loss": 0.4318, "step": 14020 }, { "epoch": 5.296338240845602, "grad_norm": 0.5722277164459229, "learning_rate": 7.724697276466645e-05, "loss": 0.426, "step": 14030 }, { "epoch": 5.300113250283125, "grad_norm": 0.595984697341919, "learning_rate": 7.721230361307738e-05, "loss": 0.4287, "step": 14040 }, { "epoch": 5.303888259720649, "grad_norm": 2.9559850692749023, "learning_rate": 7.71776158621354e-05, "loss": 0.4298, "step": 14050 }, { "epoch": 5.307663269158173, "grad_norm": 0.5847383737564087, "learning_rate": 7.714290953554925e-05, "loss": 0.4304, "step": 14060 }, { "epoch": 5.311438278595697, "grad_norm": 0.5809390544891357, "learning_rate": 7.710818465704037e-05, "loss": 0.4305, "step": 14070 }, { "epoch": 5.31521328803322, "grad_norm": 0.5706303119659424, "learning_rate": 7.707344125034288e-05, "loss": 0.4256, "step": 14080 }, { "epoch": 5.318988297470744, "grad_norm": 0.5360699892044067, "learning_rate": 7.703867933920359e-05, "loss": 0.4197, "step": 14090 }, { "epoch": 5.322763306908267, "grad_norm": 0.5971339344978333, "learning_rate": 7.700389894738194e-05, "loss": 0.4181, "step": 14100 }, { "epoch": 5.326538316345791, "grad_norm": 0.6393430233001709, "learning_rate": 7.696910009864999e-05, "loss": 0.4214, "step": 14110 }, { "epoch": 5.330313325783314, "grad_norm": 0.5459288358688354, "learning_rate": 7.693428281679241e-05, "loss": 0.422, "step": 14120 }, { "epoch": 5.334088335220838, "grad_norm": 0.516596794128418, "learning_rate": 7.689944712560652e-05, "loss": 0.4297, "step": 14130 }, { "epoch": 5.337863344658362, "grad_norm": 0.5618916153907776, "learning_rate": 7.686459304890214e-05, "loss": 0.426, "step": 14140 }, { "epoch": 5.341638354095886, "grad_norm": 0.6064931750297546, "learning_rate": 7.682972061050175e-05, "loss": 0.424, "step": 14150 }, { "epoch": 5.3454133635334085, "grad_norm": 0.6051250696182251, "learning_rate": 7.679482983424032e-05, "loss": 0.4242, "step": 14160 }, { "epoch": 5.349188372970932, "grad_norm": 0.6006349921226501, "learning_rate": 7.675992074396534e-05, "loss": 0.4295, "step": 14170 }, { "epoch": 5.352963382408456, "grad_norm": 1.0653759241104126, "learning_rate": 7.672499336353687e-05, "loss": 0.4233, "step": 14180 }, { "epoch": 5.35673839184598, "grad_norm": 0.6030228734016418, "learning_rate": 7.669004771682744e-05, "loss": 0.4259, "step": 14190 }, { "epoch": 5.360513401283503, "grad_norm": 0.6684322357177734, "learning_rate": 7.665508382772206e-05, "loss": 0.4259, "step": 14200 }, { "epoch": 5.364288410721027, "grad_norm": 0.6103203892707825, "learning_rate": 7.662010172011824e-05, "loss": 0.4218, "step": 14210 }, { "epoch": 5.36806342015855, "grad_norm": 0.5854779481887817, "learning_rate": 7.658510141792588e-05, "loss": 0.4223, "step": 14220 }, { "epoch": 5.371838429596074, "grad_norm": 0.6130976676940918, "learning_rate": 7.65500829450674e-05, "loss": 0.4235, "step": 14230 }, { "epoch": 5.3756134390335975, "grad_norm": 0.5876962542533875, "learning_rate": 7.651504632547759e-05, "loss": 0.4262, "step": 14240 }, { "epoch": 5.379388448471121, "grad_norm": 0.5816875100135803, "learning_rate": 7.647999158310364e-05, "loss": 0.4238, "step": 14250 }, { "epoch": 5.383163457908645, "grad_norm": 0.6344910860061646, "learning_rate": 7.644491874190512e-05, "loss": 0.4277, "step": 14260 }, { "epoch": 5.386938467346169, "grad_norm": 0.5912450551986694, "learning_rate": 7.6409827825854e-05, "loss": 0.4217, "step": 14270 }, { "epoch": 5.3907134767836915, "grad_norm": 0.5558596849441528, "learning_rate": 7.637471885893459e-05, "loss": 0.4291, "step": 14280 }, { "epoch": 5.394488486221215, "grad_norm": 0.6248513460159302, "learning_rate": 7.633959186514354e-05, "loss": 0.4281, "step": 14290 }, { "epoch": 5.398263495658739, "grad_norm": 0.623848557472229, "learning_rate": 7.630444686848984e-05, "loss": 0.4189, "step": 14300 }, { "epoch": 5.402038505096263, "grad_norm": 0.5933666825294495, "learning_rate": 7.626928389299471e-05, "loss": 0.4238, "step": 14310 }, { "epoch": 5.4058135145337864, "grad_norm": 0.6335635781288147, "learning_rate": 7.623410296269175e-05, "loss": 0.4352, "step": 14320 }, { "epoch": 5.40958852397131, "grad_norm": 0.5906968116760254, "learning_rate": 7.61989041016268e-05, "loss": 0.4217, "step": 14330 }, { "epoch": 5.413363533408834, "grad_norm": 1.1786829233169556, "learning_rate": 7.616368733385793e-05, "loss": 0.4292, "step": 14340 }, { "epoch": 5.417138542846357, "grad_norm": 0.6028378009796143, "learning_rate": 7.612845268345547e-05, "loss": 0.4245, "step": 14350 }, { "epoch": 5.4209135522838805, "grad_norm": 1.1031107902526855, "learning_rate": 7.609320017450199e-05, "loss": 0.4235, "step": 14360 }, { "epoch": 5.424688561721404, "grad_norm": 0.6411410570144653, "learning_rate": 7.605792983109222e-05, "loss": 0.4187, "step": 14370 }, { "epoch": 5.428463571158928, "grad_norm": 0.547097384929657, "learning_rate": 7.602264167733313e-05, "loss": 0.4282, "step": 14380 }, { "epoch": 5.432238580596452, "grad_norm": 0.6103999018669128, "learning_rate": 7.598733573734384e-05, "loss": 0.4256, "step": 14390 }, { "epoch": 5.436013590033975, "grad_norm": 0.5741267800331116, "learning_rate": 7.595201203525561e-05, "loss": 0.4254, "step": 14400 }, { "epoch": 5.439788599471498, "grad_norm": 0.5826230645179749, "learning_rate": 7.591667059521187e-05, "loss": 0.4279, "step": 14410 }, { "epoch": 5.443563608909022, "grad_norm": 0.5484496355056763, "learning_rate": 7.588131144136815e-05, "loss": 0.429, "step": 14420 }, { "epoch": 5.447338618346546, "grad_norm": 0.566040575504303, "learning_rate": 7.584593459789212e-05, "loss": 0.4213, "step": 14430 }, { "epoch": 5.4511136277840695, "grad_norm": 0.5821641087532043, "learning_rate": 7.58105400889635e-05, "loss": 0.4327, "step": 14440 }, { "epoch": 5.454888637221593, "grad_norm": 0.6450483798980713, "learning_rate": 7.57751279387741e-05, "loss": 0.4227, "step": 14450 }, { "epoch": 5.458663646659117, "grad_norm": 0.5825949907302856, "learning_rate": 7.573969817152782e-05, "loss": 0.423, "step": 14460 }, { "epoch": 5.462438656096641, "grad_norm": 0.5413886308670044, "learning_rate": 7.570425081144052e-05, "loss": 0.425, "step": 14470 }, { "epoch": 5.4662136655341635, "grad_norm": 0.5597252249717712, "learning_rate": 7.56687858827402e-05, "loss": 0.4192, "step": 14480 }, { "epoch": 5.469988674971687, "grad_norm": 0.5918856263160706, "learning_rate": 7.563330340966675e-05, "loss": 0.421, "step": 14490 }, { "epoch": 5.473763684409211, "grad_norm": 0.574697732925415, "learning_rate": 7.559780341647212e-05, "loss": 0.4286, "step": 14500 }, { "epoch": 5.477538693846735, "grad_norm": 0.7420601844787598, "learning_rate": 7.556228592742026e-05, "loss": 0.4226, "step": 14510 }, { "epoch": 5.4813137032842585, "grad_norm": 0.8182733058929443, "learning_rate": 7.552675096678696e-05, "loss": 0.4199, "step": 14520 }, { "epoch": 5.485088712721781, "grad_norm": 0.6243227124214172, "learning_rate": 7.549119855886012e-05, "loss": 0.4234, "step": 14530 }, { "epoch": 5.488863722159305, "grad_norm": 0.5414767265319824, "learning_rate": 7.545562872793941e-05, "loss": 0.4243, "step": 14540 }, { "epoch": 5.492638731596829, "grad_norm": 0.6131302714347839, "learning_rate": 7.542004149833648e-05, "loss": 0.4277, "step": 14550 }, { "epoch": 5.4964137410343525, "grad_norm": 0.6857340335845947, "learning_rate": 7.538443689437492e-05, "loss": 0.4259, "step": 14560 }, { "epoch": 5.500188750471876, "grad_norm": 0.5744028091430664, "learning_rate": 7.53488149403901e-05, "loss": 0.4208, "step": 14570 }, { "epoch": 5.5039637599094, "grad_norm": 0.602278470993042, "learning_rate": 7.531317566072929e-05, "loss": 0.4211, "step": 14580 }, { "epoch": 5.507738769346924, "grad_norm": 0.6976267695426941, "learning_rate": 7.527751907975158e-05, "loss": 0.4207, "step": 14590 }, { "epoch": 5.511513778784447, "grad_norm": 0.5509871244430542, "learning_rate": 7.524184522182793e-05, "loss": 0.4165, "step": 14600 }, { "epoch": 5.51528878822197, "grad_norm": 0.6053850650787354, "learning_rate": 7.520615411134112e-05, "loss": 0.4168, "step": 14610 }, { "epoch": 5.519063797659494, "grad_norm": 0.592635989189148, "learning_rate": 7.517044577268564e-05, "loss": 0.4263, "step": 14620 }, { "epoch": 5.522838807097018, "grad_norm": 0.6511247158050537, "learning_rate": 7.513472023026782e-05, "loss": 0.4141, "step": 14630 }, { "epoch": 5.5266138165345415, "grad_norm": 0.5748429298400879, "learning_rate": 7.509897750850572e-05, "loss": 0.4176, "step": 14640 }, { "epoch": 5.530388825972065, "grad_norm": 0.6025387048721313, "learning_rate": 7.506321763182918e-05, "loss": 0.4144, "step": 14650 }, { "epoch": 5.534163835409588, "grad_norm": 0.6291730999946594, "learning_rate": 7.50274406246797e-05, "loss": 0.4273, "step": 14660 }, { "epoch": 5.537938844847112, "grad_norm": 0.8842790722846985, "learning_rate": 7.499164651151056e-05, "loss": 0.4211, "step": 14670 }, { "epoch": 5.541713854284636, "grad_norm": 0.6468051075935364, "learning_rate": 7.495583531678669e-05, "loss": 0.4204, "step": 14680 }, { "epoch": 5.545488863722159, "grad_norm": 0.649259090423584, "learning_rate": 7.492000706498469e-05, "loss": 0.4179, "step": 14690 }, { "epoch": 5.549263873159683, "grad_norm": 0.5695837140083313, "learning_rate": 7.488416178059284e-05, "loss": 0.4255, "step": 14700 }, { "epoch": 5.553038882597207, "grad_norm": 0.66133713722229, "learning_rate": 7.484829948811107e-05, "loss": 0.4263, "step": 14710 }, { "epoch": 5.5568138920347305, "grad_norm": 0.6037610173225403, "learning_rate": 7.48124202120509e-05, "loss": 0.4221, "step": 14720 }, { "epoch": 5.560588901472253, "grad_norm": 0.6156513094902039, "learning_rate": 7.477652397693549e-05, "loss": 0.4173, "step": 14730 }, { "epoch": 5.564363910909777, "grad_norm": 0.5722167491912842, "learning_rate": 7.474061080729955e-05, "loss": 0.4218, "step": 14740 }, { "epoch": 5.568138920347301, "grad_norm": 1.3760823011398315, "learning_rate": 7.470468072768941e-05, "loss": 0.4296, "step": 14750 }, { "epoch": 5.571913929784825, "grad_norm": 0.7008207440376282, "learning_rate": 7.466873376266297e-05, "loss": 0.4214, "step": 14760 }, { "epoch": 5.575688939222348, "grad_norm": 0.5777148604393005, "learning_rate": 7.46327699367896e-05, "loss": 0.4208, "step": 14770 }, { "epoch": 5.579463948659871, "grad_norm": 0.6137166023254395, "learning_rate": 7.459678927465026e-05, "loss": 0.4207, "step": 14780 }, { "epoch": 5.583238958097395, "grad_norm": 0.6016892194747925, "learning_rate": 7.456079180083737e-05, "loss": 0.42, "step": 14790 }, { "epoch": 5.587013967534919, "grad_norm": 0.6223365068435669, "learning_rate": 7.452477753995489e-05, "loss": 0.4188, "step": 14800 }, { "epoch": 5.590788976972442, "grad_norm": 0.5763020515441895, "learning_rate": 7.448874651661823e-05, "loss": 0.4192, "step": 14810 }, { "epoch": 5.594563986409966, "grad_norm": 0.5202614665031433, "learning_rate": 7.445269875545423e-05, "loss": 0.4213, "step": 14820 }, { "epoch": 5.59833899584749, "grad_norm": 0.6435422301292419, "learning_rate": 7.44166342811012e-05, "loss": 0.4166, "step": 14830 }, { "epoch": 5.602114005285014, "grad_norm": 0.6565425992012024, "learning_rate": 7.438055311820886e-05, "loss": 0.42, "step": 14840 }, { "epoch": 5.605889014722536, "grad_norm": 0.5886382460594177, "learning_rate": 7.434445529143837e-05, "loss": 0.4143, "step": 14850 }, { "epoch": 5.60966402416006, "grad_norm": 0.5876275897026062, "learning_rate": 7.430834082546225e-05, "loss": 0.4178, "step": 14860 }, { "epoch": 5.613439033597584, "grad_norm": 0.5644301772117615, "learning_rate": 7.427220974496438e-05, "loss": 0.4271, "step": 14870 }, { "epoch": 5.617214043035108, "grad_norm": 0.6446945667266846, "learning_rate": 7.423606207464005e-05, "loss": 0.4173, "step": 14880 }, { "epoch": 5.620989052472631, "grad_norm": 0.6406450271606445, "learning_rate": 7.419989783919578e-05, "loss": 0.4194, "step": 14890 }, { "epoch": 5.624764061910155, "grad_norm": 0.6164763569831848, "learning_rate": 7.416371706334956e-05, "loss": 0.4221, "step": 14900 }, { "epoch": 5.628539071347678, "grad_norm": 0.6049771904945374, "learning_rate": 7.412751977183056e-05, "loss": 0.4181, "step": 14910 }, { "epoch": 5.632314080785202, "grad_norm": 0.6096076965332031, "learning_rate": 7.409130598937932e-05, "loss": 0.4198, "step": 14920 }, { "epoch": 5.636089090222725, "grad_norm": 0.5442625880241394, "learning_rate": 7.40550757407476e-05, "loss": 0.4373, "step": 14930 }, { "epoch": 5.639864099660249, "grad_norm": 0.6001285910606384, "learning_rate": 7.401882905069843e-05, "loss": 0.4248, "step": 14940 }, { "epoch": 5.643639109097773, "grad_norm": 0.60331791639328, "learning_rate": 7.39825659440061e-05, "loss": 0.4246, "step": 14950 }, { "epoch": 5.647414118535297, "grad_norm": 0.5998172760009766, "learning_rate": 7.394628644545609e-05, "loss": 0.4225, "step": 14960 }, { "epoch": 5.65118912797282, "grad_norm": 0.6505115628242493, "learning_rate": 7.390999057984507e-05, "loss": 0.4245, "step": 14970 }, { "epoch": 5.654964137410343, "grad_norm": 0.6152286529541016, "learning_rate": 7.387367837198097e-05, "loss": 0.4314, "step": 14980 }, { "epoch": 5.658739146847867, "grad_norm": 0.595940113067627, "learning_rate": 7.383734984668281e-05, "loss": 0.4192, "step": 14990 }, { "epoch": 5.662514156285391, "grad_norm": 0.6145548820495605, "learning_rate": 7.38010050287808e-05, "loss": 0.414, "step": 15000 }, { "epoch": 5.666289165722914, "grad_norm": 1.0223385095596313, "learning_rate": 7.376464394311628e-05, "loss": 0.4289, "step": 15010 }, { "epoch": 5.670064175160438, "grad_norm": 0.6325558423995972, "learning_rate": 7.372826661454172e-05, "loss": 0.4227, "step": 15020 }, { "epoch": 5.673839184597962, "grad_norm": 0.6463314294815063, "learning_rate": 7.369187306792068e-05, "loss": 0.4183, "step": 15030 }, { "epoch": 5.677614194035485, "grad_norm": 0.6034806370735168, "learning_rate": 7.365546332812779e-05, "loss": 0.4133, "step": 15040 }, { "epoch": 5.6813892034730085, "grad_norm": 0.5699142813682556, "learning_rate": 7.361903742004876e-05, "loss": 0.4185, "step": 15050 }, { "epoch": 5.685164212910532, "grad_norm": 0.6562982797622681, "learning_rate": 7.358259536858039e-05, "loss": 0.4136, "step": 15060 }, { "epoch": 5.688939222348056, "grad_norm": 0.691800594329834, "learning_rate": 7.354613719863044e-05, "loss": 0.4263, "step": 15070 }, { "epoch": 5.69271423178558, "grad_norm": 0.6193991303443909, "learning_rate": 7.350966293511776e-05, "loss": 0.422, "step": 15080 }, { "epoch": 5.696489241223103, "grad_norm": 0.6109247803688049, "learning_rate": 7.347317260297212e-05, "loss": 0.4272, "step": 15090 }, { "epoch": 5.700264250660627, "grad_norm": 0.5521445274353027, "learning_rate": 7.343666622713437e-05, "loss": 0.4142, "step": 15100 }, { "epoch": 5.70403926009815, "grad_norm": 0.5884120464324951, "learning_rate": 7.340014383255624e-05, "loss": 0.4103, "step": 15110 }, { "epoch": 5.707814269535674, "grad_norm": 0.7118320465087891, "learning_rate": 7.336360544420044e-05, "loss": 0.4234, "step": 15120 }, { "epoch": 5.7115892789731975, "grad_norm": 0.6134358644485474, "learning_rate": 7.332705108704064e-05, "loss": 0.4077, "step": 15130 }, { "epoch": 5.715364288410721, "grad_norm": 0.5904216170310974, "learning_rate": 7.329048078606138e-05, "loss": 0.4202, "step": 15140 }, { "epoch": 5.719139297848245, "grad_norm": 0.6160265803337097, "learning_rate": 7.32538945662581e-05, "loss": 0.4229, "step": 15150 }, { "epoch": 5.722914307285768, "grad_norm": 0.6015536189079285, "learning_rate": 7.321729245263718e-05, "loss": 0.458, "step": 15160 }, { "epoch": 5.7266893167232915, "grad_norm": 0.5921651721000671, "learning_rate": 7.318067447021578e-05, "loss": 0.42, "step": 15170 }, { "epoch": 5.730464326160815, "grad_norm": 0.5869306921958923, "learning_rate": 7.314404064402198e-05, "loss": 0.4119, "step": 15180 }, { "epoch": 5.734239335598339, "grad_norm": 0.5936274528503418, "learning_rate": 7.310739099909461e-05, "loss": 0.4105, "step": 15190 }, { "epoch": 5.738014345035863, "grad_norm": 0.6002879738807678, "learning_rate": 7.307072556048339e-05, "loss": 0.4178, "step": 15200 }, { "epoch": 5.7417893544733865, "grad_norm": 0.6081198453903198, "learning_rate": 7.30340443532488e-05, "loss": 0.4202, "step": 15210 }, { "epoch": 5.74556436391091, "grad_norm": 0.6480271816253662, "learning_rate": 7.299734740246208e-05, "loss": 0.4105, "step": 15220 }, { "epoch": 5.749339373348433, "grad_norm": 0.5739256739616394, "learning_rate": 7.296063473320528e-05, "loss": 0.4184, "step": 15230 }, { "epoch": 5.753114382785957, "grad_norm": 0.749381422996521, "learning_rate": 7.292390637057113e-05, "loss": 0.4158, "step": 15240 }, { "epoch": 5.7568893922234805, "grad_norm": 0.6204841732978821, "learning_rate": 7.288716233966314e-05, "loss": 0.4133, "step": 15250 }, { "epoch": 5.760664401661004, "grad_norm": 1.603737235069275, "learning_rate": 7.285040266559551e-05, "loss": 0.4199, "step": 15260 }, { "epoch": 5.764439411098528, "grad_norm": 0.6340755820274353, "learning_rate": 7.281362737349312e-05, "loss": 0.4139, "step": 15270 }, { "epoch": 5.768214420536052, "grad_norm": 0.6750447154045105, "learning_rate": 7.277683648849153e-05, "loss": 0.4176, "step": 15280 }, { "epoch": 5.771989429973575, "grad_norm": 0.5785810351371765, "learning_rate": 7.2740030035737e-05, "loss": 0.4137, "step": 15290 }, { "epoch": 5.775764439411098, "grad_norm": 0.5631486773490906, "learning_rate": 7.270320804038634e-05, "loss": 0.4183, "step": 15300 }, { "epoch": 5.779539448848622, "grad_norm": 0.6204792857170105, "learning_rate": 7.266637052760708e-05, "loss": 0.4185, "step": 15310 }, { "epoch": 5.783314458286146, "grad_norm": 0.6543823480606079, "learning_rate": 7.262951752257728e-05, "loss": 0.4132, "step": 15320 }, { "epoch": 5.7870894677236695, "grad_norm": 0.6588866114616394, "learning_rate": 7.259264905048564e-05, "loss": 0.4141, "step": 15330 }, { "epoch": 5.790864477161193, "grad_norm": 0.5547715425491333, "learning_rate": 7.255576513653142e-05, "loss": 0.4132, "step": 15340 }, { "epoch": 5.794639486598717, "grad_norm": 0.642467200756073, "learning_rate": 7.251886580592439e-05, "loss": 0.4144, "step": 15350 }, { "epoch": 5.79841449603624, "grad_norm": 0.6495168209075928, "learning_rate": 7.248195108388496e-05, "loss": 0.4233, "step": 15360 }, { "epoch": 5.8021895054737636, "grad_norm": 0.6559742093086243, "learning_rate": 7.244502099564395e-05, "loss": 0.415, "step": 15370 }, { "epoch": 5.805964514911287, "grad_norm": 0.6655861139297485, "learning_rate": 7.240807556644271e-05, "loss": 0.4143, "step": 15380 }, { "epoch": 5.809739524348811, "grad_norm": 0.5971380472183228, "learning_rate": 7.237111482153314e-05, "loss": 0.4144, "step": 15390 }, { "epoch": 5.813514533786335, "grad_norm": 0.6410373449325562, "learning_rate": 7.233413878617751e-05, "loss": 0.4212, "step": 15400 }, { "epoch": 5.817289543223858, "grad_norm": 0.5639101266860962, "learning_rate": 7.229714748564864e-05, "loss": 0.4118, "step": 15410 }, { "epoch": 5.821064552661381, "grad_norm": 0.631568193435669, "learning_rate": 7.22601409452297e-05, "loss": 0.4152, "step": 15420 }, { "epoch": 5.824839562098905, "grad_norm": 0.6407223343849182, "learning_rate": 7.222311919021433e-05, "loss": 0.4209, "step": 15430 }, { "epoch": 5.828614571536429, "grad_norm": 0.8518263101577759, "learning_rate": 7.218608224590655e-05, "loss": 0.4134, "step": 15440 }, { "epoch": 5.8323895809739525, "grad_norm": 0.6463886499404907, "learning_rate": 7.214903013762074e-05, "loss": 0.4219, "step": 15450 }, { "epoch": 5.836164590411476, "grad_norm": 0.6677048206329346, "learning_rate": 7.21119628906817e-05, "loss": 0.4104, "step": 15460 }, { "epoch": 5.839939599849, "grad_norm": 0.6562634706497192, "learning_rate": 7.207488053042454e-05, "loss": 0.4146, "step": 15470 }, { "epoch": 5.843714609286523, "grad_norm": 0.6001513600349426, "learning_rate": 7.203778308219467e-05, "loss": 0.4196, "step": 15480 }, { "epoch": 5.847489618724047, "grad_norm": 0.571530818939209, "learning_rate": 7.200067057134787e-05, "loss": 0.4142, "step": 15490 }, { "epoch": 5.85126462816157, "grad_norm": 0.6734705567359924, "learning_rate": 7.196354302325019e-05, "loss": 0.4137, "step": 15500 }, { "epoch": 5.855039637599094, "grad_norm": 0.7626957297325134, "learning_rate": 7.192640046327795e-05, "loss": 0.4193, "step": 15510 }, { "epoch": 5.858814647036618, "grad_norm": 0.589971125125885, "learning_rate": 7.188924291681777e-05, "loss": 0.4144, "step": 15520 }, { "epoch": 5.8625896564741415, "grad_norm": 0.6109906435012817, "learning_rate": 7.185207040926643e-05, "loss": 0.4154, "step": 15530 }, { "epoch": 5.866364665911664, "grad_norm": 0.6139832139015198, "learning_rate": 7.181488296603103e-05, "loss": 0.415, "step": 15540 }, { "epoch": 5.870139675349188, "grad_norm": 0.5642750263214111, "learning_rate": 7.177768061252885e-05, "loss": 0.4155, "step": 15550 }, { "epoch": 5.873914684786712, "grad_norm": 0.6058228015899658, "learning_rate": 7.174046337418729e-05, "loss": 0.4193, "step": 15560 }, { "epoch": 5.877689694224236, "grad_norm": 0.6035988926887512, "learning_rate": 7.170323127644403e-05, "loss": 0.4152, "step": 15570 }, { "epoch": 5.881464703661759, "grad_norm": 0.5951765179634094, "learning_rate": 7.166598434474683e-05, "loss": 0.4125, "step": 15580 }, { "epoch": 5.885239713099283, "grad_norm": 0.6099646091461182, "learning_rate": 7.162872260455364e-05, "loss": 0.4121, "step": 15590 }, { "epoch": 5.889014722536807, "grad_norm": 0.6386427879333496, "learning_rate": 7.159144608133248e-05, "loss": 0.4079, "step": 15600 }, { "epoch": 5.89278973197433, "grad_norm": 0.6007016897201538, "learning_rate": 7.155415480056153e-05, "loss": 0.411, "step": 15610 }, { "epoch": 5.896564741411853, "grad_norm": 0.6118659973144531, "learning_rate": 7.151684878772902e-05, "loss": 0.4271, "step": 15620 }, { "epoch": 5.900339750849377, "grad_norm": 0.6127933859825134, "learning_rate": 7.147952806833324e-05, "loss": 0.4097, "step": 15630 }, { "epoch": 5.904114760286901, "grad_norm": 0.6609582304954529, "learning_rate": 7.14421926678826e-05, "loss": 0.4102, "step": 15640 }, { "epoch": 5.907889769724425, "grad_norm": 0.6080712080001831, "learning_rate": 7.140484261189543e-05, "loss": 0.4141, "step": 15650 }, { "epoch": 5.911664779161947, "grad_norm": 0.621368408203125, "learning_rate": 7.136747792590017e-05, "loss": 0.4098, "step": 15660 }, { "epoch": 5.915439788599471, "grad_norm": 0.6989858746528625, "learning_rate": 7.133009863543524e-05, "loss": 0.4273, "step": 15670 }, { "epoch": 5.919214798036995, "grad_norm": 0.5902137160301208, "learning_rate": 7.129270476604901e-05, "loss": 0.4121, "step": 15680 }, { "epoch": 5.922989807474519, "grad_norm": 0.6309996843338013, "learning_rate": 7.125529634329988e-05, "loss": 0.4136, "step": 15690 }, { "epoch": 5.926764816912042, "grad_norm": 0.6389601230621338, "learning_rate": 7.12178733927561e-05, "loss": 0.4177, "step": 15700 }, { "epoch": 5.930539826349566, "grad_norm": 0.5543851852416992, "learning_rate": 7.118043593999593e-05, "loss": 0.4172, "step": 15710 }, { "epoch": 5.93431483578709, "grad_norm": 0.5611990094184875, "learning_rate": 7.114298401060752e-05, "loss": 0.4153, "step": 15720 }, { "epoch": 5.938089845224613, "grad_norm": 0.6465316414833069, "learning_rate": 7.11055176301889e-05, "loss": 0.4102, "step": 15730 }, { "epoch": 5.941864854662136, "grad_norm": 0.6097255349159241, "learning_rate": 7.1068036824348e-05, "loss": 0.4169, "step": 15740 }, { "epoch": 5.94563986409966, "grad_norm": 0.5505424737930298, "learning_rate": 7.10305416187026e-05, "loss": 0.4134, "step": 15750 }, { "epoch": 5.949414873537184, "grad_norm": 0.554263710975647, "learning_rate": 7.099303203888029e-05, "loss": 0.4127, "step": 15760 }, { "epoch": 5.953189882974708, "grad_norm": 0.6733128428459167, "learning_rate": 7.095550811051855e-05, "loss": 0.4059, "step": 15770 }, { "epoch": 5.956964892412231, "grad_norm": 0.6691094040870667, "learning_rate": 7.09179698592646e-05, "loss": 0.416, "step": 15780 }, { "epoch": 5.960739901849754, "grad_norm": 0.581087052822113, "learning_rate": 7.088041731077551e-05, "loss": 0.4081, "step": 15790 }, { "epoch": 5.964514911287278, "grad_norm": 0.6719875931739807, "learning_rate": 7.084285049071806e-05, "loss": 0.4086, "step": 15800 }, { "epoch": 5.968289920724802, "grad_norm": 0.6516507267951965, "learning_rate": 7.080526942476886e-05, "loss": 0.4166, "step": 15810 }, { "epoch": 5.972064930162325, "grad_norm": 0.6685953140258789, "learning_rate": 7.076767413861418e-05, "loss": 0.4253, "step": 15820 }, { "epoch": 5.975839939599849, "grad_norm": 0.6204284429550171, "learning_rate": 7.073006465795005e-05, "loss": 0.4134, "step": 15830 }, { "epoch": 5.979614949037373, "grad_norm": 0.5573307275772095, "learning_rate": 7.06924410084822e-05, "loss": 0.4112, "step": 15840 }, { "epoch": 5.983389958474897, "grad_norm": 0.5689889192581177, "learning_rate": 7.065480321592604e-05, "loss": 0.4107, "step": 15850 }, { "epoch": 5.9871649679124195, "grad_norm": 0.6300421953201294, "learning_rate": 7.061715130600663e-05, "loss": 0.4019, "step": 15860 }, { "epoch": 5.990939977349943, "grad_norm": 0.5939746499061584, "learning_rate": 7.057948530445873e-05, "loss": 0.4093, "step": 15870 }, { "epoch": 5.994714986787467, "grad_norm": 0.576476514339447, "learning_rate": 7.054180523702668e-05, "loss": 0.4084, "step": 15880 }, { "epoch": 5.998489996224991, "grad_norm": 0.6696039438247681, "learning_rate": 7.050411112946442e-05, "loss": 0.412, "step": 15890 }, { "epoch": 6.002265005662514, "grad_norm": 0.6292182803153992, "learning_rate": 7.046640300753557e-05, "loss": 0.4151, "step": 15900 }, { "epoch": 6.006040015100038, "grad_norm": 0.5818023085594177, "learning_rate": 7.042868089701325e-05, "loss": 0.409, "step": 15910 }, { "epoch": 6.009815024537561, "grad_norm": 0.601872444152832, "learning_rate": 7.039094482368016e-05, "loss": 0.4074, "step": 15920 }, { "epoch": 6.013590033975085, "grad_norm": 0.629568874835968, "learning_rate": 7.035319481332858e-05, "loss": 0.4241, "step": 15930 }, { "epoch": 6.0173650434126085, "grad_norm": 0.6180824637413025, "learning_rate": 7.031543089176023e-05, "loss": 0.413, "step": 15940 }, { "epoch": 6.021140052850132, "grad_norm": 0.7090234160423279, "learning_rate": 7.027765308478644e-05, "loss": 0.4159, "step": 15950 }, { "epoch": 6.024915062287656, "grad_norm": 0.8287419676780701, "learning_rate": 7.023986141822798e-05, "loss": 0.4117, "step": 15960 }, { "epoch": 6.02869007172518, "grad_norm": 0.6538956761360168, "learning_rate": 7.02020559179151e-05, "loss": 0.4134, "step": 15970 }, { "epoch": 6.0324650811627025, "grad_norm": 0.6877662539482117, "learning_rate": 7.016423660968748e-05, "loss": 0.4132, "step": 15980 }, { "epoch": 6.036240090600226, "grad_norm": 0.9112296104431152, "learning_rate": 7.012640351939428e-05, "loss": 0.4225, "step": 15990 }, { "epoch": 6.04001510003775, "grad_norm": 0.6331153512001038, "learning_rate": 7.008855667289404e-05, "loss": 0.4109, "step": 16000 }, { "epoch": 6.043790109475274, "grad_norm": 0.6433661580085754, "learning_rate": 7.005069609605476e-05, "loss": 0.4129, "step": 16010 }, { "epoch": 6.0475651189127975, "grad_norm": 0.5703040361404419, "learning_rate": 7.001282181475377e-05, "loss": 0.422, "step": 16020 }, { "epoch": 6.051340128350321, "grad_norm": 0.5761269330978394, "learning_rate": 6.997493385487775e-05, "loss": 0.4079, "step": 16030 }, { "epoch": 6.055115137787844, "grad_norm": 0.6320202350616455, "learning_rate": 6.99370322423228e-05, "loss": 0.405, "step": 16040 }, { "epoch": 6.058890147225368, "grad_norm": 0.5953516960144043, "learning_rate": 6.989911700299433e-05, "loss": 0.4178, "step": 16050 }, { "epoch": 6.0626651566628915, "grad_norm": 0.6841408014297485, "learning_rate": 6.9861188162807e-05, "loss": 0.4176, "step": 16060 }, { "epoch": 6.066440166100415, "grad_norm": 0.6144746541976929, "learning_rate": 6.982324574768487e-05, "loss": 0.4184, "step": 16070 }, { "epoch": 6.070215175537939, "grad_norm": 0.6368954181671143, "learning_rate": 6.978528978356117e-05, "loss": 0.4169, "step": 16080 }, { "epoch": 6.073990184975463, "grad_norm": 0.6184800863265991, "learning_rate": 6.974732029637846e-05, "loss": 0.4123, "step": 16090 }, { "epoch": 6.0777651944129865, "grad_norm": 0.6910690665245056, "learning_rate": 6.970933731208855e-05, "loss": 0.413, "step": 16100 }, { "epoch": 6.081540203850509, "grad_norm": 0.5969312787055969, "learning_rate": 6.967134085665244e-05, "loss": 0.4145, "step": 16110 }, { "epoch": 6.085315213288033, "grad_norm": 0.5405552983283997, "learning_rate": 6.963333095604034e-05, "loss": 0.4136, "step": 16120 }, { "epoch": 6.089090222725557, "grad_norm": 0.5933461785316467, "learning_rate": 6.959530763623166e-05, "loss": 0.4144, "step": 16130 }, { "epoch": 6.0928652321630805, "grad_norm": 0.6700628995895386, "learning_rate": 6.955727092321497e-05, "loss": 0.4092, "step": 16140 }, { "epoch": 6.096640241600604, "grad_norm": 0.6101819276809692, "learning_rate": 6.951922084298803e-05, "loss": 0.418, "step": 16150 }, { "epoch": 6.100415251038128, "grad_norm": 0.5824779272079468, "learning_rate": 6.948115742155769e-05, "loss": 0.4169, "step": 16160 }, { "epoch": 6.104190260475651, "grad_norm": 0.5644822716712952, "learning_rate": 6.944308068493996e-05, "loss": 0.42, "step": 16170 }, { "epoch": 6.107965269913175, "grad_norm": 0.6036456823348999, "learning_rate": 6.940499065915992e-05, "loss": 0.4083, "step": 16180 }, { "epoch": 6.111740279350698, "grad_norm": 0.5669860243797302, "learning_rate": 6.936688737025173e-05, "loss": 0.4195, "step": 16190 }, { "epoch": 6.115515288788222, "grad_norm": 0.5517399311065674, "learning_rate": 6.932877084425867e-05, "loss": 0.414, "step": 16200 }, { "epoch": 6.119290298225746, "grad_norm": 0.5780723094940186, "learning_rate": 6.929064110723297e-05, "loss": 0.4133, "step": 16210 }, { "epoch": 6.1230653076632695, "grad_norm": 0.5744929909706116, "learning_rate": 6.925249818523598e-05, "loss": 0.4055, "step": 16220 }, { "epoch": 6.126840317100792, "grad_norm": 0.5747778415679932, "learning_rate": 6.921434210433801e-05, "loss": 0.4068, "step": 16230 }, { "epoch": 6.130615326538316, "grad_norm": 0.6638509631156921, "learning_rate": 6.917617289061841e-05, "loss": 0.4097, "step": 16240 }, { "epoch": 6.13439033597584, "grad_norm": 0.6439339518547058, "learning_rate": 6.913799057016547e-05, "loss": 0.4181, "step": 16250 }, { "epoch": 6.138165345413364, "grad_norm": 1.5434315204620361, "learning_rate": 6.909979516907641e-05, "loss": 0.4165, "step": 16260 }, { "epoch": 6.141940354850887, "grad_norm": 0.7095974683761597, "learning_rate": 6.906158671345746e-05, "loss": 0.4082, "step": 16270 }, { "epoch": 6.145715364288411, "grad_norm": 0.5394503474235535, "learning_rate": 6.902336522942374e-05, "loss": 0.4141, "step": 16280 }, { "epoch": 6.149490373725934, "grad_norm": 1.054152011871338, "learning_rate": 6.898513074309924e-05, "loss": 0.4109, "step": 16290 }, { "epoch": 6.153265383163458, "grad_norm": 0.6516214609146118, "learning_rate": 6.894688328061693e-05, "loss": 0.4187, "step": 16300 }, { "epoch": 6.157040392600981, "grad_norm": 0.5728663802146912, "learning_rate": 6.890862286811853e-05, "loss": 0.4231, "step": 16310 }, { "epoch": 6.160815402038505, "grad_norm": 0.6336321830749512, "learning_rate": 6.88703495317547e-05, "loss": 0.4066, "step": 16320 }, { "epoch": 6.164590411476029, "grad_norm": 0.6610395908355713, "learning_rate": 6.883206329768492e-05, "loss": 0.4141, "step": 16330 }, { "epoch": 6.1683654209135526, "grad_norm": 0.6659107208251953, "learning_rate": 6.879376419207743e-05, "loss": 0.4055, "step": 16340 }, { "epoch": 6.172140430351076, "grad_norm": 0.6790131330490112, "learning_rate": 6.875545224110935e-05, "loss": 0.4178, "step": 16350 }, { "epoch": 6.175915439788599, "grad_norm": 0.5676288604736328, "learning_rate": 6.871712747096651e-05, "loss": 0.4282, "step": 16360 }, { "epoch": 6.179690449226123, "grad_norm": 0.6402314901351929, "learning_rate": 6.867878990784353e-05, "loss": 0.4146, "step": 16370 }, { "epoch": 6.183465458663647, "grad_norm": 0.6174314022064209, "learning_rate": 6.864043957794377e-05, "loss": 0.414, "step": 16380 }, { "epoch": 6.18724046810117, "grad_norm": 0.6669917702674866, "learning_rate": 6.860207650747934e-05, "loss": 0.4154, "step": 16390 }, { "epoch": 6.191015477538694, "grad_norm": 0.6218419671058655, "learning_rate": 6.856370072267104e-05, "loss": 0.4125, "step": 16400 }, { "epoch": 6.194790486976218, "grad_norm": 0.5821741223335266, "learning_rate": 6.852531224974831e-05, "loss": 0.4081, "step": 16410 }, { "epoch": 6.198565496413741, "grad_norm": 0.6421241164207458, "learning_rate": 6.848691111494936e-05, "loss": 0.4067, "step": 16420 }, { "epoch": 6.202340505851264, "grad_norm": 0.7413673996925354, "learning_rate": 6.844849734452097e-05, "loss": 0.4059, "step": 16430 }, { "epoch": 6.206115515288788, "grad_norm": 0.715653657913208, "learning_rate": 6.841007096471862e-05, "loss": 0.414, "step": 16440 }, { "epoch": 6.209890524726312, "grad_norm": 0.9158740639686584, "learning_rate": 6.837163200180636e-05, "loss": 0.4127, "step": 16450 }, { "epoch": 6.213665534163836, "grad_norm": 0.6491902470588684, "learning_rate": 6.833318048205684e-05, "loss": 0.4046, "step": 16460 }, { "epoch": 6.217440543601359, "grad_norm": 0.5672116875648499, "learning_rate": 6.829471643175136e-05, "loss": 0.4076, "step": 16470 }, { "epoch": 6.221215553038882, "grad_norm": 0.7057968378067017, "learning_rate": 6.825623987717969e-05, "loss": 0.4077, "step": 16480 }, { "epoch": 6.224990562476406, "grad_norm": 0.6522260308265686, "learning_rate": 6.821775084464022e-05, "loss": 0.4096, "step": 16490 }, { "epoch": 6.22876557191393, "grad_norm": 0.6316922903060913, "learning_rate": 6.817924936043982e-05, "loss": 0.4026, "step": 16500 }, { "epoch": 6.232540581351453, "grad_norm": 0.6496508717536926, "learning_rate": 6.81407354508939e-05, "loss": 0.4087, "step": 16510 }, { "epoch": 6.236315590788977, "grad_norm": 0.5875167846679688, "learning_rate": 6.810220914232636e-05, "loss": 0.4075, "step": 16520 }, { "epoch": 6.240090600226501, "grad_norm": 0.6535125970840454, "learning_rate": 6.806367046106959e-05, "loss": 0.4127, "step": 16530 }, { "epoch": 6.243865609664024, "grad_norm": 0.6569979786872864, "learning_rate": 6.802511943346435e-05, "loss": 0.4154, "step": 16540 }, { "epoch": 6.247640619101547, "grad_norm": 0.6199260354042053, "learning_rate": 6.798655608585997e-05, "loss": 0.4116, "step": 16550 }, { "epoch": 6.251415628539071, "grad_norm": 0.6537362337112427, "learning_rate": 6.79479804446141e-05, "loss": 0.4057, "step": 16560 }, { "epoch": 6.255190637976595, "grad_norm": 0.6324570775032043, "learning_rate": 6.790939253609284e-05, "loss": 0.4028, "step": 16570 }, { "epoch": 6.258965647414119, "grad_norm": 0.6022115349769592, "learning_rate": 6.787079238667065e-05, "loss": 0.4087, "step": 16580 }, { "epoch": 6.262740656851642, "grad_norm": 0.6702650189399719, "learning_rate": 6.783218002273039e-05, "loss": 0.4071, "step": 16590 }, { "epoch": 6.266515666289166, "grad_norm": 0.6910784244537354, "learning_rate": 6.779355547066322e-05, "loss": 0.4038, "step": 16600 }, { "epoch": 6.270290675726689, "grad_norm": 0.6253206729888916, "learning_rate": 6.775491875686865e-05, "loss": 0.4126, "step": 16610 }, { "epoch": 6.274065685164213, "grad_norm": 0.5710755586624146, "learning_rate": 6.771626990775457e-05, "loss": 0.4108, "step": 16620 }, { "epoch": 6.277840694601736, "grad_norm": 0.6698145866394043, "learning_rate": 6.767760894973704e-05, "loss": 0.4424, "step": 16630 }, { "epoch": 6.28161570403926, "grad_norm": 1.365607500076294, "learning_rate": 6.763893590924048e-05, "loss": 0.4173, "step": 16640 }, { "epoch": 6.285390713476784, "grad_norm": 0.6131840944290161, "learning_rate": 6.760025081269756e-05, "loss": 0.4056, "step": 16650 }, { "epoch": 6.289165722914308, "grad_norm": 0.5999181866645813, "learning_rate": 6.756155368654915e-05, "loss": 0.4133, "step": 16660 }, { "epoch": 6.2929407323518305, "grad_norm": 0.6875386238098145, "learning_rate": 6.752284455724442e-05, "loss": 0.4055, "step": 16670 }, { "epoch": 6.296715741789354, "grad_norm": 0.5545741319656372, "learning_rate": 6.748412345124065e-05, "loss": 0.4132, "step": 16680 }, { "epoch": 6.300490751226878, "grad_norm": 0.623226523399353, "learning_rate": 6.744539039500335e-05, "loss": 0.4087, "step": 16690 }, { "epoch": 6.304265760664402, "grad_norm": 0.6600329279899597, "learning_rate": 6.740664541500625e-05, "loss": 0.4003, "step": 16700 }, { "epoch": 6.308040770101925, "grad_norm": 0.5790383219718933, "learning_rate": 6.736788853773112e-05, "loss": 0.4076, "step": 16710 }, { "epoch": 6.311815779539449, "grad_norm": 0.6672446131706238, "learning_rate": 6.732911978966796e-05, "loss": 0.4089, "step": 16720 }, { "epoch": 6.315590788976973, "grad_norm": 0.5762278437614441, "learning_rate": 6.729033919731482e-05, "loss": 0.4116, "step": 16730 }, { "epoch": 6.319365798414496, "grad_norm": 0.5791032314300537, "learning_rate": 6.725154678717787e-05, "loss": 0.407, "step": 16740 }, { "epoch": 6.3231408078520195, "grad_norm": 0.6549326777458191, "learning_rate": 6.721274258577138e-05, "loss": 0.4114, "step": 16750 }, { "epoch": 6.326915817289543, "grad_norm": 0.6120374202728271, "learning_rate": 6.717392661961763e-05, "loss": 0.4089, "step": 16760 }, { "epoch": 6.330690826727067, "grad_norm": 0.6165977716445923, "learning_rate": 6.713509891524697e-05, "loss": 0.4128, "step": 16770 }, { "epoch": 6.334465836164591, "grad_norm": 0.6072388887405396, "learning_rate": 6.709625949919777e-05, "loss": 0.4091, "step": 16780 }, { "epoch": 6.3382408456021135, "grad_norm": 0.6050112843513489, "learning_rate": 6.705740839801642e-05, "loss": 0.4031, "step": 16790 }, { "epoch": 6.342015855039637, "grad_norm": 0.7295091152191162, "learning_rate": 6.701854563825727e-05, "loss": 0.4105, "step": 16800 }, { "epoch": 6.345790864477161, "grad_norm": 0.6450115442276001, "learning_rate": 6.697967124648266e-05, "loss": 0.4026, "step": 16810 }, { "epoch": 6.349565873914685, "grad_norm": 0.6749234199523926, "learning_rate": 6.694078524926285e-05, "loss": 0.4106, "step": 16820 }, { "epoch": 6.3533408833522085, "grad_norm": 0.6108365058898926, "learning_rate": 6.690188767317607e-05, "loss": 0.403, "step": 16830 }, { "epoch": 6.357115892789732, "grad_norm": 0.6435592174530029, "learning_rate": 6.686297854480843e-05, "loss": 0.4091, "step": 16840 }, { "epoch": 6.360890902227256, "grad_norm": 0.5970647931098938, "learning_rate": 6.682405789075398e-05, "loss": 0.404, "step": 16850 }, { "epoch": 6.364665911664779, "grad_norm": 0.5717617869377136, "learning_rate": 6.67851257376146e-05, "loss": 0.3988, "step": 16860 }, { "epoch": 6.3684409211023025, "grad_norm": 0.6971402168273926, "learning_rate": 6.674618211200004e-05, "loss": 0.4016, "step": 16870 }, { "epoch": 6.372215930539826, "grad_norm": 0.5539186000823975, "learning_rate": 6.670722704052792e-05, "loss": 0.4109, "step": 16880 }, { "epoch": 6.37599093997735, "grad_norm": 0.5876161456108093, "learning_rate": 6.666826054982365e-05, "loss": 0.406, "step": 16890 }, { "epoch": 6.379765949414874, "grad_norm": 0.6357724666595459, "learning_rate": 6.662928266652048e-05, "loss": 0.4059, "step": 16900 }, { "epoch": 6.3835409588523975, "grad_norm": 0.7432243227958679, "learning_rate": 6.659029341725941e-05, "loss": 0.4126, "step": 16910 }, { "epoch": 6.38731596828992, "grad_norm": 0.5893529057502747, "learning_rate": 6.655129282868923e-05, "loss": 0.4053, "step": 16920 }, { "epoch": 6.391090977727444, "grad_norm": 0.5809387564659119, "learning_rate": 6.651228092746646e-05, "loss": 0.41, "step": 16930 }, { "epoch": 6.394865987164968, "grad_norm": 0.6010847687721252, "learning_rate": 6.647325774025539e-05, "loss": 0.42, "step": 16940 }, { "epoch": 6.3986409966024915, "grad_norm": 0.6078194975852966, "learning_rate": 6.643422329372798e-05, "loss": 0.4069, "step": 16950 }, { "epoch": 6.402416006040015, "grad_norm": 0.637675940990448, "learning_rate": 6.639517761456392e-05, "loss": 0.407, "step": 16960 }, { "epoch": 6.406191015477539, "grad_norm": 0.5883272886276245, "learning_rate": 6.635612072945054e-05, "loss": 0.4043, "step": 16970 }, { "epoch": 6.409966024915063, "grad_norm": 0.5617126226425171, "learning_rate": 6.631705266508289e-05, "loss": 0.4079, "step": 16980 }, { "epoch": 6.413741034352586, "grad_norm": 0.6175063252449036, "learning_rate": 6.62779734481636e-05, "loss": 0.4003, "step": 16990 }, { "epoch": 6.417516043790109, "grad_norm": 0.6425361037254333, "learning_rate": 6.623888310540294e-05, "loss": 0.4033, "step": 17000 }, { "epoch": 6.421291053227633, "grad_norm": 0.5959002375602722, "learning_rate": 6.619978166351882e-05, "loss": 0.4037, "step": 17010 }, { "epoch": 6.425066062665157, "grad_norm": 0.6390359997749329, "learning_rate": 6.616066914923666e-05, "loss": 0.4058, "step": 17020 }, { "epoch": 6.4288410721026805, "grad_norm": 0.5752782225608826, "learning_rate": 6.612154558928955e-05, "loss": 0.4081, "step": 17030 }, { "epoch": 6.432616081540204, "grad_norm": 0.684485137462616, "learning_rate": 6.608241101041804e-05, "loss": 0.4, "step": 17040 }, { "epoch": 6.436391090977727, "grad_norm": 0.6156774759292603, "learning_rate": 6.604326543937025e-05, "loss": 0.4071, "step": 17050 }, { "epoch": 6.440166100415251, "grad_norm": 0.6333221197128296, "learning_rate": 6.60041089029018e-05, "loss": 0.4074, "step": 17060 }, { "epoch": 6.443941109852775, "grad_norm": 0.6324211955070496, "learning_rate": 6.596494142777583e-05, "loss": 0.4126, "step": 17070 }, { "epoch": 6.447716119290298, "grad_norm": 0.6180362701416016, "learning_rate": 6.592576304076294e-05, "loss": 0.4113, "step": 17080 }, { "epoch": 6.451491128727822, "grad_norm": 0.5131887197494507, "learning_rate": 6.588657376864119e-05, "loss": 0.4074, "step": 17090 }, { "epoch": 6.455266138165346, "grad_norm": 0.570022702217102, "learning_rate": 6.584737363819605e-05, "loss": 0.4051, "step": 17100 }, { "epoch": 6.459041147602869, "grad_norm": 0.5947955846786499, "learning_rate": 6.580816267622048e-05, "loss": 0.3996, "step": 17110 }, { "epoch": 6.462816157040392, "grad_norm": 0.5993560552597046, "learning_rate": 6.576894090951478e-05, "loss": 0.4017, "step": 17120 }, { "epoch": 6.466591166477916, "grad_norm": 0.5770085453987122, "learning_rate": 6.572970836488665e-05, "loss": 0.4095, "step": 17130 }, { "epoch": 6.47036617591544, "grad_norm": 0.5673043727874756, "learning_rate": 6.569046506915119e-05, "loss": 0.3987, "step": 17140 }, { "epoch": 6.474141185352964, "grad_norm": 0.6231826543807983, "learning_rate": 6.56512110491308e-05, "loss": 0.3943, "step": 17150 }, { "epoch": 6.477916194790487, "grad_norm": 0.6213942766189575, "learning_rate": 6.561194633165523e-05, "loss": 0.4093, "step": 17160 }, { "epoch": 6.48169120422801, "grad_norm": 0.6126642823219299, "learning_rate": 6.557267094356155e-05, "loss": 0.4164, "step": 17170 }, { "epoch": 6.485466213665534, "grad_norm": 0.5821068286895752, "learning_rate": 6.553338491169414e-05, "loss": 0.4137, "step": 17180 }, { "epoch": 6.489241223103058, "grad_norm": 0.5541254878044128, "learning_rate": 6.54940882629046e-05, "loss": 0.4062, "step": 17190 }, { "epoch": 6.493016232540581, "grad_norm": 6.614664554595947, "learning_rate": 6.545478102405184e-05, "loss": 0.4263, "step": 17200 }, { "epoch": 6.496791241978105, "grad_norm": 0.6379892230033875, "learning_rate": 6.541546322200199e-05, "loss": 0.4085, "step": 17210 }, { "epoch": 6.500566251415629, "grad_norm": 0.5639785528182983, "learning_rate": 6.537613488362837e-05, "loss": 0.3997, "step": 17220 }, { "epoch": 6.504341260853153, "grad_norm": 0.6088283061981201, "learning_rate": 6.533679603581155e-05, "loss": 0.4141, "step": 17230 }, { "epoch": 6.508116270290675, "grad_norm": 0.7054754495620728, "learning_rate": 6.529744670543926e-05, "loss": 0.4144, "step": 17240 }, { "epoch": 6.511891279728199, "grad_norm": 0.6340017914772034, "learning_rate": 6.52580869194064e-05, "loss": 0.4131, "step": 17250 }, { "epoch": 6.515666289165723, "grad_norm": 0.6918211579322815, "learning_rate": 6.521871670461499e-05, "loss": 0.411, "step": 17260 }, { "epoch": 6.519441298603247, "grad_norm": 0.6365815997123718, "learning_rate": 6.517933608797422e-05, "loss": 0.4104, "step": 17270 }, { "epoch": 6.52321630804077, "grad_norm": 0.599440336227417, "learning_rate": 6.513994509640038e-05, "loss": 0.4075, "step": 17280 }, { "epoch": 6.526991317478293, "grad_norm": 0.6258647441864014, "learning_rate": 6.510054375681682e-05, "loss": 0.4004, "step": 17290 }, { "epoch": 6.530766326915817, "grad_norm": 0.7459936141967773, "learning_rate": 6.506113209615398e-05, "loss": 0.4078, "step": 17300 }, { "epoch": 6.534541336353341, "grad_norm": 0.5855075120925903, "learning_rate": 6.502171014134938e-05, "loss": 0.4061, "step": 17310 }, { "epoch": 6.538316345790864, "grad_norm": 0.6313618421554565, "learning_rate": 6.498227791934755e-05, "loss": 0.4085, "step": 17320 }, { "epoch": 6.542091355228388, "grad_norm": 0.714318573474884, "learning_rate": 6.494283545710003e-05, "loss": 0.4015, "step": 17330 }, { "epoch": 6.545866364665912, "grad_norm": 0.6481465697288513, "learning_rate": 6.490338278156538e-05, "loss": 0.4065, "step": 17340 }, { "epoch": 6.549641374103436, "grad_norm": 0.7039319276809692, "learning_rate": 6.486391991970913e-05, "loss": 0.4136, "step": 17350 }, { "epoch": 6.553416383540959, "grad_norm": 0.642465353012085, "learning_rate": 6.482444689850377e-05, "loss": 0.4026, "step": 17360 }, { "epoch": 6.557191392978482, "grad_norm": 0.5987950563430786, "learning_rate": 6.478496374492875e-05, "loss": 0.4091, "step": 17370 }, { "epoch": 6.560966402416006, "grad_norm": 0.6216861009597778, "learning_rate": 6.474547048597042e-05, "loss": 0.4006, "step": 17380 }, { "epoch": 6.56474141185353, "grad_norm": 0.6449426412582397, "learning_rate": 6.470596714862205e-05, "loss": 0.4062, "step": 17390 }, { "epoch": 6.568516421291053, "grad_norm": 0.5876918435096741, "learning_rate": 6.46664537598838e-05, "loss": 0.4061, "step": 17400 }, { "epoch": 6.572291430728577, "grad_norm": 0.5914142727851868, "learning_rate": 6.462693034676271e-05, "loss": 0.4012, "step": 17410 }, { "epoch": 6.5760664401661, "grad_norm": 0.561425507068634, "learning_rate": 6.458739693627265e-05, "loss": 0.3989, "step": 17420 }, { "epoch": 6.579841449603624, "grad_norm": 0.6401806473731995, "learning_rate": 6.454785355543432e-05, "loss": 0.3995, "step": 17430 }, { "epoch": 6.5836164590411475, "grad_norm": 0.660929262638092, "learning_rate": 6.450830023127528e-05, "loss": 0.4131, "step": 17440 }, { "epoch": 6.587391468478671, "grad_norm": 0.6242609620094299, "learning_rate": 6.446873699082982e-05, "loss": 0.403, "step": 17450 }, { "epoch": 6.591166477916195, "grad_norm": 0.5753269195556641, "learning_rate": 6.44291638611391e-05, "loss": 0.4042, "step": 17460 }, { "epoch": 6.594941487353719, "grad_norm": 0.6394173502922058, "learning_rate": 6.43895808692509e-05, "loss": 0.4081, "step": 17470 }, { "epoch": 6.598716496791242, "grad_norm": 0.6524138450622559, "learning_rate": 6.434998804221986e-05, "loss": 0.4026, "step": 17480 }, { "epoch": 6.602491506228765, "grad_norm": 0.5924917459487915, "learning_rate": 6.431038540710732e-05, "loss": 0.4033, "step": 17490 }, { "epoch": 6.606266515666289, "grad_norm": 2.0175888538360596, "learning_rate": 6.427077299098129e-05, "loss": 0.4034, "step": 17500 }, { "epoch": 6.610041525103813, "grad_norm": 0.6277138590812683, "learning_rate": 6.423115082091651e-05, "loss": 0.3949, "step": 17510 }, { "epoch": 6.6138165345413364, "grad_norm": 0.5728242993354797, "learning_rate": 6.419151892399429e-05, "loss": 0.3917, "step": 17520 }, { "epoch": 6.61759154397886, "grad_norm": 0.6258454918861389, "learning_rate": 6.415187732730273e-05, "loss": 0.4026, "step": 17530 }, { "epoch": 6.621366553416384, "grad_norm": 0.574672281742096, "learning_rate": 6.411222605793645e-05, "loss": 0.4017, "step": 17540 }, { "epoch": 6.625141562853907, "grad_norm": 0.6237361431121826, "learning_rate": 6.407256514299674e-05, "loss": 0.3992, "step": 17550 }, { "epoch": 6.6289165722914305, "grad_norm": 0.6078923940658569, "learning_rate": 6.403289460959147e-05, "loss": 0.4027, "step": 17560 }, { "epoch": 6.632691581728954, "grad_norm": 0.6238769888877869, "learning_rate": 6.399321448483501e-05, "loss": 0.3993, "step": 17570 }, { "epoch": 6.636466591166478, "grad_norm": 0.732795238494873, "learning_rate": 6.395352479584844e-05, "loss": 0.4017, "step": 17580 }, { "epoch": 6.640241600604002, "grad_norm": 0.6137060523033142, "learning_rate": 6.391382556975923e-05, "loss": 0.3989, "step": 17590 }, { "epoch": 6.644016610041525, "grad_norm": 0.6174196004867554, "learning_rate": 6.387411683370144e-05, "loss": 0.4061, "step": 17600 }, { "epoch": 6.647791619479049, "grad_norm": 0.6093177199363708, "learning_rate": 6.383439861481562e-05, "loss": 0.3993, "step": 17610 }, { "epoch": 6.651566628916572, "grad_norm": 0.624611496925354, "learning_rate": 6.379467094024879e-05, "loss": 0.4061, "step": 17620 }, { "epoch": 6.655341638354096, "grad_norm": 0.7298497557640076, "learning_rate": 6.375493383715445e-05, "loss": 0.3959, "step": 17630 }, { "epoch": 6.6591166477916195, "grad_norm": 0.5991964936256409, "learning_rate": 6.371518733269254e-05, "loss": 0.3992, "step": 17640 }, { "epoch": 6.662891657229143, "grad_norm": 0.6233243942260742, "learning_rate": 6.367543145402942e-05, "loss": 0.4132, "step": 17650 }, { "epoch": 6.666666666666667, "grad_norm": 0.6464060544967651, "learning_rate": 6.363566622833785e-05, "loss": 0.399, "step": 17660 }, { "epoch": 6.67044167610419, "grad_norm": 0.7238352298736572, "learning_rate": 6.359589168279698e-05, "loss": 0.3989, "step": 17670 }, { "epoch": 6.6742166855417135, "grad_norm": 0.7250881791114807, "learning_rate": 6.355610784459235e-05, "loss": 0.4039, "step": 17680 }, { "epoch": 6.677991694979237, "grad_norm": 0.5912888646125793, "learning_rate": 6.351631474091585e-05, "loss": 0.4017, "step": 17690 }, { "epoch": 6.681766704416761, "grad_norm": 0.625900149345398, "learning_rate": 6.347651239896566e-05, "loss": 0.4034, "step": 17700 }, { "epoch": 6.685541713854285, "grad_norm": 0.6302257180213928, "learning_rate": 6.343670084594633e-05, "loss": 0.4045, "step": 17710 }, { "epoch": 6.6893167232918085, "grad_norm": 0.6661359667778015, "learning_rate": 6.339688010906866e-05, "loss": 0.4092, "step": 17720 }, { "epoch": 6.693091732729332, "grad_norm": 0.5921412706375122, "learning_rate": 6.335705021554975e-05, "loss": 0.3951, "step": 17730 }, { "epoch": 6.696866742166855, "grad_norm": 0.6364487409591675, "learning_rate": 6.3317211192613e-05, "loss": 0.4093, "step": 17740 }, { "epoch": 6.700641751604379, "grad_norm": 0.6267909407615662, "learning_rate": 6.327736306748795e-05, "loss": 0.4035, "step": 17750 }, { "epoch": 6.7044167610419025, "grad_norm": 0.5861899852752686, "learning_rate": 6.323750586741047e-05, "loss": 0.4027, "step": 17760 }, { "epoch": 6.708191770479426, "grad_norm": 0.647732138633728, "learning_rate": 6.319763961962252e-05, "loss": 0.4064, "step": 17770 }, { "epoch": 6.71196677991695, "grad_norm": 0.618825376033783, "learning_rate": 6.315776435137233e-05, "loss": 0.4008, "step": 17780 }, { "epoch": 6.715741789354474, "grad_norm": 0.7474274039268494, "learning_rate": 6.311788008991432e-05, "loss": 0.4089, "step": 17790 }, { "epoch": 6.719516798791997, "grad_norm": 0.6482340693473816, "learning_rate": 6.307798686250891e-05, "loss": 0.4037, "step": 17800 }, { "epoch": 6.72329180822952, "grad_norm": 0.6014053821563721, "learning_rate": 6.303808469642284e-05, "loss": 0.4031, "step": 17810 }, { "epoch": 6.727066817667044, "grad_norm": 0.6458423137664795, "learning_rate": 6.29981736189288e-05, "loss": 0.3998, "step": 17820 }, { "epoch": 6.730841827104568, "grad_norm": 0.6253558397293091, "learning_rate": 6.295825365730567e-05, "loss": 0.4044, "step": 17830 }, { "epoch": 6.7346168365420915, "grad_norm": 0.6395803093910217, "learning_rate": 6.291832483883835e-05, "loss": 0.4026, "step": 17840 }, { "epoch": 6.738391845979615, "grad_norm": 0.6076302528381348, "learning_rate": 6.28783871908178e-05, "loss": 0.3979, "step": 17850 }, { "epoch": 6.742166855417139, "grad_norm": 0.591390073299408, "learning_rate": 6.283844074054107e-05, "loss": 0.3967, "step": 17860 }, { "epoch": 6.745941864854662, "grad_norm": 0.5707950592041016, "learning_rate": 6.279848551531112e-05, "loss": 0.3941, "step": 17870 }, { "epoch": 6.749716874292186, "grad_norm": 0.6005119681358337, "learning_rate": 6.275852154243702e-05, "loss": 0.407, "step": 17880 }, { "epoch": 6.753491883729709, "grad_norm": 0.5978226065635681, "learning_rate": 6.271854884923377e-05, "loss": 0.3962, "step": 17890 }, { "epoch": 6.757266893167233, "grad_norm": 0.637346088886261, "learning_rate": 6.267856746302228e-05, "loss": 0.3962, "step": 17900 }, { "epoch": 6.761041902604757, "grad_norm": 0.7648394107818604, "learning_rate": 6.263857741112948e-05, "loss": 0.4004, "step": 17910 }, { "epoch": 6.76481691204228, "grad_norm": 0.6177964806556702, "learning_rate": 6.259857872088821e-05, "loss": 0.3901, "step": 17920 }, { "epoch": 6.768591921479803, "grad_norm": 0.6134040355682373, "learning_rate": 6.255857141963719e-05, "loss": 0.3985, "step": 17930 }, { "epoch": 6.772366930917327, "grad_norm": 0.5989758968353271, "learning_rate": 6.251855553472101e-05, "loss": 0.398, "step": 17940 }, { "epoch": 6.776141940354851, "grad_norm": 0.6749169230461121, "learning_rate": 6.247853109349016e-05, "loss": 0.3976, "step": 17950 }, { "epoch": 6.779916949792375, "grad_norm": 0.6461236476898193, "learning_rate": 6.243849812330098e-05, "loss": 0.3982, "step": 17960 }, { "epoch": 6.783691959229898, "grad_norm": 1.9516565799713135, "learning_rate": 6.239845665151563e-05, "loss": 0.3939, "step": 17970 }, { "epoch": 6.787466968667422, "grad_norm": 0.6713792681694031, "learning_rate": 6.235840670550204e-05, "loss": 0.4018, "step": 17980 }, { "epoch": 6.791241978104945, "grad_norm": 0.6273426413536072, "learning_rate": 6.231834831263403e-05, "loss": 0.401, "step": 17990 }, { "epoch": 6.795016987542469, "grad_norm": 0.6532526612281799, "learning_rate": 6.22782815002911e-05, "loss": 0.4034, "step": 18000 }, { "epoch": 6.798791996979992, "grad_norm": 0.6998550891876221, "learning_rate": 6.223820629585852e-05, "loss": 0.4019, "step": 18010 }, { "epoch": 6.802567006417516, "grad_norm": 0.6373661756515503, "learning_rate": 6.219812272672737e-05, "loss": 0.4007, "step": 18020 }, { "epoch": 6.80634201585504, "grad_norm": 0.6695569157600403, "learning_rate": 6.215803082029434e-05, "loss": 0.3977, "step": 18030 }, { "epoch": 6.810117025292564, "grad_norm": 0.7880366444587708, "learning_rate": 6.211793060396188e-05, "loss": 0.3985, "step": 18040 }, { "epoch": 6.813892034730086, "grad_norm": 0.6441872119903564, "learning_rate": 6.207782210513811e-05, "loss": 0.4013, "step": 18050 }, { "epoch": 6.81766704416761, "grad_norm": 0.6251243948936462, "learning_rate": 6.203770535123683e-05, "loss": 0.4084, "step": 18060 }, { "epoch": 6.821442053605134, "grad_norm": 0.710367739200592, "learning_rate": 6.199758036967747e-05, "loss": 0.4, "step": 18070 }, { "epoch": 6.825217063042658, "grad_norm": 0.568421483039856, "learning_rate": 6.195744718788503e-05, "loss": 0.3971, "step": 18080 }, { "epoch": 6.828992072480181, "grad_norm": 0.6238227486610413, "learning_rate": 6.191730583329021e-05, "loss": 0.3973, "step": 18090 }, { "epoch": 6.832767081917705, "grad_norm": 0.6105861067771912, "learning_rate": 6.187715633332921e-05, "loss": 0.3994, "step": 18100 }, { "epoch": 6.836542091355229, "grad_norm": 0.5954229235649109, "learning_rate": 6.183699871544386e-05, "loss": 0.4027, "step": 18110 }, { "epoch": 6.840317100792752, "grad_norm": 0.5816265344619751, "learning_rate": 6.179683300708152e-05, "loss": 0.3998, "step": 18120 }, { "epoch": 6.844092110230275, "grad_norm": 0.609210729598999, "learning_rate": 6.175665923569503e-05, "loss": 0.3993, "step": 18130 }, { "epoch": 6.847867119667799, "grad_norm": 0.6947378516197205, "learning_rate": 6.171647742874281e-05, "loss": 0.3964, "step": 18140 }, { "epoch": 6.851642129105323, "grad_norm": 0.6516598463058472, "learning_rate": 6.167628761368875e-05, "loss": 0.4027, "step": 18150 }, { "epoch": 6.855417138542847, "grad_norm": 0.5862056612968445, "learning_rate": 6.163608981800222e-05, "loss": 0.4079, "step": 18160 }, { "epoch": 6.8591921479803695, "grad_norm": 0.626376211643219, "learning_rate": 6.159588406915803e-05, "loss": 0.4003, "step": 18170 }, { "epoch": 6.862967157417893, "grad_norm": 0.6769831776618958, "learning_rate": 6.155567039463639e-05, "loss": 0.4041, "step": 18180 }, { "epoch": 6.866742166855417, "grad_norm": 0.6997120380401611, "learning_rate": 6.151544882192302e-05, "loss": 0.4028, "step": 18190 }, { "epoch": 6.870517176292941, "grad_norm": 0.5860679745674133, "learning_rate": 6.147521937850895e-05, "loss": 0.4132, "step": 18200 }, { "epoch": 6.874292185730464, "grad_norm": 0.5797774791717529, "learning_rate": 6.143498209189066e-05, "loss": 0.4019, "step": 18210 }, { "epoch": 6.878067195167988, "grad_norm": 0.6230421662330627, "learning_rate": 6.139473698956993e-05, "loss": 0.4041, "step": 18220 }, { "epoch": 6.881842204605512, "grad_norm": 0.5917826890945435, "learning_rate": 6.13544840990539e-05, "loss": 0.402, "step": 18230 }, { "epoch": 6.885617214043035, "grad_norm": 1.265276312828064, "learning_rate": 6.131422344785507e-05, "loss": 0.3984, "step": 18240 }, { "epoch": 6.8893922234805585, "grad_norm": 0.5542703866958618, "learning_rate": 6.127395506349119e-05, "loss": 0.3988, "step": 18250 }, { "epoch": 6.893167232918082, "grad_norm": 0.736250102519989, "learning_rate": 6.123367897348533e-05, "loss": 0.3968, "step": 18260 }, { "epoch": 6.896942242355606, "grad_norm": 0.8266544342041016, "learning_rate": 6.119339520536584e-05, "loss": 0.4013, "step": 18270 }, { "epoch": 6.90071725179313, "grad_norm": 0.6568412780761719, "learning_rate": 6.115310378666625e-05, "loss": 0.4019, "step": 18280 }, { "epoch": 6.904492261230653, "grad_norm": 0.5549454689025879, "learning_rate": 6.11128047449254e-05, "loss": 0.3982, "step": 18290 }, { "epoch": 6.908267270668176, "grad_norm": 0.6285567283630371, "learning_rate": 6.107249810768729e-05, "loss": 0.4012, "step": 18300 }, { "epoch": 6.9120422801057, "grad_norm": 0.6208999156951904, "learning_rate": 6.1032183902501125e-05, "loss": 0.3999, "step": 18310 }, { "epoch": 6.915817289543224, "grad_norm": 0.5961741805076599, "learning_rate": 6.099186215692131e-05, "loss": 0.4093, "step": 18320 }, { "epoch": 6.9195922989807475, "grad_norm": 0.5774527192115784, "learning_rate": 6.095153289850734e-05, "loss": 0.401, "step": 18330 }, { "epoch": 6.923367308418271, "grad_norm": 0.6928045749664307, "learning_rate": 6.0911196154823904e-05, "loss": 0.3907, "step": 18340 }, { "epoch": 6.927142317855795, "grad_norm": 0.7921097874641418, "learning_rate": 6.087085195344079e-05, "loss": 0.395, "step": 18350 }, { "epoch": 6.930917327293319, "grad_norm": 0.6023475527763367, "learning_rate": 6.083050032193286e-05, "loss": 0.4031, "step": 18360 }, { "epoch": 6.9346923367308415, "grad_norm": 1.3727141618728638, "learning_rate": 6.0790141287880097e-05, "loss": 0.4059, "step": 18370 }, { "epoch": 6.938467346168365, "grad_norm": 0.606326699256897, "learning_rate": 6.0749774878867496e-05, "loss": 0.3957, "step": 18380 }, { "epoch": 6.942242355605889, "grad_norm": 0.6082478165626526, "learning_rate": 6.0709401122485146e-05, "loss": 0.3975, "step": 18390 }, { "epoch": 6.946017365043413, "grad_norm": 0.7824208736419678, "learning_rate": 6.066902004632811e-05, "loss": 0.3967, "step": 18400 }, { "epoch": 6.9497923744809365, "grad_norm": 0.6232104301452637, "learning_rate": 6.062863167799646e-05, "loss": 0.4024, "step": 18410 }, { "epoch": 6.953567383918459, "grad_norm": 0.6417713165283203, "learning_rate": 6.058823604509529e-05, "loss": 0.3967, "step": 18420 }, { "epoch": 6.957342393355983, "grad_norm": 0.6295148730278015, "learning_rate": 6.054783317523462e-05, "loss": 0.3991, "step": 18430 }, { "epoch": 6.961117402793507, "grad_norm": 0.6527547836303711, "learning_rate": 6.050742309602944e-05, "loss": 0.3988, "step": 18440 }, { "epoch": 6.9648924122310305, "grad_norm": 0.6303055882453918, "learning_rate": 6.046700583509965e-05, "loss": 0.3933, "step": 18450 }, { "epoch": 6.968667421668554, "grad_norm": 0.637154757976532, "learning_rate": 6.042658142007007e-05, "loss": 0.4024, "step": 18460 }, { "epoch": 6.972442431106078, "grad_norm": 0.6554228663444519, "learning_rate": 6.038614987857041e-05, "loss": 0.3999, "step": 18470 }, { "epoch": 6.976217440543602, "grad_norm": 0.6131272315979004, "learning_rate": 6.0345711238235224e-05, "loss": 0.3946, "step": 18480 }, { "epoch": 6.9799924499811254, "grad_norm": 0.5949199199676514, "learning_rate": 6.030526552670399e-05, "loss": 0.4006, "step": 18490 }, { "epoch": 6.983767459418648, "grad_norm": 0.5946569442749023, "learning_rate": 6.0264812771620925e-05, "loss": 0.3971, "step": 18500 }, { "epoch": 6.987542468856172, "grad_norm": 0.7052402496337891, "learning_rate": 6.022435300063512e-05, "loss": 0.3864, "step": 18510 }, { "epoch": 6.991317478293696, "grad_norm": 0.5748345255851746, "learning_rate": 6.0183886241400466e-05, "loss": 0.4049, "step": 18520 }, { "epoch": 6.9950924877312195, "grad_norm": 0.6450919508934021, "learning_rate": 6.0143412521575584e-05, "loss": 0.4042, "step": 18530 }, { "epoch": 6.998867497168743, "grad_norm": 0.7220432162284851, "learning_rate": 6.010293186882389e-05, "loss": 0.4025, "step": 18540 }, { "epoch": 7.002642506606266, "grad_norm": 0.6318272352218628, "learning_rate": 6.0062444310813525e-05, "loss": 0.4026, "step": 18550 }, { "epoch": 7.00641751604379, "grad_norm": 0.5739586353302002, "learning_rate": 6.0021949875217355e-05, "loss": 0.3905, "step": 18560 }, { "epoch": 7.0101925254813136, "grad_norm": 0.6182407140731812, "learning_rate": 5.998144858971295e-05, "loss": 0.3936, "step": 18570 }, { "epoch": 7.013967534918837, "grad_norm": 0.6267920732498169, "learning_rate": 5.994094048198257e-05, "loss": 0.4063, "step": 18580 }, { "epoch": 7.017742544356361, "grad_norm": 0.7207063436508179, "learning_rate": 5.990042557971307e-05, "loss": 0.3992, "step": 18590 }, { "epoch": 7.021517553793885, "grad_norm": 0.6681888103485107, "learning_rate": 5.985990391059607e-05, "loss": 0.3976, "step": 18600 }, { "epoch": 7.0252925632314085, "grad_norm": 0.6886135935783386, "learning_rate": 5.981937550232771e-05, "loss": 0.3898, "step": 18610 }, { "epoch": 7.029067572668931, "grad_norm": 0.7061811685562134, "learning_rate": 5.9778840382608794e-05, "loss": 0.4119, "step": 18620 }, { "epoch": 7.032842582106455, "grad_norm": 0.5984750986099243, "learning_rate": 5.9738298579144695e-05, "loss": 0.3965, "step": 18630 }, { "epoch": 7.036617591543979, "grad_norm": 0.6792024374008179, "learning_rate": 5.9697750119645314e-05, "loss": 0.4087, "step": 18640 }, { "epoch": 7.0403926009815025, "grad_norm": 0.5802191495895386, "learning_rate": 5.96571950318252e-05, "loss": 0.4003, "step": 18650 }, { "epoch": 7.044167610419026, "grad_norm": 0.6558259129524231, "learning_rate": 5.9616633343403316e-05, "loss": 0.408, "step": 18660 }, { "epoch": 7.04794261985655, "grad_norm": 0.7390403151512146, "learning_rate": 5.957606508210324e-05, "loss": 0.404, "step": 18670 }, { "epoch": 7.051717629294073, "grad_norm": 0.6561256051063538, "learning_rate": 5.953549027565297e-05, "loss": 0.3936, "step": 18680 }, { "epoch": 7.055492638731597, "grad_norm": 0.6152408123016357, "learning_rate": 5.949490895178501e-05, "loss": 0.3959, "step": 18690 }, { "epoch": 7.05926764816912, "grad_norm": 1.134881854057312, "learning_rate": 5.945432113823632e-05, "loss": 0.3955, "step": 18700 }, { "epoch": 7.063042657606644, "grad_norm": 0.611450731754303, "learning_rate": 5.9413726862748276e-05, "loss": 0.3999, "step": 18710 }, { "epoch": 7.066817667044168, "grad_norm": 1.1153512001037598, "learning_rate": 5.9373126153066694e-05, "loss": 0.3965, "step": 18720 }, { "epoch": 7.0705926764816915, "grad_norm": 0.6631627678871155, "learning_rate": 5.933251903694177e-05, "loss": 0.4053, "step": 18730 }, { "epoch": 7.074367685919214, "grad_norm": 0.6014063954353333, "learning_rate": 5.929190554212807e-05, "loss": 0.3936, "step": 18740 }, { "epoch": 7.078142695356738, "grad_norm": 0.6215736269950867, "learning_rate": 5.9251285696384565e-05, "loss": 0.404, "step": 18750 }, { "epoch": 7.081917704794262, "grad_norm": 0.791002094745636, "learning_rate": 5.921065952747451e-05, "loss": 0.3992, "step": 18760 }, { "epoch": 7.085692714231786, "grad_norm": 0.9129757881164551, "learning_rate": 5.917002706316552e-05, "loss": 0.3984, "step": 18770 }, { "epoch": 7.089467723669309, "grad_norm": 0.5969563722610474, "learning_rate": 5.912938833122952e-05, "loss": 0.3988, "step": 18780 }, { "epoch": 7.093242733106833, "grad_norm": 1.1693058013916016, "learning_rate": 5.908874335944265e-05, "loss": 0.3969, "step": 18790 }, { "epoch": 7.097017742544356, "grad_norm": 0.6244138479232788, "learning_rate": 5.904809217558542e-05, "loss": 0.3992, "step": 18800 }, { "epoch": 7.10079275198188, "grad_norm": 0.6120061874389648, "learning_rate": 5.90074348074425e-05, "loss": 0.3984, "step": 18810 }, { "epoch": 7.104567761419403, "grad_norm": 0.6857101917266846, "learning_rate": 5.8966771282802814e-05, "loss": 0.3949, "step": 18820 }, { "epoch": 7.108342770856927, "grad_norm": 0.6439679265022278, "learning_rate": 5.892610162945952e-05, "loss": 0.401, "step": 18830 }, { "epoch": 7.112117780294451, "grad_norm": 0.686291515827179, "learning_rate": 5.8885425875209924e-05, "loss": 0.3991, "step": 18840 }, { "epoch": 7.115892789731975, "grad_norm": 0.893528938293457, "learning_rate": 5.884474404785553e-05, "loss": 0.3903, "step": 18850 }, { "epoch": 7.119667799169498, "grad_norm": 0.6380524635314941, "learning_rate": 5.8804056175201983e-05, "loss": 0.3936, "step": 18860 }, { "epoch": 7.123442808607021, "grad_norm": 0.5932635068893433, "learning_rate": 5.876336228505904e-05, "loss": 0.3988, "step": 18870 }, { "epoch": 7.127217818044545, "grad_norm": 0.6427797079086304, "learning_rate": 5.872266240524062e-05, "loss": 0.4023, "step": 18880 }, { "epoch": 7.130992827482069, "grad_norm": 0.625927209854126, "learning_rate": 5.86819565635647e-05, "loss": 0.3963, "step": 18890 }, { "epoch": 7.134767836919592, "grad_norm": 0.5989848971366882, "learning_rate": 5.8641244787853334e-05, "loss": 0.3958, "step": 18900 }, { "epoch": 7.138542846357116, "grad_norm": 0.6157662868499756, "learning_rate": 5.860052710593265e-05, "loss": 0.3974, "step": 18910 }, { "epoch": 7.14231785579464, "grad_norm": 0.6159812211990356, "learning_rate": 5.855980354563276e-05, "loss": 0.3868, "step": 18920 }, { "epoch": 7.146092865232163, "grad_norm": 0.5764108300209045, "learning_rate": 5.8519074134787874e-05, "loss": 0.3937, "step": 18930 }, { "epoch": 7.149867874669686, "grad_norm": 0.657748281955719, "learning_rate": 5.847833890123614e-05, "loss": 0.3937, "step": 18940 }, { "epoch": 7.15364288410721, "grad_norm": 0.6909071207046509, "learning_rate": 5.8437597872819737e-05, "loss": 0.397, "step": 18950 }, { "epoch": 7.157417893544734, "grad_norm": 0.6924566030502319, "learning_rate": 5.839685107738473e-05, "loss": 0.397, "step": 18960 }, { "epoch": 7.161192902982258, "grad_norm": 0.6301664710044861, "learning_rate": 5.835609854278118e-05, "loss": 0.3921, "step": 18970 }, { "epoch": 7.164967912419781, "grad_norm": 0.6503030061721802, "learning_rate": 5.831534029686308e-05, "loss": 0.3958, "step": 18980 }, { "epoch": 7.168742921857304, "grad_norm": 0.6070290207862854, "learning_rate": 5.82745763674883e-05, "loss": 0.3943, "step": 18990 }, { "epoch": 7.172517931294828, "grad_norm": 0.6799952387809753, "learning_rate": 5.823380678251861e-05, "loss": 0.3899, "step": 19000 }, { "epoch": 7.176292940732352, "grad_norm": 0.6328525543212891, "learning_rate": 5.81930315698196e-05, "loss": 0.3935, "step": 19010 }, { "epoch": 7.180067950169875, "grad_norm": 0.6666650176048279, "learning_rate": 5.815225075726076e-05, "loss": 0.3931, "step": 19020 }, { "epoch": 7.183842959607399, "grad_norm": 0.6060150861740112, "learning_rate": 5.811146437271543e-05, "loss": 0.3976, "step": 19030 }, { "epoch": 7.187617969044923, "grad_norm": 0.6627577543258667, "learning_rate": 5.807067244406066e-05, "loss": 0.4025, "step": 19040 }, { "epoch": 7.191392978482447, "grad_norm": 0.5842311978340149, "learning_rate": 5.8029874999177405e-05, "loss": 0.3941, "step": 19050 }, { "epoch": 7.1951679879199695, "grad_norm": 1.3617955446243286, "learning_rate": 5.798907206595029e-05, "loss": 0.3934, "step": 19060 }, { "epoch": 7.198942997357493, "grad_norm": 0.5990311503410339, "learning_rate": 5.794826367226773e-05, "loss": 0.3947, "step": 19070 }, { "epoch": 7.202718006795017, "grad_norm": 0.5806323885917664, "learning_rate": 5.790744984602193e-05, "loss": 0.4021, "step": 19080 }, { "epoch": 7.206493016232541, "grad_norm": 0.6318025588989258, "learning_rate": 5.786663061510872e-05, "loss": 0.3978, "step": 19090 }, { "epoch": 7.210268025670064, "grad_norm": 0.6155949234962463, "learning_rate": 5.782580600742765e-05, "loss": 0.395, "step": 19100 }, { "epoch": 7.214043035107588, "grad_norm": 0.6419726014137268, "learning_rate": 5.7784976050881965e-05, "loss": 0.395, "step": 19110 }, { "epoch": 7.217818044545111, "grad_norm": 0.6743911504745483, "learning_rate": 5.774414077337855e-05, "loss": 0.3907, "step": 19120 }, { "epoch": 7.221593053982635, "grad_norm": 0.6767613887786865, "learning_rate": 5.770330020282796e-05, "loss": 0.3911, "step": 19130 }, { "epoch": 7.2253680634201585, "grad_norm": 0.6418374180793762, "learning_rate": 5.7662454367144317e-05, "loss": 0.3978, "step": 19140 }, { "epoch": 7.229143072857682, "grad_norm": 0.5886910557746887, "learning_rate": 5.762160329424536e-05, "loss": 0.398, "step": 19150 }, { "epoch": 7.232918082295206, "grad_norm": 0.56149822473526, "learning_rate": 5.7580747012052416e-05, "loss": 0.3917, "step": 19160 }, { "epoch": 7.23669309173273, "grad_norm": 0.8316841125488281, "learning_rate": 5.753988554849037e-05, "loss": 0.3967, "step": 19170 }, { "epoch": 7.2404681011702525, "grad_norm": 0.5947608351707458, "learning_rate": 5.749901893148766e-05, "loss": 0.395, "step": 19180 }, { "epoch": 7.244243110607776, "grad_norm": 0.6548729538917542, "learning_rate": 5.745814718897621e-05, "loss": 0.3925, "step": 19190 }, { "epoch": 7.2480181200453, "grad_norm": 0.5887550115585327, "learning_rate": 5.74172703488915e-05, "loss": 0.3922, "step": 19200 }, { "epoch": 7.251793129482824, "grad_norm": 0.6358749270439148, "learning_rate": 5.737638843917242e-05, "loss": 0.3876, "step": 19210 }, { "epoch": 7.2555681389203475, "grad_norm": 0.7427713871002197, "learning_rate": 5.73355014877614e-05, "loss": 0.3955, "step": 19220 }, { "epoch": 7.259343148357871, "grad_norm": 0.6359094977378845, "learning_rate": 5.7294609522604316e-05, "loss": 0.3964, "step": 19230 }, { "epoch": 7.263118157795395, "grad_norm": 0.5932873487472534, "learning_rate": 5.7253712571650376e-05, "loss": 0.3966, "step": 19240 }, { "epoch": 7.266893167232918, "grad_norm": 0.6501133441925049, "learning_rate": 5.721281066285229e-05, "loss": 0.3909, "step": 19250 }, { "epoch": 7.2706681766704415, "grad_norm": 0.6515145301818848, "learning_rate": 5.717190382416615e-05, "loss": 0.3967, "step": 19260 }, { "epoch": 7.274443186107965, "grad_norm": 0.6237747073173523, "learning_rate": 5.713099208355135e-05, "loss": 0.3959, "step": 19270 }, { "epoch": 7.278218195545489, "grad_norm": 0.5545824766159058, "learning_rate": 5.709007546897074e-05, "loss": 0.396, "step": 19280 }, { "epoch": 7.281993204983013, "grad_norm": 0.6651487350463867, "learning_rate": 5.704915400839037e-05, "loss": 0.3994, "step": 19290 }, { "epoch": 7.2857682144205365, "grad_norm": 0.701859176158905, "learning_rate": 5.700822772977971e-05, "loss": 0.4039, "step": 19300 }, { "epoch": 7.289543223858059, "grad_norm": 0.9425735473632812, "learning_rate": 5.696729666111148e-05, "loss": 0.4047, "step": 19310 }, { "epoch": 7.293318233295583, "grad_norm": 0.6198208928108215, "learning_rate": 5.692636083036168e-05, "loss": 0.4123, "step": 19320 }, { "epoch": 7.297093242733107, "grad_norm": 0.7142148613929749, "learning_rate": 5.688542026550958e-05, "loss": 0.3903, "step": 19330 }, { "epoch": 7.3008682521706305, "grad_norm": 0.6616962552070618, "learning_rate": 5.684447499453763e-05, "loss": 0.3973, "step": 19340 }, { "epoch": 7.304643261608154, "grad_norm": 0.6994644403457642, "learning_rate": 5.680352504543156e-05, "loss": 0.3936, "step": 19350 }, { "epoch": 7.308418271045678, "grad_norm": 0.6935665011405945, "learning_rate": 5.67625704461803e-05, "loss": 0.3882, "step": 19360 }, { "epoch": 7.312193280483201, "grad_norm": 0.6593263745307922, "learning_rate": 5.672161122477589e-05, "loss": 0.4025, "step": 19370 }, { "epoch": 7.315968289920725, "grad_norm": 1.2950445413589478, "learning_rate": 5.668064740921359e-05, "loss": 0.3984, "step": 19380 }, { "epoch": 7.319743299358248, "grad_norm": 0.6627549529075623, "learning_rate": 5.663967902749179e-05, "loss": 0.3958, "step": 19390 }, { "epoch": 7.323518308795772, "grad_norm": 0.7319017648696899, "learning_rate": 5.6598706107611965e-05, "loss": 0.3883, "step": 19400 }, { "epoch": 7.327293318233296, "grad_norm": 0.6958197951316833, "learning_rate": 5.655772867757876e-05, "loss": 0.3962, "step": 19410 }, { "epoch": 7.3310683276708195, "grad_norm": 0.6247055530548096, "learning_rate": 5.651674676539982e-05, "loss": 0.3911, "step": 19420 }, { "epoch": 7.334843337108342, "grad_norm": 0.6021857857704163, "learning_rate": 5.647576039908593e-05, "loss": 0.3909, "step": 19430 }, { "epoch": 7.338618346545866, "grad_norm": 0.6767669916152954, "learning_rate": 5.6434769606650864e-05, "loss": 0.3908, "step": 19440 }, { "epoch": 7.34239335598339, "grad_norm": 0.8901523351669312, "learning_rate": 5.639377441611143e-05, "loss": 0.3988, "step": 19450 }, { "epoch": 7.346168365420914, "grad_norm": 0.6947163939476013, "learning_rate": 5.635277485548751e-05, "loss": 0.3917, "step": 19460 }, { "epoch": 7.349943374858437, "grad_norm": 0.7375092506408691, "learning_rate": 5.631177095280186e-05, "loss": 0.3965, "step": 19470 }, { "epoch": 7.353718384295961, "grad_norm": 1.315040111541748, "learning_rate": 5.627076273608027e-05, "loss": 0.3952, "step": 19480 }, { "epoch": 7.357493393733485, "grad_norm": 0.6428912878036499, "learning_rate": 5.622975023335148e-05, "loss": 0.3936, "step": 19490 }, { "epoch": 7.361268403171008, "grad_norm": 0.7309068441390991, "learning_rate": 5.618873347264716e-05, "loss": 0.3972, "step": 19500 }, { "epoch": 7.365043412608531, "grad_norm": 0.6258814930915833, "learning_rate": 5.614771248200188e-05, "loss": 0.3858, "step": 19510 }, { "epoch": 7.368818422046055, "grad_norm": 0.64792400598526, "learning_rate": 5.6106687289453066e-05, "loss": 0.398, "step": 19520 }, { "epoch": 7.372593431483579, "grad_norm": 0.6821390986442566, "learning_rate": 5.606565792304108e-05, "loss": 0.4027, "step": 19530 }, { "epoch": 7.3763684409211026, "grad_norm": 0.7018711566925049, "learning_rate": 5.602462441080909e-05, "loss": 0.3907, "step": 19540 }, { "epoch": 7.380143450358626, "grad_norm": 0.6835519671440125, "learning_rate": 5.5983586780803135e-05, "loss": 0.3912, "step": 19550 }, { "epoch": 7.383918459796149, "grad_norm": 0.6249481439590454, "learning_rate": 5.594254506107205e-05, "loss": 0.3903, "step": 19560 }, { "epoch": 7.387693469233673, "grad_norm": 0.6832369565963745, "learning_rate": 5.590149927966743e-05, "loss": 0.3882, "step": 19570 }, { "epoch": 7.391468478671197, "grad_norm": 0.6031800508499146, "learning_rate": 5.58604494646437e-05, "loss": 0.3954, "step": 19580 }, { "epoch": 7.39524348810872, "grad_norm": 0.6349557638168335, "learning_rate": 5.5819395644058025e-05, "loss": 0.3894, "step": 19590 }, { "epoch": 7.399018497546244, "grad_norm": 0.6155382990837097, "learning_rate": 5.577833784597031e-05, "loss": 0.3928, "step": 19600 }, { "epoch": 7.402793506983768, "grad_norm": 0.6620678305625916, "learning_rate": 5.573727609844316e-05, "loss": 0.3906, "step": 19610 }, { "epoch": 7.406568516421291, "grad_norm": 0.6608960628509521, "learning_rate": 5.5696210429541884e-05, "loss": 0.3975, "step": 19620 }, { "epoch": 7.410343525858814, "grad_norm": 0.6048488020896912, "learning_rate": 5.565514086733451e-05, "loss": 0.3969, "step": 19630 }, { "epoch": 7.414118535296338, "grad_norm": 0.7584668397903442, "learning_rate": 5.5614067439891657e-05, "loss": 0.3942, "step": 19640 }, { "epoch": 7.417893544733862, "grad_norm": 0.6703759431838989, "learning_rate": 5.557299017528666e-05, "loss": 0.3879, "step": 19650 }, { "epoch": 7.421668554171386, "grad_norm": 0.6314090490341187, "learning_rate": 5.5531909101595436e-05, "loss": 0.391, "step": 19660 }, { "epoch": 7.425443563608909, "grad_norm": 0.6210800409317017, "learning_rate": 5.549082424689649e-05, "loss": 0.4007, "step": 19670 }, { "epoch": 7.429218573046432, "grad_norm": 0.6025400161743164, "learning_rate": 5.544973563927095e-05, "loss": 0.3935, "step": 19680 }, { "epoch": 7.432993582483956, "grad_norm": 0.7302088141441345, "learning_rate": 5.540864330680249e-05, "loss": 0.4046, "step": 19690 }, { "epoch": 7.43676859192148, "grad_norm": 0.6146891117095947, "learning_rate": 5.536754727757733e-05, "loss": 0.396, "step": 19700 }, { "epoch": 7.440543601359003, "grad_norm": 0.6504469513893127, "learning_rate": 5.532644757968422e-05, "loss": 0.3957, "step": 19710 }, { "epoch": 7.444318610796527, "grad_norm": 0.6892192363739014, "learning_rate": 5.528534424121441e-05, "loss": 0.5172, "step": 19720 }, { "epoch": 7.448093620234051, "grad_norm": 0.673173189163208, "learning_rate": 5.524423729026165e-05, "loss": 0.3959, "step": 19730 }, { "epoch": 7.451868629671575, "grad_norm": 0.689117431640625, "learning_rate": 5.5203126754922164e-05, "loss": 0.3885, "step": 19740 }, { "epoch": 7.455643639109097, "grad_norm": 0.6626030802726746, "learning_rate": 5.5162012663294585e-05, "loss": 0.3877, "step": 19750 }, { "epoch": 7.459418648546621, "grad_norm": 0.7295408248901367, "learning_rate": 5.512089504348003e-05, "loss": 0.392, "step": 19760 }, { "epoch": 7.463193657984145, "grad_norm": 0.6033636331558228, "learning_rate": 5.5079773923582e-05, "loss": 0.3939, "step": 19770 }, { "epoch": 7.466968667421669, "grad_norm": 0.6206973791122437, "learning_rate": 5.50386493317064e-05, "loss": 0.3915, "step": 19780 }, { "epoch": 7.470743676859192, "grad_norm": 0.6662564277648926, "learning_rate": 5.49975212959615e-05, "loss": 0.3923, "step": 19790 }, { "epoch": 7.474518686296716, "grad_norm": 0.6757304072380066, "learning_rate": 5.4956389844457904e-05, "loss": 0.3936, "step": 19800 }, { "epoch": 7.478293695734239, "grad_norm": 0.717755913734436, "learning_rate": 5.491525500530859e-05, "loss": 0.3932, "step": 19810 }, { "epoch": 7.482068705171763, "grad_norm": 0.6577935218811035, "learning_rate": 5.487411680662882e-05, "loss": 0.3985, "step": 19820 }, { "epoch": 7.485843714609286, "grad_norm": 0.6445197463035583, "learning_rate": 5.483297527653618e-05, "loss": 0.3979, "step": 19830 }, { "epoch": 7.48961872404681, "grad_norm": 0.6646354794502258, "learning_rate": 5.4791830443150516e-05, "loss": 0.3896, "step": 19840 }, { "epoch": 7.493393733484334, "grad_norm": 0.6168741583824158, "learning_rate": 5.475068233459392e-05, "loss": 0.398, "step": 19850 }, { "epoch": 7.497168742921858, "grad_norm": 0.7372630834579468, "learning_rate": 5.470953097899075e-05, "loss": 0.3972, "step": 19860 }, { "epoch": 7.500943752359381, "grad_norm": 0.7275758981704712, "learning_rate": 5.466837640446756e-05, "loss": 0.3923, "step": 19870 }, { "epoch": 7.504718761796904, "grad_norm": 0.6699514389038086, "learning_rate": 5.462721863915312e-05, "loss": 0.3958, "step": 19880 }, { "epoch": 7.508493771234428, "grad_norm": 0.6782812476158142, "learning_rate": 5.4586057711178374e-05, "loss": 0.3939, "step": 19890 }, { "epoch": 7.512268780671952, "grad_norm": 0.6210011839866638, "learning_rate": 5.454489364867642e-05, "loss": 0.3935, "step": 19900 }, { "epoch": 7.516043790109475, "grad_norm": 0.6767253279685974, "learning_rate": 5.4503726479782523e-05, "loss": 0.3922, "step": 19910 }, { "epoch": 7.519818799546999, "grad_norm": 0.5947399735450745, "learning_rate": 5.446255623263403e-05, "loss": 0.3867, "step": 19920 }, { "epoch": 7.523593808984522, "grad_norm": 0.6279717683792114, "learning_rate": 5.4421382935370445e-05, "loss": 0.3933, "step": 19930 }, { "epoch": 7.527368818422046, "grad_norm": 0.6436152458190918, "learning_rate": 5.438020661613331e-05, "loss": 0.3879, "step": 19940 }, { "epoch": 7.5311438278595695, "grad_norm": 0.6227508187294006, "learning_rate": 5.433902730306625e-05, "loss": 0.3899, "step": 19950 }, { "epoch": 7.534918837297093, "grad_norm": 0.7720093131065369, "learning_rate": 5.429784502431495e-05, "loss": 0.3946, "step": 19960 }, { "epoch": 7.538693846734617, "grad_norm": 0.7744880318641663, "learning_rate": 5.42566598080271e-05, "loss": 0.3946, "step": 19970 }, { "epoch": 7.542468856172141, "grad_norm": 1.112042784690857, "learning_rate": 5.421547168235241e-05, "loss": 0.392, "step": 19980 }, { "epoch": 7.546243865609664, "grad_norm": 0.6036012172698975, "learning_rate": 5.417428067544258e-05, "loss": 0.3993, "step": 19990 }, { "epoch": 7.550018875047187, "grad_norm": 0.6224648952484131, "learning_rate": 5.413308681545126e-05, "loss": 0.3947, "step": 20000 }, { "epoch": 7.553793884484711, "grad_norm": 0.6193528771400452, "learning_rate": 5.409189013053408e-05, "loss": 0.3918, "step": 20010 }, { "epoch": 7.557568893922235, "grad_norm": 0.5908026099205017, "learning_rate": 5.4050690648848576e-05, "loss": 0.3911, "step": 20020 }, { "epoch": 7.5613439033597585, "grad_norm": 0.6350475549697876, "learning_rate": 5.400948839855421e-05, "loss": 0.3936, "step": 20030 }, { "epoch": 7.565118912797282, "grad_norm": 0.6277138590812683, "learning_rate": 5.396828340781234e-05, "loss": 0.3937, "step": 20040 }, { "epoch": 7.568893922234806, "grad_norm": 0.8242809176445007, "learning_rate": 5.392707570478617e-05, "loss": 0.407, "step": 20050 }, { "epoch": 7.572668931672329, "grad_norm": 0.7098134756088257, "learning_rate": 5.388586531764078e-05, "loss": 0.3781, "step": 20060 }, { "epoch": 7.5764439411098525, "grad_norm": 0.9367830753326416, "learning_rate": 5.384465227454311e-05, "loss": 0.3971, "step": 20070 }, { "epoch": 7.580218950547376, "grad_norm": 0.5797680616378784, "learning_rate": 5.380343660366184e-05, "loss": 0.3923, "step": 20080 }, { "epoch": 7.5839939599849, "grad_norm": 0.6200445890426636, "learning_rate": 5.376221833316752e-05, "loss": 0.3958, "step": 20090 }, { "epoch": 7.587768969422424, "grad_norm": 0.651992678642273, "learning_rate": 5.3720997491232436e-05, "loss": 0.3893, "step": 20100 }, { "epoch": 7.5915439788599475, "grad_norm": 0.6731122732162476, "learning_rate": 5.367977410603068e-05, "loss": 0.3918, "step": 20110 }, { "epoch": 7.595318988297471, "grad_norm": 0.5849369764328003, "learning_rate": 5.3638548205738004e-05, "loss": 0.3909, "step": 20120 }, { "epoch": 7.599093997734994, "grad_norm": 0.6614940762519836, "learning_rate": 5.359731981853194e-05, "loss": 0.3949, "step": 20130 }, { "epoch": 7.602869007172518, "grad_norm": 3.908480644226074, "learning_rate": 5.35560889725917e-05, "loss": 0.3875, "step": 20140 }, { "epoch": 7.6066440166100415, "grad_norm": 0.6389253735542297, "learning_rate": 5.3514855696098176e-05, "loss": 0.3883, "step": 20150 }, { "epoch": 7.610419026047565, "grad_norm": 0.6248882412910461, "learning_rate": 5.347362001723394e-05, "loss": 0.3848, "step": 20160 }, { "epoch": 7.614194035485089, "grad_norm": 0.6055360436439514, "learning_rate": 5.3432381964183176e-05, "loss": 0.3936, "step": 20170 }, { "epoch": 7.617969044922612, "grad_norm": 0.7338094115257263, "learning_rate": 5.3391141565131685e-05, "loss": 0.3904, "step": 20180 }, { "epoch": 7.621744054360136, "grad_norm": 0.8716061115264893, "learning_rate": 5.3349898848266935e-05, "loss": 0.3902, "step": 20190 }, { "epoch": 7.625519063797659, "grad_norm": 1.380915641784668, "learning_rate": 5.330865384177789e-05, "loss": 0.3927, "step": 20200 }, { "epoch": 7.629294073235183, "grad_norm": 0.6246406435966492, "learning_rate": 5.326740657385515e-05, "loss": 0.389, "step": 20210 }, { "epoch": 7.633069082672707, "grad_norm": 0.6647712588310242, "learning_rate": 5.322615707269083e-05, "loss": 0.3956, "step": 20220 }, { "epoch": 7.6368440921102305, "grad_norm": 0.6063628792762756, "learning_rate": 5.318490536647856e-05, "loss": 0.3887, "step": 20230 }, { "epoch": 7.640619101547754, "grad_norm": 0.625741720199585, "learning_rate": 5.3143651483413524e-05, "loss": 0.3936, "step": 20240 }, { "epoch": 7.644394110985277, "grad_norm": 0.676576554775238, "learning_rate": 5.310239545169232e-05, "loss": 0.3916, "step": 20250 }, { "epoch": 7.648169120422801, "grad_norm": 0.5982394218444824, "learning_rate": 5.30611372995131e-05, "loss": 0.3868, "step": 20260 }, { "epoch": 7.651944129860325, "grad_norm": 0.6712155342102051, "learning_rate": 5.30198770550754e-05, "loss": 0.3903, "step": 20270 }, { "epoch": 7.655719139297848, "grad_norm": 0.5862340331077576, "learning_rate": 5.297861474658019e-05, "loss": 0.389, "step": 20280 }, { "epoch": 7.659494148735372, "grad_norm": 0.6452902555465698, "learning_rate": 5.29373504022299e-05, "loss": 0.3847, "step": 20290 }, { "epoch": 7.663269158172896, "grad_norm": 0.6566249132156372, "learning_rate": 5.28960840502283e-05, "loss": 0.3969, "step": 20300 }, { "epoch": 7.667044167610419, "grad_norm": 0.6458132863044739, "learning_rate": 5.285481571878056e-05, "loss": 0.3942, "step": 20310 }, { "epoch": 7.670819177047942, "grad_norm": 0.6279521584510803, "learning_rate": 5.281354543609321e-05, "loss": 0.3903, "step": 20320 }, { "epoch": 7.674594186485466, "grad_norm": 0.5799058079719543, "learning_rate": 5.277227323037406e-05, "loss": 0.3871, "step": 20330 }, { "epoch": 7.67836919592299, "grad_norm": 0.5981761813163757, "learning_rate": 5.273099912983233e-05, "loss": 0.3854, "step": 20340 }, { "epoch": 7.682144205360514, "grad_norm": 0.6565678119659424, "learning_rate": 5.268972316267843e-05, "loss": 0.3878, "step": 20350 }, { "epoch": 7.685919214798037, "grad_norm": 0.6143926382064819, "learning_rate": 5.26484453571241e-05, "loss": 0.3997, "step": 20360 }, { "epoch": 7.689694224235561, "grad_norm": 0.630825936794281, "learning_rate": 5.260716574138235e-05, "loss": 0.3865, "step": 20370 }, { "epoch": 7.693469233673084, "grad_norm": 0.636688768863678, "learning_rate": 5.256588434366739e-05, "loss": 0.3895, "step": 20380 }, { "epoch": 7.697244243110608, "grad_norm": 0.675250768661499, "learning_rate": 5.25246011921947e-05, "loss": 0.3901, "step": 20390 }, { "epoch": 7.701019252548131, "grad_norm": 0.6246991157531738, "learning_rate": 5.248331631518089e-05, "loss": 0.385, "step": 20400 }, { "epoch": 7.704794261985655, "grad_norm": 0.6083090901374817, "learning_rate": 5.244202974084379e-05, "loss": 0.3906, "step": 20410 }, { "epoch": 7.708569271423179, "grad_norm": 0.63591068983078, "learning_rate": 5.240074149740239e-05, "loss": 0.3842, "step": 20420 }, { "epoch": 7.712344280860702, "grad_norm": 0.7097567319869995, "learning_rate": 5.2359451613076814e-05, "loss": 0.3777, "step": 20430 }, { "epoch": 7.716119290298225, "grad_norm": 0.6984173059463501, "learning_rate": 5.231816011608832e-05, "loss": 0.3889, "step": 20440 }, { "epoch": 7.719894299735749, "grad_norm": 0.6347232460975647, "learning_rate": 5.227686703465924e-05, "loss": 0.3888, "step": 20450 }, { "epoch": 7.723669309173273, "grad_norm": 0.5677266120910645, "learning_rate": 5.2235572397013e-05, "loss": 0.3899, "step": 20460 }, { "epoch": 7.727444318610797, "grad_norm": 0.579582929611206, "learning_rate": 5.2194276231374114e-05, "loss": 0.3896, "step": 20470 }, { "epoch": 7.73121932804832, "grad_norm": 0.5820940732955933, "learning_rate": 5.21529785659681e-05, "loss": 0.3921, "step": 20480 }, { "epoch": 7.734994337485844, "grad_norm": 0.6141514778137207, "learning_rate": 5.2111679429021565e-05, "loss": 0.3986, "step": 20490 }, { "epoch": 7.738769346923367, "grad_norm": 0.6320546269416809, "learning_rate": 5.207037884876205e-05, "loss": 0.3882, "step": 20500 }, { "epoch": 7.742544356360891, "grad_norm": 0.6534587740898132, "learning_rate": 5.202907685341809e-05, "loss": 0.3885, "step": 20510 }, { "epoch": 7.746319365798414, "grad_norm": 0.8131640553474426, "learning_rate": 5.198777347121926e-05, "loss": 0.3791, "step": 20520 }, { "epoch": 7.750094375235938, "grad_norm": 0.5980712175369263, "learning_rate": 5.194646873039598e-05, "loss": 0.3823, "step": 20530 }, { "epoch": 7.753869384673462, "grad_norm": 0.6374219059944153, "learning_rate": 5.1905162659179696e-05, "loss": 0.3967, "step": 20540 }, { "epoch": 7.757644394110986, "grad_norm": 0.6131422519683838, "learning_rate": 5.18638552858027e-05, "loss": 0.3865, "step": 20550 }, { "epoch": 7.7614194035485085, "grad_norm": 1.034074306488037, "learning_rate": 5.182254663849818e-05, "loss": 0.3907, "step": 20560 }, { "epoch": 7.765194412986032, "grad_norm": 0.6826948523521423, "learning_rate": 5.178123674550023e-05, "loss": 0.3857, "step": 20570 }, { "epoch": 7.768969422423556, "grad_norm": 0.5983694195747375, "learning_rate": 5.173992563504375e-05, "loss": 0.3914, "step": 20580 }, { "epoch": 7.77274443186108, "grad_norm": 0.5786053538322449, "learning_rate": 5.169861333536451e-05, "loss": 0.3863, "step": 20590 }, { "epoch": 7.776519441298603, "grad_norm": 0.607843816280365, "learning_rate": 5.165729987469907e-05, "loss": 0.3928, "step": 20600 }, { "epoch": 7.780294450736127, "grad_norm": 0.597486138343811, "learning_rate": 5.161598528128478e-05, "loss": 0.3845, "step": 20610 }, { "epoch": 7.784069460173651, "grad_norm": 0.6441140174865723, "learning_rate": 5.157466958335981e-05, "loss": 0.3856, "step": 20620 }, { "epoch": 7.787844469611174, "grad_norm": 0.8996423482894897, "learning_rate": 5.1533352809163025e-05, "loss": 0.3818, "step": 20630 }, { "epoch": 7.7916194790486974, "grad_norm": 0.6202008724212646, "learning_rate": 5.1492034986934046e-05, "loss": 0.3865, "step": 20640 }, { "epoch": 7.795394488486221, "grad_norm": 0.6499298810958862, "learning_rate": 5.1450716144913225e-05, "loss": 0.3887, "step": 20650 }, { "epoch": 7.799169497923745, "grad_norm": 0.6905611753463745, "learning_rate": 5.1409396311341595e-05, "loss": 0.3841, "step": 20660 }, { "epoch": 7.802944507361269, "grad_norm": 0.6405202150344849, "learning_rate": 5.136807551446089e-05, "loss": 0.3883, "step": 20670 }, { "epoch": 7.8067195167987915, "grad_norm": 0.6478602290153503, "learning_rate": 5.132675378251346e-05, "loss": 0.3856, "step": 20680 }, { "epoch": 7.810494526236315, "grad_norm": 0.6184169054031372, "learning_rate": 5.1285431143742325e-05, "loss": 0.3901, "step": 20690 }, { "epoch": 7.814269535673839, "grad_norm": 0.6250163316726685, "learning_rate": 5.1244107626391136e-05, "loss": 0.4039, "step": 20700 }, { "epoch": 7.818044545111363, "grad_norm": 0.6539636254310608, "learning_rate": 5.12027832587041e-05, "loss": 0.3866, "step": 20710 }, { "epoch": 7.821819554548886, "grad_norm": 0.6491449475288391, "learning_rate": 5.116145806892607e-05, "loss": 0.3965, "step": 20720 }, { "epoch": 7.82559456398641, "grad_norm": 0.653968870639801, "learning_rate": 5.1120132085302384e-05, "loss": 0.3825, "step": 20730 }, { "epoch": 7.829369573423934, "grad_norm": 0.7389346957206726, "learning_rate": 5.107880533607898e-05, "loss": 0.3935, "step": 20740 }, { "epoch": 7.833144582861458, "grad_norm": 0.6781933903694153, "learning_rate": 5.103747784950231e-05, "loss": 0.3852, "step": 20750 }, { "epoch": 7.8369195922989805, "grad_norm": 0.6142136454582214, "learning_rate": 5.09961496538193e-05, "loss": 0.3956, "step": 20760 }, { "epoch": 7.840694601736504, "grad_norm": 0.5901447534561157, "learning_rate": 5.095482077727742e-05, "loss": 0.3849, "step": 20770 }, { "epoch": 7.844469611174028, "grad_norm": 0.6106001138687134, "learning_rate": 5.091349124812452e-05, "loss": 0.3852, "step": 20780 }, { "epoch": 7.848244620611552, "grad_norm": 0.7389256954193115, "learning_rate": 5.087216109460897e-05, "loss": 0.3912, "step": 20790 }, { "epoch": 7.852019630049075, "grad_norm": 0.7008233070373535, "learning_rate": 5.083083034497954e-05, "loss": 0.3856, "step": 20800 }, { "epoch": 7.855794639486598, "grad_norm": 0.5946593880653381, "learning_rate": 5.07894990274854e-05, "loss": 0.3878, "step": 20810 }, { "epoch": 7.859569648924122, "grad_norm": 0.5986919403076172, "learning_rate": 5.074816717037614e-05, "loss": 0.3842, "step": 20820 }, { "epoch": 7.863344658361646, "grad_norm": 0.7096017003059387, "learning_rate": 5.070683480190165e-05, "loss": 0.3923, "step": 20830 }, { "epoch": 7.8671196677991695, "grad_norm": 0.6921922564506531, "learning_rate": 5.066550195031223e-05, "loss": 0.3864, "step": 20840 }, { "epoch": 7.870894677236693, "grad_norm": 0.5950612425804138, "learning_rate": 5.062416864385852e-05, "loss": 0.384, "step": 20850 }, { "epoch": 7.874669686674217, "grad_norm": 0.6273109912872314, "learning_rate": 5.058283491079142e-05, "loss": 0.3778, "step": 20860 }, { "epoch": 7.878444696111741, "grad_norm": 0.7356699109077454, "learning_rate": 5.054150077936216e-05, "loss": 0.3905, "step": 20870 }, { "epoch": 7.8822197055492635, "grad_norm": 0.6300787329673767, "learning_rate": 5.0500166277822214e-05, "loss": 0.387, "step": 20880 }, { "epoch": 7.885994714986787, "grad_norm": 0.6487736105918884, "learning_rate": 5.0458831434423334e-05, "loss": 0.3935, "step": 20890 }, { "epoch": 7.889769724424311, "grad_norm": 0.6694430708885193, "learning_rate": 5.0417496277417506e-05, "loss": 0.3982, "step": 20900 }, { "epoch": 7.893544733861835, "grad_norm": 0.6112816333770752, "learning_rate": 5.037616083505691e-05, "loss": 0.3769, "step": 20910 }, { "epoch": 7.8973197432993585, "grad_norm": 0.5769428014755249, "learning_rate": 5.0334825135593935e-05, "loss": 0.387, "step": 20920 }, { "epoch": 7.901094752736882, "grad_norm": 0.623289167881012, "learning_rate": 5.029348920728111e-05, "loss": 0.3874, "step": 20930 }, { "epoch": 7.904869762174405, "grad_norm": 0.6469810605049133, "learning_rate": 5.0252153078371186e-05, "loss": 0.3863, "step": 20940 }, { "epoch": 7.908644771611929, "grad_norm": 0.6396733522415161, "learning_rate": 5.021081677711704e-05, "loss": 0.3933, "step": 20950 }, { "epoch": 7.9124197810494525, "grad_norm": 0.6472081542015076, "learning_rate": 5.016948033177159e-05, "loss": 0.3804, "step": 20960 }, { "epoch": 7.916194790486976, "grad_norm": 0.670315146446228, "learning_rate": 5.012814377058793e-05, "loss": 0.3922, "step": 20970 }, { "epoch": 7.9199697999245, "grad_norm": 0.6452091932296753, "learning_rate": 5.008680712181921e-05, "loss": 0.3906, "step": 20980 }, { "epoch": 7.923744809362024, "grad_norm": 0.6105216145515442, "learning_rate": 5.0045470413718645e-05, "loss": 0.3864, "step": 20990 }, { "epoch": 7.9275198187995475, "grad_norm": 0.6681371331214905, "learning_rate": 5.00041336745395e-05, "loss": 0.3806, "step": 21000 }, { "epoch": 7.93129482823707, "grad_norm": 0.6200874447822571, "learning_rate": 4.996279693253499e-05, "loss": 0.3895, "step": 21010 }, { "epoch": 7.935069837674594, "grad_norm": 0.620955765247345, "learning_rate": 4.992146021595847e-05, "loss": 0.3872, "step": 21020 }, { "epoch": 7.938844847112118, "grad_norm": 0.6408780217170715, "learning_rate": 4.988012355306313e-05, "loss": 0.3821, "step": 21030 }, { "epoch": 7.9426198565496415, "grad_norm": 0.6567278504371643, "learning_rate": 4.98387869721022e-05, "loss": 0.3987, "step": 21040 }, { "epoch": 7.946394865987165, "grad_norm": 0.6321324706077576, "learning_rate": 4.9797450501328866e-05, "loss": 0.382, "step": 21050 }, { "epoch": 7.950169875424688, "grad_norm": 0.6207996606826782, "learning_rate": 4.97561141689962e-05, "loss": 0.3921, "step": 21060 }, { "epoch": 7.953944884862212, "grad_norm": 0.6274701952934265, "learning_rate": 4.971477800335721e-05, "loss": 0.3803, "step": 21070 }, { "epoch": 7.957719894299736, "grad_norm": 0.6380013227462769, "learning_rate": 4.967344203266475e-05, "loss": 0.3857, "step": 21080 }, { "epoch": 7.961494903737259, "grad_norm": 0.6619207859039307, "learning_rate": 4.9632106285171584e-05, "loss": 0.3828, "step": 21090 }, { "epoch": 7.965269913174783, "grad_norm": 0.6286654472351074, "learning_rate": 4.959077078913031e-05, "loss": 0.3857, "step": 21100 }, { "epoch": 7.969044922612307, "grad_norm": 0.7368075847625732, "learning_rate": 4.954943557279333e-05, "loss": 0.3843, "step": 21110 }, { "epoch": 7.9728199320498305, "grad_norm": 0.6597092747688293, "learning_rate": 4.9508100664412916e-05, "loss": 0.3883, "step": 21120 }, { "epoch": 7.976594941487353, "grad_norm": 0.673981249332428, "learning_rate": 4.946676609224105e-05, "loss": 0.3855, "step": 21130 }, { "epoch": 7.980369950924877, "grad_norm": 0.6356333494186401, "learning_rate": 4.942543188452952e-05, "loss": 0.3922, "step": 21140 }, { "epoch": 7.984144960362401, "grad_norm": 0.6984882354736328, "learning_rate": 4.938409806952988e-05, "loss": 0.3834, "step": 21150 }, { "epoch": 7.987919969799925, "grad_norm": 0.6634735465049744, "learning_rate": 4.93427646754934e-05, "loss": 0.3838, "step": 21160 }, { "epoch": 7.991694979237448, "grad_norm": 0.6481113433837891, "learning_rate": 4.930143173067108e-05, "loss": 0.3863, "step": 21170 }, { "epoch": 7.995469988674972, "grad_norm": 0.684773325920105, "learning_rate": 4.9260099263313565e-05, "loss": 0.379, "step": 21180 }, { "epoch": 7.999244998112495, "grad_norm": 0.6397860646247864, "learning_rate": 4.921876730167123e-05, "loss": 0.3962, "step": 21190 }, { "epoch": 8.003020007550019, "grad_norm": 0.6320487260818481, "learning_rate": 4.917743587399409e-05, "loss": 0.3839, "step": 21200 }, { "epoch": 8.006795016987542, "grad_norm": 0.6578797101974487, "learning_rate": 4.913610500853178e-05, "loss": 0.3841, "step": 21210 }, { "epoch": 8.010570026425066, "grad_norm": 0.6178431510925293, "learning_rate": 4.909477473353354e-05, "loss": 0.3848, "step": 21220 }, { "epoch": 8.01434503586259, "grad_norm": 0.6298217177391052, "learning_rate": 4.9053445077248236e-05, "loss": 0.3876, "step": 21230 }, { "epoch": 8.018120045300114, "grad_norm": 0.6131578087806702, "learning_rate": 4.901211606792429e-05, "loss": 0.3914, "step": 21240 }, { "epoch": 8.021895054737637, "grad_norm": 0.6626631021499634, "learning_rate": 4.89707877338097e-05, "loss": 0.3864, "step": 21250 }, { "epoch": 8.025670064175161, "grad_norm": 0.6320914626121521, "learning_rate": 4.892946010315199e-05, "loss": 0.3853, "step": 21260 }, { "epoch": 8.029445073612685, "grad_norm": 0.657503068447113, "learning_rate": 4.8888133204198204e-05, "loss": 0.3864, "step": 21270 }, { "epoch": 8.033220083050207, "grad_norm": 0.6045080423355103, "learning_rate": 4.8846807065194886e-05, "loss": 0.3859, "step": 21280 }, { "epoch": 8.03699509248773, "grad_norm": 0.6179243326187134, "learning_rate": 4.880548171438806e-05, "loss": 0.39, "step": 21290 }, { "epoch": 8.040770101925254, "grad_norm": 0.7276541590690613, "learning_rate": 4.8764157180023245e-05, "loss": 0.3869, "step": 21300 }, { "epoch": 8.044545111362778, "grad_norm": 0.5852580666542053, "learning_rate": 4.872283349034533e-05, "loss": 0.3847, "step": 21310 }, { "epoch": 8.048320120800302, "grad_norm": 0.8796269297599792, "learning_rate": 4.8681510673598674e-05, "loss": 0.3854, "step": 21320 }, { "epoch": 8.052095130237825, "grad_norm": 0.5888720750808716, "learning_rate": 4.8640188758027046e-05, "loss": 0.3767, "step": 21330 }, { "epoch": 8.05587013967535, "grad_norm": 0.6211332678794861, "learning_rate": 4.859886777187357e-05, "loss": 0.392, "step": 21340 }, { "epoch": 8.059645149112873, "grad_norm": 0.6416304111480713, "learning_rate": 4.855754774338077e-05, "loss": 0.3906, "step": 21350 }, { "epoch": 8.063420158550397, "grad_norm": 0.6397213339805603, "learning_rate": 4.851622870079048e-05, "loss": 0.3865, "step": 21360 }, { "epoch": 8.06719516798792, "grad_norm": 0.655972957611084, "learning_rate": 4.847491067234389e-05, "loss": 0.3845, "step": 21370 }, { "epoch": 8.070970177425444, "grad_norm": 0.7151166796684265, "learning_rate": 4.843359368628146e-05, "loss": 0.379, "step": 21380 }, { "epoch": 8.074745186862968, "grad_norm": 0.7403479814529419, "learning_rate": 4.8392277770842975e-05, "loss": 0.3857, "step": 21390 }, { "epoch": 8.078520196300492, "grad_norm": 0.8259317278862, "learning_rate": 4.83509629542675e-05, "loss": 0.3833, "step": 21400 }, { "epoch": 8.082295205738014, "grad_norm": 0.5347198843955994, "learning_rate": 4.830964926479329e-05, "loss": 0.3884, "step": 21410 }, { "epoch": 8.086070215175537, "grad_norm": 0.6426698565483093, "learning_rate": 4.826833673065785e-05, "loss": 0.3822, "step": 21420 }, { "epoch": 8.089845224613061, "grad_norm": 0.6192202568054199, "learning_rate": 4.822702538009794e-05, "loss": 0.386, "step": 21430 }, { "epoch": 8.093620234050585, "grad_norm": 0.622089684009552, "learning_rate": 4.818571524134945e-05, "loss": 0.3854, "step": 21440 }, { "epoch": 8.097395243488108, "grad_norm": 0.6114779114723206, "learning_rate": 4.8144406342647496e-05, "loss": 0.391, "step": 21450 }, { "epoch": 8.101170252925632, "grad_norm": 0.6789860129356384, "learning_rate": 4.81030987122263e-05, "loss": 0.3818, "step": 21460 }, { "epoch": 8.104945262363156, "grad_norm": 0.6163697242736816, "learning_rate": 4.806179237831926e-05, "loss": 0.3791, "step": 21470 }, { "epoch": 8.10872027180068, "grad_norm": 0.7106571793556213, "learning_rate": 4.802048736915884e-05, "loss": 0.3811, "step": 21480 }, { "epoch": 8.112495281238203, "grad_norm": 0.6782732009887695, "learning_rate": 4.797918371297666e-05, "loss": 0.3845, "step": 21490 }, { "epoch": 8.116270290675727, "grad_norm": 0.662436842918396, "learning_rate": 4.793788143800334e-05, "loss": 0.39, "step": 21500 }, { "epoch": 8.12004530011325, "grad_norm": 0.6009316444396973, "learning_rate": 4.789658057246862e-05, "loss": 0.3929, "step": 21510 }, { "epoch": 8.123820309550775, "grad_norm": 0.628048300743103, "learning_rate": 4.7855281144601227e-05, "loss": 0.386, "step": 21520 }, { "epoch": 8.127595318988298, "grad_norm": 0.6932970881462097, "learning_rate": 4.781398318262897e-05, "loss": 0.3889, "step": 21530 }, { "epoch": 8.13137032842582, "grad_norm": 0.6603420376777649, "learning_rate": 4.777268671477858e-05, "loss": 0.3847, "step": 21540 }, { "epoch": 8.135145337863344, "grad_norm": 0.7748346328735352, "learning_rate": 4.773139176927582e-05, "loss": 0.3916, "step": 21550 }, { "epoch": 8.138920347300868, "grad_norm": 0.7710997462272644, "learning_rate": 4.769009837434539e-05, "loss": 0.3834, "step": 21560 }, { "epoch": 8.142695356738392, "grad_norm": 0.6665760278701782, "learning_rate": 4.764880655821095e-05, "loss": 0.3885, "step": 21570 }, { "epoch": 8.146470366175915, "grad_norm": 0.6791707873344421, "learning_rate": 4.760751634909508e-05, "loss": 0.3798, "step": 21580 }, { "epoch": 8.150245375613439, "grad_norm": 0.6167095899581909, "learning_rate": 4.756622777521919e-05, "loss": 0.3799, "step": 21590 }, { "epoch": 8.154020385050963, "grad_norm": 0.7474955320358276, "learning_rate": 4.752494086480368e-05, "loss": 0.3873, "step": 21600 }, { "epoch": 8.157795394488486, "grad_norm": 0.6586377024650574, "learning_rate": 4.7483655646067744e-05, "loss": 0.386, "step": 21610 }, { "epoch": 8.16157040392601, "grad_norm": 0.6137120127677917, "learning_rate": 4.744237214722944e-05, "loss": 0.3842, "step": 21620 }, { "epoch": 8.165345413363534, "grad_norm": 0.613304615020752, "learning_rate": 4.740109039650567e-05, "loss": 0.3886, "step": 21630 }, { "epoch": 8.169120422801058, "grad_norm": 0.6361426115036011, "learning_rate": 4.73598104221121e-05, "loss": 0.3856, "step": 21640 }, { "epoch": 8.172895432238581, "grad_norm": 1.1138379573822021, "learning_rate": 4.731853225226322e-05, "loss": 0.3922, "step": 21650 }, { "epoch": 8.176670441676103, "grad_norm": 0.666292667388916, "learning_rate": 4.727725591517225e-05, "loss": 0.3847, "step": 21660 }, { "epoch": 8.180445451113627, "grad_norm": 0.6527111530303955, "learning_rate": 4.723598143905119e-05, "loss": 0.3905, "step": 21670 }, { "epoch": 8.18422046055115, "grad_norm": 0.634685754776001, "learning_rate": 4.719470885211077e-05, "loss": 0.3734, "step": 21680 }, { "epoch": 8.187995469988675, "grad_norm": 0.7132300734519958, "learning_rate": 4.7153438182560387e-05, "loss": 0.3891, "step": 21690 }, { "epoch": 8.191770479426198, "grad_norm": 0.6142542958259583, "learning_rate": 4.711216945860815e-05, "loss": 0.3857, "step": 21700 }, { "epoch": 8.195545488863722, "grad_norm": 0.6071975827217102, "learning_rate": 4.707090270846088e-05, "loss": 0.3857, "step": 21710 }, { "epoch": 8.199320498301246, "grad_norm": 0.7059184312820435, "learning_rate": 4.702963796032397e-05, "loss": 0.3856, "step": 21720 }, { "epoch": 8.20309550773877, "grad_norm": 0.6421445608139038, "learning_rate": 4.6988375242401514e-05, "loss": 0.3778, "step": 21730 }, { "epoch": 8.206870517176293, "grad_norm": 0.7011498808860779, "learning_rate": 4.694711458289618e-05, "loss": 0.3898, "step": 21740 }, { "epoch": 8.210645526613817, "grad_norm": 0.7598146796226501, "learning_rate": 4.690585601000925e-05, "loss": 0.3778, "step": 21750 }, { "epoch": 8.21442053605134, "grad_norm": 0.6215041875839233, "learning_rate": 4.686459955194055e-05, "loss": 0.3854, "step": 21760 }, { "epoch": 8.218195545488864, "grad_norm": 0.7370675206184387, "learning_rate": 4.6823345236888504e-05, "loss": 0.3823, "step": 21770 }, { "epoch": 8.221970554926388, "grad_norm": 0.6150240898132324, "learning_rate": 4.678209309305002e-05, "loss": 0.3955, "step": 21780 }, { "epoch": 8.22574556436391, "grad_norm": 0.5834974050521851, "learning_rate": 4.674084314862057e-05, "loss": 0.3852, "step": 21790 }, { "epoch": 8.229520573801434, "grad_norm": 0.6348870992660522, "learning_rate": 4.669959543179409e-05, "loss": 0.3903, "step": 21800 }, { "epoch": 8.233295583238958, "grad_norm": 0.7599389553070068, "learning_rate": 4.665834997076303e-05, "loss": 0.3849, "step": 21810 }, { "epoch": 8.237070592676481, "grad_norm": 0.7290443181991577, "learning_rate": 4.661710679371823e-05, "loss": 0.3846, "step": 21820 }, { "epoch": 8.240845602114005, "grad_norm": 0.6612056493759155, "learning_rate": 4.657586592884905e-05, "loss": 0.3849, "step": 21830 }, { "epoch": 8.244620611551529, "grad_norm": 0.5885748863220215, "learning_rate": 4.653462740434322e-05, "loss": 0.3801, "step": 21840 }, { "epoch": 8.248395620989053, "grad_norm": 0.7608574032783508, "learning_rate": 4.649339124838689e-05, "loss": 0.3873, "step": 21850 }, { "epoch": 8.252170630426576, "grad_norm": 0.5708411335945129, "learning_rate": 4.6452157489164574e-05, "loss": 0.3858, "step": 21860 }, { "epoch": 8.2559456398641, "grad_norm": 0.6522685289382935, "learning_rate": 4.6410926154859155e-05, "loss": 0.3869, "step": 21870 }, { "epoch": 8.259720649301624, "grad_norm": 0.688291609287262, "learning_rate": 4.636969727365186e-05, "loss": 0.3839, "step": 21880 }, { "epoch": 8.263495658739147, "grad_norm": 0.7088874578475952, "learning_rate": 4.632847087372226e-05, "loss": 0.3872, "step": 21890 }, { "epoch": 8.267270668176671, "grad_norm": 0.6327393054962158, "learning_rate": 4.628724698324818e-05, "loss": 0.3819, "step": 21900 }, { "epoch": 8.271045677614193, "grad_norm": 0.6789700388908386, "learning_rate": 4.6246025630405795e-05, "loss": 0.384, "step": 21910 }, { "epoch": 8.274820687051717, "grad_norm": 0.6304154396057129, "learning_rate": 4.6204806843369474e-05, "loss": 0.3883, "step": 21920 }, { "epoch": 8.27859569648924, "grad_norm": 0.6723567843437195, "learning_rate": 4.616359065031191e-05, "loss": 0.386, "step": 21930 }, { "epoch": 8.282370705926764, "grad_norm": 0.801613450050354, "learning_rate": 4.6122377079403946e-05, "loss": 0.3893, "step": 21940 }, { "epoch": 8.286145715364288, "grad_norm": 0.6835589408874512, "learning_rate": 4.6081166158814695e-05, "loss": 0.3801, "step": 21950 }, { "epoch": 8.289920724801812, "grad_norm": 0.6579275131225586, "learning_rate": 4.603995791671144e-05, "loss": 0.3877, "step": 21960 }, { "epoch": 8.293695734239336, "grad_norm": 0.6721914410591125, "learning_rate": 4.599875238125957e-05, "loss": 0.382, "step": 21970 }, { "epoch": 8.29747074367686, "grad_norm": 0.6746684312820435, "learning_rate": 4.595754958062273e-05, "loss": 0.3833, "step": 21980 }, { "epoch": 8.301245753114383, "grad_norm": 0.6919302940368652, "learning_rate": 4.591634954296265e-05, "loss": 0.3862, "step": 21990 }, { "epoch": 8.305020762551907, "grad_norm": 0.6918922662734985, "learning_rate": 4.587515229643913e-05, "loss": 0.3834, "step": 22000 }, { "epoch": 8.30879577198943, "grad_norm": 0.7155516743659973, "learning_rate": 4.583395786921013e-05, "loss": 0.3745, "step": 22010 }, { "epoch": 8.312570781426954, "grad_norm": 0.607598602771759, "learning_rate": 4.579276628943164e-05, "loss": 0.3794, "step": 22020 }, { "epoch": 8.316345790864478, "grad_norm": 0.673291027545929, "learning_rate": 4.575157758525772e-05, "loss": 0.3824, "step": 22030 }, { "epoch": 8.320120800302, "grad_norm": 0.7992842197418213, "learning_rate": 4.571039178484046e-05, "loss": 0.3756, "step": 22040 }, { "epoch": 8.323895809739524, "grad_norm": 0.7000048756599426, "learning_rate": 4.566920891632998e-05, "loss": 0.38, "step": 22050 }, { "epoch": 8.327670819177047, "grad_norm": 0.6638155579566956, "learning_rate": 4.562802900787436e-05, "loss": 0.3841, "step": 22060 }, { "epoch": 8.331445828614571, "grad_norm": 0.6273589730262756, "learning_rate": 4.558685208761968e-05, "loss": 0.3826, "step": 22070 }, { "epoch": 8.335220838052095, "grad_norm": 0.6618062853813171, "learning_rate": 4.554567818370998e-05, "loss": 0.3874, "step": 22080 }, { "epoch": 8.338995847489619, "grad_norm": 0.6083613038063049, "learning_rate": 4.550450732428726e-05, "loss": 0.3755, "step": 22090 }, { "epoch": 8.342770856927142, "grad_norm": 0.6382494568824768, "learning_rate": 4.546333953749137e-05, "loss": 0.3794, "step": 22100 }, { "epoch": 8.346545866364666, "grad_norm": 0.6135038733482361, "learning_rate": 4.5422174851460154e-05, "loss": 0.3803, "step": 22110 }, { "epoch": 8.35032087580219, "grad_norm": 0.6180480718612671, "learning_rate": 4.538101329432924e-05, "loss": 0.3844, "step": 22120 }, { "epoch": 8.354095885239714, "grad_norm": 0.6306928396224976, "learning_rate": 4.5339854894232195e-05, "loss": 0.3843, "step": 22130 }, { "epoch": 8.357870894677237, "grad_norm": 0.594245433807373, "learning_rate": 4.52986996793004e-05, "loss": 0.3936, "step": 22140 }, { "epoch": 8.361645904114761, "grad_norm": 0.6511313319206238, "learning_rate": 4.5257547677663024e-05, "loss": 0.3795, "step": 22150 }, { "epoch": 8.365420913552285, "grad_norm": 0.6905748248100281, "learning_rate": 4.52163989174471e-05, "loss": 0.378, "step": 22160 }, { "epoch": 8.369195922989807, "grad_norm": 0.6889414191246033, "learning_rate": 4.51752534267774e-05, "loss": 0.3779, "step": 22170 }, { "epoch": 8.37297093242733, "grad_norm": 0.7038361430168152, "learning_rate": 4.513411123377649e-05, "loss": 0.3789, "step": 22180 }, { "epoch": 8.376745941864854, "grad_norm": 0.8612080216407776, "learning_rate": 4.5092972366564675e-05, "loss": 0.3814, "step": 22190 }, { "epoch": 8.380520951302378, "grad_norm": 0.6390385627746582, "learning_rate": 4.505183685325997e-05, "loss": 0.3787, "step": 22200 }, { "epoch": 8.384295960739902, "grad_norm": 0.7001070976257324, "learning_rate": 4.5010704721978125e-05, "loss": 0.3821, "step": 22210 }, { "epoch": 8.388070970177425, "grad_norm": 0.6782472729682922, "learning_rate": 4.496957600083255e-05, "loss": 0.3777, "step": 22220 }, { "epoch": 8.39184597961495, "grad_norm": 0.6302990317344666, "learning_rate": 4.4928450717934343e-05, "loss": 0.379, "step": 22230 }, { "epoch": 8.395620989052473, "grad_norm": 0.6895720958709717, "learning_rate": 4.488732890139227e-05, "loss": 0.3809, "step": 22240 }, { "epoch": 8.399395998489997, "grad_norm": 0.6838264465332031, "learning_rate": 4.4846210579312665e-05, "loss": 0.3814, "step": 22250 }, { "epoch": 8.40317100792752, "grad_norm": 0.6539376974105835, "learning_rate": 4.480509577979953e-05, "loss": 0.3828, "step": 22260 }, { "epoch": 8.406946017365044, "grad_norm": 0.6490859985351562, "learning_rate": 4.476398453095445e-05, "loss": 0.387, "step": 22270 }, { "epoch": 8.410721026802568, "grad_norm": 0.6515709161758423, "learning_rate": 4.472287686087656e-05, "loss": 0.389, "step": 22280 }, { "epoch": 8.41449603624009, "grad_norm": 0.6394408941268921, "learning_rate": 4.468177279766259e-05, "loss": 0.381, "step": 22290 }, { "epoch": 8.418271045677614, "grad_norm": 0.6975342631340027, "learning_rate": 4.4640672369406746e-05, "loss": 0.373, "step": 22300 }, { "epoch": 8.422046055115137, "grad_norm": 0.6658486723899841, "learning_rate": 4.459957560420082e-05, "loss": 0.3773, "step": 22310 }, { "epoch": 8.425821064552661, "grad_norm": 0.649869441986084, "learning_rate": 4.455848253013403e-05, "loss": 0.3816, "step": 22320 }, { "epoch": 8.429596073990185, "grad_norm": 0.7822151184082031, "learning_rate": 4.4517393175293146e-05, "loss": 0.3872, "step": 22330 }, { "epoch": 8.433371083427708, "grad_norm": 0.6239002346992493, "learning_rate": 4.447630756776232e-05, "loss": 0.3834, "step": 22340 }, { "epoch": 8.437146092865232, "grad_norm": 0.6562296748161316, "learning_rate": 4.443522573562318e-05, "loss": 0.3933, "step": 22350 }, { "epoch": 8.440921102302756, "grad_norm": 0.6250916123390198, "learning_rate": 4.4394147706954776e-05, "loss": 0.383, "step": 22360 }, { "epoch": 8.44469611174028, "grad_norm": 0.6783263683319092, "learning_rate": 4.435307350983355e-05, "loss": 0.382, "step": 22370 }, { "epoch": 8.448471121177803, "grad_norm": 0.5855938792228699, "learning_rate": 4.4312003172333326e-05, "loss": 0.3753, "step": 22380 }, { "epoch": 8.452246130615327, "grad_norm": 0.6959243416786194, "learning_rate": 4.427093672252531e-05, "loss": 0.3769, "step": 22390 }, { "epoch": 8.45602114005285, "grad_norm": 0.750425398349762, "learning_rate": 4.422987418847802e-05, "loss": 0.3819, "step": 22400 }, { "epoch": 8.459796149490373, "grad_norm": 0.6782823204994202, "learning_rate": 4.4188815598257325e-05, "loss": 0.3751, "step": 22410 }, { "epoch": 8.463571158927897, "grad_norm": 0.7516863346099854, "learning_rate": 4.414776097992638e-05, "loss": 0.3794, "step": 22420 }, { "epoch": 8.46734616836542, "grad_norm": 0.694094717502594, "learning_rate": 4.4106710361545595e-05, "loss": 0.3835, "step": 22430 }, { "epoch": 8.471121177802944, "grad_norm": 0.7622425556182861, "learning_rate": 4.406566377117272e-05, "loss": 0.3817, "step": 22440 }, { "epoch": 8.474896187240468, "grad_norm": 0.6150171160697937, "learning_rate": 4.40246212368627e-05, "loss": 0.3821, "step": 22450 }, { "epoch": 8.478671196677992, "grad_norm": 0.5998985171318054, "learning_rate": 4.3983582786667715e-05, "loss": 0.3769, "step": 22460 }, { "epoch": 8.482446206115515, "grad_norm": 0.6367303133010864, "learning_rate": 4.394254844863716e-05, "loss": 0.3853, "step": 22470 }, { "epoch": 8.486221215553039, "grad_norm": 0.7072235941886902, "learning_rate": 4.390151825081762e-05, "loss": 0.3844, "step": 22480 }, { "epoch": 8.489996224990563, "grad_norm": 0.678936779499054, "learning_rate": 4.386049222125286e-05, "loss": 0.3695, "step": 22490 }, { "epoch": 8.493771234428086, "grad_norm": 0.6695368885993958, "learning_rate": 4.3819470387983774e-05, "loss": 0.3838, "step": 22500 }, { "epoch": 8.49754624386561, "grad_norm": 0.6475124359130859, "learning_rate": 4.377845277904841e-05, "loss": 0.3847, "step": 22510 }, { "epoch": 8.501321253303134, "grad_norm": 0.6073307991027832, "learning_rate": 4.37374394224819e-05, "loss": 0.3805, "step": 22520 }, { "epoch": 8.505096262740658, "grad_norm": 0.6134095191955566, "learning_rate": 4.369643034631648e-05, "loss": 0.388, "step": 22530 }, { "epoch": 8.50887127217818, "grad_norm": 1.0386673212051392, "learning_rate": 4.365542557858149e-05, "loss": 0.3853, "step": 22540 }, { "epoch": 8.512646281615703, "grad_norm": 0.6191480755805969, "learning_rate": 4.361442514730329e-05, "loss": 0.3866, "step": 22550 }, { "epoch": 8.516421291053227, "grad_norm": 0.6350696682929993, "learning_rate": 4.357342908050528e-05, "loss": 0.3817, "step": 22560 }, { "epoch": 8.52019630049075, "grad_norm": 0.6684324145317078, "learning_rate": 4.3532437406207895e-05, "loss": 0.382, "step": 22570 }, { "epoch": 8.523971309928275, "grad_norm": 0.6637814044952393, "learning_rate": 4.349145015242856e-05, "loss": 0.3775, "step": 22580 }, { "epoch": 8.527746319365798, "grad_norm": 0.7314398884773254, "learning_rate": 4.345046734718168e-05, "loss": 0.3755, "step": 22590 }, { "epoch": 8.531521328803322, "grad_norm": 0.6954400539398193, "learning_rate": 4.34094890184786e-05, "loss": 0.3816, "step": 22600 }, { "epoch": 8.535296338240846, "grad_norm": 0.6581496596336365, "learning_rate": 4.336851519432765e-05, "loss": 0.3869, "step": 22610 }, { "epoch": 8.53907134767837, "grad_norm": 0.6038832068443298, "learning_rate": 4.332754590273403e-05, "loss": 0.3821, "step": 22620 }, { "epoch": 8.542846357115893, "grad_norm": 0.6651731133460999, "learning_rate": 4.3286581171699855e-05, "loss": 0.3916, "step": 22630 }, { "epoch": 8.546621366553417, "grad_norm": 0.6608867645263672, "learning_rate": 4.324562102922416e-05, "loss": 0.3764, "step": 22640 }, { "epoch": 8.55039637599094, "grad_norm": 0.6170617938041687, "learning_rate": 4.320466550330278e-05, "loss": 0.3805, "step": 22650 }, { "epoch": 8.554171385428464, "grad_norm": 0.6135375499725342, "learning_rate": 4.3163714621928466e-05, "loss": 0.3792, "step": 22660 }, { "epoch": 8.557946394865986, "grad_norm": 0.6618521809577942, "learning_rate": 4.312276841309074e-05, "loss": 0.3768, "step": 22670 }, { "epoch": 8.56172140430351, "grad_norm": 0.6322848200798035, "learning_rate": 4.3081826904775945e-05, "loss": 0.3733, "step": 22680 }, { "epoch": 8.565496413741034, "grad_norm": 0.6854828000068665, "learning_rate": 4.3040890124967246e-05, "loss": 0.3895, "step": 22690 }, { "epoch": 8.569271423178558, "grad_norm": 0.6372854113578796, "learning_rate": 4.2999958101644537e-05, "loss": 0.3766, "step": 22700 }, { "epoch": 8.573046432616081, "grad_norm": 0.6355828642845154, "learning_rate": 4.2959030862784435e-05, "loss": 0.3769, "step": 22710 }, { "epoch": 8.576821442053605, "grad_norm": 0.6458565592765808, "learning_rate": 4.291810843636036e-05, "loss": 0.3925, "step": 22720 }, { "epoch": 8.580596451491129, "grad_norm": 0.6839278936386108, "learning_rate": 4.2877190850342375e-05, "loss": 0.388, "step": 22730 }, { "epoch": 8.584371460928653, "grad_norm": 0.6284090876579285, "learning_rate": 4.2836278132697294e-05, "loss": 0.3827, "step": 22740 }, { "epoch": 8.588146470366176, "grad_norm": 0.6752241253852844, "learning_rate": 4.279537031138855e-05, "loss": 0.3969, "step": 22750 }, { "epoch": 8.5919214798037, "grad_norm": 0.6706361770629883, "learning_rate": 4.275446741437625e-05, "loss": 0.3845, "step": 22760 }, { "epoch": 8.595696489241224, "grad_norm": 0.6637896299362183, "learning_rate": 4.2713569469617176e-05, "loss": 0.3749, "step": 22770 }, { "epoch": 8.599471498678747, "grad_norm": 0.7166405916213989, "learning_rate": 4.267267650506465e-05, "loss": 0.3803, "step": 22780 }, { "epoch": 8.603246508116271, "grad_norm": 1.1399887800216675, "learning_rate": 4.263178854866866e-05, "loss": 0.3818, "step": 22790 }, { "epoch": 8.607021517553793, "grad_norm": 0.5608764886856079, "learning_rate": 4.259090562837571e-05, "loss": 0.3855, "step": 22800 }, { "epoch": 8.610796526991317, "grad_norm": 0.6282104253768921, "learning_rate": 4.255002777212888e-05, "loss": 0.3775, "step": 22810 }, { "epoch": 8.61457153642884, "grad_norm": 0.7265238165855408, "learning_rate": 4.250915500786783e-05, "loss": 0.38, "step": 22820 }, { "epoch": 8.618346545866364, "grad_norm": 0.6828112006187439, "learning_rate": 4.24682873635287e-05, "loss": 0.3821, "step": 22830 }, { "epoch": 8.622121555303888, "grad_norm": 0.6247879862785339, "learning_rate": 4.242742486704414e-05, "loss": 0.388, "step": 22840 }, { "epoch": 8.625896564741412, "grad_norm": 0.6014358997344971, "learning_rate": 4.238656754634327e-05, "loss": 0.3832, "step": 22850 }, { "epoch": 8.629671574178936, "grad_norm": 0.6423870921134949, "learning_rate": 4.234571542935168e-05, "loss": 0.3951, "step": 22860 }, { "epoch": 8.63344658361646, "grad_norm": 3.2498388290405273, "learning_rate": 4.230486854399144e-05, "loss": 0.3945, "step": 22870 }, { "epoch": 8.637221593053983, "grad_norm": 0.9906812906265259, "learning_rate": 4.226402691818098e-05, "loss": 0.3884, "step": 22880 }, { "epoch": 8.640996602491507, "grad_norm": 0.6815230250358582, "learning_rate": 4.2223190579835196e-05, "loss": 0.3842, "step": 22890 }, { "epoch": 8.64477161192903, "grad_norm": 0.6938421726226807, "learning_rate": 4.218235955686531e-05, "loss": 0.3797, "step": 22900 }, { "epoch": 8.648546621366553, "grad_norm": 1.4999902248382568, "learning_rate": 4.214153387717894e-05, "loss": 0.3921, "step": 22910 }, { "epoch": 8.652321630804076, "grad_norm": 1.1413220167160034, "learning_rate": 4.210071356868007e-05, "loss": 0.3825, "step": 22920 }, { "epoch": 8.6560966402416, "grad_norm": 0.633385419845581, "learning_rate": 4.205989865926898e-05, "loss": 0.3819, "step": 22930 }, { "epoch": 8.659871649679124, "grad_norm": 0.680814802646637, "learning_rate": 4.2019089176842294e-05, "loss": 0.3818, "step": 22940 }, { "epoch": 8.663646659116647, "grad_norm": 0.5858637094497681, "learning_rate": 4.1978285149292894e-05, "loss": 0.3804, "step": 22950 }, { "epoch": 8.667421668554171, "grad_norm": 0.6605408191680908, "learning_rate": 4.193748660450996e-05, "loss": 0.389, "step": 22960 }, { "epoch": 8.671196677991695, "grad_norm": 0.6238226890563965, "learning_rate": 4.189669357037891e-05, "loss": 0.378, "step": 22970 }, { "epoch": 8.674971687429219, "grad_norm": 0.7179723978042603, "learning_rate": 4.1855906074781405e-05, "loss": 0.3778, "step": 22980 }, { "epoch": 8.678746696866742, "grad_norm": 0.6245715618133545, "learning_rate": 4.1815124145595285e-05, "loss": 0.3823, "step": 22990 }, { "epoch": 8.682521706304266, "grad_norm": 0.6498119831085205, "learning_rate": 4.1774347810694644e-05, "loss": 0.3801, "step": 23000 }, { "epoch": 8.68629671574179, "grad_norm": 0.8200873136520386, "learning_rate": 4.17335770979497e-05, "loss": 0.3793, "step": 23010 }, { "epoch": 8.690071725179314, "grad_norm": 0.6485256552696228, "learning_rate": 4.169281203522687e-05, "loss": 0.3811, "step": 23020 }, { "epoch": 8.693846734616837, "grad_norm": 0.5570192933082581, "learning_rate": 4.1652052650388674e-05, "loss": 0.3783, "step": 23030 }, { "epoch": 8.69762174405436, "grad_norm": 0.7071910500526428, "learning_rate": 4.1611298971293786e-05, "loss": 0.3839, "step": 23040 }, { "epoch": 8.701396753491883, "grad_norm": 0.6622625589370728, "learning_rate": 4.1570551025796935e-05, "loss": 0.3754, "step": 23050 }, { "epoch": 8.705171762929407, "grad_norm": 0.680098831653595, "learning_rate": 4.152980884174897e-05, "loss": 0.3802, "step": 23060 }, { "epoch": 8.70894677236693, "grad_norm": 0.6106149554252625, "learning_rate": 4.148907244699682e-05, "loss": 0.3755, "step": 23070 }, { "epoch": 8.712721781804454, "grad_norm": 0.5729265809059143, "learning_rate": 4.1448341869383395e-05, "loss": 0.38, "step": 23080 }, { "epoch": 8.716496791241978, "grad_norm": 0.7992091774940491, "learning_rate": 4.140761713674765e-05, "loss": 0.3833, "step": 23090 }, { "epoch": 8.720271800679502, "grad_norm": 0.6829302906990051, "learning_rate": 4.1366898276924574e-05, "loss": 0.3739, "step": 23100 }, { "epoch": 8.724046810117025, "grad_norm": 0.6297142505645752, "learning_rate": 4.132618531774512e-05, "loss": 0.3754, "step": 23110 }, { "epoch": 8.72782181955455, "grad_norm": 0.6362572908401489, "learning_rate": 4.128547828703622e-05, "loss": 0.3764, "step": 23120 }, { "epoch": 8.731596828992073, "grad_norm": 0.7810057997703552, "learning_rate": 4.1244777212620725e-05, "loss": 0.3822, "step": 23130 }, { "epoch": 8.735371838429597, "grad_norm": 0.6120405197143555, "learning_rate": 4.120408212231746e-05, "loss": 0.3755, "step": 23140 }, { "epoch": 8.73914684786712, "grad_norm": 0.6071982979774475, "learning_rate": 4.116339304394111e-05, "loss": 0.3795, "step": 23150 }, { "epoch": 8.742921857304644, "grad_norm": 0.7162799835205078, "learning_rate": 4.112271000530229e-05, "loss": 0.3815, "step": 23160 }, { "epoch": 8.746696866742166, "grad_norm": 0.5881337523460388, "learning_rate": 4.10820330342075e-05, "loss": 0.3798, "step": 23170 }, { "epoch": 8.75047187617969, "grad_norm": 0.6906888484954834, "learning_rate": 4.1041362158459027e-05, "loss": 0.3732, "step": 23180 }, { "epoch": 8.754246885617214, "grad_norm": 0.6561579704284668, "learning_rate": 4.1000697405855024e-05, "loss": 0.3825, "step": 23190 }, { "epoch": 8.758021895054737, "grad_norm": 0.5920860171318054, "learning_rate": 4.096003880418951e-05, "loss": 0.3819, "step": 23200 }, { "epoch": 8.761796904492261, "grad_norm": 1.4881322383880615, "learning_rate": 4.0919386381252215e-05, "loss": 0.3763, "step": 23210 }, { "epoch": 8.765571913929785, "grad_norm": 0.7468271255493164, "learning_rate": 4.087874016482872e-05, "loss": 0.3806, "step": 23220 }, { "epoch": 8.769346923367308, "grad_norm": 0.5837022066116333, "learning_rate": 4.0838100182700295e-05, "loss": 0.382, "step": 23230 }, { "epoch": 8.773121932804832, "grad_norm": 0.590907633304596, "learning_rate": 4.079746646264402e-05, "loss": 0.3886, "step": 23240 }, { "epoch": 8.776896942242356, "grad_norm": 0.9434071779251099, "learning_rate": 4.075683903243262e-05, "loss": 0.3737, "step": 23250 }, { "epoch": 8.78067195167988, "grad_norm": 0.7620961666107178, "learning_rate": 4.071621791983462e-05, "loss": 0.3776, "step": 23260 }, { "epoch": 8.784446961117403, "grad_norm": 0.5980284214019775, "learning_rate": 4.06756031526141e-05, "loss": 0.3804, "step": 23270 }, { "epoch": 8.788221970554927, "grad_norm": 0.6051507592201233, "learning_rate": 4.063499475853092e-05, "loss": 0.3804, "step": 23280 }, { "epoch": 8.791996979992451, "grad_norm": 0.6261473894119263, "learning_rate": 4.0594392765340506e-05, "loss": 0.3658, "step": 23290 }, { "epoch": 8.795771989429973, "grad_norm": 0.6253399848937988, "learning_rate": 4.0553797200793954e-05, "loss": 0.3819, "step": 23300 }, { "epoch": 8.799546998867497, "grad_norm": 0.6369916200637817, "learning_rate": 4.0513208092637926e-05, "loss": 0.3783, "step": 23310 }, { "epoch": 8.80332200830502, "grad_norm": 0.6382907032966614, "learning_rate": 4.0472625468614735e-05, "loss": 0.3774, "step": 23320 }, { "epoch": 8.807097017742544, "grad_norm": 1.013975977897644, "learning_rate": 4.043204935646218e-05, "loss": 0.3976, "step": 23330 }, { "epoch": 8.810872027180068, "grad_norm": 0.7413721084594727, "learning_rate": 4.0391479783913675e-05, "loss": 0.3813, "step": 23340 }, { "epoch": 8.814647036617592, "grad_norm": 0.7551056146621704, "learning_rate": 4.0350916778698155e-05, "loss": 0.3754, "step": 23350 }, { "epoch": 8.818422046055115, "grad_norm": 0.6768625378608704, "learning_rate": 4.031036036854001e-05, "loss": 0.3672, "step": 23360 }, { "epoch": 8.822197055492639, "grad_norm": 0.7512798309326172, "learning_rate": 4.026981058115918e-05, "loss": 0.3806, "step": 23370 }, { "epoch": 8.825972064930163, "grad_norm": 0.6166073679924011, "learning_rate": 4.022926744427108e-05, "loss": 0.3818, "step": 23380 }, { "epoch": 8.829747074367686, "grad_norm": 0.6651231050491333, "learning_rate": 4.018873098558654e-05, "loss": 0.379, "step": 23390 }, { "epoch": 8.83352208380521, "grad_norm": 0.6436861157417297, "learning_rate": 4.014820123281186e-05, "loss": 0.3816, "step": 23400 }, { "epoch": 8.837297093242734, "grad_norm": 0.6689978837966919, "learning_rate": 4.0107678213648735e-05, "loss": 0.3756, "step": 23410 }, { "epoch": 8.841072102680256, "grad_norm": 0.6367429494857788, "learning_rate": 4.006716195579428e-05, "loss": 0.3759, "step": 23420 }, { "epoch": 8.84484711211778, "grad_norm": 0.649011492729187, "learning_rate": 4.002665248694096e-05, "loss": 0.3727, "step": 23430 }, { "epoch": 8.848622121555303, "grad_norm": 0.7108133435249329, "learning_rate": 3.998614983477664e-05, "loss": 0.3741, "step": 23440 }, { "epoch": 8.852397130992827, "grad_norm": 0.6738606691360474, "learning_rate": 3.994565402698448e-05, "loss": 0.3754, "step": 23450 }, { "epoch": 8.85617214043035, "grad_norm": 0.6250699758529663, "learning_rate": 3.9905165091242975e-05, "loss": 0.3832, "step": 23460 }, { "epoch": 8.859947149867875, "grad_norm": 0.6175537705421448, "learning_rate": 3.9864683055225936e-05, "loss": 0.3705, "step": 23470 }, { "epoch": 8.863722159305398, "grad_norm": 0.6735917925834656, "learning_rate": 3.982420794660247e-05, "loss": 0.3799, "step": 23480 }, { "epoch": 8.867497168742922, "grad_norm": 0.7157145738601685, "learning_rate": 3.978373979303691e-05, "loss": 0.3749, "step": 23490 }, { "epoch": 8.871272178180446, "grad_norm": 0.6760987639427185, "learning_rate": 3.974327862218888e-05, "loss": 0.3851, "step": 23500 }, { "epoch": 8.87504718761797, "grad_norm": 0.6828089356422424, "learning_rate": 3.970282446171318e-05, "loss": 0.3787, "step": 23510 }, { "epoch": 8.878822197055493, "grad_norm": 0.6380763053894043, "learning_rate": 3.966237733925988e-05, "loss": 0.3748, "step": 23520 }, { "epoch": 8.882597206493017, "grad_norm": 0.6100180745124817, "learning_rate": 3.962193728247418e-05, "loss": 0.3767, "step": 23530 }, { "epoch": 8.886372215930539, "grad_norm": 0.6232542991638184, "learning_rate": 3.958150431899651e-05, "loss": 0.381, "step": 23540 }, { "epoch": 8.890147225368063, "grad_norm": 0.636439323425293, "learning_rate": 3.954107847646238e-05, "loss": 0.3755, "step": 23550 }, { "epoch": 8.893922234805586, "grad_norm": 0.611513078212738, "learning_rate": 3.950065978250249e-05, "loss": 0.3761, "step": 23560 }, { "epoch": 8.89769724424311, "grad_norm": 0.6682894825935364, "learning_rate": 3.9460248264742624e-05, "loss": 0.3789, "step": 23570 }, { "epoch": 8.901472253680634, "grad_norm": 0.6221024990081787, "learning_rate": 3.941984395080371e-05, "loss": 0.3734, "step": 23580 }, { "epoch": 8.905247263118158, "grad_norm": 0.6244845390319824, "learning_rate": 3.937944686830167e-05, "loss": 0.3826, "step": 23590 }, { "epoch": 8.909022272555681, "grad_norm": 0.6587332487106323, "learning_rate": 3.933905704484756e-05, "loss": 0.3759, "step": 23600 }, { "epoch": 8.912797281993205, "grad_norm": 0.6589406728744507, "learning_rate": 3.929867450804743e-05, "loss": 0.3756, "step": 23610 }, { "epoch": 8.916572291430729, "grad_norm": 0.6549832224845886, "learning_rate": 3.925829928550237e-05, "loss": 0.3883, "step": 23620 }, { "epoch": 8.920347300868253, "grad_norm": 0.7647480964660645, "learning_rate": 3.921793140480847e-05, "loss": 0.3709, "step": 23630 }, { "epoch": 8.924122310305776, "grad_norm": 0.6407207250595093, "learning_rate": 3.917757089355677e-05, "loss": 0.3736, "step": 23640 }, { "epoch": 8.9278973197433, "grad_norm": 1.7529853582382202, "learning_rate": 3.9137217779333326e-05, "loss": 0.3909, "step": 23650 }, { "epoch": 8.931672329180824, "grad_norm": 0.7098379135131836, "learning_rate": 3.9096872089719083e-05, "loss": 0.3768, "step": 23660 }, { "epoch": 8.935447338618346, "grad_norm": 0.5911573767662048, "learning_rate": 3.905653385228996e-05, "loss": 0.3764, "step": 23670 }, { "epoch": 8.93922234805587, "grad_norm": 0.652247965335846, "learning_rate": 3.901620309461677e-05, "loss": 0.3774, "step": 23680 }, { "epoch": 8.942997357493393, "grad_norm": 0.6572662591934204, "learning_rate": 3.897587984426518e-05, "loss": 0.3836, "step": 23690 }, { "epoch": 8.946772366930917, "grad_norm": 0.6210899949073792, "learning_rate": 3.893556412879577e-05, "loss": 0.3724, "step": 23700 }, { "epoch": 8.95054737636844, "grad_norm": 0.6767255067825317, "learning_rate": 3.889525597576395e-05, "loss": 0.3752, "step": 23710 }, { "epoch": 8.954322385805964, "grad_norm": 0.6062052249908447, "learning_rate": 3.8854955412719965e-05, "loss": 0.3703, "step": 23720 }, { "epoch": 8.958097395243488, "grad_norm": 0.7351146936416626, "learning_rate": 3.881466246720887e-05, "loss": 0.3779, "step": 23730 }, { "epoch": 8.961872404681012, "grad_norm": 0.6535313725471497, "learning_rate": 3.8774377166770484e-05, "loss": 0.3723, "step": 23740 }, { "epoch": 8.965647414118536, "grad_norm": 0.6908565163612366, "learning_rate": 3.8734099538939474e-05, "loss": 0.374, "step": 23750 }, { "epoch": 8.96942242355606, "grad_norm": 0.6240267157554626, "learning_rate": 3.869382961124518e-05, "loss": 0.3791, "step": 23760 }, { "epoch": 8.973197432993583, "grad_norm": 0.6270217895507812, "learning_rate": 3.8653567411211736e-05, "loss": 0.3766, "step": 23770 }, { "epoch": 8.976972442431107, "grad_norm": 0.6301257014274597, "learning_rate": 3.8613312966357987e-05, "loss": 0.3816, "step": 23780 }, { "epoch": 8.98074745186863, "grad_norm": 0.6038647890090942, "learning_rate": 3.857306630419745e-05, "loss": 0.3798, "step": 23790 }, { "epoch": 8.984522461306153, "grad_norm": 0.6066597104072571, "learning_rate": 3.853282745223834e-05, "loss": 0.3679, "step": 23800 }, { "epoch": 8.988297470743676, "grad_norm": 0.7371004819869995, "learning_rate": 3.8492596437983546e-05, "loss": 0.3713, "step": 23810 }, { "epoch": 8.9920724801812, "grad_norm": 0.7040576934814453, "learning_rate": 3.8452373288930586e-05, "loss": 0.3827, "step": 23820 }, { "epoch": 8.995847489618724, "grad_norm": 0.6456139087677002, "learning_rate": 3.841215803257159e-05, "loss": 0.3755, "step": 23830 }, { "epoch": 8.999622499056247, "grad_norm": 0.7062108516693115, "learning_rate": 3.83719506963933e-05, "loss": 0.374, "step": 23840 }, { "epoch": 9.003397508493771, "grad_norm": 0.6744314432144165, "learning_rate": 3.8331751307877087e-05, "loss": 0.372, "step": 23850 }, { "epoch": 9.007172517931295, "grad_norm": 0.6453410983085632, "learning_rate": 3.82915598944988e-05, "loss": 0.3676, "step": 23860 }, { "epoch": 9.010947527368819, "grad_norm": 0.7047562599182129, "learning_rate": 3.825137648372893e-05, "loss": 0.3786, "step": 23870 }, { "epoch": 9.014722536806342, "grad_norm": 0.6693297624588013, "learning_rate": 3.8211201103032465e-05, "loss": 0.3704, "step": 23880 }, { "epoch": 9.018497546243866, "grad_norm": 0.6609267592430115, "learning_rate": 3.817103377986887e-05, "loss": 0.3851, "step": 23890 }, { "epoch": 9.02227255568139, "grad_norm": 0.6702944040298462, "learning_rate": 3.813087454169215e-05, "loss": 0.3761, "step": 23900 }, { "epoch": 9.026047565118914, "grad_norm": 0.61748206615448, "learning_rate": 3.809072341595078e-05, "loss": 0.3738, "step": 23910 }, { "epoch": 9.029822574556436, "grad_norm": 0.6437124609947205, "learning_rate": 3.8050580430087636e-05, "loss": 0.3763, "step": 23920 }, { "epoch": 9.03359758399396, "grad_norm": 0.7444436550140381, "learning_rate": 3.8010445611540096e-05, "loss": 0.3817, "step": 23930 }, { "epoch": 9.037372593431483, "grad_norm": 0.6082374453544617, "learning_rate": 3.797031898773992e-05, "loss": 0.3766, "step": 23940 }, { "epoch": 9.041147602869007, "grad_norm": 0.6564713716506958, "learning_rate": 3.793020058611329e-05, "loss": 0.3716, "step": 23950 }, { "epoch": 9.04492261230653, "grad_norm": 0.5503236651420593, "learning_rate": 3.789009043408074e-05, "loss": 0.3771, "step": 23960 }, { "epoch": 9.048697621744054, "grad_norm": 0.6116209030151367, "learning_rate": 3.7849988559057194e-05, "loss": 0.3796, "step": 23970 }, { "epoch": 9.052472631181578, "grad_norm": 0.6737613081932068, "learning_rate": 3.78098949884519e-05, "loss": 0.3765, "step": 23980 }, { "epoch": 9.056247640619102, "grad_norm": 0.6291811466217041, "learning_rate": 3.776980974966843e-05, "loss": 0.3685, "step": 23990 }, { "epoch": 9.060022650056625, "grad_norm": 0.6767562031745911, "learning_rate": 3.772973287010468e-05, "loss": 0.3759, "step": 24000 }, { "epoch": 9.06379765949415, "grad_norm": 0.6803905963897705, "learning_rate": 3.768966437715283e-05, "loss": 0.372, "step": 24010 }, { "epoch": 9.067572668931673, "grad_norm": 0.6342119574546814, "learning_rate": 3.7649604298199274e-05, "loss": 0.3688, "step": 24020 }, { "epoch": 9.071347678369197, "grad_norm": 0.6667430996894836, "learning_rate": 3.760955266062473e-05, "loss": 0.3792, "step": 24030 }, { "epoch": 9.07512268780672, "grad_norm": 0.6684991717338562, "learning_rate": 3.75695094918041e-05, "loss": 0.3715, "step": 24040 }, { "epoch": 9.078897697244242, "grad_norm": 0.6000959277153015, "learning_rate": 3.752947481910652e-05, "loss": 0.3751, "step": 24050 }, { "epoch": 9.082672706681766, "grad_norm": 0.6405602693557739, "learning_rate": 3.7489448669895324e-05, "loss": 0.3863, "step": 24060 }, { "epoch": 9.08644771611929, "grad_norm": 0.6103823781013489, "learning_rate": 3.744943107152798e-05, "loss": 0.3703, "step": 24070 }, { "epoch": 9.090222725556814, "grad_norm": 0.6388285756111145, "learning_rate": 3.7409422051356165e-05, "loss": 0.3784, "step": 24080 }, { "epoch": 9.093997734994337, "grad_norm": 0.7293974161148071, "learning_rate": 3.736942163672564e-05, "loss": 0.393, "step": 24090 }, { "epoch": 9.097772744431861, "grad_norm": 0.7579730749130249, "learning_rate": 3.732942985497636e-05, "loss": 0.3702, "step": 24100 }, { "epoch": 9.101547753869385, "grad_norm": 0.6169874668121338, "learning_rate": 3.728944673344228e-05, "loss": 0.3748, "step": 24110 }, { "epoch": 9.105322763306908, "grad_norm": 0.6163791418075562, "learning_rate": 3.72494722994515e-05, "loss": 0.3786, "step": 24120 }, { "epoch": 9.109097772744432, "grad_norm": 0.7307747602462769, "learning_rate": 3.720950658032617e-05, "loss": 0.3765, "step": 24130 }, { "epoch": 9.112872782181956, "grad_norm": 0.6685255765914917, "learning_rate": 3.716954960338249e-05, "loss": 0.3714, "step": 24140 }, { "epoch": 9.11664779161948, "grad_norm": 0.6877933144569397, "learning_rate": 3.712960139593066e-05, "loss": 0.3769, "step": 24150 }, { "epoch": 9.120422801057003, "grad_norm": 0.6554901599884033, "learning_rate": 3.708966198527493e-05, "loss": 0.3784, "step": 24160 }, { "epoch": 9.124197810494525, "grad_norm": 0.7071563005447388, "learning_rate": 3.704973139871349e-05, "loss": 0.3673, "step": 24170 }, { "epoch": 9.12797281993205, "grad_norm": 0.5478484630584717, "learning_rate": 3.700980966353853e-05, "loss": 0.3713, "step": 24180 }, { "epoch": 9.131747829369573, "grad_norm": 0.6506239175796509, "learning_rate": 3.696989680703619e-05, "loss": 0.3786, "step": 24190 }, { "epoch": 9.135522838807097, "grad_norm": 0.6592217087745667, "learning_rate": 3.69299928564865e-05, "loss": 0.3782, "step": 24200 }, { "epoch": 9.13929784824462, "grad_norm": 0.6553889513015747, "learning_rate": 3.689009783916345e-05, "loss": 0.3705, "step": 24210 }, { "epoch": 9.143072857682144, "grad_norm": 0.6616779565811157, "learning_rate": 3.6850211782334895e-05, "loss": 0.3685, "step": 24220 }, { "epoch": 9.146847867119668, "grad_norm": 0.6655992269515991, "learning_rate": 3.681033471326261e-05, "loss": 0.3741, "step": 24230 }, { "epoch": 9.150622876557192, "grad_norm": 0.6539639830589294, "learning_rate": 3.677046665920216e-05, "loss": 0.3741, "step": 24240 }, { "epoch": 9.154397885994715, "grad_norm": 0.6297611594200134, "learning_rate": 3.6730607647403005e-05, "loss": 0.3766, "step": 24250 }, { "epoch": 9.158172895432239, "grad_norm": 0.6319004893302917, "learning_rate": 3.6690757705108416e-05, "loss": 0.3773, "step": 24260 }, { "epoch": 9.161947904869763, "grad_norm": 1.1053649187088013, "learning_rate": 3.665091685955542e-05, "loss": 0.3735, "step": 24270 }, { "epoch": 9.165722914307286, "grad_norm": 0.6567557454109192, "learning_rate": 3.6611085137974896e-05, "loss": 0.3817, "step": 24280 }, { "epoch": 9.16949792374481, "grad_norm": 0.6980771422386169, "learning_rate": 3.657126256759143e-05, "loss": 0.3696, "step": 24290 }, { "epoch": 9.173272933182332, "grad_norm": 0.6510254144668579, "learning_rate": 3.653144917562335e-05, "loss": 0.3991, "step": 24300 }, { "epoch": 9.177047942619856, "grad_norm": 0.5957704782485962, "learning_rate": 3.649164498928277e-05, "loss": 0.3752, "step": 24310 }, { "epoch": 9.18082295205738, "grad_norm": 0.7079484462738037, "learning_rate": 3.645185003577546e-05, "loss": 0.3757, "step": 24320 }, { "epoch": 9.184597961494903, "grad_norm": 0.6439237594604492, "learning_rate": 3.6412064342300906e-05, "loss": 0.3793, "step": 24330 }, { "epoch": 9.188372970932427, "grad_norm": 0.6560051441192627, "learning_rate": 3.637228793605224e-05, "loss": 0.3781, "step": 24340 }, { "epoch": 9.19214798036995, "grad_norm": 0.6205071210861206, "learning_rate": 3.6332520844216264e-05, "loss": 0.373, "step": 24350 }, { "epoch": 9.195922989807475, "grad_norm": 0.8601105213165283, "learning_rate": 3.6292763093973425e-05, "loss": 0.3697, "step": 24360 }, { "epoch": 9.199697999244998, "grad_norm": 0.6811267733573914, "learning_rate": 3.6253014712497754e-05, "loss": 0.3826, "step": 24370 }, { "epoch": 9.203473008682522, "grad_norm": 0.6935433745384216, "learning_rate": 3.621327572695692e-05, "loss": 0.3782, "step": 24380 }, { "epoch": 9.207248018120046, "grad_norm": 0.5961577892303467, "learning_rate": 3.617354616451211e-05, "loss": 0.375, "step": 24390 }, { "epoch": 9.21102302755757, "grad_norm": 0.5627548694610596, "learning_rate": 3.6133826052318116e-05, "loss": 0.3731, "step": 24400 }, { "epoch": 9.214798036995093, "grad_norm": 0.6557216048240662, "learning_rate": 3.609411541752327e-05, "loss": 0.3741, "step": 24410 }, { "epoch": 9.218573046432615, "grad_norm": 0.7413145899772644, "learning_rate": 3.6054414287269405e-05, "loss": 0.3712, "step": 24420 }, { "epoch": 9.222348055870139, "grad_norm": 0.6858183145523071, "learning_rate": 3.601472268869188e-05, "loss": 0.3814, "step": 24430 }, { "epoch": 9.226123065307663, "grad_norm": 0.6973282098770142, "learning_rate": 3.597504064891952e-05, "loss": 0.377, "step": 24440 }, { "epoch": 9.229898074745186, "grad_norm": 0.6426398158073425, "learning_rate": 3.5935368195074636e-05, "loss": 0.3757, "step": 24450 }, { "epoch": 9.23367308418271, "grad_norm": 0.6414214372634888, "learning_rate": 3.589570535427297e-05, "loss": 0.3642, "step": 24460 }, { "epoch": 9.237448093620234, "grad_norm": 0.6278491020202637, "learning_rate": 3.585605215362371e-05, "loss": 0.3778, "step": 24470 }, { "epoch": 9.241223103057758, "grad_norm": 0.695353090763092, "learning_rate": 3.581640862022941e-05, "loss": 0.3763, "step": 24480 }, { "epoch": 9.244998112495281, "grad_norm": 0.6020937561988831, "learning_rate": 3.57767747811861e-05, "loss": 0.3761, "step": 24490 }, { "epoch": 9.248773121932805, "grad_norm": 0.7220026254653931, "learning_rate": 3.573715066358308e-05, "loss": 0.3751, "step": 24500 }, { "epoch": 9.252548131370329, "grad_norm": 0.6463865041732788, "learning_rate": 3.569753629450311e-05, "loss": 0.3783, "step": 24510 }, { "epoch": 9.256323140807853, "grad_norm": 0.6139450073242188, "learning_rate": 3.565793170102221e-05, "loss": 0.3754, "step": 24520 }, { "epoch": 9.260098150245376, "grad_norm": 0.6605015993118286, "learning_rate": 3.561833691020976e-05, "loss": 0.3685, "step": 24530 }, { "epoch": 9.2638731596829, "grad_norm": 0.6926522254943848, "learning_rate": 3.5578751949128415e-05, "loss": 0.374, "step": 24540 }, { "epoch": 9.267648169120422, "grad_norm": 0.6264640092849731, "learning_rate": 3.5539176844834125e-05, "loss": 0.3783, "step": 24550 }, { "epoch": 9.271423178557946, "grad_norm": 0.9688851237297058, "learning_rate": 3.5499611624376125e-05, "loss": 0.3795, "step": 24560 }, { "epoch": 9.27519818799547, "grad_norm": 0.5947738885879517, "learning_rate": 3.546005631479684e-05, "loss": 0.3662, "step": 24570 }, { "epoch": 9.278973197432993, "grad_norm": 0.6355003714561462, "learning_rate": 3.542051094313196e-05, "loss": 0.3785, "step": 24580 }, { "epoch": 9.282748206870517, "grad_norm": 0.6188831329345703, "learning_rate": 3.5380975536410364e-05, "loss": 0.3804, "step": 24590 }, { "epoch": 9.28652321630804, "grad_norm": 0.6349472403526306, "learning_rate": 3.534145012165415e-05, "loss": 0.3757, "step": 24600 }, { "epoch": 9.290298225745564, "grad_norm": 0.6269373297691345, "learning_rate": 3.5301934725878546e-05, "loss": 0.3698, "step": 24610 }, { "epoch": 9.294073235183088, "grad_norm": 0.6995264291763306, "learning_rate": 3.526242937609197e-05, "loss": 0.3798, "step": 24620 }, { "epoch": 9.297848244620612, "grad_norm": 0.6309302449226379, "learning_rate": 3.522293409929595e-05, "loss": 0.3652, "step": 24630 }, { "epoch": 9.301623254058136, "grad_norm": 0.8019225597381592, "learning_rate": 3.518344892248513e-05, "loss": 0.3808, "step": 24640 }, { "epoch": 9.30539826349566, "grad_norm": 0.6818584203720093, "learning_rate": 3.514397387264725e-05, "loss": 0.3726, "step": 24650 }, { "epoch": 9.309173272933183, "grad_norm": 0.6697105169296265, "learning_rate": 3.5104508976763176e-05, "loss": 0.3678, "step": 24660 }, { "epoch": 9.312948282370705, "grad_norm": 0.5741595029830933, "learning_rate": 3.506505426180674e-05, "loss": 0.38, "step": 24670 }, { "epoch": 9.316723291808229, "grad_norm": 0.6436684131622314, "learning_rate": 3.502560975474488e-05, "loss": 0.3747, "step": 24680 }, { "epoch": 9.320498301245753, "grad_norm": 0.6101210117340088, "learning_rate": 3.4986175482537566e-05, "loss": 0.3678, "step": 24690 }, { "epoch": 9.324273310683276, "grad_norm": 0.8269137144088745, "learning_rate": 3.4946751472137725e-05, "loss": 0.3721, "step": 24700 }, { "epoch": 9.3280483201208, "grad_norm": 0.6154971718788147, "learning_rate": 3.490733775049132e-05, "loss": 0.379, "step": 24710 }, { "epoch": 9.331823329558324, "grad_norm": 0.642996072769165, "learning_rate": 3.4867934344537236e-05, "loss": 0.3823, "step": 24720 }, { "epoch": 9.335598338995847, "grad_norm": 0.6304564476013184, "learning_rate": 3.482854128120735e-05, "loss": 0.3742, "step": 24730 }, { "epoch": 9.339373348433371, "grad_norm": 0.6240081787109375, "learning_rate": 3.478915858742643e-05, "loss": 0.3698, "step": 24740 }, { "epoch": 9.343148357870895, "grad_norm": 0.6386808156967163, "learning_rate": 3.4749786290112205e-05, "loss": 0.3709, "step": 24750 }, { "epoch": 9.346923367308419, "grad_norm": 0.6619960069656372, "learning_rate": 3.471042441617524e-05, "loss": 0.3689, "step": 24760 }, { "epoch": 9.350698376745942, "grad_norm": 0.8115395903587341, "learning_rate": 3.467107299251902e-05, "loss": 0.3721, "step": 24770 }, { "epoch": 9.354473386183466, "grad_norm": 0.6916590929031372, "learning_rate": 3.463173204603984e-05, "loss": 0.3736, "step": 24780 }, { "epoch": 9.35824839562099, "grad_norm": 0.6382922530174255, "learning_rate": 3.4592401603626924e-05, "loss": 0.3715, "step": 24790 }, { "epoch": 9.362023405058512, "grad_norm": 0.6073492765426636, "learning_rate": 3.45530816921622e-05, "loss": 0.3714, "step": 24800 }, { "epoch": 9.365798414496036, "grad_norm": 0.6381345987319946, "learning_rate": 3.451377233852051e-05, "loss": 0.3758, "step": 24810 }, { "epoch": 9.36957342393356, "grad_norm": 0.5805862545967102, "learning_rate": 3.4474473569569385e-05, "loss": 0.37, "step": 24820 }, { "epoch": 9.373348433371083, "grad_norm": 0.7061961889266968, "learning_rate": 3.443518541216918e-05, "loss": 0.3721, "step": 24830 }, { "epoch": 9.377123442808607, "grad_norm": 0.6575196981430054, "learning_rate": 3.439590789317299e-05, "loss": 0.3731, "step": 24840 }, { "epoch": 9.38089845224613, "grad_norm": 0.6737225651741028, "learning_rate": 3.4356641039426607e-05, "loss": 0.3712, "step": 24850 }, { "epoch": 9.384673461683654, "grad_norm": 0.7008658051490784, "learning_rate": 3.431738487776857e-05, "loss": 0.3713, "step": 24860 }, { "epoch": 9.388448471121178, "grad_norm": 0.6766116619110107, "learning_rate": 3.4278139435030084e-05, "loss": 0.3772, "step": 24870 }, { "epoch": 9.392223480558702, "grad_norm": 0.6456305384635925, "learning_rate": 3.423890473803504e-05, "loss": 0.3804, "step": 24880 }, { "epoch": 9.395998489996225, "grad_norm": 0.7182512879371643, "learning_rate": 3.41996808136e-05, "loss": 0.37, "step": 24890 }, { "epoch": 9.39977349943375, "grad_norm": 0.6886579990386963, "learning_rate": 3.416046768853413e-05, "loss": 0.3696, "step": 24900 }, { "epoch": 9.403548508871273, "grad_norm": 0.6319290399551392, "learning_rate": 3.412126538963925e-05, "loss": 0.3771, "step": 24910 }, { "epoch": 9.407323518308797, "grad_norm": 0.5859635472297668, "learning_rate": 3.4082073943709727e-05, "loss": 0.3729, "step": 24920 }, { "epoch": 9.411098527746319, "grad_norm": 0.7094352841377258, "learning_rate": 3.404289337753258e-05, "loss": 0.374, "step": 24930 }, { "epoch": 9.414873537183842, "grad_norm": 0.6192387938499451, "learning_rate": 3.400372371788736e-05, "loss": 0.3749, "step": 24940 }, { "epoch": 9.418648546621366, "grad_norm": 0.5935202836990356, "learning_rate": 3.3964564991546124e-05, "loss": 0.3707, "step": 24950 }, { "epoch": 9.42242355605889, "grad_norm": 0.6603958606719971, "learning_rate": 3.392541722527351e-05, "loss": 0.3825, "step": 24960 }, { "epoch": 9.426198565496414, "grad_norm": 0.6225345134735107, "learning_rate": 3.3886280445826644e-05, "loss": 0.3669, "step": 24970 }, { "epoch": 9.429973574933937, "grad_norm": 0.6600792407989502, "learning_rate": 3.3847154679955154e-05, "loss": 0.373, "step": 24980 }, { "epoch": 9.433748584371461, "grad_norm": 0.6622351408004761, "learning_rate": 3.380803995440113e-05, "loss": 0.3696, "step": 24990 }, { "epoch": 9.437523593808985, "grad_norm": 0.6164969801902771, "learning_rate": 3.3768936295899115e-05, "loss": 0.3682, "step": 25000 }, { "epoch": 9.441298603246508, "grad_norm": 0.704637348651886, "learning_rate": 3.3729843731176094e-05, "loss": 0.3734, "step": 25010 }, { "epoch": 9.445073612684032, "grad_norm": 0.6271415948867798, "learning_rate": 3.369076228695146e-05, "loss": 0.3691, "step": 25020 }, { "epoch": 9.448848622121556, "grad_norm": 0.6947619318962097, "learning_rate": 3.365169198993703e-05, "loss": 0.3781, "step": 25030 }, { "epoch": 9.45262363155908, "grad_norm": 0.6423031091690063, "learning_rate": 3.361263286683697e-05, "loss": 0.3713, "step": 25040 }, { "epoch": 9.456398640996602, "grad_norm": 0.6936047673225403, "learning_rate": 3.35735849443478e-05, "loss": 0.3714, "step": 25050 }, { "epoch": 9.460173650434125, "grad_norm": 0.6276016235351562, "learning_rate": 3.3534548249158435e-05, "loss": 0.3695, "step": 25060 }, { "epoch": 9.46394865987165, "grad_norm": 0.6831814646720886, "learning_rate": 3.3495522807950086e-05, "loss": 0.3706, "step": 25070 }, { "epoch": 9.467723669309173, "grad_norm": 0.6282784342765808, "learning_rate": 3.345650864739627e-05, "loss": 0.3717, "step": 25080 }, { "epoch": 9.471498678746697, "grad_norm": 0.6343705654144287, "learning_rate": 3.3417505794162794e-05, "loss": 0.3685, "step": 25090 }, { "epoch": 9.47527368818422, "grad_norm": 0.6878871917724609, "learning_rate": 3.3378514274907745e-05, "loss": 0.3773, "step": 25100 }, { "epoch": 9.479048697621744, "grad_norm": 0.7008819580078125, "learning_rate": 3.333953411628147e-05, "loss": 0.3681, "step": 25110 }, { "epoch": 9.482823707059268, "grad_norm": 0.653729259967804, "learning_rate": 3.330056534492653e-05, "loss": 0.3794, "step": 25120 }, { "epoch": 9.486598716496792, "grad_norm": 0.7341142296791077, "learning_rate": 3.32616079874777e-05, "loss": 0.3691, "step": 25130 }, { "epoch": 9.490373725934315, "grad_norm": 0.9911676049232483, "learning_rate": 3.322266207056197e-05, "loss": 0.37, "step": 25140 }, { "epoch": 9.494148735371839, "grad_norm": 0.6838816404342651, "learning_rate": 3.318372762079852e-05, "loss": 0.3757, "step": 25150 }, { "epoch": 9.497923744809363, "grad_norm": 0.7315078973770142, "learning_rate": 3.3144804664798666e-05, "loss": 0.3732, "step": 25160 }, { "epoch": 9.501698754246885, "grad_norm": 0.6944950222969055, "learning_rate": 3.3105893229165894e-05, "loss": 0.3777, "step": 25170 }, { "epoch": 9.505473763684408, "grad_norm": 0.6807419657707214, "learning_rate": 3.30669933404958e-05, "loss": 0.3688, "step": 25180 }, { "epoch": 9.509248773121932, "grad_norm": 0.7115993499755859, "learning_rate": 3.302810502537609e-05, "loss": 0.3691, "step": 25190 }, { "epoch": 9.513023782559456, "grad_norm": 0.8097192645072937, "learning_rate": 3.298922831038655e-05, "loss": 0.3776, "step": 25200 }, { "epoch": 9.51679879199698, "grad_norm": 0.7089084386825562, "learning_rate": 3.2950363222099073e-05, "loss": 0.3708, "step": 25210 }, { "epoch": 9.520573801434503, "grad_norm": 0.6845264434814453, "learning_rate": 3.291150978707758e-05, "loss": 0.3715, "step": 25220 }, { "epoch": 9.524348810872027, "grad_norm": 0.6236247420310974, "learning_rate": 3.287266803187798e-05, "loss": 0.3671, "step": 25230 }, { "epoch": 9.52812382030955, "grad_norm": 0.7404671907424927, "learning_rate": 3.283383798304829e-05, "loss": 0.3735, "step": 25240 }, { "epoch": 9.531898829747075, "grad_norm": 0.6793319582939148, "learning_rate": 3.279501966712847e-05, "loss": 0.3674, "step": 25250 }, { "epoch": 9.535673839184598, "grad_norm": 0.7164490222930908, "learning_rate": 3.275621311065047e-05, "loss": 0.3726, "step": 25260 }, { "epoch": 9.539448848622122, "grad_norm": 0.8006794452667236, "learning_rate": 3.271741834013822e-05, "loss": 0.3671, "step": 25270 }, { "epoch": 9.543223858059646, "grad_norm": 0.6206616759300232, "learning_rate": 3.267863538210756e-05, "loss": 0.3736, "step": 25280 }, { "epoch": 9.54699886749717, "grad_norm": 0.6397393941879272, "learning_rate": 3.2639864263066296e-05, "loss": 0.3697, "step": 25290 }, { "epoch": 9.550773876934691, "grad_norm": 1.0796784162521362, "learning_rate": 3.26011050095141e-05, "loss": 0.3907, "step": 25300 }, { "epoch": 9.554548886372215, "grad_norm": 2.355830192565918, "learning_rate": 3.256235764794259e-05, "loss": 0.3707, "step": 25310 }, { "epoch": 9.558323895809739, "grad_norm": 0.6870741844177246, "learning_rate": 3.2523622204835194e-05, "loss": 0.3682, "step": 25320 }, { "epoch": 9.562098905247263, "grad_norm": 0.7428625822067261, "learning_rate": 3.2484898706667214e-05, "loss": 0.369, "step": 25330 }, { "epoch": 9.565873914684786, "grad_norm": 0.7070319652557373, "learning_rate": 3.2446187179905806e-05, "loss": 0.3685, "step": 25340 }, { "epoch": 9.56964892412231, "grad_norm": 0.6212445497512817, "learning_rate": 3.240748765100995e-05, "loss": 0.3802, "step": 25350 }, { "epoch": 9.573423933559834, "grad_norm": 0.6794653534889221, "learning_rate": 3.236880014643039e-05, "loss": 0.3665, "step": 25360 }, { "epoch": 9.577198942997358, "grad_norm": 0.6495829224586487, "learning_rate": 3.233012469260969e-05, "loss": 0.3645, "step": 25370 }, { "epoch": 9.580973952434881, "grad_norm": 0.6042360067367554, "learning_rate": 3.229146131598213e-05, "loss": 0.3745, "step": 25380 }, { "epoch": 9.584748961872405, "grad_norm": 2.035146713256836, "learning_rate": 3.2252810042973794e-05, "loss": 0.3827, "step": 25390 }, { "epoch": 9.588523971309929, "grad_norm": 0.5512405633926392, "learning_rate": 3.2214170900002456e-05, "loss": 0.3873, "step": 25400 }, { "epoch": 9.592298980747453, "grad_norm": 0.7587074041366577, "learning_rate": 3.217554391347758e-05, "loss": 0.3671, "step": 25410 }, { "epoch": 9.596073990184976, "grad_norm": 0.7429296970367432, "learning_rate": 3.213692910980037e-05, "loss": 0.3708, "step": 25420 }, { "epoch": 9.599848999622498, "grad_norm": 0.6271414160728455, "learning_rate": 3.2098326515363666e-05, "loss": 0.3715, "step": 25430 }, { "epoch": 9.603624009060022, "grad_norm": 0.6165705919265747, "learning_rate": 3.205973615655199e-05, "loss": 0.3669, "step": 25440 }, { "epoch": 9.607399018497546, "grad_norm": 0.5890395641326904, "learning_rate": 3.202115805974149e-05, "loss": 0.3681, "step": 25450 }, { "epoch": 9.61117402793507, "grad_norm": 0.6870768070220947, "learning_rate": 3.1982592251299916e-05, "loss": 0.3736, "step": 25460 }, { "epoch": 9.614949037372593, "grad_norm": 0.7155221700668335, "learning_rate": 3.1944038757586656e-05, "loss": 0.3716, "step": 25470 }, { "epoch": 9.618724046810117, "grad_norm": 0.6171225905418396, "learning_rate": 3.190549760495263e-05, "loss": 0.3712, "step": 25480 }, { "epoch": 9.62249905624764, "grad_norm": 0.7019498944282532, "learning_rate": 3.186696881974039e-05, "loss": 0.3641, "step": 25490 }, { "epoch": 9.626274065685164, "grad_norm": 0.7143846154212952, "learning_rate": 3.1828452428283986e-05, "loss": 0.3769, "step": 25500 }, { "epoch": 9.630049075122688, "grad_norm": 0.6226781606674194, "learning_rate": 3.178994845690898e-05, "loss": 0.3636, "step": 25510 }, { "epoch": 9.633824084560212, "grad_norm": 0.6864441633224487, "learning_rate": 3.17514569319325e-05, "loss": 0.3741, "step": 25520 }, { "epoch": 9.637599093997736, "grad_norm": 0.6476948261260986, "learning_rate": 3.171297787966312e-05, "loss": 0.3763, "step": 25530 }, { "epoch": 9.64137410343526, "grad_norm": 0.6583799123764038, "learning_rate": 3.167451132640093e-05, "loss": 0.3715, "step": 25540 }, { "epoch": 9.645149112872783, "grad_norm": 0.8483378291130066, "learning_rate": 3.163605729843746e-05, "loss": 0.3742, "step": 25550 }, { "epoch": 9.648924122310305, "grad_norm": 0.7014634609222412, "learning_rate": 3.159761582205565e-05, "loss": 0.3693, "step": 25560 }, { "epoch": 9.652699131747829, "grad_norm": 0.8034195899963379, "learning_rate": 3.155918692352992e-05, "loss": 0.3697, "step": 25570 }, { "epoch": 9.656474141185353, "grad_norm": 0.7062366008758545, "learning_rate": 3.152077062912602e-05, "loss": 0.3776, "step": 25580 }, { "epoch": 9.660249150622876, "grad_norm": 0.6510414481163025, "learning_rate": 3.148236696510117e-05, "loss": 0.3764, "step": 25590 }, { "epoch": 9.6640241600604, "grad_norm": 0.7279618978500366, "learning_rate": 3.144397595770388e-05, "loss": 0.3773, "step": 25600 }, { "epoch": 9.667799169497924, "grad_norm": 0.6099146008491516, "learning_rate": 3.1405597633174036e-05, "loss": 0.376, "step": 25610 }, { "epoch": 9.671574178935447, "grad_norm": 0.6417626142501831, "learning_rate": 3.136723201774289e-05, "loss": 0.3717, "step": 25620 }, { "epoch": 9.675349188372971, "grad_norm": 0.7016961574554443, "learning_rate": 3.132887913763295e-05, "loss": 0.3698, "step": 25630 }, { "epoch": 9.679124197810495, "grad_norm": 0.7685357928276062, "learning_rate": 3.129053901905806e-05, "loss": 0.3648, "step": 25640 }, { "epoch": 9.682899207248019, "grad_norm": 0.7496814131736755, "learning_rate": 3.125221168822335e-05, "loss": 0.3781, "step": 25650 }, { "epoch": 9.686674216685542, "grad_norm": 0.6221609711647034, "learning_rate": 3.1213897171325154e-05, "loss": 0.3694, "step": 25660 }, { "epoch": 9.690449226123066, "grad_norm": 0.6706365942955017, "learning_rate": 3.1175595494551116e-05, "loss": 0.3709, "step": 25670 }, { "epoch": 9.694224235560588, "grad_norm": 0.5938177704811096, "learning_rate": 3.1137306684080045e-05, "loss": 0.3659, "step": 25680 }, { "epoch": 9.697999244998112, "grad_norm": 0.69608074426651, "learning_rate": 3.1099030766081985e-05, "loss": 0.3712, "step": 25690 }, { "epoch": 9.701774254435636, "grad_norm": 0.6536232829093933, "learning_rate": 3.106076776671818e-05, "loss": 0.3678, "step": 25700 }, { "epoch": 9.70554926387316, "grad_norm": 0.6710280179977417, "learning_rate": 3.102251771214101e-05, "loss": 0.3667, "step": 25710 }, { "epoch": 9.709324273310683, "grad_norm": 0.6235374808311462, "learning_rate": 3.098428062849404e-05, "loss": 0.3684, "step": 25720 }, { "epoch": 9.713099282748207, "grad_norm": 0.6490816473960876, "learning_rate": 3.094605654191195e-05, "loss": 0.3715, "step": 25730 }, { "epoch": 9.71687429218573, "grad_norm": 0.6581624746322632, "learning_rate": 3.090784547852055e-05, "loss": 0.3714, "step": 25740 }, { "epoch": 9.720649301623254, "grad_norm": 0.7839778065681458, "learning_rate": 3.0869647464436746e-05, "loss": 0.373, "step": 25750 }, { "epoch": 9.724424311060778, "grad_norm": 0.6403361558914185, "learning_rate": 3.0831462525768496e-05, "loss": 0.3684, "step": 25760 }, { "epoch": 9.728199320498302, "grad_norm": 0.6090444326400757, "learning_rate": 3.079329068861488e-05, "loss": 0.3671, "step": 25770 }, { "epoch": 9.731974329935825, "grad_norm": 0.6840675473213196, "learning_rate": 3.075513197906597e-05, "loss": 0.3682, "step": 25780 }, { "epoch": 9.73574933937335, "grad_norm": 0.6021028757095337, "learning_rate": 3.071698642320286e-05, "loss": 0.3723, "step": 25790 }, { "epoch": 9.739524348810871, "grad_norm": 0.6243388652801514, "learning_rate": 3.067885404709772e-05, "loss": 0.371, "step": 25800 }, { "epoch": 9.743299358248395, "grad_norm": 0.6200324296951294, "learning_rate": 3.0640734876813636e-05, "loss": 0.3729, "step": 25810 }, { "epoch": 9.747074367685919, "grad_norm": 1.0078961849212646, "learning_rate": 3.060262893840473e-05, "loss": 0.3695, "step": 25820 }, { "epoch": 9.750849377123442, "grad_norm": 0.6697998642921448, "learning_rate": 3.056453625791603e-05, "loss": 0.3641, "step": 25830 }, { "epoch": 9.754624386560966, "grad_norm": 0.8702282905578613, "learning_rate": 3.052645686138353e-05, "loss": 0.3734, "step": 25840 }, { "epoch": 9.75839939599849, "grad_norm": 0.6085755825042725, "learning_rate": 3.0488390774834153e-05, "loss": 0.3702, "step": 25850 }, { "epoch": 9.762174405436014, "grad_norm": 0.7779916524887085, "learning_rate": 3.0450338024285684e-05, "loss": 0.3718, "step": 25860 }, { "epoch": 9.765949414873537, "grad_norm": 0.6143949627876282, "learning_rate": 3.0412298635746855e-05, "loss": 0.3663, "step": 25870 }, { "epoch": 9.769724424311061, "grad_norm": 0.7032211422920227, "learning_rate": 3.03742726352172e-05, "loss": 0.3688, "step": 25880 }, { "epoch": 9.773499433748585, "grad_norm": 0.6979837417602539, "learning_rate": 3.0336260048687125e-05, "loss": 0.3648, "step": 25890 }, { "epoch": 9.777274443186108, "grad_norm": 0.6320093274116516, "learning_rate": 3.0298260902137897e-05, "loss": 0.3666, "step": 25900 }, { "epoch": 9.781049452623632, "grad_norm": 0.6914851665496826, "learning_rate": 3.0260275221541566e-05, "loss": 0.3743, "step": 25910 }, { "epoch": 9.784824462061156, "grad_norm": 0.6460880041122437, "learning_rate": 3.0222303032860987e-05, "loss": 0.376, "step": 25920 }, { "epoch": 9.788599471498678, "grad_norm": 0.7050313353538513, "learning_rate": 3.018434436204979e-05, "loss": 0.3655, "step": 25930 }, { "epoch": 9.792374480936202, "grad_norm": 0.7031101584434509, "learning_rate": 3.014639923505237e-05, "loss": 0.367, "step": 25940 }, { "epoch": 9.796149490373725, "grad_norm": 0.6252807378768921, "learning_rate": 3.0108467677803863e-05, "loss": 0.3669, "step": 25950 }, { "epoch": 9.79992449981125, "grad_norm": 0.6728451251983643, "learning_rate": 3.0070549716230156e-05, "loss": 0.3685, "step": 25960 }, { "epoch": 9.803699509248773, "grad_norm": 0.6584720015525818, "learning_rate": 3.003264537624777e-05, "loss": 0.3704, "step": 25970 }, { "epoch": 9.807474518686297, "grad_norm": 0.5711626410484314, "learning_rate": 2.9994754683764e-05, "loss": 0.3833, "step": 25980 }, { "epoch": 9.81124952812382, "grad_norm": 0.7549001574516296, "learning_rate": 2.9956877664676754e-05, "loss": 0.3661, "step": 25990 }, { "epoch": 9.815024537561344, "grad_norm": 0.6147065162658691, "learning_rate": 2.9919014344874636e-05, "loss": 0.3647, "step": 26000 }, { "epoch": 9.818799546998868, "grad_norm": 0.7019761800765991, "learning_rate": 2.9881164750236857e-05, "loss": 0.3668, "step": 26010 }, { "epoch": 9.822574556436392, "grad_norm": 0.6753397583961487, "learning_rate": 2.984332890663326e-05, "loss": 0.3729, "step": 26020 }, { "epoch": 9.826349565873915, "grad_norm": 0.6772192120552063, "learning_rate": 2.9805506839924292e-05, "loss": 0.3735, "step": 26030 }, { "epoch": 9.830124575311439, "grad_norm": 0.6899954080581665, "learning_rate": 2.9767698575960968e-05, "loss": 0.3628, "step": 26040 }, { "epoch": 9.833899584748963, "grad_norm": 0.6952143907546997, "learning_rate": 2.9729904140584913e-05, "loss": 0.3724, "step": 26050 }, { "epoch": 9.837674594186485, "grad_norm": 0.6387107968330383, "learning_rate": 2.9692123559628234e-05, "loss": 0.368, "step": 26060 }, { "epoch": 9.841449603624008, "grad_norm": 0.6607224345207214, "learning_rate": 2.9654356858913596e-05, "loss": 0.3642, "step": 26070 }, { "epoch": 9.845224613061532, "grad_norm": 0.6394532322883606, "learning_rate": 2.9616604064254206e-05, "loss": 0.3641, "step": 26080 }, { "epoch": 9.848999622499056, "grad_norm": 0.6913762092590332, "learning_rate": 2.9578865201453732e-05, "loss": 0.3614, "step": 26090 }, { "epoch": 9.85277463193658, "grad_norm": 0.7300879955291748, "learning_rate": 2.9541140296306335e-05, "loss": 0.3701, "step": 26100 }, { "epoch": 9.856549641374103, "grad_norm": 0.6649528741836548, "learning_rate": 2.9503429374596627e-05, "loss": 0.3746, "step": 26110 }, { "epoch": 9.860324650811627, "grad_norm": 0.693498432636261, "learning_rate": 2.946573246209967e-05, "loss": 0.3665, "step": 26120 }, { "epoch": 9.86409966024915, "grad_norm": 0.8667472004890442, "learning_rate": 2.942804958458094e-05, "loss": 0.3788, "step": 26130 }, { "epoch": 9.867874669686675, "grad_norm": 0.6629686951637268, "learning_rate": 2.9390380767796343e-05, "loss": 0.3673, "step": 26140 }, { "epoch": 9.871649679124198, "grad_norm": 0.6008899211883545, "learning_rate": 2.9352726037492174e-05, "loss": 0.3666, "step": 26150 }, { "epoch": 9.875424688561722, "grad_norm": 0.7541015148162842, "learning_rate": 2.9315085419405052e-05, "loss": 0.364, "step": 26160 }, { "epoch": 9.879199697999246, "grad_norm": 0.6057034730911255, "learning_rate": 2.927745893926199e-05, "loss": 0.3711, "step": 26170 }, { "epoch": 9.88297470743677, "grad_norm": 0.6768243908882141, "learning_rate": 2.9239846622780358e-05, "loss": 0.3707, "step": 26180 }, { "epoch": 9.886749716874291, "grad_norm": 0.7239327430725098, "learning_rate": 2.9202248495667788e-05, "loss": 0.3791, "step": 26190 }, { "epoch": 9.890524726311815, "grad_norm": 0.5899133086204529, "learning_rate": 2.916466458362227e-05, "loss": 0.3661, "step": 26200 }, { "epoch": 9.894299735749339, "grad_norm": 0.5926426649093628, "learning_rate": 2.9127094912332033e-05, "loss": 0.3753, "step": 26210 }, { "epoch": 9.898074745186863, "grad_norm": 2.781858205795288, "learning_rate": 2.9089539507475606e-05, "loss": 0.3749, "step": 26220 }, { "epoch": 9.901849754624386, "grad_norm": 0.6369478106498718, "learning_rate": 2.9051998394721748e-05, "loss": 0.3652, "step": 26230 }, { "epoch": 9.90562476406191, "grad_norm": 0.6876001954078674, "learning_rate": 2.901447159972948e-05, "loss": 0.3613, "step": 26240 }, { "epoch": 9.909399773499434, "grad_norm": 0.6942611932754517, "learning_rate": 2.8976959148148e-05, "loss": 0.372, "step": 26250 }, { "epoch": 9.913174782936958, "grad_norm": 0.7924140095710754, "learning_rate": 2.8939461065616674e-05, "loss": 0.3695, "step": 26260 }, { "epoch": 9.916949792374481, "grad_norm": 0.6020333766937256, "learning_rate": 2.8901977377765127e-05, "loss": 0.3688, "step": 26270 }, { "epoch": 9.920724801812005, "grad_norm": 0.6307262778282166, "learning_rate": 2.8864508110213094e-05, "loss": 0.3649, "step": 26280 }, { "epoch": 9.924499811249529, "grad_norm": 0.6264989972114563, "learning_rate": 2.8827053288570503e-05, "loss": 0.3745, "step": 26290 }, { "epoch": 9.92827482068705, "grad_norm": 0.7633573412895203, "learning_rate": 2.8789612938437315e-05, "loss": 0.3692, "step": 26300 }, { "epoch": 9.932049830124575, "grad_norm": 0.9851648211479187, "learning_rate": 2.8752187085403683e-05, "loss": 0.3746, "step": 26310 }, { "epoch": 9.935824839562098, "grad_norm": 0.6989134550094604, "learning_rate": 2.8714775755049818e-05, "loss": 0.3723, "step": 26320 }, { "epoch": 9.939599848999622, "grad_norm": 0.6544426679611206, "learning_rate": 2.867737897294604e-05, "loss": 0.3744, "step": 26330 }, { "epoch": 9.943374858437146, "grad_norm": 0.6202349662780762, "learning_rate": 2.8639996764652653e-05, "loss": 0.3753, "step": 26340 }, { "epoch": 9.94714986787467, "grad_norm": 0.6523613929748535, "learning_rate": 2.8602629155720084e-05, "loss": 0.3651, "step": 26350 }, { "epoch": 9.950924877312193, "grad_norm": 0.6727064251899719, "learning_rate": 2.8565276171688703e-05, "loss": 0.3732, "step": 26360 }, { "epoch": 9.954699886749717, "grad_norm": 0.7078951001167297, "learning_rate": 2.8527937838088943e-05, "loss": 0.3648, "step": 26370 }, { "epoch": 9.95847489618724, "grad_norm": 0.7242457270622253, "learning_rate": 2.84906141804412e-05, "loss": 0.3718, "step": 26380 }, { "epoch": 9.962249905624764, "grad_norm": 0.6599700450897217, "learning_rate": 2.8453305224255867e-05, "loss": 0.364, "step": 26390 }, { "epoch": 9.966024915062288, "grad_norm": 0.6147913336753845, "learning_rate": 2.8416010995033216e-05, "loss": 0.3598, "step": 26400 }, { "epoch": 9.969799924499812, "grad_norm": 0.6449320316314697, "learning_rate": 2.8378731518263524e-05, "loss": 0.3718, "step": 26410 }, { "epoch": 9.973574933937336, "grad_norm": 0.6980818510055542, "learning_rate": 2.834146681942696e-05, "loss": 0.3695, "step": 26420 }, { "epoch": 9.977349943374858, "grad_norm": 0.6812340617179871, "learning_rate": 2.8304216923993622e-05, "loss": 0.3734, "step": 26430 }, { "epoch": 9.981124952812381, "grad_norm": 0.644469678401947, "learning_rate": 2.8266981857423413e-05, "loss": 0.3704, "step": 26440 }, { "epoch": 9.984899962249905, "grad_norm": 0.6593320369720459, "learning_rate": 2.8229761645166197e-05, "loss": 0.3713, "step": 26450 }, { "epoch": 9.988674971687429, "grad_norm": 0.5889549255371094, "learning_rate": 2.81925563126616e-05, "loss": 0.3609, "step": 26460 }, { "epoch": 9.992449981124953, "grad_norm": 0.6138685345649719, "learning_rate": 2.8155365885339124e-05, "loss": 0.3712, "step": 26470 }, { "epoch": 9.996224990562476, "grad_norm": 0.6867060661315918, "learning_rate": 2.8118190388618093e-05, "loss": 0.3694, "step": 26480 }, { "epoch": 10.0, "grad_norm": 0.6916232705116272, "learning_rate": 2.8081029847907614e-05, "loss": 0.375, "step": 26490 }, { "epoch": 10.003775009437524, "grad_norm": 0.6438423991203308, "learning_rate": 2.8043884288606525e-05, "loss": 0.3668, "step": 26500 }, { "epoch": 10.007550018875047, "grad_norm": 0.5959725975990295, "learning_rate": 2.8006753736103496e-05, "loss": 0.3637, "step": 26510 }, { "epoch": 10.011325028312571, "grad_norm": 0.5706154108047485, "learning_rate": 2.7969638215776918e-05, "loss": 0.3712, "step": 26520 }, { "epoch": 10.015100037750095, "grad_norm": 0.6247308254241943, "learning_rate": 2.793253775299487e-05, "loss": 0.3686, "step": 26530 }, { "epoch": 10.018875047187619, "grad_norm": 0.6258418560028076, "learning_rate": 2.7895452373115184e-05, "loss": 0.3733, "step": 26540 }, { "epoch": 10.022650056625142, "grad_norm": 0.6243855357170105, "learning_rate": 2.785838210148539e-05, "loss": 0.3663, "step": 26550 }, { "epoch": 10.026425066062664, "grad_norm": 0.6902307868003845, "learning_rate": 2.782132696344263e-05, "loss": 0.3783, "step": 26560 }, { "epoch": 10.030200075500188, "grad_norm": 0.650276243686676, "learning_rate": 2.7784286984313745e-05, "loss": 0.368, "step": 26570 }, { "epoch": 10.033975084937712, "grad_norm": 0.6885844469070435, "learning_rate": 2.7747262189415236e-05, "loss": 0.3603, "step": 26580 }, { "epoch": 10.037750094375236, "grad_norm": 0.6648432612419128, "learning_rate": 2.7710252604053205e-05, "loss": 0.3738, "step": 26590 }, { "epoch": 10.04152510381276, "grad_norm": 0.6929357051849365, "learning_rate": 2.767325825352332e-05, "loss": 0.3692, "step": 26600 }, { "epoch": 10.045300113250283, "grad_norm": 0.6674989461898804, "learning_rate": 2.7636279163110913e-05, "loss": 0.3668, "step": 26610 }, { "epoch": 10.049075122687807, "grad_norm": 0.6698505282402039, "learning_rate": 2.7599315358090795e-05, "loss": 0.3727, "step": 26620 }, { "epoch": 10.05285013212533, "grad_norm": 0.6277247071266174, "learning_rate": 2.7562366863727407e-05, "loss": 0.3664, "step": 26630 }, { "epoch": 10.056625141562854, "grad_norm": 0.6725890636444092, "learning_rate": 2.7525433705274695e-05, "loss": 0.3661, "step": 26640 }, { "epoch": 10.060400151000378, "grad_norm": 0.6743810772895813, "learning_rate": 2.748851590797614e-05, "loss": 0.3725, "step": 26650 }, { "epoch": 10.064175160437902, "grad_norm": 0.7060372829437256, "learning_rate": 2.7451613497064675e-05, "loss": 0.3743, "step": 26660 }, { "epoch": 10.067950169875425, "grad_norm": 0.6378623843193054, "learning_rate": 2.7414726497762765e-05, "loss": 0.3688, "step": 26670 }, { "epoch": 10.071725179312947, "grad_norm": 0.6196885704994202, "learning_rate": 2.737785493528232e-05, "loss": 0.3732, "step": 26680 }, { "epoch": 10.075500188750471, "grad_norm": 0.770141065120697, "learning_rate": 2.7340998834824745e-05, "loss": 0.3697, "step": 26690 }, { "epoch": 10.079275198187995, "grad_norm": 0.6975536942481995, "learning_rate": 2.7304158221580777e-05, "loss": 0.3688, "step": 26700 }, { "epoch": 10.083050207625519, "grad_norm": 0.6799348592758179, "learning_rate": 2.7267333120730675e-05, "loss": 0.3716, "step": 26710 }, { "epoch": 10.086825217063042, "grad_norm": 0.6629582643508911, "learning_rate": 2.7230523557444017e-05, "loss": 0.3653, "step": 26720 }, { "epoch": 10.090600226500566, "grad_norm": 0.612046480178833, "learning_rate": 2.7193729556879798e-05, "loss": 0.3646, "step": 26730 }, { "epoch": 10.09437523593809, "grad_norm": 0.8744679093360901, "learning_rate": 2.715695114418637e-05, "loss": 0.3704, "step": 26740 }, { "epoch": 10.098150245375614, "grad_norm": 0.6572837829589844, "learning_rate": 2.7120188344501475e-05, "loss": 0.368, "step": 26750 }, { "epoch": 10.101925254813137, "grad_norm": 0.680645763874054, "learning_rate": 2.7083441182952067e-05, "loss": 0.3746, "step": 26760 }, { "epoch": 10.105700264250661, "grad_norm": 0.6419723629951477, "learning_rate": 2.7046709684654527e-05, "loss": 0.366, "step": 26770 }, { "epoch": 10.109475273688185, "grad_norm": 0.6527169346809387, "learning_rate": 2.700999387471448e-05, "loss": 0.3715, "step": 26780 }, { "epoch": 10.113250283125709, "grad_norm": 0.6019080281257629, "learning_rate": 2.6973293778226854e-05, "loss": 0.3676, "step": 26790 }, { "epoch": 10.117025292563232, "grad_norm": 0.6170861721038818, "learning_rate": 2.6936609420275804e-05, "loss": 0.381, "step": 26800 }, { "epoch": 10.120800302000754, "grad_norm": 0.8691985607147217, "learning_rate": 2.689994082593472e-05, "loss": 0.3656, "step": 26810 }, { "epoch": 10.124575311438278, "grad_norm": 0.6585285067558289, "learning_rate": 2.6863288020266264e-05, "loss": 0.368, "step": 26820 }, { "epoch": 10.128350320875802, "grad_norm": 0.6726837158203125, "learning_rate": 2.682665102832228e-05, "loss": 0.3652, "step": 26830 }, { "epoch": 10.132125330313325, "grad_norm": 0.7205042839050293, "learning_rate": 2.67900298751438e-05, "loss": 0.3722, "step": 26840 }, { "epoch": 10.13590033975085, "grad_norm": 0.6744375228881836, "learning_rate": 2.6753424585761067e-05, "loss": 0.3751, "step": 26850 }, { "epoch": 10.139675349188373, "grad_norm": 3.0684523582458496, "learning_rate": 2.671683518519341e-05, "loss": 0.3647, "step": 26860 }, { "epoch": 10.143450358625897, "grad_norm": 0.585762083530426, "learning_rate": 2.668026169844936e-05, "loss": 0.3621, "step": 26870 }, { "epoch": 10.14722536806342, "grad_norm": 0.6915805339813232, "learning_rate": 2.6643704150526538e-05, "loss": 0.3648, "step": 26880 }, { "epoch": 10.151000377500944, "grad_norm": 0.7378256916999817, "learning_rate": 2.6607162566411716e-05, "loss": 0.3619, "step": 26890 }, { "epoch": 10.154775386938468, "grad_norm": 0.750008761882782, "learning_rate": 2.6570636971080697e-05, "loss": 0.3755, "step": 26900 }, { "epoch": 10.158550396375992, "grad_norm": 0.6996826529502869, "learning_rate": 2.6534127389498364e-05, "loss": 0.3635, "step": 26910 }, { "epoch": 10.162325405813515, "grad_norm": 0.6714365482330322, "learning_rate": 2.6497633846618696e-05, "loss": 0.3672, "step": 26920 }, { "epoch": 10.166100415251037, "grad_norm": 3.8461427688598633, "learning_rate": 2.6461156367384677e-05, "loss": 0.3705, "step": 26930 }, { "epoch": 10.169875424688561, "grad_norm": 0.6795220375061035, "learning_rate": 2.6424694976728316e-05, "loss": 0.3727, "step": 26940 }, { "epoch": 10.173650434126085, "grad_norm": 0.7694344520568848, "learning_rate": 2.6388249699570667e-05, "loss": 0.3705, "step": 26950 }, { "epoch": 10.177425443563608, "grad_norm": 0.6592543721199036, "learning_rate": 2.6351820560821672e-05, "loss": 0.3717, "step": 26960 }, { "epoch": 10.181200453001132, "grad_norm": 0.7184135317802429, "learning_rate": 2.631540758538034e-05, "loss": 0.3713, "step": 26970 }, { "epoch": 10.184975462438656, "grad_norm": 0.6492671966552734, "learning_rate": 2.6279010798134597e-05, "loss": 0.3637, "step": 26980 }, { "epoch": 10.18875047187618, "grad_norm": 0.6501529812812805, "learning_rate": 2.6242630223961305e-05, "loss": 0.3774, "step": 26990 }, { "epoch": 10.192525481313703, "grad_norm": 0.8947603106498718, "learning_rate": 2.6206265887726244e-05, "loss": 0.369, "step": 27000 }, { "epoch": 10.196300490751227, "grad_norm": 0.7419344186782837, "learning_rate": 2.6169917814284066e-05, "loss": 0.3672, "step": 27010 }, { "epoch": 10.20007550018875, "grad_norm": 0.703224241733551, "learning_rate": 2.6133586028478364e-05, "loss": 0.3683, "step": 27020 }, { "epoch": 10.203850509626275, "grad_norm": 0.6347489356994629, "learning_rate": 2.609727055514155e-05, "loss": 0.367, "step": 27030 }, { "epoch": 10.207625519063798, "grad_norm": 0.6552822589874268, "learning_rate": 2.606097141909494e-05, "loss": 0.3647, "step": 27040 }, { "epoch": 10.211400528501322, "grad_norm": 0.6825968623161316, "learning_rate": 2.6024688645148644e-05, "loss": 0.3713, "step": 27050 }, { "epoch": 10.215175537938844, "grad_norm": 0.7292717695236206, "learning_rate": 2.5988422258101564e-05, "loss": 0.3673, "step": 27060 }, { "epoch": 10.218950547376368, "grad_norm": 1.0935133695602417, "learning_rate": 2.5952172282741453e-05, "loss": 0.3688, "step": 27070 }, { "epoch": 10.222725556813892, "grad_norm": 0.6263819932937622, "learning_rate": 2.5915938743844853e-05, "loss": 0.3764, "step": 27080 }, { "epoch": 10.226500566251415, "grad_norm": 0.6874544024467468, "learning_rate": 2.5879721666177003e-05, "loss": 0.3675, "step": 27090 }, { "epoch": 10.230275575688939, "grad_norm": 0.640953004360199, "learning_rate": 2.5843521074491972e-05, "loss": 0.3691, "step": 27100 }, { "epoch": 10.234050585126463, "grad_norm": 0.6374151110649109, "learning_rate": 2.5807336993532487e-05, "loss": 0.3623, "step": 27110 }, { "epoch": 10.237825594563986, "grad_norm": 0.6693562269210815, "learning_rate": 2.577116944803004e-05, "loss": 0.3702, "step": 27120 }, { "epoch": 10.24160060400151, "grad_norm": 0.6881791949272156, "learning_rate": 2.5735018462704818e-05, "loss": 0.3695, "step": 27130 }, { "epoch": 10.245375613439034, "grad_norm": 0.6473117470741272, "learning_rate": 2.5698884062265665e-05, "loss": 0.3652, "step": 27140 }, { "epoch": 10.249150622876558, "grad_norm": 0.7239118218421936, "learning_rate": 2.5662766271410134e-05, "loss": 0.3715, "step": 27150 }, { "epoch": 10.252925632314081, "grad_norm": 0.6931194067001343, "learning_rate": 2.5626665114824343e-05, "loss": 0.3679, "step": 27160 }, { "epoch": 10.256700641751605, "grad_norm": 0.7126404643058777, "learning_rate": 2.5590580617183148e-05, "loss": 0.3729, "step": 27170 }, { "epoch": 10.260475651189129, "grad_norm": 0.612730085849762, "learning_rate": 2.5554512803149912e-05, "loss": 0.3626, "step": 27180 }, { "epoch": 10.26425066062665, "grad_norm": 0.631647527217865, "learning_rate": 2.5518461697376662e-05, "loss": 0.3656, "step": 27190 }, { "epoch": 10.268025670064175, "grad_norm": 0.692253053188324, "learning_rate": 2.548242732450402e-05, "loss": 0.3658, "step": 27200 }, { "epoch": 10.271800679501698, "grad_norm": 0.8286940455436707, "learning_rate": 2.5446409709161095e-05, "loss": 0.3708, "step": 27210 }, { "epoch": 10.275575688939222, "grad_norm": 1.055858850479126, "learning_rate": 2.541040887596561e-05, "loss": 0.362, "step": 27220 }, { "epoch": 10.279350698376746, "grad_norm": 0.6639923453330994, "learning_rate": 2.537442484952378e-05, "loss": 0.3684, "step": 27230 }, { "epoch": 10.28312570781427, "grad_norm": 0.6410587430000305, "learning_rate": 2.533845765443037e-05, "loss": 0.3736, "step": 27240 }, { "epoch": 10.286900717251793, "grad_norm": 0.6842541694641113, "learning_rate": 2.530250731526863e-05, "loss": 0.3624, "step": 27250 }, { "epoch": 10.290675726689317, "grad_norm": 0.6937667727470398, "learning_rate": 2.5266573856610253e-05, "loss": 0.3588, "step": 27260 }, { "epoch": 10.29445073612684, "grad_norm": 0.6150068044662476, "learning_rate": 2.5230657303015403e-05, "loss": 0.3604, "step": 27270 }, { "epoch": 10.298225745564364, "grad_norm": 0.6743102669715881, "learning_rate": 2.5194757679032728e-05, "loss": 0.3679, "step": 27280 }, { "epoch": 10.302000755001888, "grad_norm": 0.6566803455352783, "learning_rate": 2.5158875009199278e-05, "loss": 0.3657, "step": 27290 }, { "epoch": 10.305775764439412, "grad_norm": 0.6627777218818665, "learning_rate": 2.5123009318040537e-05, "loss": 0.3677, "step": 27300 }, { "epoch": 10.309550773876934, "grad_norm": 0.7576314806938171, "learning_rate": 2.508716063007034e-05, "loss": 0.3617, "step": 27310 }, { "epoch": 10.313325783314458, "grad_norm": 1.1635149717330933, "learning_rate": 2.5051328969790934e-05, "loss": 0.367, "step": 27320 }, { "epoch": 10.317100792751981, "grad_norm": 0.6241893768310547, "learning_rate": 2.501551436169292e-05, "loss": 0.3721, "step": 27330 }, { "epoch": 10.320875802189505, "grad_norm": 0.5901798605918884, "learning_rate": 2.4979716830255255e-05, "loss": 0.3629, "step": 27340 }, { "epoch": 10.324650811627029, "grad_norm": 0.6239721179008484, "learning_rate": 2.4943936399945233e-05, "loss": 0.3647, "step": 27350 }, { "epoch": 10.328425821064553, "grad_norm": 0.6076030135154724, "learning_rate": 2.4908173095218412e-05, "loss": 0.3602, "step": 27360 }, { "epoch": 10.332200830502076, "grad_norm": 0.6556723117828369, "learning_rate": 2.4872426940518663e-05, "loss": 0.369, "step": 27370 }, { "epoch": 10.3359758399396, "grad_norm": 0.6872203350067139, "learning_rate": 2.4836697960278156e-05, "loss": 0.3669, "step": 27380 }, { "epoch": 10.339750849377124, "grad_norm": 0.6859211921691895, "learning_rate": 2.480098617891732e-05, "loss": 0.3678, "step": 27390 }, { "epoch": 10.343525858814647, "grad_norm": 0.6686061024665833, "learning_rate": 2.4765291620844837e-05, "loss": 0.3558, "step": 27400 }, { "epoch": 10.347300868252171, "grad_norm": 0.6350464820861816, "learning_rate": 2.472961431045756e-05, "loss": 0.3808, "step": 27410 }, { "epoch": 10.351075877689695, "grad_norm": 0.6436896324157715, "learning_rate": 2.4693954272140622e-05, "loss": 0.3644, "step": 27420 }, { "epoch": 10.354850887127217, "grad_norm": 0.7389644980430603, "learning_rate": 2.4658311530267315e-05, "loss": 0.368, "step": 27430 }, { "epoch": 10.35862589656474, "grad_norm": 0.7080246210098267, "learning_rate": 2.4622686109199124e-05, "loss": 0.368, "step": 27440 }, { "epoch": 10.362400906002264, "grad_norm": 0.6405688524246216, "learning_rate": 2.4587078033285695e-05, "loss": 0.3731, "step": 27450 }, { "epoch": 10.366175915439788, "grad_norm": 0.6456767320632935, "learning_rate": 2.45514873268648e-05, "loss": 0.3615, "step": 27460 }, { "epoch": 10.369950924877312, "grad_norm": 0.6688347458839417, "learning_rate": 2.4515914014262336e-05, "loss": 0.3618, "step": 27470 }, { "epoch": 10.373725934314836, "grad_norm": 0.7001455426216125, "learning_rate": 2.4480358119792345e-05, "loss": 0.3669, "step": 27480 }, { "epoch": 10.37750094375236, "grad_norm": 0.6570714712142944, "learning_rate": 2.4444819667756942e-05, "loss": 0.3574, "step": 27490 }, { "epoch": 10.381275953189883, "grad_norm": 0.6306593418121338, "learning_rate": 2.4409298682446346e-05, "loss": 0.3646, "step": 27500 }, { "epoch": 10.385050962627407, "grad_norm": 0.6927255988121033, "learning_rate": 2.437379518813877e-05, "loss": 0.3671, "step": 27510 }, { "epoch": 10.38882597206493, "grad_norm": 0.7374937534332275, "learning_rate": 2.4338309209100547e-05, "loss": 0.3626, "step": 27520 }, { "epoch": 10.392600981502454, "grad_norm": 1.0554702281951904, "learning_rate": 2.4302840769586004e-05, "loss": 0.3635, "step": 27530 }, { "epoch": 10.396375990939978, "grad_norm": 0.6217535734176636, "learning_rate": 2.42673898938375e-05, "loss": 0.3728, "step": 27540 }, { "epoch": 10.400151000377502, "grad_norm": 0.654321551322937, "learning_rate": 2.4231956606085343e-05, "loss": 0.3616, "step": 27550 }, { "epoch": 10.403926009815024, "grad_norm": 0.7654052972793579, "learning_rate": 2.419654093054789e-05, "loss": 0.3723, "step": 27560 }, { "epoch": 10.407701019252547, "grad_norm": 0.6208662390708923, "learning_rate": 2.4161142891431375e-05, "loss": 0.3639, "step": 27570 }, { "epoch": 10.411476028690071, "grad_norm": 0.7695934772491455, "learning_rate": 2.412576251293005e-05, "loss": 0.3657, "step": 27580 }, { "epoch": 10.415251038127595, "grad_norm": 0.6474165916442871, "learning_rate": 2.4090399819226068e-05, "loss": 0.3664, "step": 27590 }, { "epoch": 10.419026047565119, "grad_norm": 0.716998815536499, "learning_rate": 2.4055054834489514e-05, "loss": 0.3695, "step": 27600 }, { "epoch": 10.422801057002642, "grad_norm": 0.6260504722595215, "learning_rate": 2.401972758287832e-05, "loss": 0.3659, "step": 27610 }, { "epoch": 10.426576066440166, "grad_norm": 0.668073296546936, "learning_rate": 2.398441808853834e-05, "loss": 0.3555, "step": 27620 }, { "epoch": 10.43035107587769, "grad_norm": 0.6870536208152771, "learning_rate": 2.3949126375603288e-05, "loss": 0.3675, "step": 27630 }, { "epoch": 10.434126085315214, "grad_norm": 0.6690881848335266, "learning_rate": 2.3913852468194724e-05, "loss": 0.3678, "step": 27640 }, { "epoch": 10.437901094752737, "grad_norm": 0.6636070013046265, "learning_rate": 2.387859639042201e-05, "loss": 0.3693, "step": 27650 }, { "epoch": 10.441676104190261, "grad_norm": 0.7105987668037415, "learning_rate": 2.3843358166382368e-05, "loss": 0.3618, "step": 27660 }, { "epoch": 10.445451113627785, "grad_norm": 0.6671977639198303, "learning_rate": 2.3808137820160757e-05, "loss": 0.3639, "step": 27670 }, { "epoch": 10.449226123065309, "grad_norm": 0.6227991580963135, "learning_rate": 2.3772935375829975e-05, "loss": 0.3587, "step": 27680 }, { "epoch": 10.45300113250283, "grad_norm": 0.5635626316070557, "learning_rate": 2.3737750857450553e-05, "loss": 0.3723, "step": 27690 }, { "epoch": 10.456776141940354, "grad_norm": 0.6364959478378296, "learning_rate": 2.3702584289070805e-05, "loss": 0.3624, "step": 27700 }, { "epoch": 10.460551151377878, "grad_norm": 0.6209636330604553, "learning_rate": 2.36674356947267e-05, "loss": 0.3678, "step": 27710 }, { "epoch": 10.464326160815402, "grad_norm": 0.6616617441177368, "learning_rate": 2.3632305098442004e-05, "loss": 0.3583, "step": 27720 }, { "epoch": 10.468101170252925, "grad_norm": 0.7123914361000061, "learning_rate": 2.3597192524228156e-05, "loss": 0.3636, "step": 27730 }, { "epoch": 10.47187617969045, "grad_norm": 0.6567232012748718, "learning_rate": 2.356209799608424e-05, "loss": 0.3654, "step": 27740 }, { "epoch": 10.475651189127973, "grad_norm": 0.6329348087310791, "learning_rate": 2.352702153799704e-05, "loss": 0.3665, "step": 27750 }, { "epoch": 10.479426198565497, "grad_norm": 0.6293632984161377, "learning_rate": 2.3491963173941018e-05, "loss": 0.3626, "step": 27760 }, { "epoch": 10.48320120800302, "grad_norm": 0.652357816696167, "learning_rate": 2.3456922927878196e-05, "loss": 0.3759, "step": 27770 }, { "epoch": 10.486976217440544, "grad_norm": 0.6584979295730591, "learning_rate": 2.3421900823758257e-05, "loss": 0.3652, "step": 27780 }, { "epoch": 10.490751226878068, "grad_norm": 0.6862519383430481, "learning_rate": 2.3386896885518496e-05, "loss": 0.3648, "step": 27790 }, { "epoch": 10.494526236315592, "grad_norm": 0.6221929788589478, "learning_rate": 2.335191113708378e-05, "loss": 0.3721, "step": 27800 }, { "epoch": 10.498301245753115, "grad_norm": 0.6476244330406189, "learning_rate": 2.331694360236651e-05, "loss": 0.3746, "step": 27810 }, { "epoch": 10.502076255190637, "grad_norm": 0.626964807510376, "learning_rate": 2.3281994305266702e-05, "loss": 0.3616, "step": 27820 }, { "epoch": 10.505851264628161, "grad_norm": 0.7498915791511536, "learning_rate": 2.3247063269671826e-05, "loss": 0.3689, "step": 27830 }, { "epoch": 10.509626274065685, "grad_norm": 0.6770737767219543, "learning_rate": 2.321215051945695e-05, "loss": 0.3672, "step": 27840 }, { "epoch": 10.513401283503208, "grad_norm": 0.7251414656639099, "learning_rate": 2.3177256078484588e-05, "loss": 0.3664, "step": 27850 }, { "epoch": 10.517176292940732, "grad_norm": 0.7046555876731873, "learning_rate": 2.3142379970604798e-05, "loss": 0.361, "step": 27860 }, { "epoch": 10.520951302378256, "grad_norm": 0.6266787648200989, "learning_rate": 2.3107522219655025e-05, "loss": 0.3649, "step": 27870 }, { "epoch": 10.52472631181578, "grad_norm": 0.6274697780609131, "learning_rate": 2.3072682849460236e-05, "loss": 0.3601, "step": 27880 }, { "epoch": 10.528501321253303, "grad_norm": 0.6082983613014221, "learning_rate": 2.303786188383281e-05, "loss": 0.3688, "step": 27890 }, { "epoch": 10.532276330690827, "grad_norm": 0.6668601632118225, "learning_rate": 2.300305934657257e-05, "loss": 0.361, "step": 27900 }, { "epoch": 10.53605134012835, "grad_norm": 0.6825070977210999, "learning_rate": 2.2968275261466677e-05, "loss": 0.3637, "step": 27910 }, { "epoch": 10.539826349565875, "grad_norm": 0.6331124901771545, "learning_rate": 2.293350965228977e-05, "loss": 0.3639, "step": 27920 }, { "epoch": 10.543601359003397, "grad_norm": 0.6340346932411194, "learning_rate": 2.2898762542803776e-05, "loss": 0.3625, "step": 27930 }, { "epoch": 10.54737636844092, "grad_norm": 0.7296791672706604, "learning_rate": 2.286403395675803e-05, "loss": 0.3667, "step": 27940 }, { "epoch": 10.551151377878444, "grad_norm": 0.6600122451782227, "learning_rate": 2.28293239178892e-05, "loss": 0.3671, "step": 27950 }, { "epoch": 10.554926387315968, "grad_norm": 0.7184299230575562, "learning_rate": 2.2794632449921287e-05, "loss": 0.3702, "step": 27960 }, { "epoch": 10.558701396753492, "grad_norm": 0.6096726655960083, "learning_rate": 2.275995957656555e-05, "loss": 0.3645, "step": 27970 }, { "epoch": 10.562476406191015, "grad_norm": 0.6670506000518799, "learning_rate": 2.272530532152058e-05, "loss": 0.3646, "step": 27980 }, { "epoch": 10.566251415628539, "grad_norm": 0.6431169509887695, "learning_rate": 2.2690669708472233e-05, "loss": 0.361, "step": 27990 }, { "epoch": 10.570026425066063, "grad_norm": 0.6406874060630798, "learning_rate": 2.2656052761093655e-05, "loss": 0.3677, "step": 28000 }, { "epoch": 10.573801434503586, "grad_norm": 0.6442010402679443, "learning_rate": 2.262145450304517e-05, "loss": 0.3661, "step": 28010 }, { "epoch": 10.57757644394111, "grad_norm": 0.6408098340034485, "learning_rate": 2.2586874957974352e-05, "loss": 0.3727, "step": 28020 }, { "epoch": 10.581351453378634, "grad_norm": 0.648128092288971, "learning_rate": 2.2552314149516012e-05, "loss": 0.3656, "step": 28030 }, { "epoch": 10.585126462816158, "grad_norm": 0.7829309105873108, "learning_rate": 2.2517772101292133e-05, "loss": 0.3665, "step": 28040 }, { "epoch": 10.588901472253681, "grad_norm": 0.768104612827301, "learning_rate": 2.248324883691188e-05, "loss": 0.3687, "step": 28050 }, { "epoch": 10.592676481691203, "grad_norm": 0.631535530090332, "learning_rate": 2.24487443799716e-05, "loss": 0.3683, "step": 28060 }, { "epoch": 10.596451491128727, "grad_norm": 0.7682189345359802, "learning_rate": 2.241425875405472e-05, "loss": 0.3705, "step": 28070 }, { "epoch": 10.60022650056625, "grad_norm": 0.6943038702011108, "learning_rate": 2.2379791982731868e-05, "loss": 0.4653, "step": 28080 }, { "epoch": 10.604001510003775, "grad_norm": 0.634013295173645, "learning_rate": 2.2345344089560756e-05, "loss": 0.3621, "step": 28090 }, { "epoch": 10.607776519441298, "grad_norm": 0.7249836325645447, "learning_rate": 2.2310915098086206e-05, "loss": 0.3655, "step": 28100 }, { "epoch": 10.611551528878822, "grad_norm": 0.8200324773788452, "learning_rate": 2.227650503184009e-05, "loss": 0.373, "step": 28110 }, { "epoch": 10.615326538316346, "grad_norm": 0.8548815250396729, "learning_rate": 2.2242113914341357e-05, "loss": 0.3581, "step": 28120 }, { "epoch": 10.61910154775387, "grad_norm": 0.6571340560913086, "learning_rate": 2.220774176909602e-05, "loss": 0.3629, "step": 28130 }, { "epoch": 10.622876557191393, "grad_norm": 0.6676787734031677, "learning_rate": 2.2173388619597114e-05, "loss": 0.3682, "step": 28140 }, { "epoch": 10.626651566628917, "grad_norm": 0.7401626706123352, "learning_rate": 2.21390544893247e-05, "loss": 0.3667, "step": 28150 }, { "epoch": 10.63042657606644, "grad_norm": 0.6297852396965027, "learning_rate": 2.210473940174585e-05, "loss": 0.3739, "step": 28160 }, { "epoch": 10.634201585503964, "grad_norm": 0.6574814319610596, "learning_rate": 2.207044338031456e-05, "loss": 0.3613, "step": 28170 }, { "epoch": 10.637976594941488, "grad_norm": 0.5910388827323914, "learning_rate": 2.203616644847186e-05, "loss": 0.3778, "step": 28180 }, { "epoch": 10.64175160437901, "grad_norm": 0.6458864808082581, "learning_rate": 2.200190862964571e-05, "loss": 0.3678, "step": 28190 }, { "epoch": 10.645526613816534, "grad_norm": 0.7278522849082947, "learning_rate": 2.1967669947251024e-05, "loss": 0.364, "step": 28200 }, { "epoch": 10.649301623254058, "grad_norm": 0.6762776970863342, "learning_rate": 2.1933450424689583e-05, "loss": 0.3594, "step": 28210 }, { "epoch": 10.653076632691581, "grad_norm": 1.226162075996399, "learning_rate": 2.1899250085350142e-05, "loss": 0.3733, "step": 28220 }, { "epoch": 10.656851642129105, "grad_norm": 0.5935925841331482, "learning_rate": 2.1865068952608277e-05, "loss": 0.3628, "step": 28230 }, { "epoch": 10.660626651566629, "grad_norm": 0.7492917776107788, "learning_rate": 2.1830907049826487e-05, "loss": 0.3611, "step": 28240 }, { "epoch": 10.664401661004153, "grad_norm": 0.7051345705986023, "learning_rate": 2.179676440035411e-05, "loss": 0.3602, "step": 28250 }, { "epoch": 10.668176670441676, "grad_norm": 0.6424268484115601, "learning_rate": 2.1762641027527337e-05, "loss": 0.366, "step": 28260 }, { "epoch": 10.6719516798792, "grad_norm": 0.8368542790412903, "learning_rate": 2.1728536954669143e-05, "loss": 0.3646, "step": 28270 }, { "epoch": 10.675726689316724, "grad_norm": 0.7216016054153442, "learning_rate": 2.169445220508936e-05, "loss": 0.36, "step": 28280 }, { "epoch": 10.679501698754247, "grad_norm": 0.6066749691963196, "learning_rate": 2.166038680208461e-05, "loss": 0.3587, "step": 28290 }, { "epoch": 10.683276708191771, "grad_norm": 0.5810279250144958, "learning_rate": 2.162634076893823e-05, "loss": 0.3697, "step": 28300 }, { "epoch": 10.687051717629295, "grad_norm": 0.6657140254974365, "learning_rate": 2.1592314128920388e-05, "loss": 0.3678, "step": 28310 }, { "epoch": 10.690826727066817, "grad_norm": 0.636150598526001, "learning_rate": 2.155830690528799e-05, "loss": 0.3654, "step": 28320 }, { "epoch": 10.69460173650434, "grad_norm": 0.6530934572219849, "learning_rate": 2.1524319121284613e-05, "loss": 0.3573, "step": 28330 }, { "epoch": 10.698376745941864, "grad_norm": 0.6528118848800659, "learning_rate": 2.1490350800140607e-05, "loss": 0.3712, "step": 28340 }, { "epoch": 10.702151755379388, "grad_norm": 0.6614232659339905, "learning_rate": 2.1456401965073002e-05, "loss": 0.3647, "step": 28350 }, { "epoch": 10.705926764816912, "grad_norm": 0.6931934356689453, "learning_rate": 2.1422472639285524e-05, "loss": 0.3641, "step": 28360 }, { "epoch": 10.709701774254436, "grad_norm": 0.6107961535453796, "learning_rate": 2.13885628459685e-05, "loss": 0.3576, "step": 28370 }, { "epoch": 10.71347678369196, "grad_norm": 0.6835388541221619, "learning_rate": 2.135467260829901e-05, "loss": 0.364, "step": 28380 }, { "epoch": 10.717251793129483, "grad_norm": 0.655625581741333, "learning_rate": 2.1320801949440654e-05, "loss": 0.3629, "step": 28390 }, { "epoch": 10.721026802567007, "grad_norm": 0.7267212867736816, "learning_rate": 2.1286950892543744e-05, "loss": 0.3641, "step": 28400 }, { "epoch": 10.72480181200453, "grad_norm": 0.6693907380104065, "learning_rate": 2.125311946074515e-05, "loss": 0.3657, "step": 28410 }, { "epoch": 10.728576821442054, "grad_norm": 0.6558740735054016, "learning_rate": 2.1219307677168355e-05, "loss": 0.3646, "step": 28420 }, { "epoch": 10.732351830879578, "grad_norm": 0.7080862522125244, "learning_rate": 2.118551556492336e-05, "loss": 0.361, "step": 28430 }, { "epoch": 10.7361268403171, "grad_norm": 0.6530542969703674, "learning_rate": 2.1151743147106774e-05, "loss": 0.3648, "step": 28440 }, { "epoch": 10.739901849754624, "grad_norm": 0.705801784992218, "learning_rate": 2.111799044680172e-05, "loss": 0.3622, "step": 28450 }, { "epoch": 10.743676859192147, "grad_norm": 0.6428266763687134, "learning_rate": 2.1084257487077873e-05, "loss": 0.3669, "step": 28460 }, { "epoch": 10.747451868629671, "grad_norm": 0.7444311380386353, "learning_rate": 2.1050544290991357e-05, "loss": 0.3596, "step": 28470 }, { "epoch": 10.751226878067195, "grad_norm": 0.6238864660263062, "learning_rate": 2.101685088158486e-05, "loss": 0.3694, "step": 28480 }, { "epoch": 10.755001887504719, "grad_norm": 0.681848406791687, "learning_rate": 2.0983177281887472e-05, "loss": 0.369, "step": 28490 }, { "epoch": 10.758776896942242, "grad_norm": 0.6995017528533936, "learning_rate": 2.0949523514914798e-05, "loss": 0.3644, "step": 28500 }, { "epoch": 10.762551906379766, "grad_norm": 0.6024167537689209, "learning_rate": 2.0915889603668876e-05, "loss": 0.357, "step": 28510 }, { "epoch": 10.76632691581729, "grad_norm": 0.6610404849052429, "learning_rate": 2.0882275571138175e-05, "loss": 0.3634, "step": 28520 }, { "epoch": 10.770101925254814, "grad_norm": 0.629758894443512, "learning_rate": 2.0848681440297545e-05, "loss": 0.3644, "step": 28530 }, { "epoch": 10.773876934692337, "grad_norm": 0.6765365600585938, "learning_rate": 2.081510723410827e-05, "loss": 0.3606, "step": 28540 }, { "epoch": 10.777651944129861, "grad_norm": 0.7556784152984619, "learning_rate": 2.0781552975518003e-05, "loss": 0.3672, "step": 28550 }, { "epoch": 10.781426953567383, "grad_norm": 0.7298506498336792, "learning_rate": 2.074801868746078e-05, "loss": 0.368, "step": 28560 }, { "epoch": 10.785201963004907, "grad_norm": 0.7526068091392517, "learning_rate": 2.0714504392856955e-05, "loss": 0.368, "step": 28570 }, { "epoch": 10.78897697244243, "grad_norm": 0.6001387238502502, "learning_rate": 2.0681010114613215e-05, "loss": 0.3641, "step": 28580 }, { "epoch": 10.792751981879954, "grad_norm": 0.662828803062439, "learning_rate": 2.0647535875622597e-05, "loss": 0.362, "step": 28590 }, { "epoch": 10.796526991317478, "grad_norm": 2.932863712310791, "learning_rate": 2.0614081698764432e-05, "loss": 0.3642, "step": 28600 }, { "epoch": 10.800302000755002, "grad_norm": 0.7236543297767639, "learning_rate": 2.0580647606904334e-05, "loss": 0.3654, "step": 28610 }, { "epoch": 10.804077010192525, "grad_norm": 0.6587764620780945, "learning_rate": 2.0547233622894208e-05, "loss": 0.3629, "step": 28620 }, { "epoch": 10.80785201963005, "grad_norm": 0.6333833932876587, "learning_rate": 2.0513839769572157e-05, "loss": 0.3685, "step": 28630 }, { "epoch": 10.811627029067573, "grad_norm": 0.596711277961731, "learning_rate": 2.0480466069762584e-05, "loss": 0.3737, "step": 28640 }, { "epoch": 10.815402038505097, "grad_norm": 0.7744527459144592, "learning_rate": 2.0447112546276104e-05, "loss": 0.3628, "step": 28650 }, { "epoch": 10.81917704794262, "grad_norm": 0.6719459295272827, "learning_rate": 2.0413779221909547e-05, "loss": 0.3714, "step": 28660 }, { "epoch": 10.822952057380144, "grad_norm": 0.6692177653312683, "learning_rate": 2.0380466119445912e-05, "loss": 0.3652, "step": 28670 }, { "epoch": 10.826727066817668, "grad_norm": 0.6114712953567505, "learning_rate": 2.0347173261654373e-05, "loss": 0.367, "step": 28680 }, { "epoch": 10.83050207625519, "grad_norm": 0.6642469763755798, "learning_rate": 2.03139006712903e-05, "loss": 0.3674, "step": 28690 }, { "epoch": 10.834277085692714, "grad_norm": 0.6511169075965881, "learning_rate": 2.028064837109519e-05, "loss": 0.3648, "step": 28700 }, { "epoch": 10.838052095130237, "grad_norm": 0.7235956788063049, "learning_rate": 2.0247416383796685e-05, "loss": 0.3712, "step": 28710 }, { "epoch": 10.841827104567761, "grad_norm": 0.6243189573287964, "learning_rate": 2.0214204732108548e-05, "loss": 0.3573, "step": 28720 }, { "epoch": 10.845602114005285, "grad_norm": 0.6611545085906982, "learning_rate": 2.0181013438730596e-05, "loss": 0.3603, "step": 28730 }, { "epoch": 10.849377123442808, "grad_norm": 0.6944743990898132, "learning_rate": 2.0147842526348783e-05, "loss": 0.3589, "step": 28740 }, { "epoch": 10.853152132880332, "grad_norm": 0.6612570881843567, "learning_rate": 2.011469201763511e-05, "loss": 0.3635, "step": 28750 }, { "epoch": 10.856927142317856, "grad_norm": 0.6813220977783203, "learning_rate": 2.0081561935247665e-05, "loss": 0.3619, "step": 28760 }, { "epoch": 10.86070215175538, "grad_norm": 0.638427734375, "learning_rate": 2.0048452301830523e-05, "loss": 0.3605, "step": 28770 }, { "epoch": 10.864477161192903, "grad_norm": 0.607269823551178, "learning_rate": 2.0015363140013788e-05, "loss": 0.3669, "step": 28780 }, { "epoch": 10.868252170630427, "grad_norm": 0.6158018708229065, "learning_rate": 1.9982294472413606e-05, "loss": 0.3642, "step": 28790 }, { "epoch": 10.87202718006795, "grad_norm": 0.650993287563324, "learning_rate": 1.9949246321632103e-05, "loss": 0.3563, "step": 28800 }, { "epoch": 10.875802189505475, "grad_norm": 0.6607373952865601, "learning_rate": 1.9916218710257377e-05, "loss": 0.3608, "step": 28810 }, { "epoch": 10.879577198942997, "grad_norm": 0.7220985889434814, "learning_rate": 1.988321166086351e-05, "loss": 0.3646, "step": 28820 }, { "epoch": 10.88335220838052, "grad_norm": 0.7125725150108337, "learning_rate": 1.9850225196010468e-05, "loss": 0.3605, "step": 28830 }, { "epoch": 10.887127217818044, "grad_norm": 0.6737269759178162, "learning_rate": 1.981725933824421e-05, "loss": 0.363, "step": 28840 }, { "epoch": 10.890902227255568, "grad_norm": 0.6922343373298645, "learning_rate": 1.978431411009661e-05, "loss": 0.3687, "step": 28850 }, { "epoch": 10.894677236693092, "grad_norm": 4.779869079589844, "learning_rate": 1.9751389534085375e-05, "loss": 0.3591, "step": 28860 }, { "epoch": 10.898452246130615, "grad_norm": 0.6346884369850159, "learning_rate": 1.9718485632714184e-05, "loss": 0.364, "step": 28870 }, { "epoch": 10.902227255568139, "grad_norm": 0.6880037188529968, "learning_rate": 1.968560242847251e-05, "loss": 0.3693, "step": 28880 }, { "epoch": 10.906002265005663, "grad_norm": 0.7235897183418274, "learning_rate": 1.965273994383573e-05, "loss": 0.3605, "step": 28890 }, { "epoch": 10.909777274443186, "grad_norm": 0.6026926636695862, "learning_rate": 1.961989820126504e-05, "loss": 0.3659, "step": 28900 }, { "epoch": 10.91355228388071, "grad_norm": 0.6524337530136108, "learning_rate": 1.958707722320746e-05, "loss": 0.3559, "step": 28910 }, { "epoch": 10.917327293318234, "grad_norm": 0.6593090295791626, "learning_rate": 1.955427703209584e-05, "loss": 0.366, "step": 28920 }, { "epoch": 10.921102302755758, "grad_norm": 0.6755902767181396, "learning_rate": 1.9521497650348764e-05, "loss": 0.3599, "step": 28930 }, { "epoch": 10.924877312193281, "grad_norm": 0.7313739061355591, "learning_rate": 1.948873910037067e-05, "loss": 0.363, "step": 28940 }, { "epoch": 10.928652321630803, "grad_norm": 0.686374306678772, "learning_rate": 1.9456001404551678e-05, "loss": 0.377, "step": 28950 }, { "epoch": 10.932427331068327, "grad_norm": 0.789500892162323, "learning_rate": 1.942328458526771e-05, "loss": 0.3672, "step": 28960 }, { "epoch": 10.93620234050585, "grad_norm": 0.6305020451545715, "learning_rate": 1.9390588664880427e-05, "loss": 0.3595, "step": 28970 }, { "epoch": 10.939977349943375, "grad_norm": 0.6587271690368652, "learning_rate": 1.9357913665737145e-05, "loss": 0.3596, "step": 28980 }, { "epoch": 10.943752359380898, "grad_norm": 0.6776351928710938, "learning_rate": 1.932525961017093e-05, "loss": 0.3646, "step": 28990 }, { "epoch": 10.947527368818422, "grad_norm": 0.6554774641990662, "learning_rate": 1.9292626520500533e-05, "loss": 0.3653, "step": 29000 }, { "epoch": 10.951302378255946, "grad_norm": 0.702714204788208, "learning_rate": 1.9260014419030354e-05, "loss": 0.3693, "step": 29010 }, { "epoch": 10.95507738769347, "grad_norm": 0.692285418510437, "learning_rate": 1.9227423328050475e-05, "loss": 0.3628, "step": 29020 }, { "epoch": 10.958852397130993, "grad_norm": 0.6136350631713867, "learning_rate": 1.9194853269836582e-05, "loss": 0.3628, "step": 29030 }, { "epoch": 10.962627406568517, "grad_norm": 0.8171948790550232, "learning_rate": 1.916230426664999e-05, "loss": 0.3799, "step": 29040 }, { "epoch": 10.96640241600604, "grad_norm": 0.6025060415267944, "learning_rate": 1.912977634073765e-05, "loss": 0.3664, "step": 29050 }, { "epoch": 10.970177425443563, "grad_norm": 0.678777277469635, "learning_rate": 1.9097269514332083e-05, "loss": 0.3663, "step": 29060 }, { "epoch": 10.973952434881086, "grad_norm": 0.6317446827888489, "learning_rate": 1.9064783809651433e-05, "loss": 0.3644, "step": 29070 }, { "epoch": 10.97772744431861, "grad_norm": 0.6572142243385315, "learning_rate": 1.9032319248899333e-05, "loss": 0.3561, "step": 29080 }, { "epoch": 10.981502453756134, "grad_norm": 0.5952383279800415, "learning_rate": 1.8999875854265015e-05, "loss": 0.3578, "step": 29090 }, { "epoch": 10.985277463193658, "grad_norm": 0.6633411645889282, "learning_rate": 1.8967453647923232e-05, "loss": 0.362, "step": 29100 }, { "epoch": 10.989052472631181, "grad_norm": 0.6560747027397156, "learning_rate": 1.893505265203427e-05, "loss": 0.3602, "step": 29110 }, { "epoch": 10.992827482068705, "grad_norm": 0.6242331862449646, "learning_rate": 1.8902672888743907e-05, "loss": 0.3589, "step": 29120 }, { "epoch": 10.996602491506229, "grad_norm": 0.6771829724311829, "learning_rate": 1.8870314380183396e-05, "loss": 0.3618, "step": 29130 }, { "epoch": 11.000377500943753, "grad_norm": 1.5967614650726318, "learning_rate": 1.8837977148469448e-05, "loss": 0.3566, "step": 29140 }, { "epoch": 11.004152510381276, "grad_norm": 0.6474411487579346, "learning_rate": 1.880566121570429e-05, "loss": 0.3596, "step": 29150 }, { "epoch": 11.0079275198188, "grad_norm": 0.6497718691825867, "learning_rate": 1.877336660397554e-05, "loss": 0.3564, "step": 29160 }, { "epoch": 11.011702529256324, "grad_norm": 0.6360942125320435, "learning_rate": 1.874109333535628e-05, "loss": 0.3681, "step": 29170 }, { "epoch": 11.015477538693847, "grad_norm": 0.7708438634872437, "learning_rate": 1.870884143190496e-05, "loss": 0.3651, "step": 29180 }, { "epoch": 11.01925254813137, "grad_norm": 0.8585805892944336, "learning_rate": 1.867661091566546e-05, "loss": 0.3656, "step": 29190 }, { "epoch": 11.023027557568893, "grad_norm": 2.161421537399292, "learning_rate": 1.864440180866704e-05, "loss": 0.3588, "step": 29200 }, { "epoch": 11.026802567006417, "grad_norm": 0.6585597991943359, "learning_rate": 1.8612214132924317e-05, "loss": 0.3567, "step": 29210 }, { "epoch": 11.03057757644394, "grad_norm": 0.6775062084197998, "learning_rate": 1.858004791043728e-05, "loss": 0.3581, "step": 29220 }, { "epoch": 11.034352585881464, "grad_norm": 0.8132985234260559, "learning_rate": 1.854790316319123e-05, "loss": 0.3758, "step": 29230 }, { "epoch": 11.038127595318988, "grad_norm": 0.6738925576210022, "learning_rate": 1.8515779913156766e-05, "loss": 0.3583, "step": 29240 }, { "epoch": 11.041902604756512, "grad_norm": 0.6279427409172058, "learning_rate": 1.848367818228986e-05, "loss": 0.3697, "step": 29250 }, { "epoch": 11.045677614194036, "grad_norm": 0.6961294412612915, "learning_rate": 1.8451597992531733e-05, "loss": 0.3619, "step": 29260 }, { "epoch": 11.04945262363156, "grad_norm": 0.8088012337684631, "learning_rate": 1.8419539365808914e-05, "loss": 0.3755, "step": 29270 }, { "epoch": 11.053227633069083, "grad_norm": 0.6103382706642151, "learning_rate": 1.838750232403313e-05, "loss": 0.3592, "step": 29280 }, { "epoch": 11.057002642506607, "grad_norm": 0.6636168956756592, "learning_rate": 1.835548688910142e-05, "loss": 0.3726, "step": 29290 }, { "epoch": 11.06077765194413, "grad_norm": 0.6306051015853882, "learning_rate": 1.8323493082896037e-05, "loss": 0.3665, "step": 29300 }, { "epoch": 11.064552661381654, "grad_norm": 0.6422659754753113, "learning_rate": 1.8291520927284454e-05, "loss": 0.3643, "step": 29310 }, { "epoch": 11.068327670819176, "grad_norm": 0.6881163716316223, "learning_rate": 1.8259570444119305e-05, "loss": 0.363, "step": 29320 }, { "epoch": 11.0721026802567, "grad_norm": 0.5860868096351624, "learning_rate": 1.8227641655238488e-05, "loss": 0.3639, "step": 29330 }, { "epoch": 11.075877689694224, "grad_norm": 0.6115444302558899, "learning_rate": 1.819573458246498e-05, "loss": 0.3687, "step": 29340 }, { "epoch": 11.079652699131747, "grad_norm": 0.6907765865325928, "learning_rate": 1.816384924760699e-05, "loss": 0.367, "step": 29350 }, { "epoch": 11.083427708569271, "grad_norm": 0.7525956034660339, "learning_rate": 1.813198567245784e-05, "loss": 0.3615, "step": 29360 }, { "epoch": 11.087202718006795, "grad_norm": 0.625262975692749, "learning_rate": 1.8100143878796006e-05, "loss": 0.3565, "step": 29370 }, { "epoch": 11.090977727444319, "grad_norm": 0.6794587969779968, "learning_rate": 1.8068323888385015e-05, "loss": 0.363, "step": 29380 }, { "epoch": 11.094752736881842, "grad_norm": 0.6202002167701721, "learning_rate": 1.803652572297355e-05, "loss": 0.3639, "step": 29390 }, { "epoch": 11.098527746319366, "grad_norm": 0.679071307182312, "learning_rate": 1.8004749404295353e-05, "loss": 0.3658, "step": 29400 }, { "epoch": 11.10230275575689, "grad_norm": 0.6545715928077698, "learning_rate": 1.797299495406926e-05, "loss": 0.3692, "step": 29410 }, { "epoch": 11.106077765194414, "grad_norm": 0.6335902810096741, "learning_rate": 1.7941262393999103e-05, "loss": 0.3591, "step": 29420 }, { "epoch": 11.109852774631937, "grad_norm": 1.0289889574050903, "learning_rate": 1.7909551745773816e-05, "loss": 0.3531, "step": 29430 }, { "epoch": 11.113627784069461, "grad_norm": 0.6615308523178101, "learning_rate": 1.7877863031067304e-05, "loss": 0.3552, "step": 29440 }, { "epoch": 11.117402793506983, "grad_norm": 0.6820453405380249, "learning_rate": 1.7846196271538516e-05, "loss": 0.3688, "step": 29450 }, { "epoch": 11.121177802944507, "grad_norm": 0.6518820524215698, "learning_rate": 1.7814551488831384e-05, "loss": 0.3598, "step": 29460 }, { "epoch": 11.12495281238203, "grad_norm": 0.725908637046814, "learning_rate": 1.7782928704574835e-05, "loss": 0.3621, "step": 29470 }, { "epoch": 11.128727821819554, "grad_norm": 2.0189263820648193, "learning_rate": 1.775132794038271e-05, "loss": 0.3616, "step": 29480 }, { "epoch": 11.132502831257078, "grad_norm": 0.6810202598571777, "learning_rate": 1.7719749217853855e-05, "loss": 0.3604, "step": 29490 }, { "epoch": 11.136277840694602, "grad_norm": 0.6146163940429688, "learning_rate": 1.7688192558572038e-05, "loss": 0.3593, "step": 29500 }, { "epoch": 11.140052850132125, "grad_norm": 0.7058233022689819, "learning_rate": 1.7656657984105906e-05, "loss": 0.3563, "step": 29510 }, { "epoch": 11.14382785956965, "grad_norm": 0.6673465967178345, "learning_rate": 1.7625145516009068e-05, "loss": 0.3613, "step": 29520 }, { "epoch": 11.147602869007173, "grad_norm": 0.6134123802185059, "learning_rate": 1.7593655175820005e-05, "loss": 0.3697, "step": 29530 }, { "epoch": 11.151377878444697, "grad_norm": 0.7190696597099304, "learning_rate": 1.7562186985062046e-05, "loss": 0.3666, "step": 29540 }, { "epoch": 11.15515288788222, "grad_norm": 0.7087527513504028, "learning_rate": 1.7530740965243403e-05, "loss": 0.3696, "step": 29550 }, { "epoch": 11.158927897319744, "grad_norm": 1.1012113094329834, "learning_rate": 1.7499317137857153e-05, "loss": 0.3582, "step": 29560 }, { "epoch": 11.162702906757266, "grad_norm": 0.6906222701072693, "learning_rate": 1.7467915524381184e-05, "loss": 0.3591, "step": 29570 }, { "epoch": 11.16647791619479, "grad_norm": 0.6418552398681641, "learning_rate": 1.7436536146278182e-05, "loss": 0.3702, "step": 29580 }, { "epoch": 11.170252925632314, "grad_norm": 1.7984684705734253, "learning_rate": 1.7405179024995688e-05, "loss": 0.3667, "step": 29590 }, { "epoch": 11.174027935069837, "grad_norm": 1.6302399635314941, "learning_rate": 1.737384418196596e-05, "loss": 0.3828, "step": 29600 }, { "epoch": 11.177802944507361, "grad_norm": 0.7185459136962891, "learning_rate": 1.734253163860609e-05, "loss": 0.3602, "step": 29610 }, { "epoch": 11.181577953944885, "grad_norm": 0.6619534492492676, "learning_rate": 1.7311241416317896e-05, "loss": 0.3618, "step": 29620 }, { "epoch": 11.185352963382408, "grad_norm": 0.6776392459869385, "learning_rate": 1.7279973536487982e-05, "loss": 0.3626, "step": 29630 }, { "epoch": 11.189127972819932, "grad_norm": 0.8006789088249207, "learning_rate": 1.724872802048761e-05, "loss": 0.3607, "step": 29640 }, { "epoch": 11.192902982257456, "grad_norm": 0.6797590255737305, "learning_rate": 1.7217504889672803e-05, "loss": 0.3616, "step": 29650 }, { "epoch": 11.19667799169498, "grad_norm": 0.6782128810882568, "learning_rate": 1.7186304165384287e-05, "loss": 0.3651, "step": 29660 }, { "epoch": 11.200453001132503, "grad_norm": 0.7670999765396118, "learning_rate": 1.7155125868947475e-05, "loss": 0.3615, "step": 29670 }, { "epoch": 11.204228010570027, "grad_norm": 0.6768984198570251, "learning_rate": 1.7123970021672404e-05, "loss": 0.3597, "step": 29680 }, { "epoch": 11.20800302000755, "grad_norm": 0.6669506430625916, "learning_rate": 1.709283664485384e-05, "loss": 0.3612, "step": 29690 }, { "epoch": 11.211778029445073, "grad_norm": 0.6505656838417053, "learning_rate": 1.7061725759771113e-05, "loss": 0.3583, "step": 29700 }, { "epoch": 11.215553038882597, "grad_norm": 0.6229871511459351, "learning_rate": 1.7030637387688248e-05, "loss": 0.364, "step": 29710 }, { "epoch": 11.21932804832012, "grad_norm": 2.578798770904541, "learning_rate": 1.6999571549853836e-05, "loss": 0.361, "step": 29720 }, { "epoch": 11.223103057757644, "grad_norm": 0.6974425911903381, "learning_rate": 1.696852826750112e-05, "loss": 0.3677, "step": 29730 }, { "epoch": 11.226878067195168, "grad_norm": 0.6888813376426697, "learning_rate": 1.6937507561847844e-05, "loss": 0.3687, "step": 29740 }, { "epoch": 11.230653076632692, "grad_norm": 0.6524753570556641, "learning_rate": 1.6906509454096385e-05, "loss": 0.3633, "step": 29750 }, { "epoch": 11.234428086070215, "grad_norm": 0.7103590965270996, "learning_rate": 1.687553396543367e-05, "loss": 0.3671, "step": 29760 }, { "epoch": 11.238203095507739, "grad_norm": 0.6661110520362854, "learning_rate": 1.6844581117031154e-05, "loss": 0.3597, "step": 29770 }, { "epoch": 11.241978104945263, "grad_norm": 1.3560032844543457, "learning_rate": 1.681365093004481e-05, "loss": 0.3731, "step": 29780 }, { "epoch": 11.245753114382786, "grad_norm": 0.6432333588600159, "learning_rate": 1.678274342561511e-05, "loss": 0.3623, "step": 29790 }, { "epoch": 11.24952812382031, "grad_norm": 0.6939373016357422, "learning_rate": 1.675185862486706e-05, "loss": 0.367, "step": 29800 }, { "epoch": 11.253303133257834, "grad_norm": 1.7712675333023071, "learning_rate": 1.6720996548910127e-05, "loss": 0.3675, "step": 29810 }, { "epoch": 11.257078142695356, "grad_norm": 0.7062354683876038, "learning_rate": 1.6690157218838247e-05, "loss": 0.3667, "step": 29820 }, { "epoch": 11.26085315213288, "grad_norm": 0.6343713402748108, "learning_rate": 1.665934065572984e-05, "loss": 0.3529, "step": 29830 }, { "epoch": 11.264628161570403, "grad_norm": 0.6509836316108704, "learning_rate": 1.6628546880647688e-05, "loss": 0.3628, "step": 29840 }, { "epoch": 11.268403171007927, "grad_norm": 0.6391473412513733, "learning_rate": 1.6597775914639076e-05, "loss": 0.3665, "step": 29850 }, { "epoch": 11.27217818044545, "grad_norm": 0.7817753553390503, "learning_rate": 1.6567027778735654e-05, "loss": 0.3619, "step": 29860 }, { "epoch": 11.275953189882975, "grad_norm": 0.5986528992652893, "learning_rate": 1.653630249395351e-05, "loss": 0.3592, "step": 29870 }, { "epoch": 11.279728199320498, "grad_norm": 0.6571201682090759, "learning_rate": 1.6505600081293072e-05, "loss": 0.3613, "step": 29880 }, { "epoch": 11.283503208758022, "grad_norm": 0.7287798523902893, "learning_rate": 1.647492056173912e-05, "loss": 0.3705, "step": 29890 }, { "epoch": 11.287278218195546, "grad_norm": 0.6623769998550415, "learning_rate": 1.6444263956260848e-05, "loss": 0.3559, "step": 29900 }, { "epoch": 11.29105322763307, "grad_norm": 0.6253150105476379, "learning_rate": 1.641363028581175e-05, "loss": 0.363, "step": 29910 }, { "epoch": 11.294828237070593, "grad_norm": 0.6631579399108887, "learning_rate": 1.638301957132965e-05, "loss": 0.3649, "step": 29920 }, { "epoch": 11.298603246508117, "grad_norm": 0.7632430791854858, "learning_rate": 1.6352431833736703e-05, "loss": 0.3576, "step": 29930 }, { "epoch": 11.30237825594564, "grad_norm": 0.8125953674316406, "learning_rate": 1.6321867093939298e-05, "loss": 0.3624, "step": 29940 }, { "epoch": 11.306153265383163, "grad_norm": 0.6205570101737976, "learning_rate": 1.629132537282817e-05, "loss": 0.3575, "step": 29950 }, { "epoch": 11.309928274820686, "grad_norm": 0.6771644949913025, "learning_rate": 1.62608066912783e-05, "loss": 0.3572, "step": 29960 }, { "epoch": 11.31370328425821, "grad_norm": 0.6554526090621948, "learning_rate": 1.623031107014893e-05, "loss": 0.3662, "step": 29970 }, { "epoch": 11.317478293695734, "grad_norm": 0.6397532820701599, "learning_rate": 1.619983853028351e-05, "loss": 0.3668, "step": 29980 }, { "epoch": 11.321253303133258, "grad_norm": 0.6637176871299744, "learning_rate": 1.6169389092509724e-05, "loss": 0.3611, "step": 29990 }, { "epoch": 11.325028312570781, "grad_norm": 0.6467320322990417, "learning_rate": 1.6138962777639494e-05, "loss": 0.3625, "step": 30000 }, { "epoch": 11.328803322008305, "grad_norm": 0.6978408694267273, "learning_rate": 1.610855960646891e-05, "loss": 0.3677, "step": 30010 }, { "epoch": 11.332578331445829, "grad_norm": 0.6771886348724365, "learning_rate": 1.607817959977826e-05, "loss": 0.3654, "step": 30020 }, { "epoch": 11.336353340883353, "grad_norm": 0.691331684589386, "learning_rate": 1.6047822778332005e-05, "loss": 0.363, "step": 30030 }, { "epoch": 11.340128350320876, "grad_norm": 0.6371560096740723, "learning_rate": 1.6017489162878713e-05, "loss": 0.3671, "step": 30040 }, { "epoch": 11.3439033597584, "grad_norm": 0.6647712588310242, "learning_rate": 1.5987178774151147e-05, "loss": 0.3557, "step": 30050 }, { "epoch": 11.347678369195924, "grad_norm": 0.6369796991348267, "learning_rate": 1.5956891632866184e-05, "loss": 0.3607, "step": 30060 }, { "epoch": 11.351453378633446, "grad_norm": 0.6717673540115356, "learning_rate": 1.592662775972476e-05, "loss": 0.3559, "step": 30070 }, { "epoch": 11.35522838807097, "grad_norm": 0.6065929532051086, "learning_rate": 1.5896387175411986e-05, "loss": 0.3572, "step": 30080 }, { "epoch": 11.359003397508493, "grad_norm": 0.7193185091018677, "learning_rate": 1.586616990059699e-05, "loss": 0.3617, "step": 30090 }, { "epoch": 11.362778406946017, "grad_norm": 0.7065017223358154, "learning_rate": 1.5835975955933002e-05, "loss": 0.3536, "step": 30100 }, { "epoch": 11.36655341638354, "grad_norm": 0.7551389336585999, "learning_rate": 1.5805805362057295e-05, "loss": 0.3586, "step": 30110 }, { "epoch": 11.370328425821064, "grad_norm": 0.6497236490249634, "learning_rate": 1.57756581395912e-05, "loss": 0.3634, "step": 30120 }, { "epoch": 11.374103435258588, "grad_norm": 0.6430761814117432, "learning_rate": 1.574553430914006e-05, "loss": 0.3627, "step": 30130 }, { "epoch": 11.377878444696112, "grad_norm": 0.6875909566879272, "learning_rate": 1.5715433891293206e-05, "loss": 0.3609, "step": 30140 }, { "epoch": 11.381653454133636, "grad_norm": 0.61098313331604, "learning_rate": 1.5685356906624016e-05, "loss": 0.3585, "step": 30150 }, { "epoch": 11.38542846357116, "grad_norm": 0.6427933573722839, "learning_rate": 1.56553033756898e-05, "loss": 0.3601, "step": 30160 }, { "epoch": 11.389203473008683, "grad_norm": 0.662592351436615, "learning_rate": 1.5625273319031873e-05, "loss": 0.364, "step": 30170 }, { "epoch": 11.392978482446207, "grad_norm": 0.7252653241157532, "learning_rate": 1.5595266757175518e-05, "loss": 0.3614, "step": 30180 }, { "epoch": 11.39675349188373, "grad_norm": 0.6587193608283997, "learning_rate": 1.55652837106299e-05, "loss": 0.3618, "step": 30190 }, { "epoch": 11.400528501321253, "grad_norm": 0.7090880274772644, "learning_rate": 1.5535324199888164e-05, "loss": 0.3572, "step": 30200 }, { "epoch": 11.404303510758776, "grad_norm": 0.6730210781097412, "learning_rate": 1.5505388245427355e-05, "loss": 0.3621, "step": 30210 }, { "epoch": 11.4080785201963, "grad_norm": 0.6814711689949036, "learning_rate": 1.5475475867708418e-05, "loss": 0.3647, "step": 30220 }, { "epoch": 11.411853529633824, "grad_norm": 0.6800294518470764, "learning_rate": 1.5445587087176195e-05, "loss": 0.3635, "step": 30230 }, { "epoch": 11.415628539071347, "grad_norm": 0.6638886332511902, "learning_rate": 1.5415721924259346e-05, "loss": 0.3634, "step": 30240 }, { "epoch": 11.419403548508871, "grad_norm": 0.6355138421058655, "learning_rate": 1.538588039937047e-05, "loss": 0.36, "step": 30250 }, { "epoch": 11.423178557946395, "grad_norm": 0.6450951099395752, "learning_rate": 1.535606253290594e-05, "loss": 0.3644, "step": 30260 }, { "epoch": 11.426953567383919, "grad_norm": 0.640677273273468, "learning_rate": 1.5326268345245993e-05, "loss": 0.3594, "step": 30270 }, { "epoch": 11.430728576821442, "grad_norm": 0.7016587257385254, "learning_rate": 1.5296497856754693e-05, "loss": 0.3645, "step": 30280 }, { "epoch": 11.434503586258966, "grad_norm": 0.6003674864768982, "learning_rate": 1.5266751087779858e-05, "loss": 0.3645, "step": 30290 }, { "epoch": 11.43827859569649, "grad_norm": 0.6319372057914734, "learning_rate": 1.5237028058653142e-05, "loss": 0.3662, "step": 30300 }, { "epoch": 11.442053605134014, "grad_norm": 0.6442000865936279, "learning_rate": 1.5207328789689946e-05, "loss": 0.3606, "step": 30310 }, { "epoch": 11.445828614571536, "grad_norm": 0.6055095791816711, "learning_rate": 1.5177653301189454e-05, "loss": 0.3652, "step": 30320 }, { "epoch": 11.44960362400906, "grad_norm": 0.7226874232292175, "learning_rate": 1.5148001613434587e-05, "loss": 0.3553, "step": 30330 }, { "epoch": 11.453378633446583, "grad_norm": 0.7227387428283691, "learning_rate": 1.5118373746691983e-05, "loss": 0.3628, "step": 30340 }, { "epoch": 11.457153642884107, "grad_norm": 0.6881961226463318, "learning_rate": 1.5088769721211992e-05, "loss": 0.359, "step": 30350 }, { "epoch": 11.46092865232163, "grad_norm": 0.7442403435707092, "learning_rate": 1.505918955722871e-05, "loss": 0.3632, "step": 30360 }, { "epoch": 11.464703661759154, "grad_norm": 0.6394659876823425, "learning_rate": 1.5029633274959886e-05, "loss": 0.3559, "step": 30370 }, { "epoch": 11.468478671196678, "grad_norm": 0.6404333114624023, "learning_rate": 1.5000100894606989e-05, "loss": 0.3605, "step": 30380 }, { "epoch": 11.472253680634202, "grad_norm": 0.7091559171676636, "learning_rate": 1.4970592436355079e-05, "loss": 0.3546, "step": 30390 }, { "epoch": 11.476028690071725, "grad_norm": 0.6807871460914612, "learning_rate": 1.4941107920372938e-05, "loss": 0.3703, "step": 30400 }, { "epoch": 11.47980369950925, "grad_norm": 0.6545217633247375, "learning_rate": 1.4911647366812947e-05, "loss": 0.3574, "step": 30410 }, { "epoch": 11.483578708946773, "grad_norm": 0.6775568723678589, "learning_rate": 1.4882210795811114e-05, "loss": 0.3653, "step": 30420 }, { "epoch": 11.487353718384297, "grad_norm": 0.7269086837768555, "learning_rate": 1.485279822748708e-05, "loss": 0.3608, "step": 30430 }, { "epoch": 11.49112872782182, "grad_norm": 0.7882838249206543, "learning_rate": 1.4823409681944034e-05, "loss": 0.3708, "step": 30440 }, { "epoch": 11.494903737259342, "grad_norm": 0.6561617255210876, "learning_rate": 1.4794045179268763e-05, "loss": 0.3586, "step": 30450 }, { "epoch": 11.498678746696866, "grad_norm": 0.6408365368843079, "learning_rate": 1.4764704739531643e-05, "loss": 0.3661, "step": 30460 }, { "epoch": 11.50245375613439, "grad_norm": 0.6767753958702087, "learning_rate": 1.473538838278658e-05, "loss": 0.3576, "step": 30470 }, { "epoch": 11.506228765571914, "grad_norm": 0.6685946583747864, "learning_rate": 1.4706096129071045e-05, "loss": 0.3657, "step": 30480 }, { "epoch": 11.510003775009437, "grad_norm": 0.7614819407463074, "learning_rate": 1.4676827998405996e-05, "loss": 0.3657, "step": 30490 }, { "epoch": 11.513778784446961, "grad_norm": 0.6719235181808472, "learning_rate": 1.4647584010795928e-05, "loss": 0.3587, "step": 30500 }, { "epoch": 11.517553793884485, "grad_norm": 0.6133652925491333, "learning_rate": 1.4618364186228834e-05, "loss": 0.3656, "step": 30510 }, { "epoch": 11.521328803322008, "grad_norm": 0.6366606950759888, "learning_rate": 1.4589168544676207e-05, "loss": 0.3592, "step": 30520 }, { "epoch": 11.525103812759532, "grad_norm": 0.6164776086807251, "learning_rate": 1.455999710609296e-05, "loss": 0.3579, "step": 30530 }, { "epoch": 11.528878822197056, "grad_norm": 0.5788912177085876, "learning_rate": 1.453084989041753e-05, "loss": 0.3612, "step": 30540 }, { "epoch": 11.53265383163458, "grad_norm": 0.6258981227874756, "learning_rate": 1.4501726917571728e-05, "loss": 0.3599, "step": 30550 }, { "epoch": 11.536428841072103, "grad_norm": 0.8489433526992798, "learning_rate": 1.4472628207460848e-05, "loss": 0.3561, "step": 30560 }, { "epoch": 11.540203850509627, "grad_norm": 0.7995439767837524, "learning_rate": 1.4443553779973584e-05, "loss": 0.3608, "step": 30570 }, { "epoch": 11.54397885994715, "grad_norm": 0.642866313457489, "learning_rate": 1.4414503654982037e-05, "loss": 0.3593, "step": 30580 }, { "epoch": 11.547753869384673, "grad_norm": 0.688869059085846, "learning_rate": 1.4385477852341706e-05, "loss": 0.3612, "step": 30590 }, { "epoch": 11.551528878822197, "grad_norm": 0.6555159091949463, "learning_rate": 1.4356476391891421e-05, "loss": 0.3577, "step": 30600 }, { "epoch": 11.55530388825972, "grad_norm": 0.8551262021064758, "learning_rate": 1.4327499293453423e-05, "loss": 0.371, "step": 30610 }, { "epoch": 11.559078897697244, "grad_norm": 0.6704701781272888, "learning_rate": 1.4298546576833304e-05, "loss": 0.3564, "step": 30620 }, { "epoch": 11.562853907134768, "grad_norm": 0.7134976387023926, "learning_rate": 1.426961826181994e-05, "loss": 0.3619, "step": 30630 }, { "epoch": 11.566628916572292, "grad_norm": 0.6610779166221619, "learning_rate": 1.4240714368185592e-05, "loss": 0.3642, "step": 30640 }, { "epoch": 11.570403926009815, "grad_norm": 0.6555269956588745, "learning_rate": 1.4211834915685767e-05, "loss": 0.3627, "step": 30650 }, { "epoch": 11.574178935447339, "grad_norm": 0.677355170249939, "learning_rate": 1.4182979924059313e-05, "loss": 0.3527, "step": 30660 }, { "epoch": 11.577953944884863, "grad_norm": 0.6868191361427307, "learning_rate": 1.415414941302835e-05, "loss": 0.3604, "step": 30670 }, { "epoch": 11.581728954322386, "grad_norm": 0.7704634666442871, "learning_rate": 1.4125343402298257e-05, "loss": 0.3605, "step": 30680 }, { "epoch": 11.58550396375991, "grad_norm": 0.6135473251342773, "learning_rate": 1.4096561911557688e-05, "loss": 0.36, "step": 30690 }, { "epoch": 11.589278973197432, "grad_norm": 1.2090810537338257, "learning_rate": 1.4067804960478481e-05, "loss": 0.3648, "step": 30700 }, { "epoch": 11.593053982634956, "grad_norm": 0.6949913501739502, "learning_rate": 1.4039072568715784e-05, "loss": 0.3637, "step": 30710 }, { "epoch": 11.59682899207248, "grad_norm": 0.6498703360557556, "learning_rate": 1.401036475590788e-05, "loss": 0.3651, "step": 30720 }, { "epoch": 11.600604001510003, "grad_norm": 0.7282187342643738, "learning_rate": 1.3981681541676301e-05, "loss": 0.3629, "step": 30730 }, { "epoch": 11.604379010947527, "grad_norm": 0.6524527668952942, "learning_rate": 1.3953022945625772e-05, "loss": 0.3581, "step": 30740 }, { "epoch": 11.60815402038505, "grad_norm": 0.6055677533149719, "learning_rate": 1.3924388987344138e-05, "loss": 0.3575, "step": 30750 }, { "epoch": 11.611929029822575, "grad_norm": 0.717749834060669, "learning_rate": 1.3895779686402465e-05, "loss": 0.3602, "step": 30760 }, { "epoch": 11.615704039260098, "grad_norm": 0.7012677788734436, "learning_rate": 1.3867195062354931e-05, "loss": 0.3574, "step": 30770 }, { "epoch": 11.619479048697622, "grad_norm": 0.7111396193504333, "learning_rate": 1.3838635134738858e-05, "loss": 0.3537, "step": 30780 }, { "epoch": 11.623254058135146, "grad_norm": 0.7217731475830078, "learning_rate": 1.3810099923074704e-05, "loss": 0.3599, "step": 30790 }, { "epoch": 11.62702906757267, "grad_norm": 0.609724760055542, "learning_rate": 1.3781589446866006e-05, "loss": 0.3743, "step": 30800 }, { "epoch": 11.630804077010193, "grad_norm": 0.6935325860977173, "learning_rate": 1.3753103725599387e-05, "loss": 0.3577, "step": 30810 }, { "epoch": 11.634579086447715, "grad_norm": 0.7039898037910461, "learning_rate": 1.3724642778744579e-05, "loss": 0.3615, "step": 30820 }, { "epoch": 11.638354095885239, "grad_norm": 0.6811739802360535, "learning_rate": 1.3696206625754376e-05, "loss": 0.3589, "step": 30830 }, { "epoch": 11.642129105322763, "grad_norm": 0.6018708348274231, "learning_rate": 1.3667795286064639e-05, "loss": 0.3626, "step": 30840 }, { "epoch": 11.645904114760286, "grad_norm": 0.6382670998573303, "learning_rate": 1.3639408779094215e-05, "loss": 0.3649, "step": 30850 }, { "epoch": 11.64967912419781, "grad_norm": 0.6437296867370605, "learning_rate": 1.361104712424503e-05, "loss": 0.364, "step": 30860 }, { "epoch": 11.653454133635334, "grad_norm": 0.6563718318939209, "learning_rate": 1.358271034090201e-05, "loss": 0.3597, "step": 30870 }, { "epoch": 11.657229143072858, "grad_norm": 0.6579415798187256, "learning_rate": 1.3554398448433076e-05, "loss": 0.3694, "step": 30880 }, { "epoch": 11.661004152510381, "grad_norm": 0.7728137373924255, "learning_rate": 1.3526111466189156e-05, "loss": 0.3608, "step": 30890 }, { "epoch": 11.664779161947905, "grad_norm": 0.649925708770752, "learning_rate": 1.3497849413504121e-05, "loss": 0.3571, "step": 30900 }, { "epoch": 11.668554171385429, "grad_norm": 0.6554239392280579, "learning_rate": 1.3469612309694802e-05, "loss": 0.3653, "step": 30910 }, { "epoch": 11.672329180822953, "grad_norm": 0.6548380851745605, "learning_rate": 1.3441400174061008e-05, "loss": 0.367, "step": 30920 }, { "epoch": 11.676104190260476, "grad_norm": 0.7009982466697693, "learning_rate": 1.3413213025885469e-05, "loss": 0.362, "step": 30930 }, { "epoch": 11.679879199698, "grad_norm": 0.6359342336654663, "learning_rate": 1.3385050884433853e-05, "loss": 0.3625, "step": 30940 }, { "epoch": 11.683654209135522, "grad_norm": 0.6045933365821838, "learning_rate": 1.3356913768954676e-05, "loss": 0.3534, "step": 30950 }, { "epoch": 11.687429218573046, "grad_norm": 0.6957305669784546, "learning_rate": 1.3328801698679415e-05, "loss": 0.3594, "step": 30960 }, { "epoch": 11.69120422801057, "grad_norm": 1.19200599193573, "learning_rate": 1.3300714692822413e-05, "loss": 0.3681, "step": 30970 }, { "epoch": 11.694979237448093, "grad_norm": 0.6715503931045532, "learning_rate": 1.3272652770580857e-05, "loss": 0.3561, "step": 30980 }, { "epoch": 11.698754246885617, "grad_norm": 0.679587721824646, "learning_rate": 1.3244615951134831e-05, "loss": 0.3577, "step": 30990 }, { "epoch": 11.70252925632314, "grad_norm": 0.624066948890686, "learning_rate": 1.3216604253647219e-05, "loss": 0.3624, "step": 31000 }, { "epoch": 11.706304265760664, "grad_norm": 0.7792338728904724, "learning_rate": 1.3188617697263738e-05, "loss": 0.3621, "step": 31010 }, { "epoch": 11.710079275198188, "grad_norm": 0.6704654693603516, "learning_rate": 1.3160656301112956e-05, "loss": 0.3506, "step": 31020 }, { "epoch": 11.713854284635712, "grad_norm": 0.6014306545257568, "learning_rate": 1.3132720084306211e-05, "loss": 0.3666, "step": 31030 }, { "epoch": 11.717629294073236, "grad_norm": 0.6405402421951294, "learning_rate": 1.3104809065937668e-05, "loss": 0.3574, "step": 31040 }, { "epoch": 11.72140430351076, "grad_norm": 0.6713264584541321, "learning_rate": 1.3076923265084212e-05, "loss": 0.3619, "step": 31050 }, { "epoch": 11.725179312948283, "grad_norm": 0.6099828481674194, "learning_rate": 1.3049062700805543e-05, "loss": 0.3555, "step": 31060 }, { "epoch": 11.728954322385807, "grad_norm": 0.6921255588531494, "learning_rate": 1.3021227392144092e-05, "loss": 0.3596, "step": 31070 }, { "epoch": 11.732729331823329, "grad_norm": 0.6106016635894775, "learning_rate": 1.2993417358125037e-05, "loss": 0.3541, "step": 31080 }, { "epoch": 11.736504341260853, "grad_norm": 0.5807941555976868, "learning_rate": 1.2965632617756257e-05, "loss": 0.3613, "step": 31090 }, { "epoch": 11.740279350698376, "grad_norm": 0.5835058093070984, "learning_rate": 1.2937873190028387e-05, "loss": 0.3577, "step": 31100 }, { "epoch": 11.7440543601359, "grad_norm": 0.5961214303970337, "learning_rate": 1.2910139093914697e-05, "loss": 0.3567, "step": 31110 }, { "epoch": 11.747829369573424, "grad_norm": 0.696262538433075, "learning_rate": 1.2882430348371205e-05, "loss": 0.3555, "step": 31120 }, { "epoch": 11.751604379010947, "grad_norm": 0.6330527663230896, "learning_rate": 1.2854746972336567e-05, "loss": 0.3644, "step": 31130 }, { "epoch": 11.755379388448471, "grad_norm": 0.6530078649520874, "learning_rate": 1.2827088984732132e-05, "loss": 0.3728, "step": 31140 }, { "epoch": 11.759154397885995, "grad_norm": 0.596555769443512, "learning_rate": 1.279945640446184e-05, "loss": 0.3646, "step": 31150 }, { "epoch": 11.762929407323519, "grad_norm": 0.6525673270225525, "learning_rate": 1.2771849250412315e-05, "loss": 0.3603, "step": 31160 }, { "epoch": 11.766704416761042, "grad_norm": 0.716068685054779, "learning_rate": 1.2744267541452792e-05, "loss": 0.3579, "step": 31170 }, { "epoch": 11.770479426198566, "grad_norm": 0.5986026525497437, "learning_rate": 1.2716711296435112e-05, "loss": 0.3612, "step": 31180 }, { "epoch": 11.77425443563609, "grad_norm": 0.6966813206672668, "learning_rate": 1.268918053419369e-05, "loss": 0.3645, "step": 31190 }, { "epoch": 11.778029445073614, "grad_norm": 0.6633387207984924, "learning_rate": 1.2661675273545564e-05, "loss": 0.3607, "step": 31200 }, { "epoch": 11.781804454511136, "grad_norm": 0.7241559028625488, "learning_rate": 1.26341955332903e-05, "loss": 0.3615, "step": 31210 }, { "epoch": 11.78557946394866, "grad_norm": 0.6759303212165833, "learning_rate": 1.2606741332210048e-05, "loss": 0.3611, "step": 31220 }, { "epoch": 11.789354473386183, "grad_norm": 0.642245888710022, "learning_rate": 1.2579312689069495e-05, "loss": 0.3591, "step": 31230 }, { "epoch": 11.793129482823707, "grad_norm": 0.6876919865608215, "learning_rate": 1.2551909622615881e-05, "loss": 0.3537, "step": 31240 }, { "epoch": 11.79690449226123, "grad_norm": 0.6188132762908936, "learning_rate": 1.252453215157891e-05, "loss": 0.3683, "step": 31250 }, { "epoch": 11.800679501698754, "grad_norm": 0.6617509722709656, "learning_rate": 1.2497180294670841e-05, "loss": 0.3575, "step": 31260 }, { "epoch": 11.804454511136278, "grad_norm": 0.6628468036651611, "learning_rate": 1.2469854070586423e-05, "loss": 0.3568, "step": 31270 }, { "epoch": 11.808229520573802, "grad_norm": 0.6406071186065674, "learning_rate": 1.2442553498002845e-05, "loss": 0.3705, "step": 31280 }, { "epoch": 11.812004530011325, "grad_norm": 0.704318106174469, "learning_rate": 1.241527859557981e-05, "loss": 0.3602, "step": 31290 }, { "epoch": 11.81577953944885, "grad_norm": 0.6852730512619019, "learning_rate": 1.238802938195946e-05, "loss": 0.3549, "step": 31300 }, { "epoch": 11.819554548886373, "grad_norm": 0.7002149820327759, "learning_rate": 1.2360805875766357e-05, "loss": 0.3582, "step": 31310 }, { "epoch": 11.823329558323895, "grad_norm": 0.6709305047988892, "learning_rate": 1.2333608095607524e-05, "loss": 0.3598, "step": 31320 }, { "epoch": 11.827104567761419, "grad_norm": 0.6867847442626953, "learning_rate": 1.230643606007238e-05, "loss": 0.3582, "step": 31330 }, { "epoch": 11.830879577198942, "grad_norm": 0.7019827961921692, "learning_rate": 1.2279289787732778e-05, "loss": 0.3513, "step": 31340 }, { "epoch": 11.834654586636466, "grad_norm": 0.5999962687492371, "learning_rate": 1.2252169297142906e-05, "loss": 0.3575, "step": 31350 }, { "epoch": 11.83842959607399, "grad_norm": 0.6360159516334534, "learning_rate": 1.2225074606839392e-05, "loss": 0.3611, "step": 31360 }, { "epoch": 11.842204605511514, "grad_norm": 0.5966977477073669, "learning_rate": 1.2198005735341173e-05, "loss": 0.364, "step": 31370 }, { "epoch": 11.845979614949037, "grad_norm": 0.6092012524604797, "learning_rate": 1.2170962701149584e-05, "loss": 0.3562, "step": 31380 }, { "epoch": 11.849754624386561, "grad_norm": 0.7129442095756531, "learning_rate": 1.2143945522748285e-05, "loss": 0.3621, "step": 31390 }, { "epoch": 11.853529633824085, "grad_norm": 0.6315842270851135, "learning_rate": 1.2116954218603266e-05, "loss": 0.3636, "step": 31400 }, { "epoch": 11.857304643261608, "grad_norm": 0.7170038223266602, "learning_rate": 1.2089988807162817e-05, "loss": 0.3601, "step": 31410 }, { "epoch": 11.861079652699132, "grad_norm": 0.5954352021217346, "learning_rate": 1.2063049306857543e-05, "loss": 0.3585, "step": 31420 }, { "epoch": 11.864854662136656, "grad_norm": 0.6874489784240723, "learning_rate": 1.2036135736100346e-05, "loss": 0.3589, "step": 31430 }, { "epoch": 11.86862967157418, "grad_norm": 0.6595900058746338, "learning_rate": 1.200924811328641e-05, "loss": 0.3526, "step": 31440 }, { "epoch": 11.872404681011702, "grad_norm": 0.6593577861785889, "learning_rate": 1.1982386456793143e-05, "loss": 0.3656, "step": 31450 }, { "epoch": 11.876179690449225, "grad_norm": 0.6922234296798706, "learning_rate": 1.1955550784980257e-05, "loss": 0.3593, "step": 31460 }, { "epoch": 11.87995469988675, "grad_norm": 0.6410038471221924, "learning_rate": 1.1928741116189663e-05, "loss": 0.3544, "step": 31470 }, { "epoch": 11.883729709324273, "grad_norm": 0.6991416811943054, "learning_rate": 1.1901957468745529e-05, "loss": 0.3648, "step": 31480 }, { "epoch": 11.887504718761797, "grad_norm": 0.6780281662940979, "learning_rate": 1.1875199860954217e-05, "loss": 0.3662, "step": 31490 }, { "epoch": 11.89127972819932, "grad_norm": 0.8736029863357544, "learning_rate": 1.1848468311104316e-05, "loss": 0.3568, "step": 31500 }, { "epoch": 11.895054737636844, "grad_norm": 0.5960100889205933, "learning_rate": 1.1821762837466566e-05, "loss": 0.3573, "step": 31510 }, { "epoch": 11.898829747074368, "grad_norm": 0.6374369263648987, "learning_rate": 1.179508345829392e-05, "loss": 0.3708, "step": 31520 }, { "epoch": 11.902604756511892, "grad_norm": 0.7211053967475891, "learning_rate": 1.1768430191821478e-05, "loss": 0.3584, "step": 31530 }, { "epoch": 11.906379765949415, "grad_norm": 0.8091753721237183, "learning_rate": 1.174180305626651e-05, "loss": 0.3602, "step": 31540 }, { "epoch": 11.910154775386939, "grad_norm": 0.6501530408859253, "learning_rate": 1.1715202069828401e-05, "loss": 0.3593, "step": 31550 }, { "epoch": 11.913929784824463, "grad_norm": 0.7378227710723877, "learning_rate": 1.1688627250688661e-05, "loss": 0.3638, "step": 31560 }, { "epoch": 11.917704794261986, "grad_norm": 0.6729764342308044, "learning_rate": 1.1662078617010935e-05, "loss": 0.3598, "step": 31570 }, { "epoch": 11.921479803699508, "grad_norm": 0.6561091542243958, "learning_rate": 1.1635556186940972e-05, "loss": 0.3649, "step": 31580 }, { "epoch": 11.925254813137032, "grad_norm": 0.6765795946121216, "learning_rate": 1.1609059978606602e-05, "loss": 0.3709, "step": 31590 }, { "epoch": 11.929029822574556, "grad_norm": 0.641923189163208, "learning_rate": 1.1582590010117745e-05, "loss": 0.3595, "step": 31600 }, { "epoch": 11.93280483201208, "grad_norm": 0.6592920422554016, "learning_rate": 1.1556146299566345e-05, "loss": 0.3543, "step": 31610 }, { "epoch": 11.936579841449603, "grad_norm": 0.6286612153053284, "learning_rate": 1.1529728865026452e-05, "loss": 0.3685, "step": 31620 }, { "epoch": 11.940354850887127, "grad_norm": 1.297844648361206, "learning_rate": 1.150333772455412e-05, "loss": 0.3626, "step": 31630 }, { "epoch": 11.94412986032465, "grad_norm": 0.6182207465171814, "learning_rate": 1.1476972896187466e-05, "loss": 0.355, "step": 31640 }, { "epoch": 11.947904869762175, "grad_norm": 0.7070791125297546, "learning_rate": 1.1450634397946592e-05, "loss": 0.359, "step": 31650 }, { "epoch": 11.951679879199698, "grad_norm": 0.6540588140487671, "learning_rate": 1.142432224783359e-05, "loss": 0.358, "step": 31660 }, { "epoch": 11.955454888637222, "grad_norm": 0.6664989590644836, "learning_rate": 1.1398036463832584e-05, "loss": 0.3555, "step": 31670 }, { "epoch": 11.959229898074746, "grad_norm": 0.6335260272026062, "learning_rate": 1.1371777063909661e-05, "loss": 0.3611, "step": 31680 }, { "epoch": 11.96300490751227, "grad_norm": 0.614297091960907, "learning_rate": 1.1345544066012874e-05, "loss": 0.3628, "step": 31690 }, { "epoch": 11.966779916949793, "grad_norm": 0.746405839920044, "learning_rate": 1.131933748807224e-05, "loss": 0.3556, "step": 31700 }, { "epoch": 11.970554926387315, "grad_norm": 0.6299577951431274, "learning_rate": 1.1293157347999683e-05, "loss": 0.3573, "step": 31710 }, { "epoch": 11.974329935824839, "grad_norm": 0.643743634223938, "learning_rate": 1.1267003663689097e-05, "loss": 0.3597, "step": 31720 }, { "epoch": 11.978104945262363, "grad_norm": 0.7774217128753662, "learning_rate": 1.1240876453016275e-05, "loss": 0.3816, "step": 31730 }, { "epoch": 11.981879954699886, "grad_norm": 0.6441808342933655, "learning_rate": 1.1214775733838933e-05, "loss": 0.3589, "step": 31740 }, { "epoch": 11.98565496413741, "grad_norm": 0.6050808429718018, "learning_rate": 1.118870152399666e-05, "loss": 0.3601, "step": 31750 }, { "epoch": 11.989429973574934, "grad_norm": 0.6663985252380371, "learning_rate": 1.1162653841310905e-05, "loss": 0.3602, "step": 31760 }, { "epoch": 11.993204983012458, "grad_norm": 0.6749792695045471, "learning_rate": 1.1136632703585043e-05, "loss": 0.3576, "step": 31770 }, { "epoch": 11.996979992449981, "grad_norm": 0.731533408164978, "learning_rate": 1.111063812860426e-05, "loss": 0.3565, "step": 31780 }, { "epoch": 12.000755001887505, "grad_norm": 0.6132811903953552, "learning_rate": 1.1084670134135616e-05, "loss": 0.3547, "step": 31790 }, { "epoch": 12.004530011325029, "grad_norm": 0.6406649947166443, "learning_rate": 1.1058728737927998e-05, "loss": 0.3593, "step": 31800 }, { "epoch": 12.008305020762553, "grad_norm": 0.6679116487503052, "learning_rate": 1.1032813957712086e-05, "loss": 0.3594, "step": 31810 }, { "epoch": 12.012080030200076, "grad_norm": 0.6831429600715637, "learning_rate": 1.1006925811200385e-05, "loss": 0.355, "step": 31820 }, { "epoch": 12.015855039637598, "grad_norm": 0.6503385305404663, "learning_rate": 1.0981064316087231e-05, "loss": 0.3616, "step": 31830 }, { "epoch": 12.019630049075122, "grad_norm": 0.6955201625823975, "learning_rate": 1.0955229490048675e-05, "loss": 0.3606, "step": 31840 }, { "epoch": 12.023405058512646, "grad_norm": 0.6384215950965881, "learning_rate": 1.0929421350742603e-05, "loss": 0.3632, "step": 31850 }, { "epoch": 12.02718006795017, "grad_norm": 0.6311360001564026, "learning_rate": 1.090363991580861e-05, "loss": 0.3628, "step": 31860 }, { "epoch": 12.030955077387693, "grad_norm": 0.8894734382629395, "learning_rate": 1.0877885202868083e-05, "loss": 0.3631, "step": 31870 }, { "epoch": 12.034730086825217, "grad_norm": 0.6360461115837097, "learning_rate": 1.0852157229524118e-05, "loss": 0.3779, "step": 31880 }, { "epoch": 12.03850509626274, "grad_norm": 0.7593185901641846, "learning_rate": 1.0826456013361542e-05, "loss": 0.3516, "step": 31890 }, { "epoch": 12.042280105700264, "grad_norm": 0.681585431098938, "learning_rate": 1.080078157194691e-05, "loss": 0.3682, "step": 31900 }, { "epoch": 12.046055115137788, "grad_norm": 0.7083263397216797, "learning_rate": 1.077513392282844e-05, "loss": 0.3588, "step": 31910 }, { "epoch": 12.049830124575312, "grad_norm": 0.7412534952163696, "learning_rate": 1.0749513083536073e-05, "loss": 0.3593, "step": 31920 }, { "epoch": 12.053605134012836, "grad_norm": 0.6712055802345276, "learning_rate": 1.072391907158139e-05, "loss": 0.3601, "step": 31930 }, { "epoch": 12.05738014345036, "grad_norm": 0.6988897323608398, "learning_rate": 1.0698351904457676e-05, "loss": 0.3553, "step": 31940 }, { "epoch": 12.061155152887883, "grad_norm": 0.7310452461242676, "learning_rate": 1.0672811599639849e-05, "loss": 0.3612, "step": 31950 }, { "epoch": 12.064930162325405, "grad_norm": 0.6422081589698792, "learning_rate": 1.0647298174584453e-05, "loss": 0.3648, "step": 31960 }, { "epoch": 12.068705171762929, "grad_norm": 0.6467820405960083, "learning_rate": 1.0621811646729673e-05, "loss": 0.3573, "step": 31970 }, { "epoch": 12.072480181200453, "grad_norm": 0.5941371917724609, "learning_rate": 1.0596352033495328e-05, "loss": 0.3646, "step": 31980 }, { "epoch": 12.076255190637976, "grad_norm": 0.6265065670013428, "learning_rate": 1.057091935228281e-05, "loss": 0.3674, "step": 31990 }, { "epoch": 12.0800302000755, "grad_norm": 0.7404178380966187, "learning_rate": 1.0545513620475134e-05, "loss": 0.3639, "step": 32000 }, { "epoch": 12.083805209513024, "grad_norm": 0.6679591536521912, "learning_rate": 1.0520134855436863e-05, "loss": 0.3705, "step": 32010 }, { "epoch": 12.087580218950547, "grad_norm": 0.6543688774108887, "learning_rate": 1.049478307451413e-05, "loss": 0.3586, "step": 32020 }, { "epoch": 12.091355228388071, "grad_norm": 0.6841951012611389, "learning_rate": 1.0469458295034656e-05, "loss": 0.3585, "step": 32030 }, { "epoch": 12.095130237825595, "grad_norm": 0.7727581858634949, "learning_rate": 1.044416053430769e-05, "loss": 0.3591, "step": 32040 }, { "epoch": 12.098905247263119, "grad_norm": 0.6680077910423279, "learning_rate": 1.0418889809624021e-05, "loss": 0.3614, "step": 32050 }, { "epoch": 12.102680256700642, "grad_norm": 0.5810276865959167, "learning_rate": 1.0393646138255925e-05, "loss": 0.3641, "step": 32060 }, { "epoch": 12.106455266138166, "grad_norm": 0.6576191782951355, "learning_rate": 1.036842953745723e-05, "loss": 0.3583, "step": 32070 }, { "epoch": 12.110230275575688, "grad_norm": 0.6342170834541321, "learning_rate": 1.034324002446324e-05, "loss": 0.355, "step": 32080 }, { "epoch": 12.114005285013212, "grad_norm": 0.6298282146453857, "learning_rate": 1.0318077616490751e-05, "loss": 0.3631, "step": 32090 }, { "epoch": 12.117780294450736, "grad_norm": 0.6232679486274719, "learning_rate": 1.0292942330738054e-05, "loss": 0.3566, "step": 32100 }, { "epoch": 12.12155530388826, "grad_norm": 0.663240373134613, "learning_rate": 1.026783418438485e-05, "loss": 0.3568, "step": 32110 }, { "epoch": 12.125330313325783, "grad_norm": 0.6801791191101074, "learning_rate": 1.0242753194592319e-05, "loss": 0.3593, "step": 32120 }, { "epoch": 12.129105322763307, "grad_norm": 0.6788033246994019, "learning_rate": 1.0217699378503092e-05, "loss": 0.3492, "step": 32130 }, { "epoch": 12.13288033220083, "grad_norm": 0.6879256963729858, "learning_rate": 1.0192672753241218e-05, "loss": 0.3575, "step": 32140 }, { "epoch": 12.136655341638354, "grad_norm": 0.623914897441864, "learning_rate": 1.0167673335912164e-05, "loss": 0.3612, "step": 32150 }, { "epoch": 12.140430351075878, "grad_norm": 0.6156293749809265, "learning_rate": 1.0142701143602779e-05, "loss": 0.3612, "step": 32160 }, { "epoch": 12.144205360513402, "grad_norm": 0.6880207657814026, "learning_rate": 1.0117756193381328e-05, "loss": 0.3544, "step": 32170 }, { "epoch": 12.147980369950925, "grad_norm": 0.6038033962249756, "learning_rate": 1.0092838502297453e-05, "loss": 0.3633, "step": 32180 }, { "epoch": 12.15175537938845, "grad_norm": 0.6807112693786621, "learning_rate": 1.0067948087382162e-05, "loss": 0.3554, "step": 32190 }, { "epoch": 12.155530388825973, "grad_norm": 0.6536206007003784, "learning_rate": 1.0043084965647826e-05, "loss": 0.3642, "step": 32200 }, { "epoch": 12.159305398263495, "grad_norm": 0.685413122177124, "learning_rate": 1.001824915408815e-05, "loss": 0.4406, "step": 32210 }, { "epoch": 12.163080407701019, "grad_norm": 0.6319077014923096, "learning_rate": 9.99344066967815e-06, "loss": 0.3569, "step": 32220 }, { "epoch": 12.166855417138542, "grad_norm": 0.6181957125663757, "learning_rate": 9.968659529374219e-06, "loss": 0.3557, "step": 32230 }, { "epoch": 12.170630426576066, "grad_norm": 0.6865522861480713, "learning_rate": 9.943905750114025e-06, "loss": 0.358, "step": 32240 }, { "epoch": 12.17440543601359, "grad_norm": 0.7125778794288635, "learning_rate": 9.919179348816559e-06, "loss": 0.3643, "step": 32250 }, { "epoch": 12.178180445451114, "grad_norm": 0.654378354549408, "learning_rate": 9.894480342382056e-06, "loss": 0.3583, "step": 32260 }, { "epoch": 12.181955454888637, "grad_norm": 0.6028674244880676, "learning_rate": 9.869808747692061e-06, "loss": 0.3586, "step": 32270 }, { "epoch": 12.185730464326161, "grad_norm": 0.6447907090187073, "learning_rate": 9.84516458160939e-06, "loss": 0.3629, "step": 32280 }, { "epoch": 12.189505473763685, "grad_norm": 0.687196671962738, "learning_rate": 9.820547860978097e-06, "loss": 0.3575, "step": 32290 }, { "epoch": 12.193280483201208, "grad_norm": 0.6808565258979797, "learning_rate": 9.795958602623461e-06, "loss": 0.3607, "step": 32300 }, { "epoch": 12.197055492638732, "grad_norm": 0.6441514492034912, "learning_rate": 9.77139682335203e-06, "loss": 0.3544, "step": 32310 }, { "epoch": 12.200830502076256, "grad_norm": 0.5570544600486755, "learning_rate": 9.74686253995153e-06, "loss": 0.3567, "step": 32320 }, { "epoch": 12.204605511513778, "grad_norm": 0.639428436756134, "learning_rate": 9.722355769190911e-06, "loss": 0.3592, "step": 32330 }, { "epoch": 12.208380520951302, "grad_norm": 0.636038064956665, "learning_rate": 9.697876527820333e-06, "loss": 0.3634, "step": 32340 }, { "epoch": 12.212155530388825, "grad_norm": 0.6335992813110352, "learning_rate": 9.673424832571132e-06, "loss": 0.3599, "step": 32350 }, { "epoch": 12.21593053982635, "grad_norm": 0.6807220578193665, "learning_rate": 9.649000700155791e-06, "loss": 0.3591, "step": 32360 }, { "epoch": 12.219705549263873, "grad_norm": 0.7252523303031921, "learning_rate": 9.624604147267984e-06, "loss": 0.3608, "step": 32370 }, { "epoch": 12.223480558701397, "grad_norm": 0.6216152906417847, "learning_rate": 9.600235190582524e-06, "loss": 0.358, "step": 32380 }, { "epoch": 12.22725556813892, "grad_norm": 0.6576191186904907, "learning_rate": 9.57589384675538e-06, "loss": 0.3652, "step": 32390 }, { "epoch": 12.231030577576444, "grad_norm": 0.626547634601593, "learning_rate": 9.551580132423598e-06, "loss": 0.3683, "step": 32400 }, { "epoch": 12.234805587013968, "grad_norm": 0.6060603857040405, "learning_rate": 9.527294064205405e-06, "loss": 0.363, "step": 32410 }, { "epoch": 12.238580596451492, "grad_norm": 0.6533463001251221, "learning_rate": 9.503035658700072e-06, "loss": 0.3593, "step": 32420 }, { "epoch": 12.242355605889015, "grad_norm": 0.6849529147148132, "learning_rate": 9.478804932488e-06, "loss": 0.3578, "step": 32430 }, { "epoch": 12.246130615326539, "grad_norm": 0.6339350938796997, "learning_rate": 9.454601902130667e-06, "loss": 0.3487, "step": 32440 }, { "epoch": 12.249905624764063, "grad_norm": 0.7915511727333069, "learning_rate": 9.430426584170626e-06, "loss": 0.3567, "step": 32450 }, { "epoch": 12.253680634201585, "grad_norm": 0.6535844206809998, "learning_rate": 9.40627899513145e-06, "loss": 0.3514, "step": 32460 }, { "epoch": 12.257455643639108, "grad_norm": 0.6630949378013611, "learning_rate": 9.38215915151781e-06, "loss": 0.3596, "step": 32470 }, { "epoch": 12.261230653076632, "grad_norm": 0.9704565405845642, "learning_rate": 9.358067069815396e-06, "loss": 0.3575, "step": 32480 }, { "epoch": 12.265005662514156, "grad_norm": 0.6713346838951111, "learning_rate": 9.334002766490901e-06, "loss": 0.3604, "step": 32490 }, { "epoch": 12.26878067195168, "grad_norm": 0.6596808433532715, "learning_rate": 9.309966257992054e-06, "loss": 0.3557, "step": 32500 }, { "epoch": 12.272555681389203, "grad_norm": 0.6223334670066833, "learning_rate": 9.2859575607476e-06, "loss": 0.3603, "step": 32510 }, { "epoch": 12.276330690826727, "grad_norm": 0.638768196105957, "learning_rate": 9.26197669116723e-06, "loss": 0.3637, "step": 32520 }, { "epoch": 12.28010570026425, "grad_norm": 0.6910896897315979, "learning_rate": 9.238023665641655e-06, "loss": 0.352, "step": 32530 }, { "epoch": 12.283880709701775, "grad_norm": 0.7031621932983398, "learning_rate": 9.21409850054254e-06, "loss": 0.3612, "step": 32540 }, { "epoch": 12.287655719139298, "grad_norm": 0.6571682095527649, "learning_rate": 9.190201212222526e-06, "loss": 0.3605, "step": 32550 }, { "epoch": 12.291430728576822, "grad_norm": 0.670776903629303, "learning_rate": 9.166331817015155e-06, "loss": 0.3516, "step": 32560 }, { "epoch": 12.295205738014346, "grad_norm": 0.5901675820350647, "learning_rate": 9.142490331234965e-06, "loss": 0.358, "step": 32570 }, { "epoch": 12.298980747451868, "grad_norm": 0.7036232948303223, "learning_rate": 9.11867677117736e-06, "loss": 0.3604, "step": 32580 }, { "epoch": 12.302755756889391, "grad_norm": 0.6272095441818237, "learning_rate": 9.09489115311869e-06, "loss": 0.3574, "step": 32590 }, { "epoch": 12.306530766326915, "grad_norm": 0.7180802226066589, "learning_rate": 9.071133493316202e-06, "loss": 0.3626, "step": 32600 }, { "epoch": 12.310305775764439, "grad_norm": 0.6323314309120178, "learning_rate": 9.047403808008053e-06, "loss": 0.3604, "step": 32610 }, { "epoch": 12.314080785201963, "grad_norm": 0.6044765114784241, "learning_rate": 9.023702113413224e-06, "loss": 0.3679, "step": 32620 }, { "epoch": 12.317855794639486, "grad_norm": 0.7297821044921875, "learning_rate": 9.00002842573162e-06, "loss": 0.3595, "step": 32630 }, { "epoch": 12.32163080407701, "grad_norm": 0.8278955221176147, "learning_rate": 8.976382761143981e-06, "loss": 0.3556, "step": 32640 }, { "epoch": 12.325405813514534, "grad_norm": 0.6725118160247803, "learning_rate": 8.952765135811909e-06, "loss": 0.3635, "step": 32650 }, { "epoch": 12.329180822952058, "grad_norm": 0.6396539807319641, "learning_rate": 8.929175565877795e-06, "loss": 0.3559, "step": 32660 }, { "epoch": 12.332955832389581, "grad_norm": 1.543081521987915, "learning_rate": 8.905614067464924e-06, "loss": 0.3599, "step": 32670 }, { "epoch": 12.336730841827105, "grad_norm": 0.6082138419151306, "learning_rate": 8.882080656677333e-06, "loss": 0.3559, "step": 32680 }, { "epoch": 12.340505851264629, "grad_norm": 0.6150057911872864, "learning_rate": 8.85857534959989e-06, "loss": 0.3602, "step": 32690 }, { "epoch": 12.344280860702153, "grad_norm": 0.8048917651176453, "learning_rate": 8.835098162298256e-06, "loss": 0.3605, "step": 32700 }, { "epoch": 12.348055870139675, "grad_norm": 0.735430896282196, "learning_rate": 8.811649110818875e-06, "loss": 0.353, "step": 32710 }, { "epoch": 12.351830879577198, "grad_norm": 0.6606729030609131, "learning_rate": 8.78822821118893e-06, "loss": 0.3521, "step": 32720 }, { "epoch": 12.355605889014722, "grad_norm": 0.6922546029090881, "learning_rate": 8.764835479416405e-06, "loss": 0.359, "step": 32730 }, { "epoch": 12.359380898452246, "grad_norm": 0.6572328805923462, "learning_rate": 8.741470931490003e-06, "loss": 0.3557, "step": 32740 }, { "epoch": 12.36315590788977, "grad_norm": 0.6792584657669067, "learning_rate": 8.718134583379189e-06, "loss": 0.354, "step": 32750 }, { "epoch": 12.366930917327293, "grad_norm": 0.6346094012260437, "learning_rate": 8.69482645103412e-06, "loss": 0.3549, "step": 32760 }, { "epoch": 12.370705926764817, "grad_norm": 0.6417624950408936, "learning_rate": 8.671546550385679e-06, "loss": 0.3598, "step": 32770 }, { "epoch": 12.37448093620234, "grad_norm": 0.7251724004745483, "learning_rate": 8.648294897345466e-06, "loss": 0.352, "step": 32780 }, { "epoch": 12.378255945639864, "grad_norm": 0.628129243850708, "learning_rate": 8.62507150780576e-06, "loss": 0.3574, "step": 32790 }, { "epoch": 12.382030955077388, "grad_norm": 0.6672044992446899, "learning_rate": 8.601876397639542e-06, "loss": 0.3662, "step": 32800 }, { "epoch": 12.385805964514912, "grad_norm": 0.614966630935669, "learning_rate": 8.578709582700456e-06, "loss": 0.3499, "step": 32810 }, { "epoch": 12.389580973952436, "grad_norm": 0.629789412021637, "learning_rate": 8.555571078822783e-06, "loss": 0.358, "step": 32820 }, { "epoch": 12.39335598338996, "grad_norm": 0.6185657978057861, "learning_rate": 8.532460901821477e-06, "loss": 0.3626, "step": 32830 }, { "epoch": 12.397130992827481, "grad_norm": 0.6272408366203308, "learning_rate": 8.509379067492128e-06, "loss": 0.3543, "step": 32840 }, { "epoch": 12.400906002265005, "grad_norm": 0.7130616903305054, "learning_rate": 8.486325591610967e-06, "loss": 0.3542, "step": 32850 }, { "epoch": 12.404681011702529, "grad_norm": 0.6526590585708618, "learning_rate": 8.463300489934811e-06, "loss": 0.3627, "step": 32860 }, { "epoch": 12.408456021140053, "grad_norm": 0.6703090071678162, "learning_rate": 8.440303778201093e-06, "loss": 0.3571, "step": 32870 }, { "epoch": 12.412231030577576, "grad_norm": 0.6281063556671143, "learning_rate": 8.417335472127858e-06, "loss": 0.3667, "step": 32880 }, { "epoch": 12.4160060400151, "grad_norm": 0.6609539985656738, "learning_rate": 8.394395587413722e-06, "loss": 0.3607, "step": 32890 }, { "epoch": 12.419781049452624, "grad_norm": 0.6523950099945068, "learning_rate": 8.371484139737889e-06, "loss": 0.3728, "step": 32900 }, { "epoch": 12.423556058890147, "grad_norm": 0.6022361516952515, "learning_rate": 8.348601144760127e-06, "loss": 0.3586, "step": 32910 }, { "epoch": 12.427331068327671, "grad_norm": 0.6702349185943604, "learning_rate": 8.325746618120716e-06, "loss": 0.3576, "step": 32920 }, { "epoch": 12.431106077765195, "grad_norm": 0.6322970986366272, "learning_rate": 8.302920575440531e-06, "loss": 0.356, "step": 32930 }, { "epoch": 12.434881087202719, "grad_norm": 0.6621061563491821, "learning_rate": 8.280123032320953e-06, "loss": 0.3598, "step": 32940 }, { "epoch": 12.438656096640242, "grad_norm": 2.3764612674713135, "learning_rate": 8.257354004343892e-06, "loss": 0.3598, "step": 32950 }, { "epoch": 12.442431106077764, "grad_norm": 0.6495422720909119, "learning_rate": 8.234613507071754e-06, "loss": 0.3675, "step": 32960 }, { "epoch": 12.446206115515288, "grad_norm": 0.5902871489524841, "learning_rate": 8.211901556047475e-06, "loss": 0.3548, "step": 32970 }, { "epoch": 12.449981124952812, "grad_norm": 0.597322404384613, "learning_rate": 8.189218166794427e-06, "loss": 0.35, "step": 32980 }, { "epoch": 12.453756134390336, "grad_norm": 0.656929075717926, "learning_rate": 8.166563354816509e-06, "loss": 0.3605, "step": 32990 }, { "epoch": 12.45753114382786, "grad_norm": 0.6605633497238159, "learning_rate": 8.143937135598074e-06, "loss": 0.3512, "step": 33000 }, { "epoch": 12.461306153265383, "grad_norm": 0.6512786149978638, "learning_rate": 8.121339524603943e-06, "loss": 0.3524, "step": 33010 }, { "epoch": 12.465081162702907, "grad_norm": 0.8367393016815186, "learning_rate": 8.098770537279338e-06, "loss": 0.3565, "step": 33020 }, { "epoch": 12.46885617214043, "grad_norm": 0.602468729019165, "learning_rate": 8.076230189049966e-06, "loss": 0.3513, "step": 33030 }, { "epoch": 12.472631181577954, "grad_norm": 0.7088872790336609, "learning_rate": 8.05371849532196e-06, "loss": 0.3584, "step": 33040 }, { "epoch": 12.476406191015478, "grad_norm": 0.7384507060050964, "learning_rate": 8.031235471481813e-06, "loss": 0.3559, "step": 33050 }, { "epoch": 12.480181200453002, "grad_norm": 0.714253842830658, "learning_rate": 8.008781132896486e-06, "loss": 0.3569, "step": 33060 }, { "epoch": 12.483956209890525, "grad_norm": 0.6598066091537476, "learning_rate": 7.986355494913312e-06, "loss": 0.3623, "step": 33070 }, { "epoch": 12.487731219328047, "grad_norm": 0.7399073839187622, "learning_rate": 7.963958572859975e-06, "loss": 0.3538, "step": 33080 }, { "epoch": 12.491506228765571, "grad_norm": 0.6438301205635071, "learning_rate": 7.941590382044583e-06, "loss": 0.3642, "step": 33090 }, { "epoch": 12.495281238203095, "grad_norm": 0.6220930814743042, "learning_rate": 7.919250937755574e-06, "loss": 0.3514, "step": 33100 }, { "epoch": 12.499056247640619, "grad_norm": 0.5921124815940857, "learning_rate": 7.896940255261764e-06, "loss": 0.3619, "step": 33110 }, { "epoch": 12.502831257078142, "grad_norm": 0.6638910174369812, "learning_rate": 7.874658349812269e-06, "loss": 0.3648, "step": 33120 }, { "epoch": 12.506606266515666, "grad_norm": 0.6646683812141418, "learning_rate": 7.85240523663659e-06, "loss": 0.3603, "step": 33130 }, { "epoch": 12.51038127595319, "grad_norm": 0.6426159739494324, "learning_rate": 7.830180930944487e-06, "loss": 0.3619, "step": 33140 }, { "epoch": 12.514156285390714, "grad_norm": 0.6459843516349792, "learning_rate": 7.80798544792608e-06, "loss": 0.3581, "step": 33150 }, { "epoch": 12.517931294828237, "grad_norm": 0.7059284448623657, "learning_rate": 7.785818802751776e-06, "loss": 0.3593, "step": 33160 }, { "epoch": 12.521706304265761, "grad_norm": 0.5886619687080383, "learning_rate": 7.763681010572277e-06, "loss": 0.3711, "step": 33170 }, { "epoch": 12.525481313703285, "grad_norm": 0.6966323852539062, "learning_rate": 7.741572086518528e-06, "loss": 0.3547, "step": 33180 }, { "epoch": 12.529256323140809, "grad_norm": 0.6798071265220642, "learning_rate": 7.719492045701782e-06, "loss": 0.358, "step": 33190 }, { "epoch": 12.533031332578332, "grad_norm": 0.7177457213401794, "learning_rate": 7.697440903213549e-06, "loss": 0.3652, "step": 33200 }, { "epoch": 12.536806342015854, "grad_norm": 0.7038207054138184, "learning_rate": 7.675418674125573e-06, "loss": 0.3558, "step": 33210 }, { "epoch": 12.540581351453378, "grad_norm": 0.7092708945274353, "learning_rate": 7.65342537348983e-06, "loss": 0.3569, "step": 33220 }, { "epoch": 12.544356360890902, "grad_norm": 0.6563994884490967, "learning_rate": 7.631461016338549e-06, "loss": 0.3582, "step": 33230 }, { "epoch": 12.548131370328425, "grad_norm": 0.6491038799285889, "learning_rate": 7.609525617684138e-06, "loss": 0.3556, "step": 33240 }, { "epoch": 12.55190637976595, "grad_norm": 0.6468068957328796, "learning_rate": 7.587619192519252e-06, "loss": 0.3535, "step": 33250 }, { "epoch": 12.555681389203473, "grad_norm": 0.6689091920852661, "learning_rate": 7.565741755816725e-06, "loss": 0.363, "step": 33260 }, { "epoch": 12.559456398640997, "grad_norm": 0.6048831939697266, "learning_rate": 7.543893322529588e-06, "loss": 0.358, "step": 33270 }, { "epoch": 12.56323140807852, "grad_norm": 0.6832661032676697, "learning_rate": 7.522073907591015e-06, "loss": 0.3602, "step": 33280 }, { "epoch": 12.567006417516044, "grad_norm": 0.6341314911842346, "learning_rate": 7.500283525914381e-06, "loss": 0.358, "step": 33290 }, { "epoch": 12.570781426953568, "grad_norm": 0.6825064420700073, "learning_rate": 7.478522192393211e-06, "loss": 0.3489, "step": 33300 }, { "epoch": 12.574556436391092, "grad_norm": 0.5636059045791626, "learning_rate": 7.456789921901181e-06, "loss": 0.3566, "step": 33310 }, { "epoch": 12.578331445828615, "grad_norm": 0.7039142847061157, "learning_rate": 7.435086729292079e-06, "loss": 0.357, "step": 33320 }, { "epoch": 12.582106455266139, "grad_norm": 0.6117770075798035, "learning_rate": 7.413412629399824e-06, "loss": 0.3609, "step": 33330 }, { "epoch": 12.585881464703661, "grad_norm": 0.6217195391654968, "learning_rate": 7.391767637038466e-06, "loss": 0.3557, "step": 33340 }, { "epoch": 12.589656474141185, "grad_norm": 0.7603038549423218, "learning_rate": 7.370151767002159e-06, "loss": 0.3557, "step": 33350 }, { "epoch": 12.593431483578708, "grad_norm": 0.660520613193512, "learning_rate": 7.348565034065147e-06, "loss": 0.3648, "step": 33360 }, { "epoch": 12.597206493016232, "grad_norm": 0.6169815063476562, "learning_rate": 7.327007452981765e-06, "loss": 0.3573, "step": 33370 }, { "epoch": 12.600981502453756, "grad_norm": 0.59542316198349, "learning_rate": 7.305479038486396e-06, "loss": 0.3576, "step": 33380 }, { "epoch": 12.60475651189128, "grad_norm": 0.6083225011825562, "learning_rate": 7.283979805293517e-06, "loss": 0.353, "step": 33390 }, { "epoch": 12.608531521328803, "grad_norm": 0.6213321089744568, "learning_rate": 7.262509768097653e-06, "loss": 0.3552, "step": 33400 }, { "epoch": 12.612306530766327, "grad_norm": 0.6928298473358154, "learning_rate": 7.241068941573381e-06, "loss": 0.3615, "step": 33410 }, { "epoch": 12.61608154020385, "grad_norm": 0.6654199957847595, "learning_rate": 7.219657340375296e-06, "loss": 0.3566, "step": 33420 }, { "epoch": 12.619856549641375, "grad_norm": 0.6036897301673889, "learning_rate": 7.198274979138014e-06, "loss": 0.3679, "step": 33430 }, { "epoch": 12.623631559078898, "grad_norm": 0.685053825378418, "learning_rate": 7.176921872476189e-06, "loss": 0.3523, "step": 33440 }, { "epoch": 12.627406568516422, "grad_norm": 0.6758593320846558, "learning_rate": 7.1555980349844674e-06, "loss": 0.3555, "step": 33450 }, { "epoch": 12.631181577953946, "grad_norm": 1.1203943490982056, "learning_rate": 7.134303481237492e-06, "loss": 0.3602, "step": 33460 }, { "epoch": 12.634956587391468, "grad_norm": 0.6802670955657959, "learning_rate": 7.113038225789892e-06, "loss": 0.3648, "step": 33470 }, { "epoch": 12.638731596828991, "grad_norm": 0.6041070222854614, "learning_rate": 7.091802283176258e-06, "loss": 0.3545, "step": 33480 }, { "epoch": 12.642506606266515, "grad_norm": 0.6177239418029785, "learning_rate": 7.070595667911162e-06, "loss": 0.3596, "step": 33490 }, { "epoch": 12.646281615704039, "grad_norm": 0.6666644215583801, "learning_rate": 7.049418394489127e-06, "loss": 0.3587, "step": 33500 }, { "epoch": 12.650056625141563, "grad_norm": 0.7733400464057922, "learning_rate": 7.028270477384636e-06, "loss": 0.3585, "step": 33510 }, { "epoch": 12.653831634579086, "grad_norm": 0.8025512099266052, "learning_rate": 7.0071519310520695e-06, "loss": 0.3509, "step": 33520 }, { "epoch": 12.65760664401661, "grad_norm": 0.6963528990745544, "learning_rate": 6.98606276992575e-06, "loss": 0.3549, "step": 33530 }, { "epoch": 12.661381653454134, "grad_norm": 0.705504298210144, "learning_rate": 6.965003008419929e-06, "loss": 0.3559, "step": 33540 }, { "epoch": 12.665156662891658, "grad_norm": 0.7005383968353271, "learning_rate": 6.943972660928755e-06, "loss": 0.3532, "step": 33550 }, { "epoch": 12.668931672329181, "grad_norm": 0.7059851884841919, "learning_rate": 6.922971741826273e-06, "loss": 0.3553, "step": 33560 }, { "epoch": 12.672706681766705, "grad_norm": 0.6162865161895752, "learning_rate": 6.902000265466424e-06, "loss": 0.3571, "step": 33570 }, { "epoch": 12.676481691204227, "grad_norm": 0.6487360596656799, "learning_rate": 6.881058246182998e-06, "loss": 0.36, "step": 33580 }, { "epoch": 12.68025670064175, "grad_norm": 0.63123619556427, "learning_rate": 6.860145698289666e-06, "loss": 0.3617, "step": 33590 }, { "epoch": 12.684031710079275, "grad_norm": 0.6987262964248657, "learning_rate": 6.839262636079985e-06, "loss": 0.3586, "step": 33600 }, { "epoch": 12.687806719516798, "grad_norm": 0.6346328258514404, "learning_rate": 6.818409073827292e-06, "loss": 0.3589, "step": 33610 }, { "epoch": 12.691581728954322, "grad_norm": 0.6731489896774292, "learning_rate": 6.79758502578483e-06, "loss": 0.361, "step": 33620 }, { "epoch": 12.695356738391846, "grad_norm": 0.6909154653549194, "learning_rate": 6.776790506185621e-06, "loss": 0.3558, "step": 33630 }, { "epoch": 12.69913174782937, "grad_norm": 0.6457541584968567, "learning_rate": 6.756025529242533e-06, "loss": 0.3564, "step": 33640 }, { "epoch": 12.702906757266893, "grad_norm": 0.6108506321907043, "learning_rate": 6.735290109148223e-06, "loss": 0.3597, "step": 33650 }, { "epoch": 12.706681766704417, "grad_norm": 0.699229896068573, "learning_rate": 6.714584260075163e-06, "loss": 0.3595, "step": 33660 }, { "epoch": 12.71045677614194, "grad_norm": 0.6555954813957214, "learning_rate": 6.693907996175619e-06, "loss": 0.3443, "step": 33670 }, { "epoch": 12.714231785579464, "grad_norm": 0.6641535758972168, "learning_rate": 6.673261331581593e-06, "loss": 0.3564, "step": 33680 }, { "epoch": 12.718006795016988, "grad_norm": 0.6798548698425293, "learning_rate": 6.652644280404912e-06, "loss": 0.3535, "step": 33690 }, { "epoch": 12.721781804454512, "grad_norm": 0.6146652698516846, "learning_rate": 6.6320568567371175e-06, "loss": 0.3541, "step": 33700 }, { "epoch": 12.725556813892034, "grad_norm": 0.6196102499961853, "learning_rate": 6.611499074649518e-06, "loss": 0.3546, "step": 33710 }, { "epoch": 12.729331823329558, "grad_norm": 0.7105995416641235, "learning_rate": 6.590970948193187e-06, "loss": 0.3663, "step": 33720 }, { "epoch": 12.733106832767081, "grad_norm": 0.6891978979110718, "learning_rate": 6.5704724913988825e-06, "loss": 0.3672, "step": 33730 }, { "epoch": 12.736881842204605, "grad_norm": 0.6553974151611328, "learning_rate": 6.550003718277109e-06, "loss": 0.356, "step": 33740 }, { "epoch": 12.740656851642129, "grad_norm": 0.699963390827179, "learning_rate": 6.529564642818087e-06, "loss": 0.3572, "step": 33750 }, { "epoch": 12.744431861079653, "grad_norm": 0.6198428869247437, "learning_rate": 6.50915527899173e-06, "loss": 0.3634, "step": 33760 }, { "epoch": 12.748206870517176, "grad_norm": 0.6034395098686218, "learning_rate": 6.4887756407476575e-06, "loss": 0.3631, "step": 33770 }, { "epoch": 12.7519818799547, "grad_norm": 0.6118060946464539, "learning_rate": 6.468425742015149e-06, "loss": 0.3553, "step": 33780 }, { "epoch": 12.755756889392224, "grad_norm": 0.6215893626213074, "learning_rate": 6.448105596703152e-06, "loss": 0.3587, "step": 33790 }, { "epoch": 12.759531898829747, "grad_norm": 0.6398429274559021, "learning_rate": 6.427815218700317e-06, "loss": 0.3582, "step": 33800 }, { "epoch": 12.763306908267271, "grad_norm": 0.6768361330032349, "learning_rate": 6.407554621874911e-06, "loss": 0.3592, "step": 33810 }, { "epoch": 12.767081917704795, "grad_norm": 0.6989614963531494, "learning_rate": 6.3873238200748864e-06, "loss": 0.3553, "step": 33820 }, { "epoch": 12.770856927142319, "grad_norm": 0.7120910882949829, "learning_rate": 6.367122827127775e-06, "loss": 0.3587, "step": 33830 }, { "epoch": 12.77463193657984, "grad_norm": 0.7099307179450989, "learning_rate": 6.346951656840777e-06, "loss": 0.3575, "step": 33840 }, { "epoch": 12.778406946017364, "grad_norm": 0.6789335012435913, "learning_rate": 6.326810323000698e-06, "loss": 0.3565, "step": 33850 }, { "epoch": 12.782181955454888, "grad_norm": 0.6464335322380066, "learning_rate": 6.306698839373948e-06, "loss": 0.3583, "step": 33860 }, { "epoch": 12.785956964892412, "grad_norm": 0.7024797201156616, "learning_rate": 6.286617219706547e-06, "loss": 0.3623, "step": 33870 }, { "epoch": 12.789731974329936, "grad_norm": 0.6883578896522522, "learning_rate": 6.26656547772409e-06, "loss": 0.3593, "step": 33880 }, { "epoch": 12.79350698376746, "grad_norm": 0.656173050403595, "learning_rate": 6.246543627131734e-06, "loss": 0.3568, "step": 33890 }, { "epoch": 12.797281993204983, "grad_norm": 0.593108594417572, "learning_rate": 6.22655168161424e-06, "loss": 0.3583, "step": 33900 }, { "epoch": 12.801057002642507, "grad_norm": 0.7652211785316467, "learning_rate": 6.20658965483591e-06, "loss": 0.3587, "step": 33910 }, { "epoch": 12.80483201208003, "grad_norm": 0.6598030924797058, "learning_rate": 6.186657560440612e-06, "loss": 0.3575, "step": 33920 }, { "epoch": 12.808607021517554, "grad_norm": 0.6479567289352417, "learning_rate": 6.166755412051734e-06, "loss": 0.3528, "step": 33930 }, { "epoch": 12.812382030955078, "grad_norm": 0.7282853722572327, "learning_rate": 6.1468832232721985e-06, "loss": 0.3615, "step": 33940 }, { "epoch": 12.816157040392602, "grad_norm": 0.670918345451355, "learning_rate": 6.127041007684476e-06, "loss": 0.364, "step": 33950 }, { "epoch": 12.819932049830125, "grad_norm": 0.6073424220085144, "learning_rate": 6.107228778850527e-06, "loss": 0.3621, "step": 33960 }, { "epoch": 12.823707059267647, "grad_norm": 0.6667119264602661, "learning_rate": 6.087446550311837e-06, "loss": 0.3633, "step": 33970 }, { "epoch": 12.827482068705171, "grad_norm": 0.6431853175163269, "learning_rate": 6.067694335589353e-06, "loss": 0.3569, "step": 33980 }, { "epoch": 12.831257078142695, "grad_norm": 0.74583899974823, "learning_rate": 6.047972148183528e-06, "loss": 0.3595, "step": 33990 }, { "epoch": 12.835032087580219, "grad_norm": 0.7069852948188782, "learning_rate": 6.0282800015742936e-06, "loss": 0.3536, "step": 34000 }, { "epoch": 12.838807097017742, "grad_norm": 0.6726012825965881, "learning_rate": 6.008617909221054e-06, "loss": 0.3586, "step": 34010 }, { "epoch": 12.842582106455266, "grad_norm": 0.704440712928772, "learning_rate": 5.988985884562664e-06, "loss": 0.357, "step": 34020 }, { "epoch": 12.84635711589279, "grad_norm": 0.6290958523750305, "learning_rate": 5.969383941017414e-06, "loss": 0.3569, "step": 34030 }, { "epoch": 12.850132125330314, "grad_norm": 0.6034786105155945, "learning_rate": 5.949812091983059e-06, "loss": 0.3541, "step": 34040 }, { "epoch": 12.853907134767837, "grad_norm": 0.6244705319404602, "learning_rate": 5.9302703508367695e-06, "loss": 0.3569, "step": 34050 }, { "epoch": 12.857682144205361, "grad_norm": 0.6468330025672913, "learning_rate": 5.910758730935157e-06, "loss": 0.3611, "step": 34060 }, { "epoch": 12.861457153642885, "grad_norm": 0.6652206182479858, "learning_rate": 5.8912772456142016e-06, "loss": 0.3511, "step": 34070 }, { "epoch": 12.865232163080409, "grad_norm": 0.6235213875770569, "learning_rate": 5.871825908189338e-06, "loss": 0.3538, "step": 34080 }, { "epoch": 12.86900717251793, "grad_norm": 0.6621882319450378, "learning_rate": 5.852404731955352e-06, "loss": 0.3609, "step": 34090 }, { "epoch": 12.872782181955454, "grad_norm": 0.6129846572875977, "learning_rate": 5.833013730186443e-06, "loss": 0.3619, "step": 34100 }, { "epoch": 12.876557191392978, "grad_norm": 0.6356860399246216, "learning_rate": 5.813652916136181e-06, "loss": 0.3535, "step": 34110 }, { "epoch": 12.880332200830502, "grad_norm": 0.6019640564918518, "learning_rate": 5.794322303037508e-06, "loss": 0.3631, "step": 34120 }, { "epoch": 12.884107210268025, "grad_norm": 0.7462950348854065, "learning_rate": 5.775021904102701e-06, "loss": 0.3548, "step": 34130 }, { "epoch": 12.88788221970555, "grad_norm": 0.8151988983154297, "learning_rate": 5.755751732523401e-06, "loss": 0.3544, "step": 34140 }, { "epoch": 12.891657229143073, "grad_norm": 0.6986622214317322, "learning_rate": 5.736511801470595e-06, "loss": 0.3494, "step": 34150 }, { "epoch": 12.895432238580597, "grad_norm": 0.6462591290473938, "learning_rate": 5.717302124094603e-06, "loss": 0.368, "step": 34160 }, { "epoch": 12.89920724801812, "grad_norm": 0.6205435991287231, "learning_rate": 5.6981227135250396e-06, "loss": 0.3582, "step": 34170 }, { "epoch": 12.902982257455644, "grad_norm": 4.8062543869018555, "learning_rate": 5.678973582870867e-06, "loss": 0.3628, "step": 34180 }, { "epoch": 12.906757266893168, "grad_norm": 0.6735921502113342, "learning_rate": 5.659854745220317e-06, "loss": 0.3619, "step": 34190 }, { "epoch": 12.910532276330692, "grad_norm": 0.5726591944694519, "learning_rate": 5.640766213640947e-06, "loss": 0.3549, "step": 34200 }, { "epoch": 12.914307285768214, "grad_norm": 0.6301986575126648, "learning_rate": 5.6217080011795934e-06, "loss": 0.3588, "step": 34210 }, { "epoch": 12.918082295205737, "grad_norm": 0.6712170243263245, "learning_rate": 5.602680120862364e-06, "loss": 0.3573, "step": 34220 }, { "epoch": 12.921857304643261, "grad_norm": 0.7288546562194824, "learning_rate": 5.583682585694616e-06, "loss": 0.3583, "step": 34230 }, { "epoch": 12.925632314080785, "grad_norm": 0.6534035801887512, "learning_rate": 5.564715408661009e-06, "loss": 0.3538, "step": 34240 }, { "epoch": 12.929407323518308, "grad_norm": 0.6488263607025146, "learning_rate": 5.5457786027254265e-06, "loss": 0.3587, "step": 34250 }, { "epoch": 12.933182332955832, "grad_norm": 0.6462862491607666, "learning_rate": 5.526872180830977e-06, "loss": 0.3626, "step": 34260 }, { "epoch": 12.936957342393356, "grad_norm": 0.6145511269569397, "learning_rate": 5.507996155900041e-06, "loss": 0.3568, "step": 34270 }, { "epoch": 12.94073235183088, "grad_norm": 0.6510082483291626, "learning_rate": 5.489150540834203e-06, "loss": 0.3537, "step": 34280 }, { "epoch": 12.944507361268403, "grad_norm": 0.7098113298416138, "learning_rate": 5.47033534851425e-06, "loss": 0.3498, "step": 34290 }, { "epoch": 12.948282370705927, "grad_norm": 0.6072222590446472, "learning_rate": 5.451550591800198e-06, "loss": 0.3544, "step": 34300 }, { "epoch": 12.95205738014345, "grad_norm": 0.6960251331329346, "learning_rate": 5.43279628353125e-06, "loss": 0.3611, "step": 34310 }, { "epoch": 12.955832389580975, "grad_norm": 0.6805595755577087, "learning_rate": 5.4140724365258075e-06, "loss": 0.3595, "step": 34320 }, { "epoch": 12.959607399018498, "grad_norm": 0.651698648929596, "learning_rate": 5.395379063581424e-06, "loss": 0.3566, "step": 34330 }, { "epoch": 12.96338240845602, "grad_norm": 0.6734188199043274, "learning_rate": 5.376716177474866e-06, "loss": 0.3475, "step": 34340 }, { "epoch": 12.967157417893544, "grad_norm": 0.6215294599533081, "learning_rate": 5.358083790962015e-06, "loss": 0.3579, "step": 34350 }, { "epoch": 12.970932427331068, "grad_norm": 0.7030113339424133, "learning_rate": 5.33948191677795e-06, "loss": 0.356, "step": 34360 }, { "epoch": 12.974707436768592, "grad_norm": 0.6129488945007324, "learning_rate": 5.320910567636866e-06, "loss": 0.3583, "step": 34370 }, { "epoch": 12.978482446206115, "grad_norm": 2.6713273525238037, "learning_rate": 5.302369756232123e-06, "loss": 0.3658, "step": 34380 }, { "epoch": 12.982257455643639, "grad_norm": 0.6619277596473694, "learning_rate": 5.283859495236165e-06, "loss": 0.3596, "step": 34390 }, { "epoch": 12.986032465081163, "grad_norm": 0.6460644006729126, "learning_rate": 5.265379797300602e-06, "loss": 0.352, "step": 34400 }, { "epoch": 12.989807474518686, "grad_norm": 0.6166976690292358, "learning_rate": 5.246930675056117e-06, "loss": 0.364, "step": 34410 }, { "epoch": 12.99358248395621, "grad_norm": 0.5918395519256592, "learning_rate": 5.228512141112535e-06, "loss": 0.3565, "step": 34420 }, { "epoch": 12.997357493393734, "grad_norm": 3.736072063446045, "learning_rate": 5.21012420805872e-06, "loss": 0.4422, "step": 34430 }, { "epoch": 13.001132502831258, "grad_norm": 0.6360698342323303, "learning_rate": 5.191766888462679e-06, "loss": 0.3608, "step": 34440 }, { "epoch": 13.004907512268781, "grad_norm": 1.15024733543396, "learning_rate": 5.173440194871437e-06, "loss": 0.3568, "step": 34450 }, { "epoch": 13.008682521706305, "grad_norm": 0.6446855068206787, "learning_rate": 5.155144139811141e-06, "loss": 0.3572, "step": 34460 }, { "epoch": 13.012457531143827, "grad_norm": 0.6619588732719421, "learning_rate": 5.13687873578696e-06, "loss": 0.3568, "step": 34470 }, { "epoch": 13.01623254058135, "grad_norm": 0.7428315281867981, "learning_rate": 5.11864399528314e-06, "loss": 0.3547, "step": 34480 }, { "epoch": 13.020007550018875, "grad_norm": 0.6570870876312256, "learning_rate": 5.1004399307629346e-06, "loss": 0.3606, "step": 34490 }, { "epoch": 13.023782559456398, "grad_norm": 0.660350501537323, "learning_rate": 5.082266554668652e-06, "loss": 0.3647, "step": 34500 }, { "epoch": 13.027557568893922, "grad_norm": 0.8094748258590698, "learning_rate": 5.06412387942164e-06, "loss": 0.3574, "step": 34510 }, { "epoch": 13.031332578331446, "grad_norm": 0.669248640537262, "learning_rate": 5.046011917422239e-06, "loss": 0.3524, "step": 34520 }, { "epoch": 13.03510758776897, "grad_norm": 0.6466529369354248, "learning_rate": 5.0279306810498026e-06, "loss": 0.3646, "step": 34530 }, { "epoch": 13.038882597206493, "grad_norm": 0.6580777168273926, "learning_rate": 5.009880182662679e-06, "loss": 0.3482, "step": 34540 }, { "epoch": 13.042657606644017, "grad_norm": 1.0861129760742188, "learning_rate": 4.991860434598217e-06, "loss": 0.3527, "step": 34550 }, { "epoch": 13.04643261608154, "grad_norm": 0.6849576830863953, "learning_rate": 4.973871449172746e-06, "loss": 0.3616, "step": 34560 }, { "epoch": 13.050207625519064, "grad_norm": 0.6503666043281555, "learning_rate": 4.9559132386815695e-06, "loss": 0.3614, "step": 34570 }, { "epoch": 13.053982634956588, "grad_norm": 0.6452770233154297, "learning_rate": 4.93798581539896e-06, "loss": 0.3598, "step": 34580 }, { "epoch": 13.05775764439411, "grad_norm": 0.6467193961143494, "learning_rate": 4.920089191578131e-06, "loss": 0.36, "step": 34590 }, { "epoch": 13.061532653831634, "grad_norm": 0.6155425906181335, "learning_rate": 4.90222337945126e-06, "loss": 0.3565, "step": 34600 }, { "epoch": 13.065307663269158, "grad_norm": 0.687741219997406, "learning_rate": 4.8843883912294605e-06, "loss": 0.3541, "step": 34610 }, { "epoch": 13.069082672706681, "grad_norm": 0.6692128777503967, "learning_rate": 4.866584239102795e-06, "loss": 0.3546, "step": 34620 }, { "epoch": 13.072857682144205, "grad_norm": 1.0037407875061035, "learning_rate": 4.8488109352402164e-06, "loss": 0.3644, "step": 34630 }, { "epoch": 13.076632691581729, "grad_norm": 0.6341879963874817, "learning_rate": 4.831068491789603e-06, "loss": 0.3578, "step": 34640 }, { "epoch": 13.080407701019253, "grad_norm": 0.6491614580154419, "learning_rate": 4.81335692087776e-06, "loss": 0.3579, "step": 34650 }, { "epoch": 13.084182710456776, "grad_norm": 0.6379111409187317, "learning_rate": 4.795676234610374e-06, "loss": 0.3643, "step": 34660 }, { "epoch": 13.0879577198943, "grad_norm": 0.678970992565155, "learning_rate": 4.778026445072037e-06, "loss": 0.3638, "step": 34670 }, { "epoch": 13.091732729331824, "grad_norm": 0.6357844471931458, "learning_rate": 4.760407564326208e-06, "loss": 0.3521, "step": 34680 }, { "epoch": 13.095507738769347, "grad_norm": 0.5853003859519958, "learning_rate": 4.742819604415222e-06, "loss": 0.3604, "step": 34690 }, { "epoch": 13.099282748206871, "grad_norm": 0.7209833264350891, "learning_rate": 4.725262577360284e-06, "loss": 0.3599, "step": 34700 }, { "epoch": 13.103057757644395, "grad_norm": 0.6747405529022217, "learning_rate": 4.707736495161458e-06, "loss": 0.3538, "step": 34710 }, { "epoch": 13.106832767081917, "grad_norm": 0.6171832084655762, "learning_rate": 4.690241369797671e-06, "loss": 0.3574, "step": 34720 }, { "epoch": 13.11060777651944, "grad_norm": 0.9365268349647522, "learning_rate": 4.672777213226659e-06, "loss": 0.3562, "step": 34730 }, { "epoch": 13.114382785956964, "grad_norm": 0.6216146945953369, "learning_rate": 4.655344037385001e-06, "loss": 0.3577, "step": 34740 }, { "epoch": 13.118157795394488, "grad_norm": 0.6404061317443848, "learning_rate": 4.6379418541881215e-06, "loss": 0.3617, "step": 34750 }, { "epoch": 13.121932804832012, "grad_norm": 0.638793408870697, "learning_rate": 4.620570675530245e-06, "loss": 0.3593, "step": 34760 }, { "epoch": 13.125707814269536, "grad_norm": 0.6440224647521973, "learning_rate": 4.603230513284407e-06, "loss": 0.3637, "step": 34770 }, { "epoch": 13.12948282370706, "grad_norm": 0.7139766216278076, "learning_rate": 4.58592137930246e-06, "loss": 0.3537, "step": 34780 }, { "epoch": 13.133257833144583, "grad_norm": 0.752666711807251, "learning_rate": 4.568643285415009e-06, "loss": 0.3647, "step": 34790 }, { "epoch": 13.137032842582107, "grad_norm": 0.7453562617301941, "learning_rate": 4.551396243431483e-06, "loss": 0.3593, "step": 34800 }, { "epoch": 13.14080785201963, "grad_norm": 0.6269282698631287, "learning_rate": 4.534180265140075e-06, "loss": 0.3643, "step": 34810 }, { "epoch": 13.144582861457154, "grad_norm": 0.7206990122795105, "learning_rate": 4.5169953623077356e-06, "loss": 0.3554, "step": 34820 }, { "epoch": 13.148357870894678, "grad_norm": 0.6602156758308411, "learning_rate": 4.499841546680189e-06, "loss": 0.3563, "step": 34830 }, { "epoch": 13.1521328803322, "grad_norm": 0.6244385838508606, "learning_rate": 4.4827188299819025e-06, "loss": 0.348, "step": 34840 }, { "epoch": 13.155907889769724, "grad_norm": 0.6173793077468872, "learning_rate": 4.465627223916086e-06, "loss": 0.359, "step": 34850 }, { "epoch": 13.159682899207247, "grad_norm": 0.6792297959327698, "learning_rate": 4.448566740164706e-06, "loss": 0.3597, "step": 34860 }, { "epoch": 13.163457908644771, "grad_norm": 0.6277993321418762, "learning_rate": 4.431537390388429e-06, "loss": 0.3617, "step": 34870 }, { "epoch": 13.167232918082295, "grad_norm": 0.6655922532081604, "learning_rate": 4.414539186226674e-06, "loss": 0.353, "step": 34880 }, { "epoch": 13.171007927519819, "grad_norm": 0.670093834400177, "learning_rate": 4.397572139297529e-06, "loss": 0.3579, "step": 34890 }, { "epoch": 13.174782936957342, "grad_norm": 0.6460064649581909, "learning_rate": 4.380636261197829e-06, "loss": 0.3587, "step": 34900 }, { "epoch": 13.178557946394866, "grad_norm": 0.5938677191734314, "learning_rate": 4.36373156350307e-06, "loss": 0.3583, "step": 34910 }, { "epoch": 13.18233295583239, "grad_norm": 0.6573899388313293, "learning_rate": 4.3468580577674665e-06, "loss": 0.3568, "step": 34920 }, { "epoch": 13.186107965269914, "grad_norm": 0.6724887490272522, "learning_rate": 4.330015755523908e-06, "loss": 0.3636, "step": 34930 }, { "epoch": 13.189882974707437, "grad_norm": 0.8144491910934448, "learning_rate": 4.313204668283926e-06, "loss": 0.3654, "step": 34940 }, { "epoch": 13.193657984144961, "grad_norm": 0.5686742663383484, "learning_rate": 4.296424807537752e-06, "loss": 0.3628, "step": 34950 }, { "epoch": 13.197432993582485, "grad_norm": 0.9146296977996826, "learning_rate": 4.279676184754267e-06, "loss": 0.3596, "step": 34960 }, { "epoch": 13.201208003020007, "grad_norm": 0.639041006565094, "learning_rate": 4.262958811380985e-06, "loss": 0.3575, "step": 34970 }, { "epoch": 13.20498301245753, "grad_norm": 0.7431603074073792, "learning_rate": 4.246272698844095e-06, "loss": 0.3555, "step": 34980 }, { "epoch": 13.208758021895054, "grad_norm": 0.6070488691329956, "learning_rate": 4.229617858548374e-06, "loss": 0.3574, "step": 34990 }, { "epoch": 13.212533031332578, "grad_norm": 1.7812660932540894, "learning_rate": 4.212994301877271e-06, "loss": 0.3628, "step": 35000 }, { "epoch": 13.216308040770102, "grad_norm": 0.763363242149353, "learning_rate": 4.196402040192804e-06, "loss": 0.3554, "step": 35010 }, { "epoch": 13.220083050207625, "grad_norm": 1.0130776166915894, "learning_rate": 4.179841084835639e-06, "loss": 0.3557, "step": 35020 }, { "epoch": 13.22385805964515, "grad_norm": 0.64314204454422, "learning_rate": 4.163311447125046e-06, "loss": 0.3546, "step": 35030 }, { "epoch": 13.227633069082673, "grad_norm": 0.6824278831481934, "learning_rate": 4.146813138358851e-06, "loss": 0.3567, "step": 35040 }, { "epoch": 13.231408078520197, "grad_norm": 0.6351913213729858, "learning_rate": 4.130346169813509e-06, "loss": 0.3514, "step": 35050 }, { "epoch": 13.23518308795772, "grad_norm": 0.6978806853294373, "learning_rate": 4.113910552744038e-06, "loss": 0.3568, "step": 35060 }, { "epoch": 13.238958097395244, "grad_norm": 0.6495206356048584, "learning_rate": 4.097506298384024e-06, "loss": 0.3525, "step": 35070 }, { "epoch": 13.242733106832768, "grad_norm": 0.6304375529289246, "learning_rate": 4.081133417945626e-06, "loss": 0.3564, "step": 35080 }, { "epoch": 13.246508116270292, "grad_norm": 0.6751596927642822, "learning_rate": 4.0647919226195484e-06, "loss": 0.3574, "step": 35090 }, { "epoch": 13.250283125707814, "grad_norm": 0.6097880005836487, "learning_rate": 4.04848182357504e-06, "loss": 0.3571, "step": 35100 }, { "epoch": 13.254058135145337, "grad_norm": 0.6437689661979675, "learning_rate": 4.032203131959911e-06, "loss": 0.3664, "step": 35110 }, { "epoch": 13.257833144582861, "grad_norm": 0.6554880142211914, "learning_rate": 4.015955858900483e-06, "loss": 0.3576, "step": 35120 }, { "epoch": 13.261608154020385, "grad_norm": 0.6221573352813721, "learning_rate": 3.999740015501641e-06, "loss": 0.3552, "step": 35130 }, { "epoch": 13.265383163457908, "grad_norm": 0.8590520024299622, "learning_rate": 3.9835556128467236e-06, "loss": 0.3618, "step": 35140 }, { "epoch": 13.269158172895432, "grad_norm": 0.6513279676437378, "learning_rate": 3.967402661997644e-06, "loss": 0.3588, "step": 35150 }, { "epoch": 13.272933182332956, "grad_norm": 0.7151180505752563, "learning_rate": 3.951281173994781e-06, "loss": 0.3559, "step": 35160 }, { "epoch": 13.27670819177048, "grad_norm": 0.6879009008407593, "learning_rate": 3.935191159857016e-06, "loss": 0.3538, "step": 35170 }, { "epoch": 13.280483201208003, "grad_norm": 0.6725070476531982, "learning_rate": 3.919132630581746e-06, "loss": 0.364, "step": 35180 }, { "epoch": 13.284258210645527, "grad_norm": 0.6446594595909119, "learning_rate": 3.903105597144802e-06, "loss": 0.3523, "step": 35190 }, { "epoch": 13.28803322008305, "grad_norm": 0.6886351704597473, "learning_rate": 3.887110070500505e-06, "loss": 0.3607, "step": 35200 }, { "epoch": 13.291808229520575, "grad_norm": 0.6238877177238464, "learning_rate": 3.871146061581654e-06, "loss": 0.3507, "step": 35210 }, { "epoch": 13.295583238958097, "grad_norm": 0.640650749206543, "learning_rate": 3.855213581299505e-06, "loss": 0.3612, "step": 35220 }, { "epoch": 13.29935824839562, "grad_norm": 0.665745198726654, "learning_rate": 3.839312640543763e-06, "loss": 0.359, "step": 35230 }, { "epoch": 13.303133257833144, "grad_norm": 0.6206352114677429, "learning_rate": 3.823443250182551e-06, "loss": 0.3547, "step": 35240 }, { "epoch": 13.306908267270668, "grad_norm": 1.118415117263794, "learning_rate": 3.80760542106246e-06, "loss": 0.3534, "step": 35250 }, { "epoch": 13.310683276708192, "grad_norm": 0.6065353751182556, "learning_rate": 3.7917991640084972e-06, "loss": 0.3532, "step": 35260 }, { "epoch": 13.314458286145715, "grad_norm": 0.6253830194473267, "learning_rate": 3.7760244898240925e-06, "loss": 0.4729, "step": 35270 }, { "epoch": 13.318233295583239, "grad_norm": 0.5790948271751404, "learning_rate": 3.760281409291078e-06, "loss": 0.3582, "step": 35280 }, { "epoch": 13.322008305020763, "grad_norm": 0.6971233487129211, "learning_rate": 3.7445699331697113e-06, "loss": 0.3591, "step": 35290 }, { "epoch": 13.325783314458286, "grad_norm": 0.7019782066345215, "learning_rate": 3.7288900721986262e-06, "loss": 0.3583, "step": 35300 }, { "epoch": 13.32955832389581, "grad_norm": 0.682455837726593, "learning_rate": 3.7132418370948663e-06, "loss": 0.3568, "step": 35310 }, { "epoch": 13.333333333333334, "grad_norm": 0.6666685938835144, "learning_rate": 3.6976252385538535e-06, "loss": 0.358, "step": 35320 }, { "epoch": 13.337108342770858, "grad_norm": 0.634299099445343, "learning_rate": 3.6820402872493832e-06, "loss": 0.3542, "step": 35330 }, { "epoch": 13.34088335220838, "grad_norm": 0.6098415851593018, "learning_rate": 3.6664869938336343e-06, "loss": 0.3598, "step": 35340 }, { "epoch": 13.344658361645903, "grad_norm": 0.6019866466522217, "learning_rate": 3.650965368937115e-06, "loss": 0.3575, "step": 35350 }, { "epoch": 13.348433371083427, "grad_norm": 0.6554697155952454, "learning_rate": 3.6354754231687214e-06, "loss": 0.3604, "step": 35360 }, { "epoch": 13.35220838052095, "grad_norm": 1.0949931144714355, "learning_rate": 3.620017167115697e-06, "loss": 0.3586, "step": 35370 }, { "epoch": 13.355983389958475, "grad_norm": 0.663237988948822, "learning_rate": 3.604590611343589e-06, "loss": 0.3582, "step": 35380 }, { "epoch": 13.359758399395998, "grad_norm": 0.6626077890396118, "learning_rate": 3.589195766396325e-06, "loss": 0.3581, "step": 35390 }, { "epoch": 13.363533408833522, "grad_norm": 0.5932807326316833, "learning_rate": 3.573832642796121e-06, "loss": 0.3545, "step": 35400 }, { "epoch": 13.367308418271046, "grad_norm": 0.6529201865196228, "learning_rate": 3.5585012510435336e-06, "loss": 0.3567, "step": 35410 }, { "epoch": 13.37108342770857, "grad_norm": 0.6512988805770874, "learning_rate": 3.5432016016174184e-06, "loss": 0.3592, "step": 35420 }, { "epoch": 13.374858437146093, "grad_norm": 0.6883192658424377, "learning_rate": 3.5279337049749506e-06, "loss": 0.3495, "step": 35430 }, { "epoch": 13.378633446583617, "grad_norm": 0.625953733921051, "learning_rate": 3.5126975715515985e-06, "loss": 0.3524, "step": 35440 }, { "epoch": 13.38240845602114, "grad_norm": 0.7120693325996399, "learning_rate": 3.4974932117610994e-06, "loss": 0.3546, "step": 35450 }, { "epoch": 13.386183465458664, "grad_norm": 0.6227659583091736, "learning_rate": 3.4823206359955065e-06, "loss": 0.3585, "step": 35460 }, { "epoch": 13.389958474896186, "grad_norm": 0.6581076979637146, "learning_rate": 3.4671798546251133e-06, "loss": 0.3584, "step": 35470 }, { "epoch": 13.39373348433371, "grad_norm": 0.6351528167724609, "learning_rate": 3.4520708779985088e-06, "loss": 0.3486, "step": 35480 }, { "epoch": 13.397508493771234, "grad_norm": 0.6142309308052063, "learning_rate": 3.4369937164425493e-06, "loss": 0.3578, "step": 35490 }, { "epoch": 13.401283503208758, "grad_norm": 0.6407093405723572, "learning_rate": 3.421948380262313e-06, "loss": 0.3559, "step": 35500 }, { "epoch": 13.405058512646281, "grad_norm": 0.6450949311256409, "learning_rate": 3.406934879741158e-06, "loss": 0.3575, "step": 35510 }, { "epoch": 13.408833522083805, "grad_norm": 0.7444202899932861, "learning_rate": 3.391953225140665e-06, "loss": 0.3688, "step": 35520 }, { "epoch": 13.412608531521329, "grad_norm": 0.706688642501831, "learning_rate": 3.3770034267006533e-06, "loss": 0.3587, "step": 35530 }, { "epoch": 13.416383540958853, "grad_norm": 0.6597870588302612, "learning_rate": 3.362085494639178e-06, "loss": 0.3541, "step": 35540 }, { "epoch": 13.420158550396376, "grad_norm": 0.5902731418609619, "learning_rate": 3.3471994391524986e-06, "loss": 0.3547, "step": 35550 }, { "epoch": 13.4239335598339, "grad_norm": 0.6152647733688354, "learning_rate": 3.3323452704150825e-06, "loss": 0.3527, "step": 35560 }, { "epoch": 13.427708569271424, "grad_norm": 0.6410410404205322, "learning_rate": 3.3175229985796253e-06, "loss": 0.3597, "step": 35570 }, { "epoch": 13.431483578708947, "grad_norm": 0.6362301111221313, "learning_rate": 3.3027326337770115e-06, "loss": 0.3563, "step": 35580 }, { "epoch": 13.435258588146471, "grad_norm": 0.6383989453315735, "learning_rate": 3.2879741861163216e-06, "loss": 0.3579, "step": 35590 }, { "epoch": 13.439033597583993, "grad_norm": 0.6493092179298401, "learning_rate": 3.2732476656847976e-06, "loss": 0.3608, "step": 35600 }, { "epoch": 13.442808607021517, "grad_norm": 0.6834067106246948, "learning_rate": 3.2585530825478883e-06, "loss": 0.3529, "step": 35610 }, { "epoch": 13.44658361645904, "grad_norm": 0.7077740430831909, "learning_rate": 3.243890446749204e-06, "loss": 0.3575, "step": 35620 }, { "epoch": 13.450358625896564, "grad_norm": 0.6362556219100952, "learning_rate": 3.2292597683105107e-06, "loss": 0.3594, "step": 35630 }, { "epoch": 13.454133635334088, "grad_norm": 0.6774179339408875, "learning_rate": 3.214661057231755e-06, "loss": 0.3668, "step": 35640 }, { "epoch": 13.457908644771612, "grad_norm": 0.7050675749778748, "learning_rate": 3.2000943234910097e-06, "loss": 0.3517, "step": 35650 }, { "epoch": 13.461683654209136, "grad_norm": 0.8577159643173218, "learning_rate": 3.1855595770444835e-06, "loss": 0.3551, "step": 35660 }, { "epoch": 13.46545866364666, "grad_norm": 0.6533629298210144, "learning_rate": 3.1710568278265574e-06, "loss": 0.347, "step": 35670 }, { "epoch": 13.469233673084183, "grad_norm": 0.63999342918396, "learning_rate": 3.156586085749719e-06, "loss": 0.3484, "step": 35680 }, { "epoch": 13.473008682521707, "grad_norm": 0.6630726456642151, "learning_rate": 3.1421473607045893e-06, "loss": 0.3556, "step": 35690 }, { "epoch": 13.47678369195923, "grad_norm": 1.4107182025909424, "learning_rate": 3.127740662559886e-06, "loss": 0.355, "step": 35700 }, { "epoch": 13.480558701396754, "grad_norm": 0.7354855537414551, "learning_rate": 3.1133660011624655e-06, "loss": 0.3602, "step": 35710 }, { "epoch": 13.484333710834276, "grad_norm": 0.6842527389526367, "learning_rate": 3.099023386337263e-06, "loss": 0.3602, "step": 35720 }, { "epoch": 13.4881087202718, "grad_norm": 0.6937289237976074, "learning_rate": 3.0847128278873317e-06, "loss": 0.3585, "step": 35730 }, { "epoch": 13.491883729709324, "grad_norm": 0.7109084725379944, "learning_rate": 3.0704343355938027e-06, "loss": 0.3568, "step": 35740 }, { "epoch": 13.495658739146847, "grad_norm": 0.7426275610923767, "learning_rate": 3.056187919215886e-06, "loss": 0.3545, "step": 35750 }, { "epoch": 13.499433748584371, "grad_norm": 0.7047889828681946, "learning_rate": 3.041973588490865e-06, "loss": 0.3583, "step": 35760 }, { "epoch": 13.503208758021895, "grad_norm": 1.050673007965088, "learning_rate": 3.0277913531341128e-06, "loss": 0.3619, "step": 35770 }, { "epoch": 13.506983767459419, "grad_norm": 0.6489849090576172, "learning_rate": 3.0136412228390476e-06, "loss": 0.3544, "step": 35780 }, { "epoch": 13.510758776896942, "grad_norm": 0.635733425617218, "learning_rate": 2.999523207277155e-06, "loss": 0.3575, "step": 35790 }, { "epoch": 13.514533786334466, "grad_norm": 0.6572074890136719, "learning_rate": 2.9854373160979552e-06, "loss": 0.3506, "step": 35800 }, { "epoch": 13.51830879577199, "grad_norm": 0.6802811622619629, "learning_rate": 2.9713835589290362e-06, "loss": 0.3496, "step": 35810 }, { "epoch": 13.522083805209514, "grad_norm": 0.6372085213661194, "learning_rate": 2.9573619453759913e-06, "loss": 0.3557, "step": 35820 }, { "epoch": 13.525858814647037, "grad_norm": 0.6938003897666931, "learning_rate": 2.943372485022489e-06, "loss": 0.3472, "step": 35830 }, { "epoch": 13.52963382408456, "grad_norm": 0.6606130599975586, "learning_rate": 2.9294151874301623e-06, "loss": 0.3722, "step": 35840 }, { "epoch": 13.533408833522083, "grad_norm": 0.6342771649360657, "learning_rate": 2.9154900621387093e-06, "loss": 0.355, "step": 35850 }, { "epoch": 13.537183842959607, "grad_norm": 0.691672682762146, "learning_rate": 2.9015971186658164e-06, "loss": 0.3586, "step": 35860 }, { "epoch": 13.54095885239713, "grad_norm": 0.5848689675331116, "learning_rate": 2.887736366507182e-06, "loss": 0.3591, "step": 35870 }, { "epoch": 13.544733861834654, "grad_norm": 0.5893767476081848, "learning_rate": 2.873907815136501e-06, "loss": 0.3476, "step": 35880 }, { "epoch": 13.548508871272178, "grad_norm": 0.6813375949859619, "learning_rate": 2.8601114740054556e-06, "loss": 0.3556, "step": 35890 }, { "epoch": 13.552283880709702, "grad_norm": 0.629166841506958, "learning_rate": 2.8463473525437155e-06, "loss": 0.357, "step": 35900 }, { "epoch": 13.556058890147225, "grad_norm": 0.6545192003250122, "learning_rate": 2.8326154601589216e-06, "loss": 0.3553, "step": 35910 }, { "epoch": 13.55983389958475, "grad_norm": 0.639487087726593, "learning_rate": 2.818915806236705e-06, "loss": 0.3558, "step": 35920 }, { "epoch": 13.563608909022273, "grad_norm": 0.6719843149185181, "learning_rate": 2.8052484001406478e-06, "loss": 0.3527, "step": 35930 }, { "epoch": 13.567383918459797, "grad_norm": 0.598436713218689, "learning_rate": 2.791613251212277e-06, "loss": 0.3582, "step": 35940 }, { "epoch": 13.57115892789732, "grad_norm": 0.6548103094100952, "learning_rate": 2.7780103687711166e-06, "loss": 0.3549, "step": 35950 }, { "epoch": 13.574933937334844, "grad_norm": 0.6896668076515198, "learning_rate": 2.764439762114579e-06, "loss": 0.3508, "step": 35960 }, { "epoch": 13.578708946772366, "grad_norm": 0.6703106164932251, "learning_rate": 2.750901440518061e-06, "loss": 0.3556, "step": 35970 }, { "epoch": 13.58248395620989, "grad_norm": 0.664481520652771, "learning_rate": 2.7373954132348776e-06, "loss": 0.356, "step": 35980 }, { "epoch": 13.586258965647414, "grad_norm": 0.6433864831924438, "learning_rate": 2.7239216894962784e-06, "loss": 0.3526, "step": 35990 }, { "epoch": 13.590033975084937, "grad_norm": 0.6484426259994507, "learning_rate": 2.710480278511407e-06, "loss": 0.3545, "step": 36000 }, { "epoch": 13.593808984522461, "grad_norm": 5.11317777633667, "learning_rate": 2.6970711894673483e-06, "loss": 0.3684, "step": 36010 }, { "epoch": 13.597583993959985, "grad_norm": 0.7673658132553101, "learning_rate": 2.6836944315291047e-06, "loss": 0.3563, "step": 36020 }, { "epoch": 13.601359003397508, "grad_norm": 0.6486364603042603, "learning_rate": 2.67035001383954e-06, "loss": 0.3549, "step": 36030 }, { "epoch": 13.605134012835032, "grad_norm": 0.5933120250701904, "learning_rate": 2.6570379455194527e-06, "loss": 0.3573, "step": 36040 }, { "epoch": 13.608909022272556, "grad_norm": 0.7512244582176208, "learning_rate": 2.6437582356675206e-06, "loss": 0.3507, "step": 36050 }, { "epoch": 13.61268403171008, "grad_norm": 0.7129798531532288, "learning_rate": 2.6305108933602775e-06, "loss": 0.362, "step": 36060 }, { "epoch": 13.616459041147603, "grad_norm": 0.5950183868408203, "learning_rate": 2.6172959276521803e-06, "loss": 0.3546, "step": 36070 }, { "epoch": 13.620234050585127, "grad_norm": 0.6639828085899353, "learning_rate": 2.6041133475755153e-06, "loss": 0.3559, "step": 36080 }, { "epoch": 13.62400906002265, "grad_norm": 0.654403567314148, "learning_rate": 2.590963162140475e-06, "loss": 0.3544, "step": 36090 }, { "epoch": 13.627784069460173, "grad_norm": 0.6647920608520508, "learning_rate": 2.5778453803350587e-06, "loss": 0.3572, "step": 36100 }, { "epoch": 13.631559078897697, "grad_norm": 0.6501339077949524, "learning_rate": 2.5647600111251723e-06, "loss": 0.3574, "step": 36110 }, { "epoch": 13.63533408833522, "grad_norm": 0.6624252200126648, "learning_rate": 2.551707063454528e-06, "loss": 0.3604, "step": 36120 }, { "epoch": 13.639109097772744, "grad_norm": 1.6550086736679077, "learning_rate": 2.5386865462446897e-06, "loss": 0.3647, "step": 36130 }, { "epoch": 13.642884107210268, "grad_norm": 0.6511169672012329, "learning_rate": 2.52569846839506e-06, "loss": 0.3546, "step": 36140 }, { "epoch": 13.646659116647792, "grad_norm": 0.6163169741630554, "learning_rate": 2.512742838782883e-06, "loss": 0.3556, "step": 36150 }, { "epoch": 13.650434126085315, "grad_norm": 0.6587780117988586, "learning_rate": 2.4998196662631924e-06, "loss": 0.3611, "step": 36160 }, { "epoch": 13.654209135522839, "grad_norm": 0.6356183290481567, "learning_rate": 2.4869289596688604e-06, "loss": 0.35, "step": 36170 }, { "epoch": 13.657984144960363, "grad_norm": 0.6465098857879639, "learning_rate": 2.4740707278105567e-06, "loss": 0.3539, "step": 36180 }, { "epoch": 13.661759154397886, "grad_norm": 0.6922428011894226, "learning_rate": 2.461244979476779e-06, "loss": 0.3567, "step": 36190 }, { "epoch": 13.66553416383541, "grad_norm": 0.6384284496307373, "learning_rate": 2.4484517234337866e-06, "loss": 0.3517, "step": 36200 }, { "epoch": 13.669309173272934, "grad_norm": 0.6639804244041443, "learning_rate": 2.4356909684256634e-06, "loss": 0.3593, "step": 36210 }, { "epoch": 13.673084182710458, "grad_norm": 0.6959599852561951, "learning_rate": 2.4229627231742545e-06, "loss": 0.3552, "step": 36220 }, { "epoch": 13.67685919214798, "grad_norm": 0.6804720163345337, "learning_rate": 2.4102669963791956e-06, "loss": 0.3578, "step": 36230 }, { "epoch": 13.680634201585503, "grad_norm": 0.6466209888458252, "learning_rate": 2.3976037967179065e-06, "loss": 0.3557, "step": 36240 }, { "epoch": 13.684409211023027, "grad_norm": 0.6484111547470093, "learning_rate": 2.3849731328455582e-06, "loss": 0.3582, "step": 36250 }, { "epoch": 13.68818422046055, "grad_norm": 0.637500524520874, "learning_rate": 2.3723750133950894e-06, "loss": 0.3571, "step": 36260 }, { "epoch": 13.691959229898075, "grad_norm": 3.0127627849578857, "learning_rate": 2.3598094469771957e-06, "loss": 0.3572, "step": 36270 }, { "epoch": 13.695734239335598, "grad_norm": 0.6502493619918823, "learning_rate": 2.347276442180324e-06, "loss": 0.3556, "step": 36280 }, { "epoch": 13.699509248773122, "grad_norm": 0.6446804404258728, "learning_rate": 2.3347760075706715e-06, "loss": 0.3613, "step": 36290 }, { "epoch": 13.703284258210646, "grad_norm": 0.6583272218704224, "learning_rate": 2.3223081516921595e-06, "loss": 0.3598, "step": 36300 }, { "epoch": 13.70705926764817, "grad_norm": 0.5799362063407898, "learning_rate": 2.3098728830664494e-06, "loss": 0.3619, "step": 36310 }, { "epoch": 13.710834277085693, "grad_norm": 0.7087305188179016, "learning_rate": 2.2974702101929314e-06, "loss": 0.3569, "step": 36320 }, { "epoch": 13.714609286523217, "grad_norm": 0.6843621730804443, "learning_rate": 2.2851001415487185e-06, "loss": 0.3589, "step": 36330 }, { "epoch": 13.71838429596074, "grad_norm": 0.71535325050354, "learning_rate": 2.272762685588631e-06, "loss": 0.3547, "step": 36340 }, { "epoch": 13.722159305398263, "grad_norm": 0.6820167303085327, "learning_rate": 2.260457850745218e-06, "loss": 0.3498, "step": 36350 }, { "epoch": 13.725934314835786, "grad_norm": 0.6012173295021057, "learning_rate": 2.2481856454286963e-06, "loss": 0.3575, "step": 36360 }, { "epoch": 13.72970932427331, "grad_norm": 0.687364399433136, "learning_rate": 2.235946078027018e-06, "loss": 0.3519, "step": 36370 }, { "epoch": 13.733484333710834, "grad_norm": 0.6643616557121277, "learning_rate": 2.223739156905802e-06, "loss": 0.3559, "step": 36380 }, { "epoch": 13.737259343148358, "grad_norm": 0.6354239583015442, "learning_rate": 2.2115648904083796e-06, "loss": 0.3561, "step": 36390 }, { "epoch": 13.741034352585881, "grad_norm": 0.6376845836639404, "learning_rate": 2.199423286855734e-06, "loss": 0.3633, "step": 36400 }, { "epoch": 13.744809362023405, "grad_norm": 0.6230825185775757, "learning_rate": 2.1873143545465323e-06, "loss": 0.3594, "step": 36410 }, { "epoch": 13.748584371460929, "grad_norm": 0.644271969795227, "learning_rate": 2.1752381017571267e-06, "loss": 0.3556, "step": 36420 }, { "epoch": 13.752359380898453, "grad_norm": 0.6420625448226929, "learning_rate": 2.1631945367415142e-06, "loss": 0.3586, "step": 36430 }, { "epoch": 13.756134390335976, "grad_norm": 0.8240686655044556, "learning_rate": 2.15118366773136e-06, "loss": 0.3611, "step": 36440 }, { "epoch": 13.7599093997735, "grad_norm": 0.6971853971481323, "learning_rate": 2.1392055029359926e-06, "loss": 0.3599, "step": 36450 }, { "epoch": 13.763684409211024, "grad_norm": 0.6388399004936218, "learning_rate": 2.127260050542357e-06, "loss": 0.3603, "step": 36460 }, { "epoch": 13.767459418648546, "grad_norm": 0.6157034039497375, "learning_rate": 2.1153473187150608e-06, "loss": 0.3599, "step": 36470 }, { "epoch": 13.77123442808607, "grad_norm": 0.6063554286956787, "learning_rate": 2.1034673155963467e-06, "loss": 0.3592, "step": 36480 }, { "epoch": 13.775009437523593, "grad_norm": 0.6081372499465942, "learning_rate": 2.0916200493060976e-06, "loss": 0.3581, "step": 36490 }, { "epoch": 13.778784446961117, "grad_norm": 0.6699353456497192, "learning_rate": 2.0798055279417916e-06, "loss": 0.3602, "step": 36500 }, { "epoch": 13.78255945639864, "grad_norm": 0.6695933938026428, "learning_rate": 2.0680237595785356e-06, "loss": 0.3552, "step": 36510 }, { "epoch": 13.786334465836164, "grad_norm": 0.5835472941398621, "learning_rate": 2.0562747522690774e-06, "loss": 0.3573, "step": 36520 }, { "epoch": 13.790109475273688, "grad_norm": 0.7515608072280884, "learning_rate": 2.0445585140437373e-06, "loss": 0.3574, "step": 36530 }, { "epoch": 13.793884484711212, "grad_norm": 0.6449537873268127, "learning_rate": 2.0328750529104602e-06, "loss": 0.3573, "step": 36540 }, { "epoch": 13.797659494148736, "grad_norm": 0.6437288522720337, "learning_rate": 2.021224376854791e-06, "loss": 0.3579, "step": 36550 }, { "epoch": 13.80143450358626, "grad_norm": 0.6661025285720825, "learning_rate": 2.0096064938398386e-06, "loss": 0.3663, "step": 36560 }, { "epoch": 13.805209513023783, "grad_norm": 0.6766248941421509, "learning_rate": 1.998021411806328e-06, "loss": 0.3595, "step": 36570 }, { "epoch": 13.808984522461307, "grad_norm": 0.6111851930618286, "learning_rate": 1.9864691386725586e-06, "loss": 0.3578, "step": 36580 }, { "epoch": 13.81275953189883, "grad_norm": 0.7111583948135376, "learning_rate": 1.9749496823343814e-06, "loss": 0.3598, "step": 36590 }, { "epoch": 13.816534541336353, "grad_norm": 0.7276173830032349, "learning_rate": 1.963463050665254e-06, "loss": 0.349, "step": 36600 }, { "epoch": 13.820309550773876, "grad_norm": 0.6221917271614075, "learning_rate": 1.9520092515161682e-06, "loss": 0.3696, "step": 36610 }, { "epoch": 13.8240845602114, "grad_norm": 0.7171021699905396, "learning_rate": 1.940588292715695e-06, "loss": 0.355, "step": 36620 }, { "epoch": 13.827859569648924, "grad_norm": 0.7668521404266357, "learning_rate": 1.929200182069951e-06, "loss": 0.3491, "step": 36630 }, { "epoch": 13.831634579086447, "grad_norm": 0.9479874968528748, "learning_rate": 1.9178449273625997e-06, "loss": 0.3594, "step": 36640 }, { "epoch": 13.835409588523971, "grad_norm": 0.6809915900230408, "learning_rate": 1.90652253635486e-06, "loss": 0.3518, "step": 36650 }, { "epoch": 13.839184597961495, "grad_norm": 0.6257880330085754, "learning_rate": 1.8952330167854693e-06, "loss": 0.358, "step": 36660 }, { "epoch": 13.842959607399019, "grad_norm": 0.6460468769073486, "learning_rate": 1.8839763763707107e-06, "loss": 0.3555, "step": 36670 }, { "epoch": 13.846734616836542, "grad_norm": 0.7003082036972046, "learning_rate": 1.8727526228043912e-06, "loss": 0.3521, "step": 36680 }, { "epoch": 13.850509626274066, "grad_norm": 0.6699946522712708, "learning_rate": 1.8615617637578408e-06, "loss": 0.3566, "step": 36690 }, { "epoch": 13.85428463571159, "grad_norm": 0.6396785378456116, "learning_rate": 1.850403806879908e-06, "loss": 0.3517, "step": 36700 }, { "epoch": 13.858059645149114, "grad_norm": 0.6421851515769958, "learning_rate": 1.8392787597969475e-06, "loss": 0.3532, "step": 36710 }, { "epoch": 13.861834654586637, "grad_norm": 0.6195537447929382, "learning_rate": 1.828186630112827e-06, "loss": 0.3673, "step": 36720 }, { "epoch": 13.86560966402416, "grad_norm": 0.5870434045791626, "learning_rate": 1.8171274254089098e-06, "loss": 0.3594, "step": 36730 }, { "epoch": 13.869384673461683, "grad_norm": 0.7309495806694031, "learning_rate": 1.806101153244061e-06, "loss": 0.36, "step": 36740 }, { "epoch": 13.873159682899207, "grad_norm": 0.6230713129043579, "learning_rate": 1.7951078211546401e-06, "loss": 0.3546, "step": 36750 }, { "epoch": 13.87693469233673, "grad_norm": 0.6043184399604797, "learning_rate": 1.7841474366544764e-06, "loss": 0.3515, "step": 36760 }, { "epoch": 13.880709701774254, "grad_norm": 0.6820389032363892, "learning_rate": 1.7732200072348825e-06, "loss": 0.3657, "step": 36770 }, { "epoch": 13.884484711211778, "grad_norm": 0.6838846206665039, "learning_rate": 1.7623255403646677e-06, "loss": 0.3519, "step": 36780 }, { "epoch": 13.888259720649302, "grad_norm": 0.6790589094161987, "learning_rate": 1.7514640434900865e-06, "loss": 0.3544, "step": 36790 }, { "epoch": 13.892034730086825, "grad_norm": 0.6140289902687073, "learning_rate": 1.7406355240348838e-06, "loss": 0.353, "step": 36800 }, { "epoch": 13.89580973952435, "grad_norm": 0.6230329871177673, "learning_rate": 1.7298399894002392e-06, "loss": 0.3547, "step": 36810 }, { "epoch": 13.899584748961873, "grad_norm": 0.674197793006897, "learning_rate": 1.7190774469647952e-06, "loss": 0.3633, "step": 36820 }, { "epoch": 13.903359758399397, "grad_norm": 0.8990030884742737, "learning_rate": 1.7083479040846617e-06, "loss": 0.3618, "step": 36830 }, { "epoch": 13.90713476783692, "grad_norm": 0.6309157609939575, "learning_rate": 1.697651368093367e-06, "loss": 0.3521, "step": 36840 }, { "epoch": 13.910909777274444, "grad_norm": 0.6408490538597107, "learning_rate": 1.6869878463019017e-06, "loss": 0.357, "step": 36850 }, { "epoch": 13.914684786711966, "grad_norm": 0.6487820148468018, "learning_rate": 1.6763573459986858e-06, "loss": 0.3681, "step": 36860 }, { "epoch": 13.91845979614949, "grad_norm": 0.663537859916687, "learning_rate": 1.6657598744495516e-06, "loss": 0.3592, "step": 36870 }, { "epoch": 13.922234805587014, "grad_norm": 0.629589319229126, "learning_rate": 1.6551954388977774e-06, "loss": 0.3481, "step": 36880 }, { "epoch": 13.926009815024537, "grad_norm": 0.6090066432952881, "learning_rate": 1.644664046564054e-06, "loss": 0.3581, "step": 36890 }, { "epoch": 13.929784824462061, "grad_norm": 0.765868067741394, "learning_rate": 1.634165704646501e-06, "loss": 0.3518, "step": 36900 }, { "epoch": 13.933559833899585, "grad_norm": 0.6522102355957031, "learning_rate": 1.6237004203206175e-06, "loss": 0.3544, "step": 36910 }, { "epoch": 13.937334843337108, "grad_norm": 0.6589536666870117, "learning_rate": 1.6132682007393373e-06, "loss": 0.3544, "step": 36920 }, { "epoch": 13.941109852774632, "grad_norm": 0.6735085248947144, "learning_rate": 1.6028690530329847e-06, "loss": 0.3427, "step": 36930 }, { "epoch": 13.944884862212156, "grad_norm": 0.6307739615440369, "learning_rate": 1.5925029843092742e-06, "loss": 0.3557, "step": 36940 }, { "epoch": 13.94865987164968, "grad_norm": 0.6922662854194641, "learning_rate": 1.5821700016533214e-06, "loss": 0.3629, "step": 36950 }, { "epoch": 13.952434881087203, "grad_norm": 0.6589555740356445, "learning_rate": 1.5718701121276268e-06, "loss": 0.3487, "step": 36960 }, { "epoch": 13.956209890524725, "grad_norm": 0.7357824444770813, "learning_rate": 1.5616033227720483e-06, "loss": 0.3495, "step": 36970 }, { "epoch": 13.95998489996225, "grad_norm": 0.6229332089424133, "learning_rate": 1.5513696406038558e-06, "loss": 0.3577, "step": 36980 }, { "epoch": 13.963759909399773, "grad_norm": 0.5904605388641357, "learning_rate": 1.5411690726176653e-06, "loss": 0.3531, "step": 36990 }, { "epoch": 13.967534918837297, "grad_norm": 0.7504417300224304, "learning_rate": 1.5310016257854832e-06, "loss": 0.3604, "step": 37000 }, { "epoch": 13.97130992827482, "grad_norm": 0.6638484001159668, "learning_rate": 1.5208673070566504e-06, "loss": 0.3546, "step": 37010 }, { "epoch": 13.975084937712344, "grad_norm": 0.6932189464569092, "learning_rate": 1.5107661233578818e-06, "loss": 0.353, "step": 37020 }, { "epoch": 13.978859947149868, "grad_norm": 0.871865451335907, "learning_rate": 1.5006980815932437e-06, "loss": 0.3684, "step": 37030 }, { "epoch": 13.982634956587392, "grad_norm": 0.6513112783432007, "learning_rate": 1.490663188644148e-06, "loss": 0.3602, "step": 37040 }, { "epoch": 13.986409966024915, "grad_norm": 0.6713985204696655, "learning_rate": 1.4806614513693418e-06, "loss": 0.3541, "step": 37050 }, { "epoch": 13.990184975462439, "grad_norm": 0.681988537311554, "learning_rate": 1.4706928766049289e-06, "loss": 0.3581, "step": 37060 }, { "epoch": 13.993959984899963, "grad_norm": 0.948577344417572, "learning_rate": 1.4607574711643257e-06, "loss": 0.3532, "step": 37070 }, { "epoch": 13.997734994337486, "grad_norm": 0.7015256881713867, "learning_rate": 1.4508552418382949e-06, "loss": 0.355, "step": 37080 }, { "epoch": 14.00151000377501, "grad_norm": 0.6600831151008606, "learning_rate": 1.4409861953949112e-06, "loss": 0.3522, "step": 37090 }, { "epoch": 14.005285013212532, "grad_norm": 0.6127675771713257, "learning_rate": 1.43115033857959e-06, "loss": 0.3566, "step": 37100 }, { "epoch": 14.009060022650056, "grad_norm": 0.6839843988418579, "learning_rate": 1.4213476781150258e-06, "loss": 0.3503, "step": 37110 }, { "epoch": 14.01283503208758, "grad_norm": 0.6490364074707031, "learning_rate": 1.411578220701254e-06, "loss": 0.3628, "step": 37120 }, { "epoch": 14.016610041525103, "grad_norm": 0.6181484460830688, "learning_rate": 1.4018419730156108e-06, "loss": 0.3498, "step": 37130 }, { "epoch": 14.020385050962627, "grad_norm": 0.638804018497467, "learning_rate": 1.3921389417127283e-06, "loss": 0.3595, "step": 37140 }, { "epoch": 14.02416006040015, "grad_norm": 0.6494945287704468, "learning_rate": 1.3824691334245298e-06, "loss": 0.36, "step": 37150 }, { "epoch": 14.027935069837675, "grad_norm": 0.6143277287483215, "learning_rate": 1.3728325547602505e-06, "loss": 0.3568, "step": 37160 }, { "epoch": 14.031710079275198, "grad_norm": 0.6732072234153748, "learning_rate": 1.3632292123063828e-06, "loss": 0.3524, "step": 37170 }, { "epoch": 14.035485088712722, "grad_norm": 0.6264946460723877, "learning_rate": 1.3536591126267317e-06, "loss": 0.3535, "step": 37180 }, { "epoch": 14.039260098150246, "grad_norm": 0.6948634386062622, "learning_rate": 1.3441222622623761e-06, "loss": 0.3541, "step": 37190 }, { "epoch": 14.04303510758777, "grad_norm": 0.6478399038314819, "learning_rate": 1.3346186677316575e-06, "loss": 0.3537, "step": 37200 }, { "epoch": 14.046810117025293, "grad_norm": 0.727102518081665, "learning_rate": 1.3251483355301852e-06, "loss": 0.3499, "step": 37210 }, { "epoch": 14.050585126462817, "grad_norm": 0.5733349919319153, "learning_rate": 1.315711272130854e-06, "loss": 0.3598, "step": 37220 }, { "epoch": 14.054360135900339, "grad_norm": 0.6830695867538452, "learning_rate": 1.306307483983804e-06, "loss": 0.3582, "step": 37230 }, { "epoch": 14.058135145337863, "grad_norm": 0.6603965163230896, "learning_rate": 1.2969369775164275e-06, "loss": 0.3528, "step": 37240 }, { "epoch": 14.061910154775386, "grad_norm": 0.672484278678894, "learning_rate": 1.28759975913339e-06, "loss": 0.353, "step": 37250 }, { "epoch": 14.06568516421291, "grad_norm": 0.6874468922615051, "learning_rate": 1.2782958352165863e-06, "loss": 0.3559, "step": 37260 }, { "epoch": 14.069460173650434, "grad_norm": 0.6835435032844543, "learning_rate": 1.2690252121251633e-06, "loss": 0.359, "step": 37270 }, { "epoch": 14.073235183087958, "grad_norm": 0.6037148237228394, "learning_rate": 1.2597878961954968e-06, "loss": 0.3595, "step": 37280 }, { "epoch": 14.077010192525481, "grad_norm": 0.7436287999153137, "learning_rate": 1.250583893741214e-06, "loss": 0.3606, "step": 37290 }, { "epoch": 14.080785201963005, "grad_norm": 0.6740867495536804, "learning_rate": 1.2414132110531607e-06, "loss": 0.3607, "step": 37300 }, { "epoch": 14.084560211400529, "grad_norm": 0.7421841025352478, "learning_rate": 1.2322758543994062e-06, "loss": 0.3554, "step": 37310 }, { "epoch": 14.088335220838053, "grad_norm": 0.7035681009292603, "learning_rate": 1.2231718300252603e-06, "loss": 0.3539, "step": 37320 }, { "epoch": 14.092110230275576, "grad_norm": 0.6054702401161194, "learning_rate": 1.214101144153229e-06, "loss": 0.3691, "step": 37330 }, { "epoch": 14.0958852397131, "grad_norm": 0.6276698112487793, "learning_rate": 1.205063802983042e-06, "loss": 0.3569, "step": 37340 }, { "epoch": 14.099660249150624, "grad_norm": 0.652378261089325, "learning_rate": 1.1960598126916301e-06, "loss": 0.3559, "step": 37350 }, { "epoch": 14.103435258588146, "grad_norm": 0.6906107664108276, "learning_rate": 1.187089179433154e-06, "loss": 0.3614, "step": 37360 }, { "epoch": 14.10721026802567, "grad_norm": 0.6413514018058777, "learning_rate": 1.178151909338937e-06, "loss": 0.3563, "step": 37370 }, { "epoch": 14.110985277463193, "grad_norm": 0.6062707901000977, "learning_rate": 1.1692480085175317e-06, "loss": 0.3571, "step": 37380 }, { "epoch": 14.114760286900717, "grad_norm": 0.6939738392829895, "learning_rate": 1.1603774830546587e-06, "loss": 0.3524, "step": 37390 }, { "epoch": 14.11853529633824, "grad_norm": 0.6625202894210815, "learning_rate": 1.1515403390132573e-06, "loss": 0.3564, "step": 37400 }, { "epoch": 14.122310305775764, "grad_norm": 0.7371194362640381, "learning_rate": 1.1427365824334124e-06, "loss": 0.3535, "step": 37410 }, { "epoch": 14.126085315213288, "grad_norm": 0.5587255954742432, "learning_rate": 1.133966219332422e-06, "loss": 0.3549, "step": 37420 }, { "epoch": 14.129860324650812, "grad_norm": 0.6779747605323792, "learning_rate": 1.1252292557047405e-06, "loss": 0.3618, "step": 37430 }, { "epoch": 14.133635334088336, "grad_norm": 0.6948127150535583, "learning_rate": 1.1165256975219974e-06, "loss": 0.3502, "step": 37440 }, { "epoch": 14.13741034352586, "grad_norm": 0.796514630317688, "learning_rate": 1.1078555507330002e-06, "loss": 0.3551, "step": 37450 }, { "epoch": 14.141185352963383, "grad_norm": 0.6783875226974487, "learning_rate": 1.0992188212637089e-06, "loss": 0.355, "step": 37460 }, { "epoch": 14.144960362400907, "grad_norm": 0.6731642484664917, "learning_rate": 1.0906155150172458e-06, "loss": 0.351, "step": 37470 }, { "epoch": 14.148735371838429, "grad_norm": 0.670313835144043, "learning_rate": 1.0820456378738852e-06, "loss": 0.3541, "step": 37480 }, { "epoch": 14.152510381275953, "grad_norm": 0.768552303314209, "learning_rate": 1.0735091956910692e-06, "loss": 0.3524, "step": 37490 }, { "epoch": 14.156285390713476, "grad_norm": 0.6623304486274719, "learning_rate": 1.0650061943033695e-06, "loss": 0.3575, "step": 37500 }, { "epoch": 14.160060400151, "grad_norm": 0.632932186126709, "learning_rate": 1.0565366395225041e-06, "loss": 0.3482, "step": 37510 }, { "epoch": 14.163835409588524, "grad_norm": 0.633401095867157, "learning_rate": 1.0481005371373366e-06, "loss": 0.3536, "step": 37520 }, { "epoch": 14.167610419026047, "grad_norm": 0.6039905548095703, "learning_rate": 1.0396978929138602e-06, "loss": 0.3505, "step": 37530 }, { "epoch": 14.171385428463571, "grad_norm": 0.6414741277694702, "learning_rate": 1.0313287125952032e-06, "loss": 0.3634, "step": 37540 }, { "epoch": 14.175160437901095, "grad_norm": 0.6331318020820618, "learning_rate": 1.0229930019016288e-06, "loss": 0.3648, "step": 37550 }, { "epoch": 14.178935447338619, "grad_norm": 0.6243364214897156, "learning_rate": 1.0146907665305183e-06, "loss": 0.3647, "step": 37560 }, { "epoch": 14.182710456776142, "grad_norm": 0.6351826786994934, "learning_rate": 1.0064220121563551e-06, "loss": 0.358, "step": 37570 }, { "epoch": 14.186485466213666, "grad_norm": 0.6927258372306824, "learning_rate": 9.98186744430768e-07, "loss": 0.3517, "step": 37580 }, { "epoch": 14.19026047565119, "grad_norm": 0.6401153206825256, "learning_rate": 9.899849689824826e-07, "loss": 0.3611, "step": 37590 }, { "epoch": 14.194035485088712, "grad_norm": 0.7551748752593994, "learning_rate": 9.818166914173365e-07, "loss": 0.3524, "step": 37600 }, { "epoch": 14.197810494526236, "grad_norm": 0.6238972544670105, "learning_rate": 9.736819173182642e-07, "loss": 0.3578, "step": 37610 }, { "epoch": 14.20158550396376, "grad_norm": 0.795691728591919, "learning_rate": 9.655806522453126e-07, "loss": 0.3544, "step": 37620 }, { "epoch": 14.205360513401283, "grad_norm": 0.558307945728302, "learning_rate": 9.575129017356187e-07, "loss": 0.3547, "step": 37630 }, { "epoch": 14.209135522838807, "grad_norm": 0.6478992104530334, "learning_rate": 9.494786713034109e-07, "loss": 0.3494, "step": 37640 }, { "epoch": 14.21291053227633, "grad_norm": 0.646564245223999, "learning_rate": 9.414779664400075e-07, "loss": 0.3572, "step": 37650 }, { "epoch": 14.216685541713854, "grad_norm": 0.6614131927490234, "learning_rate": 9.335107926138342e-07, "loss": 0.3513, "step": 37660 }, { "epoch": 14.220460551151378, "grad_norm": 0.7705824375152588, "learning_rate": 9.255771552703519e-07, "loss": 0.3504, "step": 37670 }, { "epoch": 14.224235560588902, "grad_norm": 0.6299923658370972, "learning_rate": 9.176770598321449e-07, "loss": 0.3535, "step": 37680 }, { "epoch": 14.228010570026425, "grad_norm": 0.6774992346763611, "learning_rate": 9.098105116988498e-07, "loss": 0.3546, "step": 37690 }, { "epoch": 14.23178557946395, "grad_norm": 0.6203901171684265, "learning_rate": 9.019775162471767e-07, "loss": 0.3477, "step": 37700 }, { "epoch": 14.235560588901473, "grad_norm": 0.590056300163269, "learning_rate": 8.94178078830904e-07, "loss": 0.3571, "step": 37710 }, { "epoch": 14.239335598338997, "grad_norm": 0.6835026741027832, "learning_rate": 8.864122047808731e-07, "loss": 0.362, "step": 37720 }, { "epoch": 14.243110607776519, "grad_norm": 0.7431061863899231, "learning_rate": 8.786798994049882e-07, "loss": 0.3595, "step": 37730 }, { "epoch": 14.246885617214042, "grad_norm": 0.6774893999099731, "learning_rate": 8.709811679882052e-07, "loss": 0.3546, "step": 37740 }, { "epoch": 14.250660626651566, "grad_norm": 0.6696170568466187, "learning_rate": 8.633160157925313e-07, "loss": 0.3597, "step": 37750 }, { "epoch": 14.25443563608909, "grad_norm": 0.592789351940155, "learning_rate": 8.556844480570369e-07, "loss": 0.3576, "step": 37760 }, { "epoch": 14.258210645526614, "grad_norm": 0.6565343141555786, "learning_rate": 8.480864699978164e-07, "loss": 0.3634, "step": 37770 }, { "epoch": 14.261985654964137, "grad_norm": 0.5975145697593689, "learning_rate": 8.405220868080154e-07, "loss": 0.3502, "step": 37780 }, { "epoch": 14.265760664401661, "grad_norm": 0.9825413227081299, "learning_rate": 8.329913036578263e-07, "loss": 0.3514, "step": 37790 }, { "epoch": 14.269535673839185, "grad_norm": 0.6201307773590088, "learning_rate": 8.254941256944593e-07, "loss": 0.358, "step": 37800 }, { "epoch": 14.273310683276708, "grad_norm": 0.6925731897354126, "learning_rate": 8.18030558042171e-07, "loss": 0.3512, "step": 37810 }, { "epoch": 14.277085692714232, "grad_norm": 0.6344226002693176, "learning_rate": 8.106006058022419e-07, "loss": 0.3585, "step": 37820 }, { "epoch": 14.280860702151756, "grad_norm": 0.6603698134422302, "learning_rate": 8.03204274052971e-07, "loss": 0.3566, "step": 37830 }, { "epoch": 14.28463571158928, "grad_norm": 0.7116165161132812, "learning_rate": 7.958415678496756e-07, "loss": 0.359, "step": 37840 }, { "epoch": 14.288410721026803, "grad_norm": 0.6361221075057983, "learning_rate": 7.885124922247022e-07, "loss": 0.3539, "step": 37850 }, { "epoch": 14.292185730464325, "grad_norm": 0.6573535203933716, "learning_rate": 7.812170521874051e-07, "loss": 0.3601, "step": 37860 }, { "epoch": 14.29596073990185, "grad_norm": 0.6462699770927429, "learning_rate": 7.739552527241456e-07, "loss": 0.3573, "step": 37870 }, { "epoch": 14.299735749339373, "grad_norm": 0.587371826171875, "learning_rate": 7.667270987983033e-07, "loss": 0.3561, "step": 37880 }, { "epoch": 14.303510758776897, "grad_norm": 0.6839852929115295, "learning_rate": 7.595325953502375e-07, "loss": 0.3578, "step": 37890 }, { "epoch": 14.30728576821442, "grad_norm": 0.5843148231506348, "learning_rate": 7.523717472973368e-07, "loss": 0.3582, "step": 37900 }, { "epoch": 14.311060777651944, "grad_norm": 0.6995810866355896, "learning_rate": 7.452445595339696e-07, "loss": 0.3567, "step": 37910 }, { "epoch": 14.314835787089468, "grad_norm": 0.6964356899261475, "learning_rate": 7.381510369315059e-07, "loss": 0.3596, "step": 37920 }, { "epoch": 14.318610796526992, "grad_norm": 1.5723952054977417, "learning_rate": 7.310911843382895e-07, "loss": 0.3567, "step": 37930 }, { "epoch": 14.322385805964515, "grad_norm": 0.7851320505142212, "learning_rate": 7.240650065796716e-07, "loss": 0.3499, "step": 37940 }, { "epoch": 14.326160815402039, "grad_norm": 0.6869962811470032, "learning_rate": 7.170725084579777e-07, "loss": 0.3588, "step": 37950 }, { "epoch": 14.329935824839563, "grad_norm": 0.6377101540565491, "learning_rate": 7.101136947525233e-07, "loss": 0.3529, "step": 37960 }, { "epoch": 14.333710834277086, "grad_norm": 0.713228166103363, "learning_rate": 7.031885702195707e-07, "loss": 0.357, "step": 37970 }, { "epoch": 14.337485843714608, "grad_norm": 0.6053051352500916, "learning_rate": 6.962971395924e-07, "loss": 0.3571, "step": 37980 }, { "epoch": 14.341260853152132, "grad_norm": 0.6226997971534729, "learning_rate": 6.894394075812217e-07, "loss": 0.3558, "step": 37990 }, { "epoch": 14.345035862589656, "grad_norm": 0.6585350036621094, "learning_rate": 6.826153788732359e-07, "loss": 0.3541, "step": 38000 }, { "epoch": 14.34881087202718, "grad_norm": 0.7358768582344055, "learning_rate": 6.758250581326009e-07, "loss": 0.3588, "step": 38010 }, { "epoch": 14.352585881464703, "grad_norm": 0.7400320172309875, "learning_rate": 6.690684500004485e-07, "loss": 0.3553, "step": 38020 }, { "epoch": 14.356360890902227, "grad_norm": 0.6965455412864685, "learning_rate": 6.623455590948458e-07, "loss": 0.3536, "step": 38030 }, { "epoch": 14.36013590033975, "grad_norm": 0.6424039006233215, "learning_rate": 6.556563900108226e-07, "loss": 0.3559, "step": 38040 }, { "epoch": 14.363910909777275, "grad_norm": 0.664570152759552, "learning_rate": 6.490009473203662e-07, "loss": 0.3596, "step": 38050 }, { "epoch": 14.367685919214798, "grad_norm": 0.6233497262001038, "learning_rate": 6.423792355724156e-07, "loss": 0.3518, "step": 38060 }, { "epoch": 14.371460928652322, "grad_norm": 0.6261218190193176, "learning_rate": 6.35791259292845e-07, "loss": 0.3594, "step": 38070 }, { "epoch": 14.375235938089846, "grad_norm": 0.7287961840629578, "learning_rate": 6.292370229844691e-07, "loss": 0.3546, "step": 38080 }, { "epoch": 14.37901094752737, "grad_norm": 0.5852963328361511, "learning_rate": 6.227165311270488e-07, "loss": 0.3556, "step": 38090 }, { "epoch": 14.382785956964893, "grad_norm": 0.6326406002044678, "learning_rate": 6.16229788177275e-07, "loss": 0.3512, "step": 38100 }, { "epoch": 14.386560966402415, "grad_norm": 0.6186873912811279, "learning_rate": 6.097767985687841e-07, "loss": 0.3541, "step": 38110 }, { "epoch": 14.390335975839939, "grad_norm": 1.111286997795105, "learning_rate": 6.033575667121317e-07, "loss": 0.3576, "step": 38120 }, { "epoch": 14.394110985277463, "grad_norm": 0.6550930738449097, "learning_rate": 5.969720969947967e-07, "loss": 0.355, "step": 38130 }, { "epoch": 14.397885994714986, "grad_norm": 0.6559245586395264, "learning_rate": 5.906203937811883e-07, "loss": 0.3628, "step": 38140 }, { "epoch": 14.40166100415251, "grad_norm": 0.6878376007080078, "learning_rate": 5.843024614126447e-07, "loss": 0.3554, "step": 38150 }, { "epoch": 14.405436013590034, "grad_norm": 0.644935667514801, "learning_rate": 5.780183042074061e-07, "loss": 0.3574, "step": 38160 }, { "epoch": 14.409211023027558, "grad_norm": 0.6619318723678589, "learning_rate": 5.717679264606312e-07, "loss": 0.3569, "step": 38170 }, { "epoch": 14.412986032465081, "grad_norm": 0.6411536931991577, "learning_rate": 5.655513324444028e-07, "loss": 0.3592, "step": 38180 }, { "epoch": 14.416761041902605, "grad_norm": 0.6617749929428101, "learning_rate": 5.593685264076942e-07, "loss": 0.3463, "step": 38190 }, { "epoch": 14.420536051340129, "grad_norm": 0.6915404796600342, "learning_rate": 5.532195125764029e-07, "loss": 0.3586, "step": 38200 }, { "epoch": 14.424311060777653, "grad_norm": 0.6411628723144531, "learning_rate": 5.471042951533167e-07, "loss": 0.3425, "step": 38210 }, { "epoch": 14.428086070215176, "grad_norm": 0.645287275314331, "learning_rate": 5.410228783181371e-07, "loss": 0.3582, "step": 38220 }, { "epoch": 14.431861079652698, "grad_norm": 0.6621134281158447, "learning_rate": 5.349752662274444e-07, "loss": 0.3567, "step": 38230 }, { "epoch": 14.435636089090222, "grad_norm": 0.6822801232337952, "learning_rate": 5.289614630147321e-07, "loss": 0.3463, "step": 38240 }, { "epoch": 14.439411098527746, "grad_norm": 0.6010865569114685, "learning_rate": 5.229814727903737e-07, "loss": 0.3603, "step": 38250 }, { "epoch": 14.44318610796527, "grad_norm": 0.7438214421272278, "learning_rate": 5.170352996416439e-07, "loss": 0.3529, "step": 38260 }, { "epoch": 14.446961117402793, "grad_norm": 0.6098829507827759, "learning_rate": 5.111229476326918e-07, "loss": 0.3516, "step": 38270 }, { "epoch": 14.450736126840317, "grad_norm": 0.6266598105430603, "learning_rate": 5.052444208045459e-07, "loss": 0.3569, "step": 38280 }, { "epoch": 14.45451113627784, "grad_norm": 0.6430060267448425, "learning_rate": 4.993997231751312e-07, "loss": 0.3495, "step": 38290 }, { "epoch": 14.458286145715364, "grad_norm": 0.6154651045799255, "learning_rate": 4.935888587392468e-07, "loss": 0.3546, "step": 38300 }, { "epoch": 14.462061155152888, "grad_norm": 0.6602094173431396, "learning_rate": 4.878118314685598e-07, "loss": 0.3539, "step": 38310 }, { "epoch": 14.465836164590412, "grad_norm": 0.7772398591041565, "learning_rate": 4.820686453116119e-07, "loss": 0.3564, "step": 38320 }, { "epoch": 14.469611174027936, "grad_norm": 0.629912257194519, "learning_rate": 4.763593041938186e-07, "loss": 0.355, "step": 38330 }, { "epoch": 14.47338618346546, "grad_norm": 0.6255464553833008, "learning_rate": 4.706838120174584e-07, "loss": 0.3516, "step": 38340 }, { "epoch": 14.477161192902983, "grad_norm": 0.6643255949020386, "learning_rate": 4.650421726616838e-07, "loss": 0.3522, "step": 38350 }, { "epoch": 14.480936202340505, "grad_norm": 0.6950973868370056, "learning_rate": 4.594343899824993e-07, "loss": 0.3541, "step": 38360 }, { "epoch": 14.484711211778029, "grad_norm": 0.7115216851234436, "learning_rate": 4.538604678127667e-07, "loss": 0.3539, "step": 38370 }, { "epoch": 14.488486221215553, "grad_norm": 0.6523749232292175, "learning_rate": 4.483204099622107e-07, "loss": 0.3555, "step": 38380 }, { "epoch": 14.492261230653076, "grad_norm": 0.6457560658454895, "learning_rate": 4.4281422021740793e-07, "loss": 0.3583, "step": 38390 }, { "epoch": 14.4960362400906, "grad_norm": 0.6862673163414001, "learning_rate": 4.373419023417924e-07, "loss": 0.3614, "step": 38400 }, { "epoch": 14.499811249528124, "grad_norm": 0.623674750328064, "learning_rate": 4.319034600756389e-07, "loss": 0.3476, "step": 38410 }, { "epoch": 14.503586258965647, "grad_norm": 0.6614101529121399, "learning_rate": 4.264988971360739e-07, "loss": 0.3646, "step": 38420 }, { "epoch": 14.507361268403171, "grad_norm": 0.650454044342041, "learning_rate": 4.211282172170594e-07, "loss": 0.3581, "step": 38430 }, { "epoch": 14.511136277840695, "grad_norm": 2.998237133026123, "learning_rate": 4.1579142398940894e-07, "loss": 0.3612, "step": 38440 }, { "epoch": 14.514911287278219, "grad_norm": 0.6962543725967407, "learning_rate": 4.1048852110076585e-07, "loss": 0.351, "step": 38450 }, { "epoch": 14.518686296715742, "grad_norm": 0.6083894371986389, "learning_rate": 4.052195121756197e-07, "loss": 0.3617, "step": 38460 }, { "epoch": 14.522461306153266, "grad_norm": 0.6192116737365723, "learning_rate": 3.9998440081528974e-07, "loss": 0.3561, "step": 38470 }, { "epoch": 14.52623631559079, "grad_norm": 0.755118727684021, "learning_rate": 3.9478319059791936e-07, "loss": 0.3502, "step": 38480 }, { "epoch": 14.530011325028312, "grad_norm": 0.606930673122406, "learning_rate": 3.8961588507848143e-07, "loss": 0.3569, "step": 38490 }, { "epoch": 14.533786334465836, "grad_norm": 0.6492167115211487, "learning_rate": 3.844824877887954e-07, "loss": 0.3622, "step": 38500 }, { "epoch": 14.53756134390336, "grad_norm": 0.771045982837677, "learning_rate": 3.7938300223748224e-07, "loss": 0.3567, "step": 38510 }, { "epoch": 14.541336353340883, "grad_norm": 0.6617088317871094, "learning_rate": 3.743174319099929e-07, "loss": 0.3589, "step": 38520 }, { "epoch": 14.545111362778407, "grad_norm": 0.7514125108718872, "learning_rate": 3.6928578026859673e-07, "loss": 0.3504, "step": 38530 }, { "epoch": 14.54888637221593, "grad_norm": 0.797964870929718, "learning_rate": 3.642880507523816e-07, "loss": 0.3609, "step": 38540 }, { "epoch": 14.552661381653454, "grad_norm": 0.6243863105773926, "learning_rate": 3.5932424677724285e-07, "loss": 0.3564, "step": 38550 }, { "epoch": 14.556436391090978, "grad_norm": 0.6841285824775696, "learning_rate": 3.5439437173589996e-07, "loss": 0.3466, "step": 38560 }, { "epoch": 14.560211400528502, "grad_norm": 0.6735568642616272, "learning_rate": 3.4949842899787423e-07, "loss": 0.3542, "step": 38570 }, { "epoch": 14.563986409966025, "grad_norm": 0.6594800353050232, "learning_rate": 3.446364219094944e-07, "loss": 0.3551, "step": 38580 }, { "epoch": 14.56776141940355, "grad_norm": 0.6403582692146301, "learning_rate": 3.3980835379389675e-07, "loss": 0.3547, "step": 38590 }, { "epoch": 14.571536428841073, "grad_norm": 0.7487730979919434, "learning_rate": 3.350142279510249e-07, "loss": 0.3494, "step": 38600 }, { "epoch": 14.575311438278595, "grad_norm": 0.6844361424446106, "learning_rate": 3.3025404765761324e-07, "loss": 0.3558, "step": 38610 }, { "epoch": 14.579086447716119, "grad_norm": 0.6371558904647827, "learning_rate": 3.255278161672037e-07, "loss": 0.354, "step": 38620 }, { "epoch": 14.582861457153642, "grad_norm": 0.6130198836326599, "learning_rate": 3.20835536710129e-07, "loss": 0.3504, "step": 38630 }, { "epoch": 14.586636466591166, "grad_norm": 0.6150995492935181, "learning_rate": 3.1617721249351804e-07, "loss": 0.3579, "step": 38640 }, { "epoch": 14.59041147602869, "grad_norm": 0.6839898824691772, "learning_rate": 3.115528467012963e-07, "loss": 0.3568, "step": 38650 }, { "epoch": 14.594186485466214, "grad_norm": 0.6125872731208801, "learning_rate": 3.0696244249416305e-07, "loss": 0.3575, "step": 38660 }, { "epoch": 14.597961494903737, "grad_norm": 0.6067168712615967, "learning_rate": 3.024060030096254e-07, "loss": 0.3578, "step": 38670 }, { "epoch": 14.601736504341261, "grad_norm": 0.6161952614784241, "learning_rate": 2.9788353136196435e-07, "loss": 0.3523, "step": 38680 }, { "epoch": 14.605511513778785, "grad_norm": 0.6272019147872925, "learning_rate": 2.9339503064224615e-07, "loss": 0.3504, "step": 38690 }, { "epoch": 14.609286523216308, "grad_norm": 1.410046935081482, "learning_rate": 2.889405039183224e-07, "loss": 0.3514, "step": 38700 }, { "epoch": 14.613061532653832, "grad_norm": 0.6313207745552063, "learning_rate": 2.8451995423481315e-07, "loss": 0.3526, "step": 38710 }, { "epoch": 14.616836542091356, "grad_norm": 0.5658713579177856, "learning_rate": 2.801333846131293e-07, "loss": 0.3529, "step": 38720 }, { "epoch": 14.620611551528878, "grad_norm": 0.6897971034049988, "learning_rate": 2.757807980514504e-07, "loss": 0.3529, "step": 38730 }, { "epoch": 14.624386560966402, "grad_norm": 0.6845794320106506, "learning_rate": 2.714621975247189e-07, "loss": 0.3556, "step": 38740 }, { "epoch": 14.628161570403925, "grad_norm": 0.6772444844245911, "learning_rate": 2.671775859846626e-07, "loss": 0.3499, "step": 38750 }, { "epoch": 14.63193657984145, "grad_norm": 0.5830336213111877, "learning_rate": 2.629269663597722e-07, "loss": 0.3561, "step": 38760 }, { "epoch": 14.635711589278973, "grad_norm": 1.1887177228927612, "learning_rate": 2.587103415553127e-07, "loss": 0.3635, "step": 38770 }, { "epoch": 14.639486598716497, "grad_norm": 0.6681121587753296, "learning_rate": 2.545277144533009e-07, "loss": 0.3584, "step": 38780 }, { "epoch": 14.64326160815402, "grad_norm": 0.6925423741340637, "learning_rate": 2.5037908791252227e-07, "loss": 0.352, "step": 38790 }, { "epoch": 14.647036617591544, "grad_norm": 0.6650035977363586, "learning_rate": 2.462644647685308e-07, "loss": 0.357, "step": 38800 }, { "epoch": 14.650811627029068, "grad_norm": 0.7153099775314331, "learning_rate": 2.421838478336269e-07, "loss": 0.3473, "step": 38810 }, { "epoch": 14.654586636466592, "grad_norm": 0.738753616809845, "learning_rate": 2.3813723989687398e-07, "loss": 0.358, "step": 38820 }, { "epoch": 14.658361645904115, "grad_norm": 0.6535517573356628, "learning_rate": 2.3412464372409293e-07, "loss": 0.36, "step": 38830 }, { "epoch": 14.662136655341639, "grad_norm": 0.6161585450172424, "learning_rate": 2.301460620578455e-07, "loss": 0.3573, "step": 38840 }, { "epoch": 14.665911664779163, "grad_norm": 0.6572152972221375, "learning_rate": 2.262014976174731e-07, "loss": 0.349, "step": 38850 }, { "epoch": 14.669686674216685, "grad_norm": 0.6994696855545044, "learning_rate": 2.2229095309903025e-07, "loss": 0.364, "step": 38860 }, { "epoch": 14.673461683654208, "grad_norm": 0.6057741641998291, "learning_rate": 2.184144311753511e-07, "loss": 0.3513, "step": 38870 }, { "epoch": 14.677236693091732, "grad_norm": 0.6180564165115356, "learning_rate": 2.14571934495994e-07, "loss": 0.3535, "step": 38880 }, { "epoch": 14.681011702529256, "grad_norm": 0.7175959944725037, "learning_rate": 2.1076346568726923e-07, "loss": 0.3528, "step": 38890 }, { "epoch": 14.68478671196678, "grad_norm": 0.7563894391059875, "learning_rate": 2.0698902735223346e-07, "loss": 0.3621, "step": 38900 }, { "epoch": 14.688561721404303, "grad_norm": 0.6452277302742004, "learning_rate": 2.0324862207067863e-07, "loss": 0.3579, "step": 38910 }, { "epoch": 14.692336730841827, "grad_norm": 0.613015353679657, "learning_rate": 1.9954225239913195e-07, "loss": 0.353, "step": 38920 }, { "epoch": 14.69611174027935, "grad_norm": 0.5917133688926697, "learning_rate": 1.9586992087087252e-07, "loss": 0.3599, "step": 38930 }, { "epoch": 14.699886749716875, "grad_norm": 0.61355060338974, "learning_rate": 1.9223162999588705e-07, "loss": 0.3512, "step": 38940 }, { "epoch": 14.703661759154398, "grad_norm": 0.6601895689964294, "learning_rate": 1.886273822609308e-07, "loss": 0.3498, "step": 38950 }, { "epoch": 14.707436768591922, "grad_norm": 0.6723988056182861, "learning_rate": 1.8505718012946094e-07, "loss": 0.3566, "step": 38960 }, { "epoch": 14.711211778029446, "grad_norm": 0.5990282893180847, "learning_rate": 1.8152102604168664e-07, "loss": 0.3544, "step": 38970 }, { "epoch": 14.71498678746697, "grad_norm": 0.7967637181282043, "learning_rate": 1.7801892241452457e-07, "loss": 0.3581, "step": 38980 }, { "epoch": 14.718761796904491, "grad_norm": 0.7710434198379517, "learning_rate": 1.745508716416433e-07, "loss": 0.3552, "step": 38990 }, { "epoch": 14.722536806342015, "grad_norm": 0.7385621666908264, "learning_rate": 1.7111687609341342e-07, "loss": 0.3571, "step": 39000 }, { "epoch": 14.726311815779539, "grad_norm": 0.6296396851539612, "learning_rate": 1.6771693811693523e-07, "loss": 0.358, "step": 39010 }, { "epoch": 14.730086825217063, "grad_norm": 0.6770869493484497, "learning_rate": 1.643510600360443e-07, "loss": 0.3515, "step": 39020 }, { "epoch": 14.733861834654586, "grad_norm": 0.6164085865020752, "learning_rate": 1.610192441512781e-07, "loss": 0.3659, "step": 39030 }, { "epoch": 14.73763684409211, "grad_norm": 1.3167994022369385, "learning_rate": 1.5772149273990956e-07, "loss": 0.3633, "step": 39040 }, { "epoch": 14.741411853529634, "grad_norm": 0.6471686959266663, "learning_rate": 1.5445780805591337e-07, "loss": 0.3429, "step": 39050 }, { "epoch": 14.745186862967158, "grad_norm": 0.6361897587776184, "learning_rate": 1.5122819232998852e-07, "loss": 0.3591, "step": 39060 }, { "epoch": 14.748961872404681, "grad_norm": 0.6572710275650024, "learning_rate": 1.4803264776955262e-07, "loss": 0.3654, "step": 39070 }, { "epoch": 14.752736881842205, "grad_norm": 0.6715047359466553, "learning_rate": 1.448711765587196e-07, "loss": 0.3587, "step": 39080 }, { "epoch": 14.756511891279729, "grad_norm": 0.6003199219703674, "learning_rate": 1.4174378085833885e-07, "loss": 0.3599, "step": 39090 }, { "epoch": 14.760286900717253, "grad_norm": 0.637787938117981, "learning_rate": 1.3865046280594486e-07, "loss": 0.3602, "step": 39100 }, { "epoch": 14.764061910154776, "grad_norm": 0.6328327655792236, "learning_rate": 1.3559122451579643e-07, "loss": 0.354, "step": 39110 }, { "epoch": 14.767836919592298, "grad_norm": 0.6023069024085999, "learning_rate": 1.3256606807884874e-07, "loss": 0.3549, "step": 39120 }, { "epoch": 14.771611929029822, "grad_norm": 0.631928563117981, "learning_rate": 1.2957499556277563e-07, "loss": 0.3608, "step": 39130 }, { "epoch": 14.775386938467346, "grad_norm": 0.6594300866127014, "learning_rate": 1.266180090119473e-07, "loss": 0.3556, "step": 39140 }, { "epoch": 14.77916194790487, "grad_norm": 0.6091174483299255, "learning_rate": 1.2369511044742488e-07, "loss": 0.3608, "step": 39150 }, { "epoch": 14.782936957342393, "grad_norm": 0.7767238020896912, "learning_rate": 1.2080630186698805e-07, "loss": 0.3599, "step": 39160 }, { "epoch": 14.786711966779917, "grad_norm": 0.6567763090133667, "learning_rate": 1.1795158524511851e-07, "loss": 0.3513, "step": 39170 }, { "epoch": 14.79048697621744, "grad_norm": 0.6258882880210876, "learning_rate": 1.1513096253297773e-07, "loss": 0.3553, "step": 39180 }, { "epoch": 14.794261985654964, "grad_norm": 0.6142001152038574, "learning_rate": 1.1234443565843466e-07, "loss": 0.3534, "step": 39190 }, { "epoch": 14.798036995092488, "grad_norm": 0.6969514489173889, "learning_rate": 1.0959200652606028e-07, "loss": 0.3575, "step": 39200 }, { "epoch": 14.801812004530012, "grad_norm": 0.729709804058075, "learning_rate": 1.0687367701711082e-07, "loss": 0.347, "step": 39210 }, { "epoch": 14.805587013967536, "grad_norm": 0.6622199416160583, "learning_rate": 1.0418944898953898e-07, "loss": 0.3567, "step": 39220 }, { "epoch": 14.809362023405058, "grad_norm": 0.7699572443962097, "learning_rate": 1.0153932427798829e-07, "loss": 0.3591, "step": 39230 }, { "epoch": 14.813137032842581, "grad_norm": 0.6490731239318848, "learning_rate": 9.892330469379874e-08, "loss": 0.359, "step": 39240 }, { "epoch": 14.816912042280105, "grad_norm": 0.6460186839103699, "learning_rate": 9.634139202498449e-08, "loss": 0.3545, "step": 39250 }, { "epoch": 14.820687051717629, "grad_norm": 0.789674699306488, "learning_rate": 9.379358803627281e-08, "loss": 0.3515, "step": 39260 }, { "epoch": 14.824462061155153, "grad_norm": 0.6353222131729126, "learning_rate": 9.12798944690485e-08, "loss": 0.354, "step": 39270 }, { "epoch": 14.828237070592676, "grad_norm": 0.6650903820991516, "learning_rate": 8.880031304141501e-08, "loss": 0.349, "step": 39280 }, { "epoch": 14.8320120800302, "grad_norm": 0.7060394287109375, "learning_rate": 8.635484544812777e-08, "loss": 0.3572, "step": 39290 }, { "epoch": 14.835787089467724, "grad_norm": 2.0902068614959717, "learning_rate": 8.394349336064422e-08, "loss": 0.3551, "step": 39300 }, { "epoch": 14.839562098905247, "grad_norm": 0.7247877717018127, "learning_rate": 8.156625842710153e-08, "loss": 0.3518, "step": 39310 }, { "epoch": 14.843337108342771, "grad_norm": 0.6578109264373779, "learning_rate": 7.92231422723222e-08, "loss": 0.3499, "step": 39320 }, { "epoch": 14.847112117780295, "grad_norm": 0.630346953868866, "learning_rate": 7.691414649780293e-08, "loss": 0.351, "step": 39330 }, { "epoch": 14.850887127217819, "grad_norm": 0.6793875098228455, "learning_rate": 7.463927268171467e-08, "loss": 0.3486, "step": 39340 }, { "epoch": 14.854662136655342, "grad_norm": 0.6113439798355103, "learning_rate": 7.239852237891365e-08, "loss": 0.3605, "step": 39350 }, { "epoch": 14.858437146092864, "grad_norm": 0.7049407958984375, "learning_rate": 7.019189712093588e-08, "loss": 0.3501, "step": 39360 }, { "epoch": 14.862212155530388, "grad_norm": 0.7017509937286377, "learning_rate": 6.801939841598603e-08, "loss": 0.3599, "step": 39370 }, { "epoch": 14.865987164967912, "grad_norm": 0.6390900015830994, "learning_rate": 6.588102774894856e-08, "loss": 0.3557, "step": 39380 }, { "epoch": 14.869762174405436, "grad_norm": 0.6362788081169128, "learning_rate": 6.377678658137653e-08, "loss": 0.3548, "step": 39390 }, { "epoch": 14.87353718384296, "grad_norm": 0.6280719041824341, "learning_rate": 6.17066763514973e-08, "loss": 0.3493, "step": 39400 }, { "epoch": 14.877312193280483, "grad_norm": 0.6526157855987549, "learning_rate": 5.967069847421792e-08, "loss": 0.3532, "step": 39410 }, { "epoch": 14.881087202718007, "grad_norm": 0.7428958415985107, "learning_rate": 5.7668854341108583e-08, "loss": 0.3563, "step": 39420 }, { "epoch": 14.88486221215553, "grad_norm": 0.6039178371429443, "learning_rate": 5.570114532040815e-08, "loss": 0.3567, "step": 39430 }, { "epoch": 14.888637221593054, "grad_norm": 1.0244649648666382, "learning_rate": 5.3767572757029706e-08, "loss": 0.3521, "step": 39440 }, { "epoch": 14.892412231030578, "grad_norm": 0.6426689028739929, "learning_rate": 5.186813797254386e-08, "loss": 0.3525, "step": 39450 }, { "epoch": 14.896187240468102, "grad_norm": 0.7617814540863037, "learning_rate": 5.000284226521213e-08, "loss": 0.3529, "step": 39460 }, { "epoch": 14.899962249905625, "grad_norm": 0.7546370625495911, "learning_rate": 4.8171686909931346e-08, "loss": 0.3519, "step": 39470 }, { "epoch": 14.90373725934315, "grad_norm": 0.6495998501777649, "learning_rate": 4.637467315828925e-08, "loss": 0.358, "step": 39480 }, { "epoch": 14.907512268780671, "grad_norm": 0.6548603177070618, "learning_rate": 4.461180223852002e-08, "loss": 0.3483, "step": 39490 }, { "epoch": 14.911287278218195, "grad_norm": 0.6339406371116638, "learning_rate": 4.288307535553204e-08, "loss": 0.3604, "step": 39500 }, { "epoch": 14.915062287655719, "grad_norm": 0.6532941460609436, "learning_rate": 4.1188493690896836e-08, "loss": 0.3598, "step": 39510 }, { "epoch": 14.918837297093242, "grad_norm": 0.6393380761146545, "learning_rate": 3.952805840284346e-08, "loss": 0.3521, "step": 39520 }, { "epoch": 14.922612306530766, "grad_norm": 0.6625932455062866, "learning_rate": 3.790177062625855e-08, "loss": 0.3567, "step": 39530 }, { "epoch": 14.92638731596829, "grad_norm": 0.6500564217567444, "learning_rate": 3.6309631472702946e-08, "loss": 0.3549, "step": 39540 }, { "epoch": 14.930162325405814, "grad_norm": 0.6394986510276794, "learning_rate": 3.475164203038395e-08, "loss": 0.3602, "step": 39550 }, { "epoch": 14.933937334843337, "grad_norm": 0.6065540313720703, "learning_rate": 3.3227803364171974e-08, "loss": 0.3435, "step": 39560 }, { "epoch": 14.937712344280861, "grad_norm": 0.6980146765708923, "learning_rate": 3.173811651560055e-08, "loss": 0.3484, "step": 39570 }, { "epoch": 14.941487353718385, "grad_norm": 0.6110873222351074, "learning_rate": 3.0282582502855204e-08, "loss": 0.3528, "step": 39580 }, { "epoch": 14.945262363155909, "grad_norm": 0.74271559715271, "learning_rate": 2.8861202320773496e-08, "loss": 0.3547, "step": 39590 }, { "epoch": 14.949037372593432, "grad_norm": 0.5941160321235657, "learning_rate": 2.7473976940867175e-08, "loss": 0.3606, "step": 39600 }, { "epoch": 14.952812382030956, "grad_norm": 0.6520451903343201, "learning_rate": 2.6120907311277808e-08, "loss": 0.351, "step": 39610 }, { "epoch": 14.956587391468478, "grad_norm": 0.6158818006515503, "learning_rate": 2.480199435682673e-08, "loss": 0.3502, "step": 39620 }, { "epoch": 14.960362400906002, "grad_norm": 0.6889585852622986, "learning_rate": 2.3517238978976174e-08, "loss": 0.3592, "step": 39630 }, { "epoch": 14.964137410343525, "grad_norm": 0.5970800518989563, "learning_rate": 2.2266642055834842e-08, "loss": 0.359, "step": 39640 }, { "epoch": 14.96791241978105, "grad_norm": 0.65416419506073, "learning_rate": 2.105020444219119e-08, "loss": 0.3527, "step": 39650 }, { "epoch": 14.971687429218573, "grad_norm": 0.664504885673523, "learning_rate": 1.9867926969452388e-08, "loss": 0.364, "step": 39660 }, { "epoch": 14.975462438656097, "grad_norm": 0.6672757863998413, "learning_rate": 1.8719810445705366e-08, "loss": 0.3531, "step": 39670 }, { "epoch": 14.97923744809362, "grad_norm": 0.6902869343757629, "learning_rate": 1.760585565567241e-08, "loss": 0.3564, "step": 39680 }, { "epoch": 14.983012457531144, "grad_norm": 0.7175024151802063, "learning_rate": 1.6526063360727816e-08, "loss": 0.3559, "step": 39690 }, { "epoch": 14.986787466968668, "grad_norm": 0.6465026140213013, "learning_rate": 1.5480434298908997e-08, "loss": 0.3539, "step": 39700 }, { "epoch": 14.990562476406192, "grad_norm": 0.6712453961372375, "learning_rate": 1.4468969184883164e-08, "loss": 0.3535, "step": 39710 }, { "epoch": 14.994337485843715, "grad_norm": 0.620090126991272, "learning_rate": 1.3491668709980643e-08, "loss": 0.3544, "step": 39720 }, { "epoch": 14.998112495281239, "grad_norm": 0.6814067363739014, "learning_rate": 1.254853354217822e-08, "loss": 0.3551, "step": 39730 }, { "epoch": 15.001887504718761, "grad_norm": 0.8120419979095459, "learning_rate": 1.1639564326099139e-08, "loss": 0.3553, "step": 39740 }, { "epoch": 15.005662514156285, "grad_norm": 0.6603171825408936, "learning_rate": 1.0764761683018653e-08, "loss": 0.3537, "step": 39750 }, { "epoch": 15.009437523593808, "grad_norm": 0.6315841674804688, "learning_rate": 9.924126210852925e-09, "loss": 0.3503, "step": 39760 }, { "epoch": 15.013212533031332, "grad_norm": 0.6496405601501465, "learning_rate": 9.117658484170122e-09, "loss": 0.3577, "step": 39770 }, { "epoch": 15.016987542468856, "grad_norm": 0.7022120356559753, "learning_rate": 8.345359054179324e-09, "loss": 0.3678, "step": 39780 }, { "epoch": 15.02076255190638, "grad_norm": 0.7122491598129272, "learning_rate": 7.60722844874162e-09, "loss": 0.3561, "step": 39790 }, { "epoch": 15.024537561343903, "grad_norm": 0.625100314617157, "learning_rate": 6.903267172364558e-09, "loss": 0.347, "step": 39800 }, { "epoch": 15.028312570781427, "grad_norm": 0.6396900415420532, "learning_rate": 6.23347570619659e-09, "loss": 0.3605, "step": 39810 }, { "epoch": 15.03208758021895, "grad_norm": 0.5994987487792969, "learning_rate": 5.59785450803818e-09, "loss": 0.3573, "step": 39820 }, { "epoch": 15.035862589656475, "grad_norm": 0.7558580636978149, "learning_rate": 4.996404012325151e-09, "loss": 0.3536, "step": 39830 }, { "epoch": 15.039637599093998, "grad_norm": 0.6943445801734924, "learning_rate": 4.429124630145331e-09, "loss": 0.3561, "step": 39840 }, { "epoch": 15.043412608531522, "grad_norm": 0.9574018716812134, "learning_rate": 3.89601674922746e-09, "loss": 0.362, "step": 39850 }, { "epoch": 15.047187617969046, "grad_norm": 0.6449063420295715, "learning_rate": 3.3970807339467336e-09, "loss": 0.3537, "step": 39860 }, { "epoch": 15.050962627406568, "grad_norm": 0.5976826548576355, "learning_rate": 2.9323169253248073e-09, "loss": 0.3513, "step": 39870 }, { "epoch": 15.054737636844091, "grad_norm": 0.6099610328674316, "learning_rate": 2.501725641018693e-09, "loss": 0.3545, "step": 39880 }, { "epoch": 15.058512646281615, "grad_norm": 0.7549223899841309, "learning_rate": 2.105307175331861e-09, "loss": 0.3567, "step": 39890 }, { "epoch": 15.062287655719139, "grad_norm": 0.6407955288887024, "learning_rate": 1.7430617992197918e-09, "loss": 0.3558, "step": 39900 }, { "epoch": 15.066062665156663, "grad_norm": 0.6533637046813965, "learning_rate": 1.4149897602677708e-09, "loss": 0.3484, "step": 39910 }, { "epoch": 15.069837674594186, "grad_norm": 0.604796826839447, "learning_rate": 1.1210912827130937e-09, "loss": 0.3582, "step": 39920 }, { "epoch": 15.07361268403171, "grad_norm": 0.5980611443519592, "learning_rate": 8.613665674284121e-10, "loss": 0.3548, "step": 39930 }, { "epoch": 15.077387693469234, "grad_norm": 0.6383752226829529, "learning_rate": 6.358157919383878e-10, "loss": 0.3544, "step": 39940 }, { "epoch": 15.081162702906758, "grad_norm": 0.7199327349662781, "learning_rate": 4.4443911040303877e-10, "loss": 0.3526, "step": 39950 }, { "epoch": 15.084937712344281, "grad_norm": 0.7080827951431274, "learning_rate": 2.8723665362329066e-10, "loss": 0.3549, "step": 39960 }, { "epoch": 15.088712721781805, "grad_norm": 0.6483982801437378, "learning_rate": 1.6420852905207895e-10, "loss": 0.3583, "step": 39970 }, { "epoch": 15.092487731219329, "grad_norm": 0.6438269019126892, "learning_rate": 7.535482076659328e-11, "loss": 0.3451, "step": 39980 }, { "epoch": 15.09626274065685, "grad_norm": 0.6255228519439697, "learning_rate": 2.0675589512686443e-11, "loss": 0.3557, "step": 39990 }, { "epoch": 15.100037750094375, "grad_norm": 0.8365612626075745, "learning_rate": 1.7087265491433358e-13, "loss": 0.353, "step": 40000 } ], "logging_steps": 10, "max_steps": 40000, "num_input_tokens_seen": 0, "num_train_epochs": 16, "save_steps": 4000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0739727642197072e+20, "train_batch_size": 32, "trial_name": null, "trial_params": null }