diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,28034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 15.100037750094375, + "eval_steps": 500, + "global_step": 40000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.003775009437523594, + "grad_norm": 12.114919662475586, + "learning_rate": 4.5e-07, + "loss": 4.3735, + "step": 10 + }, + { + "epoch": 0.007550018875047188, + "grad_norm": 9.25322437286377, + "learning_rate": 9.5e-07, + "loss": 4.2737, + "step": 20 + }, + { + "epoch": 0.011325028312570781, + "grad_norm": 11.06075668334961, + "learning_rate": 1.45e-06, + "loss": 4.3709, + "step": 30 + }, + { + "epoch": 0.015100037750094376, + "grad_norm": 11.528477668762207, + "learning_rate": 1.95e-06, + "loss": 4.5857, + "step": 40 + }, + { + "epoch": 0.01887504718761797, + "grad_norm": 9.225029945373535, + "learning_rate": 2.4500000000000003e-06, + "loss": 4.1594, + "step": 50 + }, + { + "epoch": 0.022650056625141562, + "grad_norm": 9.194082260131836, + "learning_rate": 2.95e-06, + "loss": 3.9774, + "step": 60 + }, + { + "epoch": 0.02642506606266516, + "grad_norm": 9.647115707397461, + "learning_rate": 3.4500000000000004e-06, + "loss": 3.7629, + "step": 70 + }, + { + "epoch": 0.03020007550018875, + "grad_norm": 7.582672595977783, + "learning_rate": 3.95e-06, + "loss": 3.3272, + "step": 80 + }, + { + "epoch": 0.03397508493771234, + "grad_norm": 6.46481466293335, + "learning_rate": 4.45e-06, + "loss": 2.9437, + "step": 90 + }, + { + "epoch": 0.03775009437523594, + "grad_norm": 4.374029159545898, + "learning_rate": 4.950000000000001e-06, + "loss": 2.6371, + "step": 100 + }, + { + "epoch": 0.041525103812759534, + "grad_norm": 2.939470052719116, + "learning_rate": 5.45e-06, + "loss": 2.2299, + "step": 110 + }, + { + "epoch": 0.045300113250283124, + "grad_norm": 2.47430682182312, + "learning_rate": 5.95e-06, + "loss": 2.0141, + "step": 120 + }, + { + "epoch": 0.04907512268780672, + "grad_norm": 1.9557963609695435, + "learning_rate": 6.45e-06, + "loss": 1.8435, + "step": 130 + }, + { + "epoch": 0.05285013212533032, + "grad_norm": 1.204042673110962, + "learning_rate": 6.950000000000001e-06, + "loss": 1.5451, + "step": 140 + }, + { + "epoch": 0.056625141562853906, + "grad_norm": 0.9715531468391418, + "learning_rate": 7.45e-06, + "loss": 1.5027, + "step": 150 + }, + { + "epoch": 0.0604001510003775, + "grad_norm": 1.261393427848816, + "learning_rate": 7.95e-06, + "loss": 1.3894, + "step": 160 + }, + { + "epoch": 0.0641751604379011, + "grad_norm": 1.091807246208191, + "learning_rate": 8.45e-06, + "loss": 1.3591, + "step": 170 + }, + { + "epoch": 0.06795016987542468, + "grad_norm": 0.8383681178092957, + "learning_rate": 8.95e-06, + "loss": 1.2757, + "step": 180 + }, + { + "epoch": 0.07172517931294828, + "grad_norm": 1.3401917219161987, + "learning_rate": 9.450000000000001e-06, + "loss": 1.2084, + "step": 190 + }, + { + "epoch": 0.07550018875047187, + "grad_norm": 0.8086752891540527, + "learning_rate": 9.950000000000001e-06, + "loss": 1.3253, + "step": 200 + }, + { + "epoch": 0.07927519818799547, + "grad_norm": 0.9537988901138306, + "learning_rate": 1.045e-05, + "loss": 1.171, + "step": 210 + }, + { + "epoch": 0.08305020762551907, + "grad_norm": 0.8162275552749634, + "learning_rate": 1.095e-05, + "loss": 1.1737, + "step": 220 + }, + { + "epoch": 0.08682521706304266, + "grad_norm": 1.144768476486206, + "learning_rate": 1.145e-05, + "loss": 1.1406, + "step": 230 + }, + { + "epoch": 0.09060022650056625, + "grad_norm": 1.2052839994430542, + "learning_rate": 1.195e-05, + "loss": 1.0841, + "step": 240 + }, + { + "epoch": 0.09437523593808984, + "grad_norm": 1.1619216203689575, + "learning_rate": 1.2450000000000001e-05, + "loss": 1.2028, + "step": 250 + }, + { + "epoch": 0.09815024537561344, + "grad_norm": 0.971839189529419, + "learning_rate": 1.2950000000000001e-05, + "loss": 1.0536, + "step": 260 + }, + { + "epoch": 0.10192525481313704, + "grad_norm": 1.1455223560333252, + "learning_rate": 1.3450000000000002e-05, + "loss": 1.0539, + "step": 270 + }, + { + "epoch": 0.10570026425066063, + "grad_norm": 0.9137556552886963, + "learning_rate": 1.3950000000000002e-05, + "loss": 1.0439, + "step": 280 + }, + { + "epoch": 0.10947527368818422, + "grad_norm": 1.4055426120758057, + "learning_rate": 1.4449999999999999e-05, + "loss": 1.147, + "step": 290 + }, + { + "epoch": 0.11325028312570781, + "grad_norm": 1.3591498136520386, + "learning_rate": 1.4950000000000001e-05, + "loss": 1.0227, + "step": 300 + }, + { + "epoch": 0.11702529256323141, + "grad_norm": 1.004621148109436, + "learning_rate": 1.545e-05, + "loss": 1.0075, + "step": 310 + }, + { + "epoch": 0.120800302000755, + "grad_norm": 1.2812256813049316, + "learning_rate": 1.595e-05, + "loss": 1.0799, + "step": 320 + }, + { + "epoch": 0.1245753114382786, + "grad_norm": 1.5927451848983765, + "learning_rate": 1.645e-05, + "loss": 0.9892, + "step": 330 + }, + { + "epoch": 0.1283503208758022, + "grad_norm": 1.2106682062149048, + "learning_rate": 1.6950000000000002e-05, + "loss": 0.9776, + "step": 340 + }, + { + "epoch": 0.1321253303133258, + "grad_norm": 1.2619518041610718, + "learning_rate": 1.745e-05, + "loss": 0.9786, + "step": 350 + }, + { + "epoch": 0.13590033975084936, + "grad_norm": 1.5045936107635498, + "learning_rate": 1.795e-05, + "loss": 0.9763, + "step": 360 + }, + { + "epoch": 0.13967534918837296, + "grad_norm": 1.4551500082015991, + "learning_rate": 1.845e-05, + "loss": 0.9857, + "step": 370 + }, + { + "epoch": 0.14345035862589656, + "grad_norm": 1.3843104839324951, + "learning_rate": 1.895e-05, + "loss": 0.9163, + "step": 380 + }, + { + "epoch": 0.14722536806342015, + "grad_norm": 1.4003466367721558, + "learning_rate": 1.9450000000000002e-05, + "loss": 0.9637, + "step": 390 + }, + { + "epoch": 0.15100037750094375, + "grad_norm": 1.3285315036773682, + "learning_rate": 1.995e-05, + "loss": 0.9353, + "step": 400 + }, + { + "epoch": 0.15477538693846735, + "grad_norm": 1.0920555591583252, + "learning_rate": 2.045e-05, + "loss": 1.0432, + "step": 410 + }, + { + "epoch": 0.15855039637599094, + "grad_norm": 1.4697812795639038, + "learning_rate": 2.095e-05, + "loss": 1.0497, + "step": 420 + }, + { + "epoch": 0.16232540581351454, + "grad_norm": 1.444948434829712, + "learning_rate": 2.145e-05, + "loss": 0.9317, + "step": 430 + }, + { + "epoch": 0.16610041525103814, + "grad_norm": 1.8489030599594116, + "learning_rate": 2.195e-05, + "loss": 0.9296, + "step": 440 + }, + { + "epoch": 0.16987542468856173, + "grad_norm": 1.27092444896698, + "learning_rate": 2.245e-05, + "loss": 0.8998, + "step": 450 + }, + { + "epoch": 0.17365043412608533, + "grad_norm": 1.3666588068008423, + "learning_rate": 2.2950000000000002e-05, + "loss": 1.0234, + "step": 460 + }, + { + "epoch": 0.1774254435636089, + "grad_norm": 0.9460456967353821, + "learning_rate": 2.345e-05, + "loss": 0.9337, + "step": 470 + }, + { + "epoch": 0.1812004530011325, + "grad_norm": 1.4684996604919434, + "learning_rate": 2.395e-05, + "loss": 0.9308, + "step": 480 + }, + { + "epoch": 0.1849754624386561, + "grad_norm": 1.284343957901001, + "learning_rate": 2.445e-05, + "loss": 0.908, + "step": 490 + }, + { + "epoch": 0.1887504718761797, + "grad_norm": 0.9511118531227112, + "learning_rate": 2.495e-05, + "loss": 0.8959, + "step": 500 + }, + { + "epoch": 0.19252548131370328, + "grad_norm": 1.701767086982727, + "learning_rate": 2.5450000000000002e-05, + "loss": 0.9207, + "step": 510 + }, + { + "epoch": 0.19630049075122688, + "grad_norm": 1.341230034828186, + "learning_rate": 2.595e-05, + "loss": 0.9837, + "step": 520 + }, + { + "epoch": 0.20007550018875048, + "grad_norm": 1.1268380880355835, + "learning_rate": 2.6450000000000003e-05, + "loss": 0.8893, + "step": 530 + }, + { + "epoch": 0.20385050962627407, + "grad_norm": 1.5450079441070557, + "learning_rate": 2.6950000000000005e-05, + "loss": 0.8875, + "step": 540 + }, + { + "epoch": 0.20762551906379767, + "grad_norm": 1.827506184577942, + "learning_rate": 2.7450000000000003e-05, + "loss": 0.9069, + "step": 550 + }, + { + "epoch": 0.21140052850132127, + "grad_norm": 1.225759506225586, + "learning_rate": 2.7950000000000005e-05, + "loss": 0.8977, + "step": 560 + }, + { + "epoch": 0.21517553793884484, + "grad_norm": 1.1317689418792725, + "learning_rate": 2.845e-05, + "loss": 0.8673, + "step": 570 + }, + { + "epoch": 0.21895054737636843, + "grad_norm": 1.3549116849899292, + "learning_rate": 2.895e-05, + "loss": 0.8613, + "step": 580 + }, + { + "epoch": 0.22272555681389203, + "grad_norm": 1.5646511316299438, + "learning_rate": 2.945e-05, + "loss": 0.8847, + "step": 590 + }, + { + "epoch": 0.22650056625141562, + "grad_norm": 1.453460693359375, + "learning_rate": 2.995e-05, + "loss": 0.8523, + "step": 600 + }, + { + "epoch": 0.23027557568893922, + "grad_norm": 1.3521499633789062, + "learning_rate": 3.045e-05, + "loss": 0.8773, + "step": 610 + }, + { + "epoch": 0.23405058512646282, + "grad_norm": 1.041245698928833, + "learning_rate": 3.095e-05, + "loss": 0.857, + "step": 620 + }, + { + "epoch": 0.23782559456398641, + "grad_norm": 1.2298246622085571, + "learning_rate": 3.145e-05, + "loss": 0.8364, + "step": 630 + }, + { + "epoch": 0.24160060400151, + "grad_norm": 1.2487760782241821, + "learning_rate": 3.1950000000000004e-05, + "loss": 0.8697, + "step": 640 + }, + { + "epoch": 0.2453756134390336, + "grad_norm": 1.3753890991210938, + "learning_rate": 3.245e-05, + "loss": 0.8579, + "step": 650 + }, + { + "epoch": 0.2491506228765572, + "grad_norm": 1.354738712310791, + "learning_rate": 3.295e-05, + "loss": 0.8574, + "step": 660 + }, + { + "epoch": 0.2529256323140808, + "grad_norm": 1.3853284120559692, + "learning_rate": 3.345000000000001e-05, + "loss": 0.8346, + "step": 670 + }, + { + "epoch": 0.2567006417516044, + "grad_norm": 1.4447784423828125, + "learning_rate": 3.3950000000000005e-05, + "loss": 0.9262, + "step": 680 + }, + { + "epoch": 0.26047565118912797, + "grad_norm": 1.2130733728408813, + "learning_rate": 3.445e-05, + "loss": 0.8176, + "step": 690 + }, + { + "epoch": 0.2642506606266516, + "grad_norm": 1.698705792427063, + "learning_rate": 3.495e-05, + "loss": 0.8424, + "step": 700 + }, + { + "epoch": 0.26802567006417516, + "grad_norm": 1.724214792251587, + "learning_rate": 3.545e-05, + "loss": 0.9077, + "step": 710 + }, + { + "epoch": 0.2718006795016987, + "grad_norm": 1.296976923942566, + "learning_rate": 3.595e-05, + "loss": 0.8253, + "step": 720 + }, + { + "epoch": 0.27557568893922235, + "grad_norm": 1.2953417301177979, + "learning_rate": 3.645e-05, + "loss": 0.8949, + "step": 730 + }, + { + "epoch": 0.2793506983767459, + "grad_norm": 1.267232894897461, + "learning_rate": 3.6950000000000004e-05, + "loss": 0.8476, + "step": 740 + }, + { + "epoch": 0.28312570781426954, + "grad_norm": 1.76276695728302, + "learning_rate": 3.745e-05, + "loss": 0.8981, + "step": 750 + }, + { + "epoch": 0.2869007172517931, + "grad_norm": 1.5421119928359985, + "learning_rate": 3.795e-05, + "loss": 0.8343, + "step": 760 + }, + { + "epoch": 0.29067572668931674, + "grad_norm": 1.1343638896942139, + "learning_rate": 3.845e-05, + "loss": 0.8316, + "step": 770 + }, + { + "epoch": 0.2944507361268403, + "grad_norm": 0.9002248048782349, + "learning_rate": 3.8950000000000005e-05, + "loss": 0.8207, + "step": 780 + }, + { + "epoch": 0.29822574556436393, + "grad_norm": 1.273121953010559, + "learning_rate": 3.9450000000000003e-05, + "loss": 0.8782, + "step": 790 + }, + { + "epoch": 0.3020007550018875, + "grad_norm": 1.6263741254806519, + "learning_rate": 3.995e-05, + "loss": 0.8359, + "step": 800 + }, + { + "epoch": 0.3057757644394111, + "grad_norm": 1.3746614456176758, + "learning_rate": 4.045000000000001e-05, + "loss": 0.7946, + "step": 810 + }, + { + "epoch": 0.3095507738769347, + "grad_norm": 1.2415761947631836, + "learning_rate": 4.095e-05, + "loss": 0.7973, + "step": 820 + }, + { + "epoch": 0.31332578331445826, + "grad_norm": 1.372182846069336, + "learning_rate": 4.145e-05, + "loss": 0.9107, + "step": 830 + }, + { + "epoch": 0.3171007927519819, + "grad_norm": 1.5173795223236084, + "learning_rate": 4.195e-05, + "loss": 0.7988, + "step": 840 + }, + { + "epoch": 0.32087580218950545, + "grad_norm": 1.5358010530471802, + "learning_rate": 4.245e-05, + "loss": 0.7724, + "step": 850 + }, + { + "epoch": 0.3246508116270291, + "grad_norm": 0.867175817489624, + "learning_rate": 4.295e-05, + "loss": 0.7816, + "step": 860 + }, + { + "epoch": 0.32842582106455265, + "grad_norm": 0.9137406349182129, + "learning_rate": 4.345e-05, + "loss": 0.7855, + "step": 870 + }, + { + "epoch": 0.33220083050207627, + "grad_norm": 0.9933797717094421, + "learning_rate": 4.3950000000000004e-05, + "loss": 0.8015, + "step": 880 + }, + { + "epoch": 0.33597583993959984, + "grad_norm": 1.1218225955963135, + "learning_rate": 4.445e-05, + "loss": 0.7944, + "step": 890 + }, + { + "epoch": 0.33975084937712347, + "grad_norm": 0.9928627610206604, + "learning_rate": 4.495e-05, + "loss": 0.785, + "step": 900 + }, + { + "epoch": 0.34352585881464703, + "grad_norm": 1.0522204637527466, + "learning_rate": 4.545000000000001e-05, + "loss": 0.7819, + "step": 910 + }, + { + "epoch": 0.34730086825217066, + "grad_norm": 1.162044644355774, + "learning_rate": 4.5950000000000006e-05, + "loss": 0.7721, + "step": 920 + }, + { + "epoch": 0.3510758776896942, + "grad_norm": 0.9964906573295593, + "learning_rate": 4.6450000000000004e-05, + "loss": 0.7716, + "step": 930 + }, + { + "epoch": 0.3548508871272178, + "grad_norm": 1.3030009269714355, + "learning_rate": 4.695e-05, + "loss": 0.7756, + "step": 940 + }, + { + "epoch": 0.3586258965647414, + "grad_norm": 1.413189172744751, + "learning_rate": 4.745e-05, + "loss": 0.7709, + "step": 950 + }, + { + "epoch": 0.362400906002265, + "grad_norm": 1.1650564670562744, + "learning_rate": 4.795e-05, + "loss": 0.7647, + "step": 960 + }, + { + "epoch": 0.3661759154397886, + "grad_norm": 1.0290861129760742, + "learning_rate": 4.845e-05, + "loss": 0.7601, + "step": 970 + }, + { + "epoch": 0.3699509248773122, + "grad_norm": 1.1548311710357666, + "learning_rate": 4.8950000000000004e-05, + "loss": 0.8407, + "step": 980 + }, + { + "epoch": 0.3737259343148358, + "grad_norm": 1.037048578262329, + "learning_rate": 4.945e-05, + "loss": 0.7576, + "step": 990 + }, + { + "epoch": 0.3775009437523594, + "grad_norm": 1.5112618207931519, + "learning_rate": 4.995e-05, + "loss": 0.7462, + "step": 1000 + }, + { + "epoch": 0.381275953189883, + "grad_norm": 1.65474534034729, + "learning_rate": 5.045e-05, + "loss": 0.7722, + "step": 1010 + }, + { + "epoch": 0.38505096262740657, + "grad_norm": 1.3295589685440063, + "learning_rate": 5.095e-05, + "loss": 0.7551, + "step": 1020 + }, + { + "epoch": 0.38882597206493014, + "grad_norm": 0.9878939390182495, + "learning_rate": 5.145e-05, + "loss": 0.7522, + "step": 1030 + }, + { + "epoch": 0.39260098150245376, + "grad_norm": 1.3945156335830688, + "learning_rate": 5.1949999999999996e-05, + "loss": 0.7827, + "step": 1040 + }, + { + "epoch": 0.39637599093997733, + "grad_norm": 1.2583463191986084, + "learning_rate": 5.245e-05, + "loss": 0.7658, + "step": 1050 + }, + { + "epoch": 0.40015100037750095, + "grad_norm": 1.0991382598876953, + "learning_rate": 5.295e-05, + "loss": 0.8056, + "step": 1060 + }, + { + "epoch": 0.4039260098150245, + "grad_norm": 1.0522119998931885, + "learning_rate": 5.345e-05, + "loss": 0.8069, + "step": 1070 + }, + { + "epoch": 0.40770101925254815, + "grad_norm": 0.8524134755134583, + "learning_rate": 5.3950000000000004e-05, + "loss": 0.8186, + "step": 1080 + }, + { + "epoch": 0.4114760286900717, + "grad_norm": 1.1254100799560547, + "learning_rate": 5.445e-05, + "loss": 0.7362, + "step": 1090 + }, + { + "epoch": 0.41525103812759534, + "grad_norm": 0.7959914207458496, + "learning_rate": 5.495e-05, + "loss": 0.7387, + "step": 1100 + }, + { + "epoch": 0.4190260475651189, + "grad_norm": 1.2034642696380615, + "learning_rate": 5.545e-05, + "loss": 0.7458, + "step": 1110 + }, + { + "epoch": 0.42280105700264253, + "grad_norm": 1.2874923944473267, + "learning_rate": 5.5950000000000005e-05, + "loss": 0.7442, + "step": 1120 + }, + { + "epoch": 0.4265760664401661, + "grad_norm": 1.2025848627090454, + "learning_rate": 5.645e-05, + "loss": 0.7624, + "step": 1130 + }, + { + "epoch": 0.43035107587768967, + "grad_norm": 1.1338868141174316, + "learning_rate": 5.695e-05, + "loss": 0.7311, + "step": 1140 + }, + { + "epoch": 0.4341260853152133, + "grad_norm": 1.1724085807800293, + "learning_rate": 5.745e-05, + "loss": 0.7383, + "step": 1150 + }, + { + "epoch": 0.43790109475273686, + "grad_norm": 1.3248220682144165, + "learning_rate": 5.7950000000000006e-05, + "loss": 0.7379, + "step": 1160 + }, + { + "epoch": 0.4416761041902605, + "grad_norm": 1.4836875200271606, + "learning_rate": 5.8450000000000005e-05, + "loss": 0.7369, + "step": 1170 + }, + { + "epoch": 0.44545111362778406, + "grad_norm": 1.1852290630340576, + "learning_rate": 5.895e-05, + "loss": 0.8117, + "step": 1180 + }, + { + "epoch": 0.4492261230653077, + "grad_norm": 2.466386079788208, + "learning_rate": 5.945000000000001e-05, + "loss": 0.7997, + "step": 1190 + }, + { + "epoch": 0.45300113250283125, + "grad_norm": 1.0344855785369873, + "learning_rate": 5.995000000000001e-05, + "loss": 0.7181, + "step": 1200 + }, + { + "epoch": 0.4567761419403549, + "grad_norm": 1.1200584173202515, + "learning_rate": 6.0450000000000006e-05, + "loss": 0.7204, + "step": 1210 + }, + { + "epoch": 0.46055115137787844, + "grad_norm": 1.0157345533370972, + "learning_rate": 6.0950000000000004e-05, + "loss": 0.7261, + "step": 1220 + }, + { + "epoch": 0.464326160815402, + "grad_norm": 0.9601739048957825, + "learning_rate": 6.145e-05, + "loss": 0.7099, + "step": 1230 + }, + { + "epoch": 0.46810117025292564, + "grad_norm": 0.8642465472221375, + "learning_rate": 6.195e-05, + "loss": 0.7242, + "step": 1240 + }, + { + "epoch": 0.4718761796904492, + "grad_norm": 0.9395157098770142, + "learning_rate": 6.245000000000001e-05, + "loss": 0.716, + "step": 1250 + }, + { + "epoch": 0.47565118912797283, + "grad_norm": 1.0848698616027832, + "learning_rate": 6.295e-05, + "loss": 0.714, + "step": 1260 + }, + { + "epoch": 0.4794261985654964, + "grad_norm": 1.1483030319213867, + "learning_rate": 6.345e-05, + "loss": 0.7224, + "step": 1270 + }, + { + "epoch": 0.48320120800302, + "grad_norm": 0.8236299753189087, + "learning_rate": 6.395e-05, + "loss": 0.7639, + "step": 1280 + }, + { + "epoch": 0.4869762174405436, + "grad_norm": 0.9821550846099854, + "learning_rate": 6.445e-05, + "loss": 0.7136, + "step": 1290 + }, + { + "epoch": 0.4907512268780672, + "grad_norm": 0.94017493724823, + "learning_rate": 6.494999999999999e-05, + "loss": 0.7056, + "step": 1300 + }, + { + "epoch": 0.4945262363155908, + "grad_norm": 0.8587389588356018, + "learning_rate": 6.545e-05, + "loss": 0.7518, + "step": 1310 + }, + { + "epoch": 0.4983012457531144, + "grad_norm": 0.9293227791786194, + "learning_rate": 6.595e-05, + "loss": 0.7058, + "step": 1320 + }, + { + "epoch": 0.5020762551906379, + "grad_norm": 1.033216118812561, + "learning_rate": 6.645e-05, + "loss": 0.7106, + "step": 1330 + }, + { + "epoch": 0.5058512646281615, + "grad_norm": 0.8532618880271912, + "learning_rate": 6.695e-05, + "loss": 0.7153, + "step": 1340 + }, + { + "epoch": 0.5096262740656852, + "grad_norm": 0.7834755182266235, + "learning_rate": 6.745e-05, + "loss": 0.705, + "step": 1350 + }, + { + "epoch": 0.5134012835032088, + "grad_norm": 0.7160763740539551, + "learning_rate": 6.795e-05, + "loss": 0.6986, + "step": 1360 + }, + { + "epoch": 0.5171762929407323, + "grad_norm": 0.9718685150146484, + "learning_rate": 6.845e-05, + "loss": 0.7002, + "step": 1370 + }, + { + "epoch": 0.5209513023782559, + "grad_norm": 0.9730781316757202, + "learning_rate": 6.895000000000001e-05, + "loss": 0.7107, + "step": 1380 + }, + { + "epoch": 0.5247263118157796, + "grad_norm": 1.1289321184158325, + "learning_rate": 6.945000000000001e-05, + "loss": 0.6985, + "step": 1390 + }, + { + "epoch": 0.5285013212533032, + "grad_norm": 0.6597205400466919, + "learning_rate": 6.995e-05, + "loss": 0.7051, + "step": 1400 + }, + { + "epoch": 0.5322763306908267, + "grad_norm": 1.0954090356826782, + "learning_rate": 7.045e-05, + "loss": 0.6965, + "step": 1410 + }, + { + "epoch": 0.5360513401283503, + "grad_norm": 0.9379609823226929, + "learning_rate": 7.095e-05, + "loss": 0.6937, + "step": 1420 + }, + { + "epoch": 0.5398263495658739, + "grad_norm": 1.048135757446289, + "learning_rate": 7.145e-05, + "loss": 0.6957, + "step": 1430 + }, + { + "epoch": 0.5436013590033975, + "grad_norm": 1.2906863689422607, + "learning_rate": 7.195e-05, + "loss": 0.6994, + "step": 1440 + }, + { + "epoch": 0.5473763684409211, + "grad_norm": 1.1085039377212524, + "learning_rate": 7.245000000000001e-05, + "loss": 0.6932, + "step": 1450 + }, + { + "epoch": 0.5511513778784447, + "grad_norm": 1.1647526025772095, + "learning_rate": 7.295000000000001e-05, + "loss": 0.7641, + "step": 1460 + }, + { + "epoch": 0.5549263873159683, + "grad_norm": 0.9016187191009521, + "learning_rate": 7.345000000000001e-05, + "loss": 0.7047, + "step": 1470 + }, + { + "epoch": 0.5587013967534918, + "grad_norm": 0.7149607539176941, + "learning_rate": 7.395000000000001e-05, + "loss": 0.6924, + "step": 1480 + }, + { + "epoch": 0.5624764061910155, + "grad_norm": 0.9485162496566772, + "learning_rate": 7.445000000000001e-05, + "loss": 0.6896, + "step": 1490 + }, + { + "epoch": 0.5662514156285391, + "grad_norm": 0.9566022157669067, + "learning_rate": 7.495e-05, + "loss": 0.6893, + "step": 1500 + }, + { + "epoch": 0.5700264250660627, + "grad_norm": 0.8616087436676025, + "learning_rate": 7.545e-05, + "loss": 0.7356, + "step": 1510 + }, + { + "epoch": 0.5738014345035862, + "grad_norm": 0.7664632201194763, + "learning_rate": 7.595e-05, + "loss": 0.6909, + "step": 1520 + }, + { + "epoch": 0.5775764439411099, + "grad_norm": 0.673774003982544, + "learning_rate": 7.645e-05, + "loss": 0.6804, + "step": 1530 + }, + { + "epoch": 0.5813514533786335, + "grad_norm": 1.0145224332809448, + "learning_rate": 7.695e-05, + "loss": 0.696, + "step": 1540 + }, + { + "epoch": 0.585126462816157, + "grad_norm": 0.805166482925415, + "learning_rate": 7.745e-05, + "loss": 0.7003, + "step": 1550 + }, + { + "epoch": 0.5889014722536806, + "grad_norm": 0.7893358469009399, + "learning_rate": 7.795e-05, + "loss": 0.6933, + "step": 1560 + }, + { + "epoch": 0.5926764816912042, + "grad_norm": 0.9435643553733826, + "learning_rate": 7.845e-05, + "loss": 0.6926, + "step": 1570 + }, + { + "epoch": 0.5964514911287279, + "grad_norm": 0.9065937399864197, + "learning_rate": 7.895000000000001e-05, + "loss": 0.685, + "step": 1580 + }, + { + "epoch": 0.6002265005662514, + "grad_norm": 1.084681749343872, + "learning_rate": 7.945e-05, + "loss": 0.7046, + "step": 1590 + }, + { + "epoch": 0.604001510003775, + "grad_norm": 0.6447507739067078, + "learning_rate": 7.995e-05, + "loss": 0.7189, + "step": 1600 + }, + { + "epoch": 0.6077765194412986, + "grad_norm": 0.855850100517273, + "learning_rate": 8.045e-05, + "loss": 0.6817, + "step": 1610 + }, + { + "epoch": 0.6115515288788222, + "grad_norm": 1.1896417140960693, + "learning_rate": 8.095e-05, + "loss": 0.6844, + "step": 1620 + }, + { + "epoch": 0.6153265383163458, + "grad_norm": 0.8419579267501831, + "learning_rate": 8.145e-05, + "loss": 0.6771, + "step": 1630 + }, + { + "epoch": 0.6191015477538694, + "grad_norm": 0.8358510732650757, + "learning_rate": 8.195e-05, + "loss": 0.6855, + "step": 1640 + }, + { + "epoch": 0.622876557191393, + "grad_norm": 0.896302342414856, + "learning_rate": 8.245e-05, + "loss": 0.6955, + "step": 1650 + }, + { + "epoch": 0.6266515666289165, + "grad_norm": 1.024187684059143, + "learning_rate": 8.295000000000001e-05, + "loss": 0.6811, + "step": 1660 + }, + { + "epoch": 0.6304265760664401, + "grad_norm": 0.8998435735702515, + "learning_rate": 8.345000000000001e-05, + "loss": 0.6807, + "step": 1670 + }, + { + "epoch": 0.6342015855039638, + "grad_norm": 0.6220159530639648, + "learning_rate": 8.395000000000001e-05, + "loss": 0.6795, + "step": 1680 + }, + { + "epoch": 0.6379765949414874, + "grad_norm": 0.9150254130363464, + "learning_rate": 8.445e-05, + "loss": 0.6695, + "step": 1690 + }, + { + "epoch": 0.6417516043790109, + "grad_norm": 0.7432071566581726, + "learning_rate": 8.495e-05, + "loss": 0.6792, + "step": 1700 + }, + { + "epoch": 0.6455266138165345, + "grad_norm": 0.808611273765564, + "learning_rate": 8.545e-05, + "loss": 0.6734, + "step": 1710 + }, + { + "epoch": 0.6493016232540582, + "grad_norm": 0.6958522200584412, + "learning_rate": 8.595e-05, + "loss": 0.6867, + "step": 1720 + }, + { + "epoch": 0.6530766326915818, + "grad_norm": 0.7854651808738708, + "learning_rate": 8.645000000000001e-05, + "loss": 0.6837, + "step": 1730 + }, + { + "epoch": 0.6568516421291053, + "grad_norm": 0.7910404205322266, + "learning_rate": 8.695000000000001e-05, + "loss": 0.6746, + "step": 1740 + }, + { + "epoch": 0.6606266515666289, + "grad_norm": 0.7932810187339783, + "learning_rate": 8.745000000000001e-05, + "loss": 0.6642, + "step": 1750 + }, + { + "epoch": 0.6644016610041525, + "grad_norm": 0.7843324542045593, + "learning_rate": 8.795e-05, + "loss": 0.6641, + "step": 1760 + }, + { + "epoch": 0.6681766704416761, + "grad_norm": 0.728706955909729, + "learning_rate": 8.845e-05, + "loss": 0.6745, + "step": 1770 + }, + { + "epoch": 0.6719516798791997, + "grad_norm": 0.8016153573989868, + "learning_rate": 8.895e-05, + "loss": 0.6671, + "step": 1780 + }, + { + "epoch": 0.6757266893167233, + "grad_norm": 0.695740282535553, + "learning_rate": 8.945e-05, + "loss": 0.6634, + "step": 1790 + }, + { + "epoch": 0.6795016987542469, + "grad_norm": 0.8751599788665771, + "learning_rate": 8.995e-05, + "loss": 0.6747, + "step": 1800 + }, + { + "epoch": 0.6832767081917704, + "grad_norm": 0.46772128343582153, + "learning_rate": 9.045e-05, + "loss": 0.6721, + "step": 1810 + }, + { + "epoch": 0.6870517176292941, + "grad_norm": 0.7569857239723206, + "learning_rate": 9.095e-05, + "loss": 0.6675, + "step": 1820 + }, + { + "epoch": 0.6908267270668177, + "grad_norm": 0.853040337562561, + "learning_rate": 9.145e-05, + "loss": 0.6685, + "step": 1830 + }, + { + "epoch": 0.6946017365043413, + "grad_norm": 0.6927493214607239, + "learning_rate": 9.195e-05, + "loss": 0.654, + "step": 1840 + }, + { + "epoch": 0.6983767459418648, + "grad_norm": 0.6520769000053406, + "learning_rate": 9.245e-05, + "loss": 0.6641, + "step": 1850 + }, + { + "epoch": 0.7021517553793885, + "grad_norm": 0.9358047246932983, + "learning_rate": 9.295000000000001e-05, + "loss": 0.6696, + "step": 1860 + }, + { + "epoch": 0.7059267648169121, + "grad_norm": 0.7453898787498474, + "learning_rate": 9.345000000000001e-05, + "loss": 0.6549, + "step": 1870 + }, + { + "epoch": 0.7097017742544356, + "grad_norm": 0.6966333389282227, + "learning_rate": 9.395000000000001e-05, + "loss": 0.6653, + "step": 1880 + }, + { + "epoch": 0.7134767836919592, + "grad_norm": 0.7843179702758789, + "learning_rate": 9.445e-05, + "loss": 0.6715, + "step": 1890 + }, + { + "epoch": 0.7172517931294828, + "grad_norm": 0.6947644948959351, + "learning_rate": 9.495e-05, + "loss": 0.6584, + "step": 1900 + }, + { + "epoch": 0.7210268025670065, + "grad_norm": 0.8173465132713318, + "learning_rate": 9.545e-05, + "loss": 0.6649, + "step": 1910 + }, + { + "epoch": 0.72480181200453, + "grad_norm": 0.9682592749595642, + "learning_rate": 9.595e-05, + "loss": 0.6549, + "step": 1920 + }, + { + "epoch": 0.7285768214420536, + "grad_norm": 1.0117123126983643, + "learning_rate": 9.645000000000001e-05, + "loss": 0.6536, + "step": 1930 + }, + { + "epoch": 0.7323518308795772, + "grad_norm": 0.9777953028678894, + "learning_rate": 9.695000000000001e-05, + "loss": 0.6578, + "step": 1940 + }, + { + "epoch": 0.7361268403171007, + "grad_norm": 0.7957656383514404, + "learning_rate": 9.745000000000001e-05, + "loss": 0.6566, + "step": 1950 + }, + { + "epoch": 0.7399018497546244, + "grad_norm": 0.7392054796218872, + "learning_rate": 9.795000000000001e-05, + "loss": 0.66, + "step": 1960 + }, + { + "epoch": 0.743676859192148, + "grad_norm": 0.811776340007782, + "learning_rate": 9.845000000000001e-05, + "loss": 0.6535, + "step": 1970 + }, + { + "epoch": 0.7474518686296716, + "grad_norm": 0.750037670135498, + "learning_rate": 9.895e-05, + "loss": 0.651, + "step": 1980 + }, + { + "epoch": 0.7512268780671951, + "grad_norm": 0.6936712861061096, + "learning_rate": 9.945e-05, + "loss": 0.6594, + "step": 1990 + }, + { + "epoch": 0.7550018875047187, + "grad_norm": 0.9359191060066223, + "learning_rate": 9.995e-05, + "loss": 0.6546, + "step": 2000 + }, + { + "epoch": 0.7587768969422424, + "grad_norm": 0.8320378065109253, + "learning_rate": 9.99999861593158e-05, + "loss": 0.6427, + "step": 2010 + }, + { + "epoch": 0.762551906379766, + "grad_norm": 0.6402168869972229, + "learning_rate": 9.999993831498517e-05, + "loss": 0.6711, + "step": 2020 + }, + { + "epoch": 0.7663269158172895, + "grad_norm": 0.6975911259651184, + "learning_rate": 9.99998562961682e-05, + "loss": 0.6551, + "step": 2030 + }, + { + "epoch": 0.7701019252548131, + "grad_norm": 0.7861037254333496, + "learning_rate": 9.99997401029209e-05, + "loss": 0.6548, + "step": 2040 + }, + { + "epoch": 0.7738769346923368, + "grad_norm": 0.7647865414619446, + "learning_rate": 9.999958973532271e-05, + "loss": 0.6553, + "step": 2050 + }, + { + "epoch": 0.7776519441298603, + "grad_norm": 1.0054905414581299, + "learning_rate": 9.999940519347642e-05, + "loss": 0.6523, + "step": 2060 + }, + { + "epoch": 0.7814269535673839, + "grad_norm": 0.9435162544250488, + "learning_rate": 9.999918647750812e-05, + "loss": 0.6573, + "step": 2070 + }, + { + "epoch": 0.7852019630049075, + "grad_norm": 0.7330409288406372, + "learning_rate": 9.999893358756736e-05, + "loss": 0.6601, + "step": 2080 + }, + { + "epoch": 0.7889769724424311, + "grad_norm": 1.0650360584259033, + "learning_rate": 9.999864652382692e-05, + "loss": 0.6481, + "step": 2090 + }, + { + "epoch": 0.7927519818799547, + "grad_norm": 0.6721388697624207, + "learning_rate": 9.999832528648307e-05, + "loss": 0.6395, + "step": 2100 + }, + { + "epoch": 0.7965269913174783, + "grad_norm": 0.8571961522102356, + "learning_rate": 9.999796987575532e-05, + "loss": 0.6459, + "step": 2110 + }, + { + "epoch": 0.8003020007550019, + "grad_norm": 0.575950026512146, + "learning_rate": 9.999758029188662e-05, + "loss": 0.6459, + "step": 2120 + }, + { + "epoch": 0.8040770101925255, + "grad_norm": 0.5099409818649292, + "learning_rate": 9.999715653514324e-05, + "loss": 0.6484, + "step": 2130 + }, + { + "epoch": 0.807852019630049, + "grad_norm": 0.6396839618682861, + "learning_rate": 9.999669860581482e-05, + "loss": 0.6504, + "step": 2140 + }, + { + "epoch": 0.8116270290675727, + "grad_norm": 0.6424649357795715, + "learning_rate": 9.999620650421433e-05, + "loss": 0.6396, + "step": 2150 + }, + { + "epoch": 0.8154020385050963, + "grad_norm": 0.7561330199241638, + "learning_rate": 9.999568023067813e-05, + "loss": 0.6481, + "step": 2160 + }, + { + "epoch": 0.8191770479426198, + "grad_norm": 0.8530335426330566, + "learning_rate": 9.999511978556595e-05, + "loss": 0.6392, + "step": 2170 + }, + { + "epoch": 0.8229520573801434, + "grad_norm": 0.8621054887771606, + "learning_rate": 9.99945251692608e-05, + "loss": 0.652, + "step": 2180 + }, + { + "epoch": 0.8267270668176671, + "grad_norm": 0.7959499359130859, + "learning_rate": 9.999389638216912e-05, + "loss": 0.6384, + "step": 2190 + }, + { + "epoch": 0.8305020762551907, + "grad_norm": 0.6927378177642822, + "learning_rate": 9.999323342472068e-05, + "loss": 0.6432, + "step": 2200 + }, + { + "epoch": 0.8342770856927142, + "grad_norm": 0.5878588557243347, + "learning_rate": 9.99925362973686e-05, + "loss": 0.6464, + "step": 2210 + }, + { + "epoch": 0.8380520951302378, + "grad_norm": 0.6220201253890991, + "learning_rate": 9.999180500058935e-05, + "loss": 0.6978, + "step": 2220 + }, + { + "epoch": 0.8418271045677614, + "grad_norm": 0.6637720465660095, + "learning_rate": 9.99910395348828e-05, + "loss": 0.6336, + "step": 2230 + }, + { + "epoch": 0.8456021140052851, + "grad_norm": 0.7113637924194336, + "learning_rate": 9.99902399007721e-05, + "loss": 0.633, + "step": 2240 + }, + { + "epoch": 0.8493771234428086, + "grad_norm": 0.7006509900093079, + "learning_rate": 9.99894060988038e-05, + "loss": 0.634, + "step": 2250 + }, + { + "epoch": 0.8531521328803322, + "grad_norm": 0.7758764624595642, + "learning_rate": 9.998853812954783e-05, + "loss": 0.6327, + "step": 2260 + }, + { + "epoch": 0.8569271423178558, + "grad_norm": 0.7731608748435974, + "learning_rate": 9.998763599359739e-05, + "loss": 0.6301, + "step": 2270 + }, + { + "epoch": 0.8607021517553793, + "grad_norm": 0.9600358009338379, + "learning_rate": 9.998669969156912e-05, + "loss": 0.6473, + "step": 2280 + }, + { + "epoch": 0.864477161192903, + "grad_norm": 0.6508148312568665, + "learning_rate": 9.998572922410294e-05, + "loss": 0.6401, + "step": 2290 + }, + { + "epoch": 0.8682521706304266, + "grad_norm": 0.6029082536697388, + "learning_rate": 9.998472459186219e-05, + "loss": 0.6355, + "step": 2300 + }, + { + "epoch": 0.8720271800679502, + "grad_norm": 0.5723931193351746, + "learning_rate": 9.99836857955335e-05, + "loss": 0.6352, + "step": 2310 + }, + { + "epoch": 0.8758021895054737, + "grad_norm": 0.5845490097999573, + "learning_rate": 9.998261283582688e-05, + "loss": 0.6348, + "step": 2320 + }, + { + "epoch": 0.8795771989429974, + "grad_norm": 0.702069103717804, + "learning_rate": 9.998150571347571e-05, + "loss": 0.7144, + "step": 2330 + }, + { + "epoch": 0.883352208380521, + "grad_norm": 0.7790460586547852, + "learning_rate": 9.998036442923667e-05, + "loss": 0.6357, + "step": 2340 + }, + { + "epoch": 0.8871272178180446, + "grad_norm": 0.6690578460693359, + "learning_rate": 9.997918898388983e-05, + "loss": 0.6327, + "step": 2350 + }, + { + "epoch": 0.8909022272555681, + "grad_norm": 0.7188066244125366, + "learning_rate": 9.997797937823859e-05, + "loss": 0.626, + "step": 2360 + }, + { + "epoch": 0.8946772366930917, + "grad_norm": 0.6960968971252441, + "learning_rate": 9.997673561310971e-05, + "loss": 0.6359, + "step": 2370 + }, + { + "epoch": 0.8984522461306154, + "grad_norm": 0.5999962091445923, + "learning_rate": 9.997545768935331e-05, + "loss": 0.642, + "step": 2380 + }, + { + "epoch": 0.9022272555681389, + "grad_norm": 0.6590844988822937, + "learning_rate": 9.997414560784282e-05, + "loss": 0.6334, + "step": 2390 + }, + { + "epoch": 0.9060022650056625, + "grad_norm": 0.5802249908447266, + "learning_rate": 9.997279936947502e-05, + "loss": 0.6319, + "step": 2400 + }, + { + "epoch": 0.9097772744431861, + "grad_norm": 0.5891656279563904, + "learning_rate": 9.997141897517006e-05, + "loss": 0.6261, + "step": 2410 + }, + { + "epoch": 0.9135522838807097, + "grad_norm": 0.6824517846107483, + "learning_rate": 9.997000442587145e-05, + "loss": 0.6261, + "step": 2420 + }, + { + "epoch": 0.9173272933182333, + "grad_norm": 0.7084891200065613, + "learning_rate": 9.9968555722546e-05, + "loss": 0.6266, + "step": 2430 + }, + { + "epoch": 0.9211023027557569, + "grad_norm": 0.5537872314453125, + "learning_rate": 9.996707286618387e-05, + "loss": 0.6271, + "step": 2440 + }, + { + "epoch": 0.9248773121932805, + "grad_norm": 0.5806447863578796, + "learning_rate": 9.996555585779861e-05, + "loss": 0.6356, + "step": 2450 + }, + { + "epoch": 0.928652321630804, + "grad_norm": 0.6291959285736084, + "learning_rate": 9.996400469842708e-05, + "loss": 0.6344, + "step": 2460 + }, + { + "epoch": 0.9324273310683276, + "grad_norm": 0.5299859046936035, + "learning_rate": 9.996241938912945e-05, + "loss": 0.6256, + "step": 2470 + }, + { + "epoch": 0.9362023405058513, + "grad_norm": 0.5358120203018188, + "learning_rate": 9.99607999309893e-05, + "loss": 0.6164, + "step": 2480 + }, + { + "epoch": 0.9399773499433749, + "grad_norm": 0.4257400929927826, + "learning_rate": 9.99591463251135e-05, + "loss": 0.6238, + "step": 2490 + }, + { + "epoch": 0.9437523593808984, + "grad_norm": 0.5488512516021729, + "learning_rate": 9.995745857263226e-05, + "loss": 0.632, + "step": 2500 + }, + { + "epoch": 0.947527368818422, + "grad_norm": 0.45664718747138977, + "learning_rate": 9.995573667469918e-05, + "loss": 0.6276, + "step": 2510 + }, + { + "epoch": 0.9513023782559457, + "grad_norm": 0.7003333568572998, + "learning_rate": 9.995398063249111e-05, + "loss": 0.6248, + "step": 2520 + }, + { + "epoch": 0.9550773876934693, + "grad_norm": 0.5362398028373718, + "learning_rate": 9.995219044720833e-05, + "loss": 0.6277, + "step": 2530 + }, + { + "epoch": 0.9588523971309928, + "grad_norm": 0.8197181224822998, + "learning_rate": 9.99503661200744e-05, + "loss": 0.6219, + "step": 2540 + }, + { + "epoch": 0.9626274065685164, + "grad_norm": 0.671510636806488, + "learning_rate": 9.994850765233624e-05, + "loss": 0.6262, + "step": 2550 + }, + { + "epoch": 0.96640241600604, + "grad_norm": 0.6340009570121765, + "learning_rate": 9.994661504526406e-05, + "loss": 0.6167, + "step": 2560 + }, + { + "epoch": 0.9701774254435636, + "grad_norm": 0.5801326632499695, + "learning_rate": 9.994468830015148e-05, + "loss": 0.6319, + "step": 2570 + }, + { + "epoch": 0.9739524348810872, + "grad_norm": 0.6170343160629272, + "learning_rate": 9.99427274183154e-05, + "loss": 0.6233, + "step": 2580 + }, + { + "epoch": 0.9777274443186108, + "grad_norm": 0.5810487866401672, + "learning_rate": 9.994073240109606e-05, + "loss": 0.6281, + "step": 2590 + }, + { + "epoch": 0.9815024537561344, + "grad_norm": 0.46137765049934387, + "learning_rate": 9.993870324985703e-05, + "loss": 0.6144, + "step": 2600 + }, + { + "epoch": 0.9852774631936579, + "grad_norm": 0.6433082222938538, + "learning_rate": 9.993663996598523e-05, + "loss": 0.6129, + "step": 2610 + }, + { + "epoch": 0.9890524726311816, + "grad_norm": 0.6449345350265503, + "learning_rate": 9.993454255089089e-05, + "loss": 0.6149, + "step": 2620 + }, + { + "epoch": 0.9928274820687052, + "grad_norm": 0.6302064061164856, + "learning_rate": 9.993241100600757e-05, + "loss": 0.6192, + "step": 2630 + }, + { + "epoch": 0.9966024915062288, + "grad_norm": 0.7228666543960571, + "learning_rate": 9.993024533279215e-05, + "loss": 0.6179, + "step": 2640 + }, + { + "epoch": 1.0003775009437523, + "grad_norm": 0.5953015685081482, + "learning_rate": 9.992804553272487e-05, + "loss": 0.6176, + "step": 2650 + }, + { + "epoch": 1.0041525103812758, + "grad_norm": 0.586175799369812, + "learning_rate": 9.992581160730926e-05, + "loss": 0.6095, + "step": 2660 + }, + { + "epoch": 1.0079275198187996, + "grad_norm": 0.6601322293281555, + "learning_rate": 9.992354355807218e-05, + "loss": 0.608, + "step": 2670 + }, + { + "epoch": 1.011702529256323, + "grad_norm": 0.7241320013999939, + "learning_rate": 9.992124138656385e-05, + "loss": 0.6204, + "step": 2680 + }, + { + "epoch": 1.0154775386938468, + "grad_norm": 0.7130767107009888, + "learning_rate": 9.991890509435774e-05, + "loss": 0.6191, + "step": 2690 + }, + { + "epoch": 1.0192525481313703, + "grad_norm": 0.5906509757041931, + "learning_rate": 9.991653468305071e-05, + "loss": 0.6216, + "step": 2700 + }, + { + "epoch": 1.0230275575688939, + "grad_norm": 0.4647708535194397, + "learning_rate": 9.991413015426292e-05, + "loss": 0.6226, + "step": 2710 + }, + { + "epoch": 1.0268025670064176, + "grad_norm": 0.5289870500564575, + "learning_rate": 9.991169150963781e-05, + "loss": 0.6204, + "step": 2720 + }, + { + "epoch": 1.030577576443941, + "grad_norm": 0.5250043272972107, + "learning_rate": 9.990921875084221e-05, + "loss": 0.6101, + "step": 2730 + }, + { + "epoch": 1.0343525858814646, + "grad_norm": 0.553863525390625, + "learning_rate": 9.990671187956622e-05, + "loss": 0.6236, + "step": 2740 + }, + { + "epoch": 1.0381275953189883, + "grad_norm": 0.6878931522369385, + "learning_rate": 9.990417089752324e-05, + "loss": 0.6211, + "step": 2750 + }, + { + "epoch": 1.0419026047565119, + "grad_norm": 0.5037298798561096, + "learning_rate": 9.990159580645001e-05, + "loss": 0.6128, + "step": 2760 + }, + { + "epoch": 1.0456776141940356, + "grad_norm": 0.5697045922279358, + "learning_rate": 9.98989866081066e-05, + "loss": 0.6088, + "step": 2770 + }, + { + "epoch": 1.0494526236315591, + "grad_norm": 0.6046783328056335, + "learning_rate": 9.989634330427636e-05, + "loss": 0.624, + "step": 2780 + }, + { + "epoch": 1.0532276330690826, + "grad_norm": 0.6193393468856812, + "learning_rate": 9.989366589676596e-05, + "loss": 0.6133, + "step": 2790 + }, + { + "epoch": 1.0570026425066064, + "grad_norm": 0.6406804323196411, + "learning_rate": 9.98909543874054e-05, + "loss": 0.6164, + "step": 2800 + }, + { + "epoch": 1.0607776519441299, + "grad_norm": 0.6087714433670044, + "learning_rate": 9.988820877804795e-05, + "loss": 0.6113, + "step": 2810 + }, + { + "epoch": 1.0645526613816534, + "grad_norm": 0.4003846347332001, + "learning_rate": 9.988542907057022e-05, + "loss": 0.6065, + "step": 2820 + }, + { + "epoch": 1.0683276708191771, + "grad_norm": 0.5162433385848999, + "learning_rate": 9.988261526687212e-05, + "loss": 0.6105, + "step": 2830 + }, + { + "epoch": 1.0721026802567006, + "grad_norm": 0.5106794238090515, + "learning_rate": 9.987976736887685e-05, + "loss": 0.6105, + "step": 2840 + }, + { + "epoch": 1.0758776896942241, + "grad_norm": 0.5757883787155151, + "learning_rate": 9.987688537853091e-05, + "loss": 0.6105, + "step": 2850 + }, + { + "epoch": 1.0796526991317479, + "grad_norm": 0.5995069742202759, + "learning_rate": 9.987396929780413e-05, + "loss": 0.6119, + "step": 2860 + }, + { + "epoch": 1.0834277085692714, + "grad_norm": 0.5145584940910339, + "learning_rate": 9.987101912868962e-05, + "loss": 0.6108, + "step": 2870 + }, + { + "epoch": 1.087202718006795, + "grad_norm": 0.418322890996933, + "learning_rate": 9.98680348732038e-05, + "loss": 0.6076, + "step": 2880 + }, + { + "epoch": 1.0909777274443186, + "grad_norm": 0.526392936706543, + "learning_rate": 9.986501653338636e-05, + "loss": 0.61, + "step": 2890 + }, + { + "epoch": 1.0947527368818422, + "grad_norm": 0.6863415837287903, + "learning_rate": 9.986196411130031e-05, + "loss": 0.6483, + "step": 2900 + }, + { + "epoch": 1.098527746319366, + "grad_norm": 0.8329252600669861, + "learning_rate": 9.985887760903197e-05, + "loss": 0.6162, + "step": 2910 + }, + { + "epoch": 1.1023027557568894, + "grad_norm": 0.6749866008758545, + "learning_rate": 9.985575702869093e-05, + "loss": 0.6157, + "step": 2920 + }, + { + "epoch": 1.106077765194413, + "grad_norm": 0.6485674381256104, + "learning_rate": 9.985260237241008e-05, + "loss": 0.6161, + "step": 2930 + }, + { + "epoch": 1.1098527746319367, + "grad_norm": 0.5751580595970154, + "learning_rate": 9.984941364234557e-05, + "loss": 0.6196, + "step": 2940 + }, + { + "epoch": 1.1136277840694602, + "grad_norm": 0.539655864238739, + "learning_rate": 9.98461908406769e-05, + "loss": 0.6143, + "step": 2950 + }, + { + "epoch": 1.1174027935069837, + "grad_norm": 0.5958253145217896, + "learning_rate": 9.98429339696068e-05, + "loss": 0.6111, + "step": 2960 + }, + { + "epoch": 1.1211778029445074, + "grad_norm": 0.6310438513755798, + "learning_rate": 9.983964303136133e-05, + "loss": 0.6088, + "step": 2970 + }, + { + "epoch": 1.124952812382031, + "grad_norm": 0.5665124654769897, + "learning_rate": 9.983631802818981e-05, + "loss": 0.6046, + "step": 2980 + }, + { + "epoch": 1.1287278218195547, + "grad_norm": 0.4563884437084198, + "learning_rate": 9.983295896236484e-05, + "loss": 0.5985, + "step": 2990 + }, + { + "epoch": 1.1325028312570782, + "grad_norm": 0.5045254230499268, + "learning_rate": 9.982956583618232e-05, + "loss": 0.6024, + "step": 3000 + }, + { + "epoch": 1.1362778406946017, + "grad_norm": 0.5779048800468445, + "learning_rate": 9.982613865196142e-05, + "loss": 0.6088, + "step": 3010 + }, + { + "epoch": 1.1400528501321254, + "grad_norm": 0.5393741726875305, + "learning_rate": 9.982267741204458e-05, + "loss": 0.6145, + "step": 3020 + }, + { + "epoch": 1.143827859569649, + "grad_norm": 0.5488253831863403, + "learning_rate": 9.981918211879753e-05, + "loss": 0.6019, + "step": 3030 + }, + { + "epoch": 1.1476028690071725, + "grad_norm": 0.5438612103462219, + "learning_rate": 9.981565277460927e-05, + "loss": 0.6165, + "step": 3040 + }, + { + "epoch": 1.1513778784446962, + "grad_norm": 0.5116326808929443, + "learning_rate": 9.981208938189206e-05, + "loss": 0.6062, + "step": 3050 + }, + { + "epoch": 1.1551528878822197, + "grad_norm": 0.48625659942626953, + "learning_rate": 9.980849194308149e-05, + "loss": 0.6113, + "step": 3060 + }, + { + "epoch": 1.1589278973197432, + "grad_norm": 0.5895056128501892, + "learning_rate": 9.98048604606363e-05, + "loss": 0.6002, + "step": 3070 + }, + { + "epoch": 1.162702906757267, + "grad_norm": 0.7021406292915344, + "learning_rate": 9.980119493703864e-05, + "loss": 0.6139, + "step": 3080 + }, + { + "epoch": 1.1664779161947905, + "grad_norm": 0.7085283398628235, + "learning_rate": 9.979749537479383e-05, + "loss": 0.6058, + "step": 3090 + }, + { + "epoch": 1.170252925632314, + "grad_norm": 0.6100856065750122, + "learning_rate": 9.979376177643051e-05, + "loss": 0.6003, + "step": 3100 + }, + { + "epoch": 1.1740279350698377, + "grad_norm": 0.6801294684410095, + "learning_rate": 9.978999414450052e-05, + "loss": 0.6171, + "step": 3110 + }, + { + "epoch": 1.1778029445073612, + "grad_norm": 0.47147953510284424, + "learning_rate": 9.978619248157904e-05, + "loss": 0.6057, + "step": 3120 + }, + { + "epoch": 1.1815779539448847, + "grad_norm": 0.5377033352851868, + "learning_rate": 9.978235679026445e-05, + "loss": 0.603, + "step": 3130 + }, + { + "epoch": 1.1853529633824085, + "grad_norm": 0.6004739999771118, + "learning_rate": 9.97784870731784e-05, + "loss": 0.5997, + "step": 3140 + }, + { + "epoch": 1.189127972819932, + "grad_norm": 0.42120230197906494, + "learning_rate": 9.977458333296584e-05, + "loss": 0.6065, + "step": 3150 + }, + { + "epoch": 1.1929029822574557, + "grad_norm": 0.4445311427116394, + "learning_rate": 9.977064557229492e-05, + "loss": 0.6049, + "step": 3160 + }, + { + "epoch": 1.1966779916949792, + "grad_norm": 0.5501314997673035, + "learning_rate": 9.976667379385705e-05, + "loss": 0.6177, + "step": 3170 + }, + { + "epoch": 1.2004530011325028, + "grad_norm": 0.6366961598396301, + "learning_rate": 9.97626680003669e-05, + "loss": 0.6134, + "step": 3180 + }, + { + "epoch": 1.2042280105700265, + "grad_norm": 0.5312647223472595, + "learning_rate": 9.975862819456242e-05, + "loss": 0.5979, + "step": 3190 + }, + { + "epoch": 1.20800302000755, + "grad_norm": 0.5397608876228333, + "learning_rate": 9.975455437920477e-05, + "loss": 0.5947, + "step": 3200 + }, + { + "epoch": 1.2117780294450737, + "grad_norm": 0.4828698933124542, + "learning_rate": 9.975044655707834e-05, + "loss": 0.6012, + "step": 3210 + }, + { + "epoch": 1.2155530388825972, + "grad_norm": 0.5150812864303589, + "learning_rate": 9.974630473099082e-05, + "loss": 0.604, + "step": 3220 + }, + { + "epoch": 1.2193280483201208, + "grad_norm": 0.6184629201889038, + "learning_rate": 9.974212890377311e-05, + "loss": 0.5917, + "step": 3230 + }, + { + "epoch": 1.2231030577576445, + "grad_norm": 0.48113495111465454, + "learning_rate": 9.973791907827931e-05, + "loss": 0.5953, + "step": 3240 + }, + { + "epoch": 1.226878067195168, + "grad_norm": 0.4420939087867737, + "learning_rate": 9.973367525738683e-05, + "loss": 0.6002, + "step": 3250 + }, + { + "epoch": 1.2306530766326915, + "grad_norm": 0.610998272895813, + "learning_rate": 9.972939744399627e-05, + "loss": 0.5944, + "step": 3260 + }, + { + "epoch": 1.2344280860702153, + "grad_norm": 0.5040599703788757, + "learning_rate": 9.97250856410315e-05, + "loss": 0.5988, + "step": 3270 + }, + { + "epoch": 1.2382030955077388, + "grad_norm": 0.5534839630126953, + "learning_rate": 9.972073985143955e-05, + "loss": 0.5982, + "step": 3280 + }, + { + "epoch": 1.2419781049452623, + "grad_norm": 0.5562023520469666, + "learning_rate": 9.971636007819074e-05, + "loss": 0.6003, + "step": 3290 + }, + { + "epoch": 1.245753114382786, + "grad_norm": 0.6595808863639832, + "learning_rate": 9.971194632427863e-05, + "loss": 0.602, + "step": 3300 + }, + { + "epoch": 1.2495281238203095, + "grad_norm": 0.7071061134338379, + "learning_rate": 9.970749859271997e-05, + "loss": 0.6124, + "step": 3310 + }, + { + "epoch": 1.253303133257833, + "grad_norm": 0.7782847285270691, + "learning_rate": 9.970301688655473e-05, + "loss": 0.5991, + "step": 3320 + }, + { + "epoch": 1.2570781426953568, + "grad_norm": 0.7165196537971497, + "learning_rate": 9.969850120884612e-05, + "loss": 0.5957, + "step": 3330 + }, + { + "epoch": 1.2608531521328803, + "grad_norm": 0.6089136004447937, + "learning_rate": 9.969395156268055e-05, + "loss": 0.5891, + "step": 3340 + }, + { + "epoch": 1.2646281615704038, + "grad_norm": 0.586713969707489, + "learning_rate": 9.968936795116768e-05, + "loss": 0.594, + "step": 3350 + }, + { + "epoch": 1.2684031710079275, + "grad_norm": 0.6137309074401855, + "learning_rate": 9.968475037744036e-05, + "loss": 0.6046, + "step": 3360 + }, + { + "epoch": 1.272178180445451, + "grad_norm": 0.5665665864944458, + "learning_rate": 9.968009884465465e-05, + "loss": 0.5905, + "step": 3370 + }, + { + "epoch": 1.2759531898829746, + "grad_norm": 0.5130990743637085, + "learning_rate": 9.967541335598984e-05, + "loss": 0.5971, + "step": 3380 + }, + { + "epoch": 1.2797281993204983, + "grad_norm": 0.6189398765563965, + "learning_rate": 9.967069391464841e-05, + "loss": 0.5988, + "step": 3390 + }, + { + "epoch": 1.2835032087580218, + "grad_norm": 0.5427692532539368, + "learning_rate": 9.966594052385608e-05, + "loss": 0.5979, + "step": 3400 + }, + { + "epoch": 1.2872782181955456, + "grad_norm": 0.5762437582015991, + "learning_rate": 9.96611531868617e-05, + "loss": 0.5932, + "step": 3410 + }, + { + "epoch": 1.291053227633069, + "grad_norm": 0.5316555500030518, + "learning_rate": 9.96563319069374e-05, + "loss": 0.5962, + "step": 3420 + }, + { + "epoch": 1.2948282370705928, + "grad_norm": 0.52821284532547, + "learning_rate": 9.965147668737847e-05, + "loss": 0.5899, + "step": 3430 + }, + { + "epoch": 1.2986032465081163, + "grad_norm": 0.6211134195327759, + "learning_rate": 9.96465875315034e-05, + "loss": 0.595, + "step": 3440 + }, + { + "epoch": 1.3023782559456398, + "grad_norm": 0.4804578125476837, + "learning_rate": 9.96416644426539e-05, + "loss": 0.5943, + "step": 3450 + }, + { + "epoch": 1.3061532653831636, + "grad_norm": 0.520340621471405, + "learning_rate": 9.963670742419485e-05, + "loss": 0.585, + "step": 3460 + }, + { + "epoch": 1.309928274820687, + "grad_norm": 0.5734260678291321, + "learning_rate": 9.96317164795143e-05, + "loss": 0.5853, + "step": 3470 + }, + { + "epoch": 1.3137032842582106, + "grad_norm": 0.6706385612487793, + "learning_rate": 9.962669161202356e-05, + "loss": 0.6087, + "step": 3480 + }, + { + "epoch": 1.3174782936957343, + "grad_norm": 0.5032151341438293, + "learning_rate": 9.962163282515705e-05, + "loss": 0.5877, + "step": 3490 + }, + { + "epoch": 1.3212533031332578, + "grad_norm": 0.5707389116287231, + "learning_rate": 9.961654012237241e-05, + "loss": 0.5801, + "step": 3500 + }, + { + "epoch": 1.3250283125707814, + "grad_norm": 0.5376269817352295, + "learning_rate": 9.961141350715044e-05, + "loss": 0.5817, + "step": 3510 + }, + { + "epoch": 1.328803322008305, + "grad_norm": 0.4683854579925537, + "learning_rate": 9.960625298299516e-05, + "loss": 0.5847, + "step": 3520 + }, + { + "epoch": 1.3325783314458286, + "grad_norm": 0.4962872266769409, + "learning_rate": 9.960105855343372e-05, + "loss": 0.5854, + "step": 3530 + }, + { + "epoch": 1.3363533408833521, + "grad_norm": 0.4632238745689392, + "learning_rate": 9.959583022201647e-05, + "loss": 0.5844, + "step": 3540 + }, + { + "epoch": 1.3401283503208758, + "grad_norm": 0.49450498819351196, + "learning_rate": 9.959056799231692e-05, + "loss": 0.5877, + "step": 3550 + }, + { + "epoch": 1.3439033597583994, + "grad_norm": 0.45359042286872864, + "learning_rate": 9.958527186793176e-05, + "loss": 0.5798, + "step": 3560 + }, + { + "epoch": 1.3476783691959229, + "grad_norm": 0.48816585540771484, + "learning_rate": 9.957994185248086e-05, + "loss": 0.587, + "step": 3570 + }, + { + "epoch": 1.3514533786334466, + "grad_norm": 0.5069116353988647, + "learning_rate": 9.957457794960718e-05, + "loss": 0.5972, + "step": 3580 + }, + { + "epoch": 1.3552283880709701, + "grad_norm": 0.4354189336299896, + "learning_rate": 9.956918016297694e-05, + "loss": 0.5872, + "step": 3590 + }, + { + "epoch": 1.3590033975084936, + "grad_norm": 0.5266293883323669, + "learning_rate": 9.956374849627948e-05, + "loss": 0.5869, + "step": 3600 + }, + { + "epoch": 1.3627784069460174, + "grad_norm": 0.4389692544937134, + "learning_rate": 9.955828295322728e-05, + "loss": 0.5792, + "step": 3610 + }, + { + "epoch": 1.3665534163835409, + "grad_norm": 0.5600664615631104, + "learning_rate": 9.955278353755598e-05, + "loss": 0.582, + "step": 3620 + }, + { + "epoch": 1.3703284258210646, + "grad_norm": 0.6405794024467468, + "learning_rate": 9.954725025302439e-05, + "loss": 0.5898, + "step": 3630 + }, + { + "epoch": 1.3741034352585881, + "grad_norm": 0.5375349521636963, + "learning_rate": 9.954168310341445e-05, + "loss": 0.5863, + "step": 3640 + }, + { + "epoch": 1.3778784446961119, + "grad_norm": 0.6790446043014526, + "learning_rate": 9.953608209253126e-05, + "loss": 0.5904, + "step": 3650 + }, + { + "epoch": 1.3816534541336354, + "grad_norm": 0.4370107352733612, + "learning_rate": 9.953044722420307e-05, + "loss": 0.5874, + "step": 3660 + }, + { + "epoch": 1.385428463571159, + "grad_norm": 0.524257242679596, + "learning_rate": 9.952477850228124e-05, + "loss": 0.5879, + "step": 3670 + }, + { + "epoch": 1.3892034730086826, + "grad_norm": 0.4980759918689728, + "learning_rate": 9.95190759306403e-05, + "loss": 0.5817, + "step": 3680 + }, + { + "epoch": 1.3929784824462061, + "grad_norm": 0.5062684416770935, + "learning_rate": 9.951333951317789e-05, + "loss": 0.5827, + "step": 3690 + }, + { + "epoch": 1.3967534918837297, + "grad_norm": 0.7371021509170532, + "learning_rate": 9.950756925381479e-05, + "loss": 0.5829, + "step": 3700 + }, + { + "epoch": 1.4005285013212534, + "grad_norm": 0.5569843649864197, + "learning_rate": 9.950176515649496e-05, + "loss": 0.5779, + "step": 3710 + }, + { + "epoch": 1.404303510758777, + "grad_norm": 0.5766958594322205, + "learning_rate": 9.949592722518542e-05, + "loss": 0.5861, + "step": 3720 + }, + { + "epoch": 1.4080785201963004, + "grad_norm": 0.5430606603622437, + "learning_rate": 9.949005546387631e-05, + "loss": 0.579, + "step": 3730 + }, + { + "epoch": 1.4118535296338242, + "grad_norm": 0.49899691343307495, + "learning_rate": 9.948414987658098e-05, + "loss": 0.5759, + "step": 3740 + }, + { + "epoch": 1.4156285390713477, + "grad_norm": 0.54237300157547, + "learning_rate": 9.94782104673358e-05, + "loss": 0.5809, + "step": 3750 + }, + { + "epoch": 1.4194035485088712, + "grad_norm": 0.6004769802093506, + "learning_rate": 9.947223724020034e-05, + "loss": 0.5743, + "step": 3760 + }, + { + "epoch": 1.423178557946395, + "grad_norm": 0.5146884322166443, + "learning_rate": 9.94662301992572e-05, + "loss": 0.5782, + "step": 3770 + }, + { + "epoch": 1.4269535673839184, + "grad_norm": 0.5596660375595093, + "learning_rate": 9.946018934861216e-05, + "loss": 0.5787, + "step": 3780 + }, + { + "epoch": 1.430728576821442, + "grad_norm": 0.5646491050720215, + "learning_rate": 9.945411469239409e-05, + "loss": 0.5871, + "step": 3790 + }, + { + "epoch": 1.4345035862589657, + "grad_norm": 0.6031035780906677, + "learning_rate": 9.944800623475497e-05, + "loss": 0.5735, + "step": 3800 + }, + { + "epoch": 1.4382785956964892, + "grad_norm": 0.6148284077644348, + "learning_rate": 9.944186397986984e-05, + "loss": 0.5807, + "step": 3810 + }, + { + "epoch": 1.4420536051340127, + "grad_norm": 0.541500449180603, + "learning_rate": 9.943568793193687e-05, + "loss": 0.5806, + "step": 3820 + }, + { + "epoch": 1.4458286145715364, + "grad_norm": 0.5454701781272888, + "learning_rate": 9.942947809517737e-05, + "loss": 0.584, + "step": 3830 + }, + { + "epoch": 1.44960362400906, + "grad_norm": 0.6133937835693359, + "learning_rate": 9.942323447383568e-05, + "loss": 0.5811, + "step": 3840 + }, + { + "epoch": 1.4533786334465837, + "grad_norm": 0.5456222295761108, + "learning_rate": 9.941695707217925e-05, + "loss": 0.5747, + "step": 3850 + }, + { + "epoch": 1.4571536428841072, + "grad_norm": 0.5243522524833679, + "learning_rate": 9.941064589449865e-05, + "loss": 0.5792, + "step": 3860 + }, + { + "epoch": 1.460928652321631, + "grad_norm": 0.5548741221427917, + "learning_rate": 9.940430094510748e-05, + "loss": 0.5756, + "step": 3870 + }, + { + "epoch": 1.4647036617591545, + "grad_norm": 0.65179044008255, + "learning_rate": 9.939792222834248e-05, + "loss": 0.5788, + "step": 3880 + }, + { + "epoch": 1.468478671196678, + "grad_norm": 0.6072384715080261, + "learning_rate": 9.939150974856343e-05, + "loss": 0.5779, + "step": 3890 + }, + { + "epoch": 1.4722536806342017, + "grad_norm": 0.6061842441558838, + "learning_rate": 9.93850635101532e-05, + "loss": 0.5735, + "step": 3900 + }, + { + "epoch": 1.4760286900717252, + "grad_norm": 0.5786122679710388, + "learning_rate": 9.937858351751775e-05, + "loss": 0.5794, + "step": 3910 + }, + { + "epoch": 1.4798036995092487, + "grad_norm": 0.5088226199150085, + "learning_rate": 9.937206977508604e-05, + "loss": 0.5737, + "step": 3920 + }, + { + "epoch": 1.4835787089467725, + "grad_norm": 0.5596620440483093, + "learning_rate": 9.936552228731022e-05, + "loss": 0.5912, + "step": 3930 + }, + { + "epoch": 1.487353718384296, + "grad_norm": 0.4660915732383728, + "learning_rate": 9.93589410586654e-05, + "loss": 0.5735, + "step": 3940 + }, + { + "epoch": 1.4911287278218195, + "grad_norm": 0.4994395971298218, + "learning_rate": 9.93523260936498e-05, + "loss": 0.5807, + "step": 3950 + }, + { + "epoch": 1.4949037372593432, + "grad_norm": 0.4547739624977112, + "learning_rate": 9.934567739678467e-05, + "loss": 0.5723, + "step": 3960 + }, + { + "epoch": 1.4986787466968667, + "grad_norm": 0.4519752264022827, + "learning_rate": 9.933899497261433e-05, + "loss": 0.5749, + "step": 3970 + }, + { + "epoch": 1.5024537561343903, + "grad_norm": 0.49861350655555725, + "learning_rate": 9.933227882570617e-05, + "loss": 0.5825, + "step": 3980 + }, + { + "epoch": 1.506228765571914, + "grad_norm": 0.5241252183914185, + "learning_rate": 9.932552896065062e-05, + "loss": 0.582, + "step": 3990 + }, + { + "epoch": 1.5100037750094375, + "grad_norm": 0.6853564977645874, + "learning_rate": 9.931874538206114e-05, + "loss": 0.5793, + "step": 4000 + }, + { + "epoch": 1.513778784446961, + "grad_norm": 0.47244563698768616, + "learning_rate": 9.931192809457423e-05, + "loss": 0.5737, + "step": 4010 + }, + { + "epoch": 1.5175537938844847, + "grad_norm": 0.5106404423713684, + "learning_rate": 9.930507710284944e-05, + "loss": 0.5773, + "step": 4020 + }, + { + "epoch": 1.5213288033220083, + "grad_norm": 0.5653221011161804, + "learning_rate": 9.92981924115694e-05, + "loss": 0.5729, + "step": 4030 + }, + { + "epoch": 1.5251038127595318, + "grad_norm": 0.7388667464256287, + "learning_rate": 9.929127402543968e-05, + "loss": 0.5695, + "step": 4040 + }, + { + "epoch": 1.5288788221970555, + "grad_norm": 0.55259770154953, + "learning_rate": 9.928432194918895e-05, + "loss": 0.5715, + "step": 4050 + }, + { + "epoch": 1.5326538316345792, + "grad_norm": 0.5290758609771729, + "learning_rate": 9.927733618756889e-05, + "loss": 0.5753, + "step": 4060 + }, + { + "epoch": 1.5364288410721025, + "grad_norm": 0.48387640714645386, + "learning_rate": 9.92703167453542e-05, + "loss": 0.5744, + "step": 4070 + }, + { + "epoch": 1.5402038505096263, + "grad_norm": 0.5005270838737488, + "learning_rate": 9.926326362734263e-05, + "loss": 0.5767, + "step": 4080 + }, + { + "epoch": 1.54397885994715, + "grad_norm": 0.5427155494689941, + "learning_rate": 9.925617683835489e-05, + "loss": 0.5726, + "step": 4090 + }, + { + "epoch": 1.5477538693846733, + "grad_norm": 0.5388875603675842, + "learning_rate": 9.924905638323472e-05, + "loss": 0.5673, + "step": 4100 + }, + { + "epoch": 1.551528878822197, + "grad_norm": 0.48597651720046997, + "learning_rate": 9.92419022668489e-05, + "loss": 0.5684, + "step": 4110 + }, + { + "epoch": 1.5553038882597208, + "grad_norm": 0.5301367044448853, + "learning_rate": 9.923471449408723e-05, + "loss": 0.5752, + "step": 4120 + }, + { + "epoch": 1.5590788976972443, + "grad_norm": 0.5001996159553528, + "learning_rate": 9.922749306986244e-05, + "loss": 0.5629, + "step": 4130 + }, + { + "epoch": 1.5628539071347678, + "grad_norm": 0.5677825212478638, + "learning_rate": 9.922023799911033e-05, + "loss": 0.5742, + "step": 4140 + }, + { + "epoch": 1.5666289165722915, + "grad_norm": 0.5715368390083313, + "learning_rate": 9.921294928678968e-05, + "loss": 0.5751, + "step": 4150 + }, + { + "epoch": 1.570403926009815, + "grad_norm": 0.632077157497406, + "learning_rate": 9.920562693788222e-05, + "loss": 0.5702, + "step": 4160 + }, + { + "epoch": 1.5741789354473386, + "grad_norm": 0.5607465505599976, + "learning_rate": 9.919827095739273e-05, + "loss": 0.5713, + "step": 4170 + }, + { + "epoch": 1.5779539448848623, + "grad_norm": 0.5231792330741882, + "learning_rate": 9.919088135034898e-05, + "loss": 0.5696, + "step": 4180 + }, + { + "epoch": 1.5817289543223858, + "grad_norm": 0.46094974875450134, + "learning_rate": 9.918345812180165e-05, + "loss": 0.5599, + "step": 4190 + }, + { + "epoch": 1.5855039637599093, + "grad_norm": 0.5033226609230042, + "learning_rate": 9.917600127682446e-05, + "loss": 0.5706, + "step": 4200 + }, + { + "epoch": 1.589278973197433, + "grad_norm": 0.5279197096824646, + "learning_rate": 9.91685108205141e-05, + "loss": 0.5674, + "step": 4210 + }, + { + "epoch": 1.5930539826349566, + "grad_norm": 0.49496957659721375, + "learning_rate": 9.916098675799024e-05, + "loss": 0.5644, + "step": 4220 + }, + { + "epoch": 1.59682899207248, + "grad_norm": 0.5188071131706238, + "learning_rate": 9.915342909439548e-05, + "loss": 0.5699, + "step": 4230 + }, + { + "epoch": 1.6006040015100038, + "grad_norm": 0.6602573394775391, + "learning_rate": 9.914583783489543e-05, + "loss": 0.5699, + "step": 4240 + }, + { + "epoch": 1.6043790109475273, + "grad_norm": 0.44954216480255127, + "learning_rate": 9.913821298467863e-05, + "loss": 0.5622, + "step": 4250 + }, + { + "epoch": 1.6081540203850508, + "grad_norm": 0.40007320046424866, + "learning_rate": 9.913055454895661e-05, + "loss": 0.5649, + "step": 4260 + }, + { + "epoch": 1.6119290298225746, + "grad_norm": 0.5066043734550476, + "learning_rate": 9.912286253296381e-05, + "loss": 0.5689, + "step": 4270 + }, + { + "epoch": 1.6157040392600983, + "grad_norm": 0.5492740273475647, + "learning_rate": 9.911513694195768e-05, + "loss": 0.5592, + "step": 4280 + }, + { + "epoch": 1.6194790486976216, + "grad_norm": 0.40257328748703003, + "learning_rate": 9.910737778121859e-05, + "loss": 0.5671, + "step": 4290 + }, + { + "epoch": 1.6232540581351453, + "grad_norm": 0.42792394757270813, + "learning_rate": 9.909958505604984e-05, + "loss": 0.5624, + "step": 4300 + }, + { + "epoch": 1.627029067572669, + "grad_norm": 0.5710054039955139, + "learning_rate": 9.909175877177767e-05, + "loss": 0.5697, + "step": 4310 + }, + { + "epoch": 1.6308040770101924, + "grad_norm": 0.47160592675209045, + "learning_rate": 9.908389893375129e-05, + "loss": 0.5684, + "step": 4320 + }, + { + "epoch": 1.634579086447716, + "grad_norm": 0.6177746653556824, + "learning_rate": 9.907600554734283e-05, + "loss": 0.5724, + "step": 4330 + }, + { + "epoch": 1.6383540958852398, + "grad_norm": 0.56391841173172, + "learning_rate": 9.906807861794734e-05, + "loss": 0.5675, + "step": 4340 + }, + { + "epoch": 1.6421291053227633, + "grad_norm": 0.6061177849769592, + "learning_rate": 9.906011815098279e-05, + "loss": 0.5607, + "step": 4350 + }, + { + "epoch": 1.6459041147602869, + "grad_norm": 0.580389678478241, + "learning_rate": 9.90521241518901e-05, + "loss": 0.563, + "step": 4360 + }, + { + "epoch": 1.6496791241978106, + "grad_norm": 0.6616781949996948, + "learning_rate": 9.904409662613308e-05, + "loss": 0.5664, + "step": 4370 + }, + { + "epoch": 1.6534541336353341, + "grad_norm": 0.5005329847335815, + "learning_rate": 9.903603557919849e-05, + "loss": 0.5592, + "step": 4380 + }, + { + "epoch": 1.6572291430728576, + "grad_norm": 0.4905034601688385, + "learning_rate": 9.902794101659594e-05, + "loss": 0.5678, + "step": 4390 + }, + { + "epoch": 1.6610041525103814, + "grad_norm": 0.44465407729148865, + "learning_rate": 9.901981294385803e-05, + "loss": 0.5684, + "step": 4400 + }, + { + "epoch": 1.6647791619479049, + "grad_norm": 0.5083572864532471, + "learning_rate": 9.901165136654018e-05, + "loss": 0.561, + "step": 4410 + }, + { + "epoch": 1.6685541713854284, + "grad_norm": 0.5734196305274963, + "learning_rate": 9.900345629022079e-05, + "loss": 0.57, + "step": 4420 + }, + { + "epoch": 1.6723291808229521, + "grad_norm": 0.568246066570282, + "learning_rate": 9.89952277205011e-05, + "loss": 0.5666, + "step": 4430 + }, + { + "epoch": 1.6761041902604756, + "grad_norm": 0.5339452624320984, + "learning_rate": 9.898696566300527e-05, + "loss": 0.5599, + "step": 4440 + }, + { + "epoch": 1.6798791996979991, + "grad_norm": 0.5378457307815552, + "learning_rate": 9.897867012338032e-05, + "loss": 0.5605, + "step": 4450 + }, + { + "epoch": 1.6836542091355229, + "grad_norm": 0.5373885035514832, + "learning_rate": 9.897034110729617e-05, + "loss": 0.5641, + "step": 4460 + }, + { + "epoch": 1.6874292185730464, + "grad_norm": 1.103470802307129, + "learning_rate": 9.896197862044564e-05, + "loss": 0.5773, + "step": 4470 + }, + { + "epoch": 1.69120422801057, + "grad_norm": 0.6465436220169067, + "learning_rate": 9.895358266854442e-05, + "loss": 0.5764, + "step": 4480 + }, + { + "epoch": 1.6949792374480936, + "grad_norm": 0.5247374176979065, + "learning_rate": 9.894515325733103e-05, + "loss": 0.5593, + "step": 4490 + }, + { + "epoch": 1.6987542468856174, + "grad_norm": 0.589596688747406, + "learning_rate": 9.893669039256693e-05, + "loss": 0.5637, + "step": 4500 + }, + { + "epoch": 1.7025292563231407, + "grad_norm": 0.4953310489654541, + "learning_rate": 9.89281940800364e-05, + "loss": 0.5622, + "step": 4510 + }, + { + "epoch": 1.7063042657606644, + "grad_norm": 0.45567750930786133, + "learning_rate": 9.891966432554655e-05, + "loss": 0.5594, + "step": 4520 + }, + { + "epoch": 1.7100792751981881, + "grad_norm": 0.4500698149204254, + "learning_rate": 9.891110113492745e-05, + "loss": 0.5619, + "step": 4530 + }, + { + "epoch": 1.7138542846357114, + "grad_norm": 0.4617537260055542, + "learning_rate": 9.89025045140319e-05, + "loss": 0.5623, + "step": 4540 + }, + { + "epoch": 1.7176292940732352, + "grad_norm": 0.5609983801841736, + "learning_rate": 9.889387446873567e-05, + "loss": 0.5629, + "step": 4550 + }, + { + "epoch": 1.721404303510759, + "grad_norm": 0.6847432851791382, + "learning_rate": 9.888521100493726e-05, + "loss": 0.57, + "step": 4560 + }, + { + "epoch": 1.7251793129482824, + "grad_norm": 0.50513756275177, + "learning_rate": 9.887651412855809e-05, + "loss": 0.5586, + "step": 4570 + }, + { + "epoch": 1.728954322385806, + "grad_norm": 0.45217183232307434, + "learning_rate": 9.88677838455424e-05, + "loss": 0.5493, + "step": 4580 + }, + { + "epoch": 1.7327293318233297, + "grad_norm": 0.445190966129303, + "learning_rate": 9.885902016185725e-05, + "loss": 0.5597, + "step": 4590 + }, + { + "epoch": 1.7365043412608532, + "grad_norm": 0.49329379200935364, + "learning_rate": 9.885022308349252e-05, + "loss": 0.5608, + "step": 4600 + }, + { + "epoch": 1.7402793506983767, + "grad_norm": 0.531954824924469, + "learning_rate": 9.884139261646095e-05, + "loss": 0.5646, + "step": 4610 + }, + { + "epoch": 1.7440543601359004, + "grad_norm": 0.5033367276191711, + "learning_rate": 9.883252876679807e-05, + "loss": 0.564, + "step": 4620 + }, + { + "epoch": 1.747829369573424, + "grad_norm": 0.8055509328842163, + "learning_rate": 9.882363154056225e-05, + "loss": 0.5519, + "step": 4630 + }, + { + "epoch": 1.7516043790109475, + "grad_norm": 0.592330813407898, + "learning_rate": 9.881470094383465e-05, + "loss": 0.5565, + "step": 4640 + }, + { + "epoch": 1.7553793884484712, + "grad_norm": 0.605402946472168, + "learning_rate": 9.880573698271924e-05, + "loss": 0.5737, + "step": 4650 + }, + { + "epoch": 1.7591543978859947, + "grad_norm": 0.6107518076896667, + "learning_rate": 9.879673966334282e-05, + "loss": 0.5569, + "step": 4660 + }, + { + "epoch": 1.7629294073235182, + "grad_norm": 0.5053714513778687, + "learning_rate": 9.878770899185496e-05, + "loss": 0.5599, + "step": 4670 + }, + { + "epoch": 1.766704416761042, + "grad_norm": 0.4542829394340515, + "learning_rate": 9.877864497442804e-05, + "loss": 0.5606, + "step": 4680 + }, + { + "epoch": 1.7704794261985655, + "grad_norm": 0.6275748610496521, + "learning_rate": 9.876954761725723e-05, + "loss": 0.5576, + "step": 4690 + }, + { + "epoch": 1.774254435636089, + "grad_norm": 0.5331715941429138, + "learning_rate": 9.876041692656052e-05, + "loss": 0.5567, + "step": 4700 + }, + { + "epoch": 1.7780294450736127, + "grad_norm": 0.49501025676727295, + "learning_rate": 9.87512529085786e-05, + "loss": 0.5617, + "step": 4710 + }, + { + "epoch": 1.7818044545111362, + "grad_norm": 0.4898965060710907, + "learning_rate": 9.874205556957503e-05, + "loss": 0.5482, + "step": 4720 + }, + { + "epoch": 1.7855794639486597, + "grad_norm": 0.47128167748451233, + "learning_rate": 9.873282491583608e-05, + "loss": 0.5579, + "step": 4730 + }, + { + "epoch": 1.7893544733861835, + "grad_norm": 0.4385862946510315, + "learning_rate": 9.872356095367084e-05, + "loss": 0.5551, + "step": 4740 + }, + { + "epoch": 1.7931294828237072, + "grad_norm": 0.5858024954795837, + "learning_rate": 9.87142636894111e-05, + "loss": 0.5474, + "step": 4750 + }, + { + "epoch": 1.7969044922612305, + "grad_norm": 0.4746013283729553, + "learning_rate": 9.870493312941148e-05, + "loss": 0.5596, + "step": 4760 + }, + { + "epoch": 1.8006795016987542, + "grad_norm": 0.5050995945930481, + "learning_rate": 9.869556928004933e-05, + "loss": 0.5451, + "step": 4770 + }, + { + "epoch": 1.804454511136278, + "grad_norm": 0.6011149287223816, + "learning_rate": 9.868617214772476e-05, + "loss": 0.5543, + "step": 4780 + }, + { + "epoch": 1.8082295205738015, + "grad_norm": 0.5801728367805481, + "learning_rate": 9.867674173886059e-05, + "loss": 0.549, + "step": 4790 + }, + { + "epoch": 1.812004530011325, + "grad_norm": 0.5891419053077698, + "learning_rate": 9.866727805990245e-05, + "loss": 0.5539, + "step": 4800 + }, + { + "epoch": 1.8157795394488487, + "grad_norm": 0.538972795009613, + "learning_rate": 9.865778111731865e-05, + "loss": 0.5506, + "step": 4810 + }, + { + "epoch": 1.8195545488863722, + "grad_norm": 0.5291750431060791, + "learning_rate": 9.864825091760028e-05, + "loss": 0.5576, + "step": 4820 + }, + { + "epoch": 1.8233295583238958, + "grad_norm": 0.5145063400268555, + "learning_rate": 9.863868746726112e-05, + "loss": 0.5523, + "step": 4830 + }, + { + "epoch": 1.8271045677614195, + "grad_norm": 0.4349626898765564, + "learning_rate": 9.862909077283773e-05, + "loss": 0.5525, + "step": 4840 + }, + { + "epoch": 1.830879577198943, + "grad_norm": 0.5479471683502197, + "learning_rate": 9.861946084088933e-05, + "loss": 0.5528, + "step": 4850 + }, + { + "epoch": 1.8346545866364665, + "grad_norm": 0.7466859817504883, + "learning_rate": 9.860979767799792e-05, + "loss": 0.5539, + "step": 4860 + }, + { + "epoch": 1.8384295960739903, + "grad_norm": 0.5633376836776733, + "learning_rate": 9.860010129076813e-05, + "loss": 0.5538, + "step": 4870 + }, + { + "epoch": 1.8422046055115138, + "grad_norm": 0.49453553557395935, + "learning_rate": 9.859037168582741e-05, + "loss": 0.5486, + "step": 4880 + }, + { + "epoch": 1.8459796149490373, + "grad_norm": 0.5297577381134033, + "learning_rate": 9.85806088698258e-05, + "loss": 0.5394, + "step": 4890 + }, + { + "epoch": 1.849754624386561, + "grad_norm": 0.5536269545555115, + "learning_rate": 9.857081284943612e-05, + "loss": 0.5494, + "step": 4900 + }, + { + "epoch": 1.8535296338240845, + "grad_norm": 0.4998897612094879, + "learning_rate": 9.856098363135385e-05, + "loss": 0.5459, + "step": 4910 + }, + { + "epoch": 1.857304643261608, + "grad_norm": 0.48842960596084595, + "learning_rate": 9.855112122229717e-05, + "loss": 0.5523, + "step": 4920 + }, + { + "epoch": 1.8610796526991318, + "grad_norm": 0.4732976257801056, + "learning_rate": 9.854122562900696e-05, + "loss": 0.568, + "step": 4930 + }, + { + "epoch": 1.8648546621366553, + "grad_norm": 0.47209542989730835, + "learning_rate": 9.853129685824673e-05, + "loss": 0.5526, + "step": 4940 + }, + { + "epoch": 1.8686296715741788, + "grad_norm": 0.4802228510379791, + "learning_rate": 9.852133491680273e-05, + "loss": 0.5431, + "step": 4950 + }, + { + "epoch": 1.8724046810117025, + "grad_norm": 0.507754385471344, + "learning_rate": 9.851133981148385e-05, + "loss": 0.5486, + "step": 4960 + }, + { + "epoch": 1.8761796904492263, + "grad_norm": 0.5926216244697571, + "learning_rate": 9.850131154912164e-05, + "loss": 0.549, + "step": 4970 + }, + { + "epoch": 1.8799546998867496, + "grad_norm": 0.49967560172080994, + "learning_rate": 9.849125013657031e-05, + "loss": 0.5459, + "step": 4980 + }, + { + "epoch": 1.8837297093242733, + "grad_norm": 0.5526815056800842, + "learning_rate": 9.848115558070676e-05, + "loss": 0.5439, + "step": 4990 + }, + { + "epoch": 1.887504718761797, + "grad_norm": 0.5583565831184387, + "learning_rate": 9.847102788843055e-05, + "loss": 0.5418, + "step": 5000 + }, + { + "epoch": 1.8912797281993206, + "grad_norm": 0.5497617125511169, + "learning_rate": 9.84608670666638e-05, + "loss": 0.5473, + "step": 5010 + }, + { + "epoch": 1.895054737636844, + "grad_norm": 0.6256689429283142, + "learning_rate": 9.845067312235138e-05, + "loss": 0.5485, + "step": 5020 + }, + { + "epoch": 1.8988297470743678, + "grad_norm": 0.5378864407539368, + "learning_rate": 9.844044606246074e-05, + "loss": 0.5432, + "step": 5030 + }, + { + "epoch": 1.9026047565118913, + "grad_norm": 0.5322272777557373, + "learning_rate": 9.843018589398199e-05, + "loss": 0.5605, + "step": 5040 + }, + { + "epoch": 1.9063797659494148, + "grad_norm": 0.47538843750953674, + "learning_rate": 9.841989262392785e-05, + "loss": 0.5452, + "step": 5050 + }, + { + "epoch": 1.9101547753869386, + "grad_norm": 0.5338093042373657, + "learning_rate": 9.840956625933367e-05, + "loss": 0.5443, + "step": 5060 + }, + { + "epoch": 1.913929784824462, + "grad_norm": 0.5679603815078735, + "learning_rate": 9.839920680725743e-05, + "loss": 0.5463, + "step": 5070 + }, + { + "epoch": 1.9177047942619856, + "grad_norm": 0.5604372024536133, + "learning_rate": 9.83888142747797e-05, + "loss": 0.5508, + "step": 5080 + }, + { + "epoch": 1.9214798036995093, + "grad_norm": 0.6041125655174255, + "learning_rate": 9.83783886690037e-05, + "loss": 0.5441, + "step": 5090 + }, + { + "epoch": 1.9252548131370328, + "grad_norm": 0.6924186944961548, + "learning_rate": 9.836792999705524e-05, + "loss": 0.546, + "step": 5100 + }, + { + "epoch": 1.9290298225745564, + "grad_norm": 0.4839281737804413, + "learning_rate": 9.835743826608271e-05, + "loss": 0.5505, + "step": 5110 + }, + { + "epoch": 1.93280483201208, + "grad_norm": 0.49724170565605164, + "learning_rate": 9.834691348325709e-05, + "loss": 0.5384, + "step": 5120 + }, + { + "epoch": 1.9365798414496036, + "grad_norm": 0.5565958619117737, + "learning_rate": 9.833635565577199e-05, + "loss": 0.5432, + "step": 5130 + }, + { + "epoch": 1.9403548508871271, + "grad_norm": 0.5241300463676453, + "learning_rate": 9.832576479084359e-05, + "loss": 0.5466, + "step": 5140 + }, + { + "epoch": 1.9441298603246508, + "grad_norm": 0.4891330599784851, + "learning_rate": 9.831514089571064e-05, + "loss": 0.5466, + "step": 5150 + }, + { + "epoch": 1.9479048697621744, + "grad_norm": 0.582870364189148, + "learning_rate": 9.830448397763447e-05, + "loss": 0.5394, + "step": 5160 + }, + { + "epoch": 1.9516798791996979, + "grad_norm": 0.5557212829589844, + "learning_rate": 9.8293794043899e-05, + "loss": 0.5426, + "step": 5170 + }, + { + "epoch": 1.9554548886372216, + "grad_norm": 0.5366427302360535, + "learning_rate": 9.828307110181066e-05, + "loss": 0.5488, + "step": 5180 + }, + { + "epoch": 1.9592298980747453, + "grad_norm": 0.512102484703064, + "learning_rate": 9.827231515869852e-05, + "loss": 0.5463, + "step": 5190 + }, + { + "epoch": 1.9630049075122686, + "grad_norm": 0.4691350758075714, + "learning_rate": 9.826152622191413e-05, + "loss": 0.5435, + "step": 5200 + }, + { + "epoch": 1.9667799169497924, + "grad_norm": 0.5275509357452393, + "learning_rate": 9.825070429883167e-05, + "loss": 0.5343, + "step": 5210 + }, + { + "epoch": 1.970554926387316, + "grad_norm": 0.4976699948310852, + "learning_rate": 9.82398493968478e-05, + "loss": 0.5423, + "step": 5220 + }, + { + "epoch": 1.9743299358248394, + "grad_norm": 0.5303046107292175, + "learning_rate": 9.822896152338173e-05, + "loss": 0.5428, + "step": 5230 + }, + { + "epoch": 1.9781049452623631, + "grad_norm": 0.47112399339675903, + "learning_rate": 9.821804068587523e-05, + "loss": 0.5357, + "step": 5240 + }, + { + "epoch": 1.9818799546998869, + "grad_norm": 0.4589209258556366, + "learning_rate": 9.820708689179259e-05, + "loss": 0.5445, + "step": 5250 + }, + { + "epoch": 1.9856549641374104, + "grad_norm": 0.47271859645843506, + "learning_rate": 9.819610014862063e-05, + "loss": 0.5339, + "step": 5260 + }, + { + "epoch": 1.989429973574934, + "grad_norm": 0.45445945858955383, + "learning_rate": 9.818508046386868e-05, + "loss": 0.5405, + "step": 5270 + }, + { + "epoch": 1.9932049830124576, + "grad_norm": 0.5319023728370667, + "learning_rate": 9.817402784506859e-05, + "loss": 0.5421, + "step": 5280 + }, + { + "epoch": 1.9969799924499811, + "grad_norm": 0.5033361315727234, + "learning_rate": 9.816294229977472e-05, + "loss": 0.5405, + "step": 5290 + }, + { + "epoch": 2.0007550018875047, + "grad_norm": 0.5742616057395935, + "learning_rate": 9.815182383556394e-05, + "loss": 0.5346, + "step": 5300 + }, + { + "epoch": 2.0045300113250284, + "grad_norm": 0.591316819190979, + "learning_rate": 9.81406724600356e-05, + "loss": 0.5396, + "step": 5310 + }, + { + "epoch": 2.0083050207625517, + "grad_norm": 0.4500562846660614, + "learning_rate": 9.812948818081161e-05, + "loss": 0.5375, + "step": 5320 + }, + { + "epoch": 2.0120800302000754, + "grad_norm": 0.5774345993995667, + "learning_rate": 9.811827100553623e-05, + "loss": 0.5345, + "step": 5330 + }, + { + "epoch": 2.015855039637599, + "grad_norm": 0.627068042755127, + "learning_rate": 9.810702094187638e-05, + "loss": 0.5432, + "step": 5340 + }, + { + "epoch": 2.019630049075123, + "grad_norm": 0.5072972178459167, + "learning_rate": 9.809573799752135e-05, + "loss": 0.5488, + "step": 5350 + }, + { + "epoch": 2.023405058512646, + "grad_norm": 0.5516794919967651, + "learning_rate": 9.808442218018288e-05, + "loss": 0.5383, + "step": 5360 + }, + { + "epoch": 2.02718006795017, + "grad_norm": 0.5118292570114136, + "learning_rate": 9.807307349759527e-05, + "loss": 0.539, + "step": 5370 + }, + { + "epoch": 2.0309550773876937, + "grad_norm": 0.5869454145431519, + "learning_rate": 9.806169195751525e-05, + "loss": 0.5374, + "step": 5380 + }, + { + "epoch": 2.034730086825217, + "grad_norm": 0.46649426221847534, + "learning_rate": 9.805027756772194e-05, + "loss": 0.5426, + "step": 5390 + }, + { + "epoch": 2.0385050962627407, + "grad_norm": 0.5441772937774658, + "learning_rate": 9.803883033601702e-05, + "loss": 0.5428, + "step": 5400 + }, + { + "epoch": 2.0422801057002644, + "grad_norm": 0.4856852889060974, + "learning_rate": 9.802735027022453e-05, + "loss": 0.5407, + "step": 5410 + }, + { + "epoch": 2.0460551151377877, + "grad_norm": 0.9828983545303345, + "learning_rate": 9.8015837378191e-05, + "loss": 0.5436, + "step": 5420 + }, + { + "epoch": 2.0498301245753114, + "grad_norm": 0.5788944959640503, + "learning_rate": 9.80042916677854e-05, + "loss": 0.54, + "step": 5430 + }, + { + "epoch": 2.053605134012835, + "grad_norm": 0.5199455618858337, + "learning_rate": 9.799271314689908e-05, + "loss": 0.5301, + "step": 5440 + }, + { + "epoch": 2.0573801434503585, + "grad_norm": 0.5322635173797607, + "learning_rate": 9.798110182344588e-05, + "loss": 0.5441, + "step": 5450 + }, + { + "epoch": 2.061155152887882, + "grad_norm": 0.4913989305496216, + "learning_rate": 9.796945770536204e-05, + "loss": 0.5466, + "step": 5460 + }, + { + "epoch": 2.064930162325406, + "grad_norm": 0.5005329847335815, + "learning_rate": 9.795778080060615e-05, + "loss": 0.5479, + "step": 5470 + }, + { + "epoch": 2.0687051717629292, + "grad_norm": 0.45034486055374146, + "learning_rate": 9.79460711171593e-05, + "loss": 0.5365, + "step": 5480 + }, + { + "epoch": 2.072480181200453, + "grad_norm": 0.530417799949646, + "learning_rate": 9.793432866302497e-05, + "loss": 0.5343, + "step": 5490 + }, + { + "epoch": 2.0762551906379767, + "grad_norm": 0.5080109238624573, + "learning_rate": 9.792255344622897e-05, + "loss": 0.5389, + "step": 5500 + }, + { + "epoch": 2.0800302000755, + "grad_norm": 0.5648706555366516, + "learning_rate": 9.791074547481957e-05, + "loss": 0.5375, + "step": 5510 + }, + { + "epoch": 2.0838052095130237, + "grad_norm": 0.4704912602901459, + "learning_rate": 9.789890475686742e-05, + "loss": 0.5314, + "step": 5520 + }, + { + "epoch": 2.0875802189505475, + "grad_norm": 0.47333037853240967, + "learning_rate": 9.788703130046552e-05, + "loss": 0.5357, + "step": 5530 + }, + { + "epoch": 2.091355228388071, + "grad_norm": 0.5688567161560059, + "learning_rate": 9.78751251137293e-05, + "loss": 0.5406, + "step": 5540 + }, + { + "epoch": 2.0951302378255945, + "grad_norm": 0.5980736613273621, + "learning_rate": 9.786318620479646e-05, + "loss": 0.541, + "step": 5550 + }, + { + "epoch": 2.0989052472631182, + "grad_norm": 0.5263574719429016, + "learning_rate": 9.78512145818272e-05, + "loss": 0.5337, + "step": 5560 + }, + { + "epoch": 2.102680256700642, + "grad_norm": 0.6168349981307983, + "learning_rate": 9.783921025300398e-05, + "loss": 0.534, + "step": 5570 + }, + { + "epoch": 2.1064552661381652, + "grad_norm": 0.5928676128387451, + "learning_rate": 9.782717322653164e-05, + "loss": 0.5298, + "step": 5580 + }, + { + "epoch": 2.110230275575689, + "grad_norm": 0.5435881018638611, + "learning_rate": 9.781510351063738e-05, + "loss": 0.5395, + "step": 5590 + }, + { + "epoch": 2.1140052850132127, + "grad_norm": 0.5674425363540649, + "learning_rate": 9.780300111357075e-05, + "loss": 0.5384, + "step": 5600 + }, + { + "epoch": 2.117780294450736, + "grad_norm": 0.60903400182724, + "learning_rate": 9.779086604360361e-05, + "loss": 0.5377, + "step": 5610 + }, + { + "epoch": 2.1215553038882597, + "grad_norm": 0.6011250615119934, + "learning_rate": 9.777869830903019e-05, + "loss": 0.5371, + "step": 5620 + }, + { + "epoch": 2.1253303133257835, + "grad_norm": 0.5067744851112366, + "learning_rate": 9.776649791816698e-05, + "loss": 0.529, + "step": 5630 + }, + { + "epoch": 2.1291053227633068, + "grad_norm": 0.5736885070800781, + "learning_rate": 9.775426487935285e-05, + "loss": 0.5298, + "step": 5640 + }, + { + "epoch": 2.1328803322008305, + "grad_norm": 0.5602228045463562, + "learning_rate": 9.774199920094898e-05, + "loss": 0.5312, + "step": 5650 + }, + { + "epoch": 2.1366553416383542, + "grad_norm": 0.5565608739852905, + "learning_rate": 9.772970089133884e-05, + "loss": 0.531, + "step": 5660 + }, + { + "epoch": 2.1404303510758775, + "grad_norm": 0.6767264604568481, + "learning_rate": 9.77173699589282e-05, + "loss": 0.53, + "step": 5670 + }, + { + "epoch": 2.1442053605134013, + "grad_norm": 0.5440645813941956, + "learning_rate": 9.770500641214513e-05, + "loss": 0.5317, + "step": 5680 + }, + { + "epoch": 2.147980369950925, + "grad_norm": 0.6627269387245178, + "learning_rate": 9.769261025944003e-05, + "loss": 0.5445, + "step": 5690 + }, + { + "epoch": 2.1517553793884483, + "grad_norm": 1.5312882661819458, + "learning_rate": 9.768018150928552e-05, + "loss": 0.5425, + "step": 5700 + }, + { + "epoch": 2.155530388825972, + "grad_norm": 0.621485710144043, + "learning_rate": 9.766772017017654e-05, + "loss": 0.5347, + "step": 5710 + }, + { + "epoch": 2.1593053982634958, + "grad_norm": 0.5858162641525269, + "learning_rate": 9.765522625063032e-05, + "loss": 0.5268, + "step": 5720 + }, + { + "epoch": 2.163080407701019, + "grad_norm": 0.4868731200695038, + "learning_rate": 9.76426997591863e-05, + "loss": 0.5361, + "step": 5730 + }, + { + "epoch": 2.166855417138543, + "grad_norm": 0.5163987874984741, + "learning_rate": 9.763014070440625e-05, + "loss": 0.5292, + "step": 5740 + }, + { + "epoch": 2.1706304265760665, + "grad_norm": 0.5586041808128357, + "learning_rate": 9.761754909487415e-05, + "loss": 0.5264, + "step": 5750 + }, + { + "epoch": 2.17440543601359, + "grad_norm": 0.45366203784942627, + "learning_rate": 9.760492493919626e-05, + "loss": 0.5405, + "step": 5760 + }, + { + "epoch": 2.1781804454511136, + "grad_norm": 0.5266132354736328, + "learning_rate": 9.759226824600105e-05, + "loss": 0.5403, + "step": 5770 + }, + { + "epoch": 2.1819554548886373, + "grad_norm": 0.516366183757782, + "learning_rate": 9.757957902393928e-05, + "loss": 0.5345, + "step": 5780 + }, + { + "epoch": 2.185730464326161, + "grad_norm": 0.4906608760356903, + "learning_rate": 9.756685728168387e-05, + "loss": 0.5437, + "step": 5790 + }, + { + "epoch": 2.1895054737636843, + "grad_norm": 0.5202317237854004, + "learning_rate": 9.755410302793004e-05, + "loss": 0.5299, + "step": 5800 + }, + { + "epoch": 2.193280483201208, + "grad_norm": 0.5337672829627991, + "learning_rate": 9.754131627139522e-05, + "loss": 0.5359, + "step": 5810 + }, + { + "epoch": 2.197055492638732, + "grad_norm": 0.5173420906066895, + "learning_rate": 9.752849702081901e-05, + "loss": 0.5345, + "step": 5820 + }, + { + "epoch": 2.200830502076255, + "grad_norm": 0.4675118625164032, + "learning_rate": 9.751564528496324e-05, + "loss": 0.531, + "step": 5830 + }, + { + "epoch": 2.204605511513779, + "grad_norm": 0.48218834400177, + "learning_rate": 9.750276107261197e-05, + "loss": 0.5279, + "step": 5840 + }, + { + "epoch": 2.2083805209513026, + "grad_norm": 0.5200138688087463, + "learning_rate": 9.748984439257142e-05, + "loss": 0.5289, + "step": 5850 + }, + { + "epoch": 2.212155530388826, + "grad_norm": 0.6025936603546143, + "learning_rate": 9.747689525367005e-05, + "loss": 0.526, + "step": 5860 + }, + { + "epoch": 2.2159305398263496, + "grad_norm": 0.4721621572971344, + "learning_rate": 9.746391366475845e-05, + "loss": 0.535, + "step": 5870 + }, + { + "epoch": 2.2197055492638733, + "grad_norm": 0.5410889983177185, + "learning_rate": 9.745089963470942e-05, + "loss": 0.5316, + "step": 5880 + }, + { + "epoch": 2.2234805587013966, + "grad_norm": 0.5121446251869202, + "learning_rate": 9.743785317241791e-05, + "loss": 0.5283, + "step": 5890 + }, + { + "epoch": 2.2272555681389203, + "grad_norm": 0.48920613527297974, + "learning_rate": 9.742477428680108e-05, + "loss": 0.5278, + "step": 5900 + }, + { + "epoch": 2.231030577576444, + "grad_norm": 0.47083306312561035, + "learning_rate": 9.741166298679821e-05, + "loss": 0.5301, + "step": 5910 + }, + { + "epoch": 2.2348055870139674, + "grad_norm": 0.45593857765197754, + "learning_rate": 9.739851928137076e-05, + "loss": 0.5278, + "step": 5920 + }, + { + "epoch": 2.238580596451491, + "grad_norm": 0.5642737746238708, + "learning_rate": 9.738534317950232e-05, + "loss": 0.5226, + "step": 5930 + }, + { + "epoch": 2.242355605889015, + "grad_norm": 0.5394318103790283, + "learning_rate": 9.737213469019864e-05, + "loss": 0.5278, + "step": 5940 + }, + { + "epoch": 2.246130615326538, + "grad_norm": 0.5335193872451782, + "learning_rate": 9.735889382248757e-05, + "loss": 0.5239, + "step": 5950 + }, + { + "epoch": 2.249905624764062, + "grad_norm": 0.5988762974739075, + "learning_rate": 9.734562058541916e-05, + "loss": 0.5418, + "step": 5960 + }, + { + "epoch": 2.2536806342015856, + "grad_norm": 0.5146274566650391, + "learning_rate": 9.733231498806552e-05, + "loss": 0.5384, + "step": 5970 + }, + { + "epoch": 2.2574556436391093, + "grad_norm": 0.5223647356033325, + "learning_rate": 9.731897703952088e-05, + "loss": 0.5321, + "step": 5980 + }, + { + "epoch": 2.2612306530766326, + "grad_norm": 0.5377148985862732, + "learning_rate": 9.730560674890165e-05, + "loss": 0.5249, + "step": 5990 + }, + { + "epoch": 2.2650056625141564, + "grad_norm": 0.6117594838142395, + "learning_rate": 9.729220412534628e-05, + "loss": 0.524, + "step": 6000 + }, + { + "epoch": 2.2687806719516797, + "grad_norm": 0.612998366355896, + "learning_rate": 9.72787691780153e-05, + "loss": 0.5314, + "step": 6010 + }, + { + "epoch": 2.2725556813892034, + "grad_norm": 0.5014058947563171, + "learning_rate": 9.726530191609142e-05, + "loss": 0.5267, + "step": 6020 + }, + { + "epoch": 2.276330690826727, + "grad_norm": 0.5796371698379517, + "learning_rate": 9.725180234877937e-05, + "loss": 0.5323, + "step": 6030 + }, + { + "epoch": 2.280105700264251, + "grad_norm": 0.6417374610900879, + "learning_rate": 9.723827048530595e-05, + "loss": 0.5308, + "step": 6040 + }, + { + "epoch": 2.283880709701774, + "grad_norm": 0.4562576711177826, + "learning_rate": 9.722470633492011e-05, + "loss": 0.5298, + "step": 6050 + }, + { + "epoch": 2.287655719139298, + "grad_norm": 0.5561509728431702, + "learning_rate": 9.721110990689278e-05, + "loss": 0.5265, + "step": 6060 + }, + { + "epoch": 2.2914307285768216, + "grad_norm": 0.5673072338104248, + "learning_rate": 9.719748121051699e-05, + "loss": 0.5195, + "step": 6070 + }, + { + "epoch": 2.295205738014345, + "grad_norm": 0.5953671336174011, + "learning_rate": 9.718382025510785e-05, + "loss": 0.5248, + "step": 6080 + }, + { + "epoch": 2.2989807474518686, + "grad_norm": 0.6203503608703613, + "learning_rate": 9.717012705000248e-05, + "loss": 0.5264, + "step": 6090 + }, + { + "epoch": 2.3027557568893924, + "grad_norm": 0.5094435811042786, + "learning_rate": 9.715640160456006e-05, + "loss": 0.5325, + "step": 6100 + }, + { + "epoch": 2.3065307663269157, + "grad_norm": 0.5434731245040894, + "learning_rate": 9.714264392816181e-05, + "loss": 0.5243, + "step": 6110 + }, + { + "epoch": 2.3103057757644394, + "grad_norm": 0.5409674048423767, + "learning_rate": 9.712885403021095e-05, + "loss": 0.523, + "step": 6120 + }, + { + "epoch": 2.314080785201963, + "grad_norm": 0.5891657471656799, + "learning_rate": 9.711503192013276e-05, + "loss": 0.529, + "step": 6130 + }, + { + "epoch": 2.3178557946394864, + "grad_norm": 0.6148802042007446, + "learning_rate": 9.710117760737454e-05, + "loss": 0.5234, + "step": 6140 + }, + { + "epoch": 2.32163080407701, + "grad_norm": 0.5786886215209961, + "learning_rate": 9.708729110140554e-05, + "loss": 0.524, + "step": 6150 + }, + { + "epoch": 2.325405813514534, + "grad_norm": 0.5547001361846924, + "learning_rate": 9.707337241171709e-05, + "loss": 0.5212, + "step": 6160 + }, + { + "epoch": 2.329180822952057, + "grad_norm": 0.6747357845306396, + "learning_rate": 9.705942154782245e-05, + "loss": 0.5274, + "step": 6170 + }, + { + "epoch": 2.332955832389581, + "grad_norm": 0.4640480875968933, + "learning_rate": 9.704543851925696e-05, + "loss": 0.5254, + "step": 6180 + }, + { + "epoch": 2.3367308418271047, + "grad_norm": 0.5004093647003174, + "learning_rate": 9.703142333557784e-05, + "loss": 0.5208, + "step": 6190 + }, + { + "epoch": 2.340505851264628, + "grad_norm": 0.5276467204093933, + "learning_rate": 9.701737600636436e-05, + "loss": 0.5205, + "step": 6200 + }, + { + "epoch": 2.3442808607021517, + "grad_norm": 0.5766600370407104, + "learning_rate": 9.700329654121771e-05, + "loss": 0.5145, + "step": 6210 + }, + { + "epoch": 2.3480558701396754, + "grad_norm": 0.4839555025100708, + "learning_rate": 9.698918494976109e-05, + "loss": 0.5234, + "step": 6220 + }, + { + "epoch": 2.351830879577199, + "grad_norm": 0.5838026404380798, + "learning_rate": 9.697504124163965e-05, + "loss": 0.5232, + "step": 6230 + }, + { + "epoch": 2.3556058890147225, + "grad_norm": 0.553741991519928, + "learning_rate": 9.696086542652045e-05, + "loss": 0.5223, + "step": 6240 + }, + { + "epoch": 2.359380898452246, + "grad_norm": 0.48992642760276794, + "learning_rate": 9.694665751409256e-05, + "loss": 0.521, + "step": 6250 + }, + { + "epoch": 2.3631559078897695, + "grad_norm": 0.511161208152771, + "learning_rate": 9.693241751406694e-05, + "loss": 0.5186, + "step": 6260 + }, + { + "epoch": 2.366930917327293, + "grad_norm": 0.480486124753952, + "learning_rate": 9.69181454361765e-05, + "loss": 0.517, + "step": 6270 + }, + { + "epoch": 2.370705926764817, + "grad_norm": 0.49377861618995667, + "learning_rate": 9.690384129017605e-05, + "loss": 0.5187, + "step": 6280 + }, + { + "epoch": 2.3744809362023407, + "grad_norm": 0.5575188398361206, + "learning_rate": 9.688950508584237e-05, + "loss": 0.5194, + "step": 6290 + }, + { + "epoch": 2.378255945639864, + "grad_norm": 0.501025378704071, + "learning_rate": 9.68751368329741e-05, + "loss": 0.519, + "step": 6300 + }, + { + "epoch": 2.3820309550773877, + "grad_norm": 0.5015434622764587, + "learning_rate": 9.68607365413918e-05, + "loss": 0.5213, + "step": 6310 + }, + { + "epoch": 2.3858059645149114, + "grad_norm": 0.4865148663520813, + "learning_rate": 9.684630422093797e-05, + "loss": 0.5212, + "step": 6320 + }, + { + "epoch": 2.3895809739524347, + "grad_norm": 0.545136034488678, + "learning_rate": 9.683183988147693e-05, + "loss": 0.5199, + "step": 6330 + }, + { + "epoch": 2.3933559833899585, + "grad_norm": 0.4947832226753235, + "learning_rate": 9.681734353289491e-05, + "loss": 0.5217, + "step": 6340 + }, + { + "epoch": 2.397130992827482, + "grad_norm": 0.554629921913147, + "learning_rate": 9.680281518510006e-05, + "loss": 0.5204, + "step": 6350 + }, + { + "epoch": 2.4009060022650055, + "grad_norm": 0.49678510427474976, + "learning_rate": 9.678825484802236e-05, + "loss": 0.5212, + "step": 6360 + }, + { + "epoch": 2.4046810117025292, + "grad_norm": 0.5372890830039978, + "learning_rate": 9.677366253161365e-05, + "loss": 0.5252, + "step": 6370 + }, + { + "epoch": 2.408456021140053, + "grad_norm": 0.4868631660938263, + "learning_rate": 9.675903824584765e-05, + "loss": 0.5309, + "step": 6380 + }, + { + "epoch": 2.4122310305775763, + "grad_norm": 0.5607171654701233, + "learning_rate": 9.674438200071991e-05, + "loss": 0.5197, + "step": 6390 + }, + { + "epoch": 2.4160060400151, + "grad_norm": 0.6623253226280212, + "learning_rate": 9.672969380624786e-05, + "loss": 0.5247, + "step": 6400 + }, + { + "epoch": 2.4197810494526237, + "grad_norm": 0.5090137720108032, + "learning_rate": 9.671497367247073e-05, + "loss": 0.5169, + "step": 6410 + }, + { + "epoch": 2.4235560588901475, + "grad_norm": 0.4954577684402466, + "learning_rate": 9.670022160944959e-05, + "loss": 0.5109, + "step": 6420 + }, + { + "epoch": 2.4273310683276708, + "grad_norm": 0.471021831035614, + "learning_rate": 9.668543762726734e-05, + "loss": 0.5147, + "step": 6430 + }, + { + "epoch": 2.4311060777651945, + "grad_norm": 0.5328172445297241, + "learning_rate": 9.667062173602868e-05, + "loss": 0.5144, + "step": 6440 + }, + { + "epoch": 2.434881087202718, + "grad_norm": 0.596956193447113, + "learning_rate": 9.665577394586013e-05, + "loss": 0.5231, + "step": 6450 + }, + { + "epoch": 2.4386560966402415, + "grad_norm": 0.5546612739562988, + "learning_rate": 9.664089426691006e-05, + "loss": 0.5233, + "step": 6460 + }, + { + "epoch": 2.4424311060777653, + "grad_norm": 0.5410290360450745, + "learning_rate": 9.662598270934852e-05, + "loss": 0.5234, + "step": 6470 + }, + { + "epoch": 2.446206115515289, + "grad_norm": 0.5477890968322754, + "learning_rate": 9.661103928336748e-05, + "loss": 0.5153, + "step": 6480 + }, + { + "epoch": 2.4499811249528123, + "grad_norm": 0.5000752210617065, + "learning_rate": 9.659606399918058e-05, + "loss": 0.5104, + "step": 6490 + }, + { + "epoch": 2.453756134390336, + "grad_norm": 0.5385310053825378, + "learning_rate": 9.658105686702334e-05, + "loss": 0.5143, + "step": 6500 + }, + { + "epoch": 2.4575311438278598, + "grad_norm": 0.48516371846199036, + "learning_rate": 9.656601789715294e-05, + "loss": 0.5117, + "step": 6510 + }, + { + "epoch": 2.461306153265383, + "grad_norm": 0.574826717376709, + "learning_rate": 9.655094709984842e-05, + "loss": 0.5114, + "step": 6520 + }, + { + "epoch": 2.465081162702907, + "grad_norm": 0.5119508504867554, + "learning_rate": 9.653584448541048e-05, + "loss": 0.519, + "step": 6530 + }, + { + "epoch": 2.4688561721404305, + "grad_norm": 0.5045502781867981, + "learning_rate": 9.652071006416166e-05, + "loss": 0.5134, + "step": 6540 + }, + { + "epoch": 2.472631181577954, + "grad_norm": 0.4496314227581024, + "learning_rate": 9.650554384644617e-05, + "loss": 0.513, + "step": 6550 + }, + { + "epoch": 2.4764061910154775, + "grad_norm": 0.45445775985717773, + "learning_rate": 9.649034584263e-05, + "loss": 0.5129, + "step": 6560 + }, + { + "epoch": 2.4801812004530013, + "grad_norm": 0.5979011058807373, + "learning_rate": 9.64751160631008e-05, + "loss": 0.5236, + "step": 6570 + }, + { + "epoch": 2.4839562098905246, + "grad_norm": 0.595025897026062, + "learning_rate": 9.645985451826803e-05, + "loss": 0.5233, + "step": 6580 + }, + { + "epoch": 2.4877312193280483, + "grad_norm": 0.5131351351737976, + "learning_rate": 9.644456121856275e-05, + "loss": 0.5186, + "step": 6590 + }, + { + "epoch": 2.491506228765572, + "grad_norm": 0.5269580483436584, + "learning_rate": 9.642923617443784e-05, + "loss": 0.527, + "step": 6600 + }, + { + "epoch": 2.4952812382030953, + "grad_norm": 0.43365490436553955, + "learning_rate": 9.641387939636782e-05, + "loss": 0.5164, + "step": 6610 + }, + { + "epoch": 2.499056247640619, + "grad_norm": 0.5464170575141907, + "learning_rate": 9.639849089484888e-05, + "loss": 0.5193, + "step": 6620 + }, + { + "epoch": 2.502831257078143, + "grad_norm": 0.5362778306007385, + "learning_rate": 9.638307068039893e-05, + "loss": 0.5204, + "step": 6630 + }, + { + "epoch": 2.506606266515666, + "grad_norm": 0.5616941452026367, + "learning_rate": 9.636761876355753e-05, + "loss": 0.5144, + "step": 6640 + }, + { + "epoch": 2.51038127595319, + "grad_norm": 0.4664197862148285, + "learning_rate": 9.635213515488593e-05, + "loss": 0.5116, + "step": 6650 + }, + { + "epoch": 2.5141562853907136, + "grad_norm": 0.5451635122299194, + "learning_rate": 9.633661986496702e-05, + "loss": 0.5167, + "step": 6660 + }, + { + "epoch": 2.5179312948282373, + "grad_norm": 0.47441262006759644, + "learning_rate": 9.632107290440535e-05, + "loss": 0.5114, + "step": 6670 + }, + { + "epoch": 2.5217063042657606, + "grad_norm": 0.5057079195976257, + "learning_rate": 9.630549428382715e-05, + "loss": 0.518, + "step": 6680 + }, + { + "epoch": 2.5254813137032843, + "grad_norm": 0.5143377184867859, + "learning_rate": 9.628988401388024e-05, + "loss": 0.5096, + "step": 6690 + }, + { + "epoch": 2.5292563231408076, + "grad_norm": 0.485270231962204, + "learning_rate": 9.627424210523407e-05, + "loss": 0.5115, + "step": 6700 + }, + { + "epoch": 2.5330313325783314, + "grad_norm": 0.5043366551399231, + "learning_rate": 9.62585685685798e-05, + "loss": 0.5142, + "step": 6710 + }, + { + "epoch": 2.536806342015855, + "grad_norm": 0.5317198038101196, + "learning_rate": 9.624286341463008e-05, + "loss": 0.5193, + "step": 6720 + }, + { + "epoch": 2.540581351453379, + "grad_norm": 0.4412330985069275, + "learning_rate": 9.622712665411927e-05, + "loss": 0.5119, + "step": 6730 + }, + { + "epoch": 2.544356360890902, + "grad_norm": 0.5754643082618713, + "learning_rate": 9.621135829780328e-05, + "loss": 0.515, + "step": 6740 + }, + { + "epoch": 2.548131370328426, + "grad_norm": 0.5370513200759888, + "learning_rate": 9.619555835645964e-05, + "loss": 0.5132, + "step": 6750 + }, + { + "epoch": 2.551906379765949, + "grad_norm": 0.5530955791473389, + "learning_rate": 9.617972684088747e-05, + "loss": 0.5167, + "step": 6760 + }, + { + "epoch": 2.555681389203473, + "grad_norm": 0.5820709466934204, + "learning_rate": 9.616386376190745e-05, + "loss": 0.5118, + "step": 6770 + }, + { + "epoch": 2.5594563986409966, + "grad_norm": 0.48913341760635376, + "learning_rate": 9.614796913036184e-05, + "loss": 0.5152, + "step": 6780 + }, + { + "epoch": 2.5632314080785203, + "grad_norm": 0.48742911219596863, + "learning_rate": 9.61320429571145e-05, + "loss": 0.5074, + "step": 6790 + }, + { + "epoch": 2.5670064175160436, + "grad_norm": 0.935467004776001, + "learning_rate": 9.611608525305078e-05, + "loss": 0.5215, + "step": 6800 + }, + { + "epoch": 2.5707814269535674, + "grad_norm": 0.4997836947441101, + "learning_rate": 9.610009602907764e-05, + "loss": 0.5082, + "step": 6810 + }, + { + "epoch": 2.574556436391091, + "grad_norm": 0.47966811060905457, + "learning_rate": 9.608407529612357e-05, + "loss": 0.5044, + "step": 6820 + }, + { + "epoch": 2.5783314458286144, + "grad_norm": 0.5587261915206909, + "learning_rate": 9.606802306513857e-05, + "loss": 0.506, + "step": 6830 + }, + { + "epoch": 2.582106455266138, + "grad_norm": 0.4997664988040924, + "learning_rate": 9.605193934709421e-05, + "loss": 0.5145, + "step": 6840 + }, + { + "epoch": 2.585881464703662, + "grad_norm": 0.5600175261497498, + "learning_rate": 9.603582415298354e-05, + "loss": 0.5111, + "step": 6850 + }, + { + "epoch": 2.5896564741411856, + "grad_norm": 0.49723485112190247, + "learning_rate": 9.601967749382117e-05, + "loss": 0.5114, + "step": 6860 + }, + { + "epoch": 2.593431483578709, + "grad_norm": 0.4990619421005249, + "learning_rate": 9.600349938064316e-05, + "loss": 0.5086, + "step": 6870 + }, + { + "epoch": 2.5972064930162326, + "grad_norm": 0.4644363522529602, + "learning_rate": 9.598728982450711e-05, + "loss": 0.5134, + "step": 6880 + }, + { + "epoch": 2.600981502453756, + "grad_norm": 0.4740346670150757, + "learning_rate": 9.597104883649211e-05, + "loss": 0.5176, + "step": 6890 + }, + { + "epoch": 2.6047565118912797, + "grad_norm": 0.5966189503669739, + "learning_rate": 9.59547764276987e-05, + "loss": 0.5044, + "step": 6900 + }, + { + "epoch": 2.6085315213288034, + "grad_norm": 0.5578002333641052, + "learning_rate": 9.593847260924892e-05, + "loss": 0.5072, + "step": 6910 + }, + { + "epoch": 2.612306530766327, + "grad_norm": 0.5736666321754456, + "learning_rate": 9.59221373922863e-05, + "loss": 0.5144, + "step": 6920 + }, + { + "epoch": 2.6160815402038504, + "grad_norm": 0.6220727562904358, + "learning_rate": 9.59057707879758e-05, + "loss": 0.5031, + "step": 6930 + }, + { + "epoch": 2.619856549641374, + "grad_norm": 0.565088152885437, + "learning_rate": 9.588937280750382e-05, + "loss": 0.5083, + "step": 6940 + }, + { + "epoch": 2.6236315590788974, + "grad_norm": 0.5720441937446594, + "learning_rate": 9.587294346207824e-05, + "loss": 0.5303, + "step": 6950 + }, + { + "epoch": 2.627406568516421, + "grad_norm": 0.6338963508605957, + "learning_rate": 9.585648276292836e-05, + "loss": 0.5149, + "step": 6960 + }, + { + "epoch": 2.631181577953945, + "grad_norm": 0.504148542881012, + "learning_rate": 9.583999072130492e-05, + "loss": 0.5045, + "step": 6970 + }, + { + "epoch": 2.6349565873914687, + "grad_norm": 0.5418031215667725, + "learning_rate": 9.582346734848005e-05, + "loss": 0.5128, + "step": 6980 + }, + { + "epoch": 2.638731596828992, + "grad_norm": 0.5693923234939575, + "learning_rate": 9.580691265574735e-05, + "loss": 0.5106, + "step": 6990 + }, + { + "epoch": 2.6425066062665157, + "grad_norm": 0.4889356195926666, + "learning_rate": 9.579032665442178e-05, + "loss": 0.5053, + "step": 7000 + }, + { + "epoch": 2.6462816157040394, + "grad_norm": 0.5362417697906494, + "learning_rate": 9.577370935583971e-05, + "loss": 0.5196, + "step": 7010 + }, + { + "epoch": 2.6500566251415627, + "grad_norm": 0.572538435459137, + "learning_rate": 9.575706077135893e-05, + "loss": 0.5022, + "step": 7020 + }, + { + "epoch": 2.6538316345790864, + "grad_norm": 0.5328758955001831, + "learning_rate": 9.574038091235857e-05, + "loss": 0.5186, + "step": 7030 + }, + { + "epoch": 2.65760664401661, + "grad_norm": 0.4795249104499817, + "learning_rate": 9.572366979023916e-05, + "loss": 0.5051, + "step": 7040 + }, + { + "epoch": 2.661381653454134, + "grad_norm": 1.024570345878601, + "learning_rate": 9.57069274164226e-05, + "loss": 0.5087, + "step": 7050 + }, + { + "epoch": 2.665156662891657, + "grad_norm": 0.641263484954834, + "learning_rate": 9.569015380235213e-05, + "loss": 0.5094, + "step": 7060 + }, + { + "epoch": 2.668931672329181, + "grad_norm": 0.5658965706825256, + "learning_rate": 9.567334895949238e-05, + "loss": 0.5089, + "step": 7070 + }, + { + "epoch": 2.6727066817667042, + "grad_norm": 0.6503294706344604, + "learning_rate": 9.565651289932928e-05, + "loss": 0.5071, + "step": 7080 + }, + { + "epoch": 2.676481691204228, + "grad_norm": 0.5400107502937317, + "learning_rate": 9.563964563337015e-05, + "loss": 0.5072, + "step": 7090 + }, + { + "epoch": 2.6802567006417517, + "grad_norm": 0.5809869766235352, + "learning_rate": 9.562274717314357e-05, + "loss": 0.5055, + "step": 7100 + }, + { + "epoch": 2.6840317100792754, + "grad_norm": 0.5175840854644775, + "learning_rate": 9.56058175301995e-05, + "loss": 0.5084, + "step": 7110 + }, + { + "epoch": 2.6878067195167987, + "grad_norm": 0.510823130607605, + "learning_rate": 9.558885671610918e-05, + "loss": 0.5142, + "step": 7120 + }, + { + "epoch": 2.6915817289543225, + "grad_norm": 0.5192266702651978, + "learning_rate": 9.557186474246517e-05, + "loss": 0.51, + "step": 7130 + }, + { + "epoch": 2.6953567383918458, + "grad_norm": 0.4871031939983368, + "learning_rate": 9.555484162088133e-05, + "loss": 0.5029, + "step": 7140 + }, + { + "epoch": 2.6991317478293695, + "grad_norm": 0.6060387492179871, + "learning_rate": 9.553778736299279e-05, + "loss": 0.5032, + "step": 7150 + }, + { + "epoch": 2.7029067572668932, + "grad_norm": 0.5182626247406006, + "learning_rate": 9.552070198045599e-05, + "loss": 0.5086, + "step": 7160 + }, + { + "epoch": 2.706681766704417, + "grad_norm": 0.509029746055603, + "learning_rate": 9.550358548494863e-05, + "loss": 0.5076, + "step": 7170 + }, + { + "epoch": 2.7104567761419402, + "grad_norm": 0.5284494757652283, + "learning_rate": 9.548643788816965e-05, + "loss": 0.4986, + "step": 7180 + }, + { + "epoch": 2.714231785579464, + "grad_norm": 0.5049721598625183, + "learning_rate": 9.546925920183929e-05, + "loss": 0.5059, + "step": 7190 + }, + { + "epoch": 2.7180067950169873, + "grad_norm": 0.5326721668243408, + "learning_rate": 9.545204943769902e-05, + "loss": 0.5078, + "step": 7200 + }, + { + "epoch": 2.721781804454511, + "grad_norm": 0.5484986305236816, + "learning_rate": 9.543480860751155e-05, + "loss": 0.5036, + "step": 7210 + }, + { + "epoch": 2.7255568138920347, + "grad_norm": 0.5126481652259827, + "learning_rate": 9.541753672306081e-05, + "loss": 0.5037, + "step": 7220 + }, + { + "epoch": 2.7293318233295585, + "grad_norm": 0.4992707669734955, + "learning_rate": 9.540023379615198e-05, + "loss": 0.499, + "step": 7230 + }, + { + "epoch": 2.7331068327670818, + "grad_norm": 0.46070438623428345, + "learning_rate": 9.538289983861146e-05, + "loss": 0.504, + "step": 7240 + }, + { + "epoch": 2.7368818422046055, + "grad_norm": 0.6228576898574829, + "learning_rate": 9.536553486228683e-05, + "loss": 0.4989, + "step": 7250 + }, + { + "epoch": 2.7406568516421292, + "grad_norm": 0.5347064137458801, + "learning_rate": 9.534813887904689e-05, + "loss": 0.5057, + "step": 7260 + }, + { + "epoch": 2.7444318610796525, + "grad_norm": 0.5296137928962708, + "learning_rate": 9.533071190078163e-05, + "loss": 0.5062, + "step": 7270 + }, + { + "epoch": 2.7482068705171763, + "grad_norm": 0.521789014339447, + "learning_rate": 9.531325393940225e-05, + "loss": 0.5067, + "step": 7280 + }, + { + "epoch": 2.7519818799547, + "grad_norm": 0.5194920897483826, + "learning_rate": 9.529576500684105e-05, + "loss": 0.4991, + "step": 7290 + }, + { + "epoch": 2.7557568893922237, + "grad_norm": 0.6369245648384094, + "learning_rate": 9.52782451150516e-05, + "loss": 0.5752, + "step": 7300 + }, + { + "epoch": 2.759531898829747, + "grad_norm": 0.6341930031776428, + "learning_rate": 9.526069427600857e-05, + "loss": 0.5042, + "step": 7310 + }, + { + "epoch": 2.7633069082672708, + "grad_norm": 0.5371559262275696, + "learning_rate": 9.524311250170776e-05, + "loss": 0.5055, + "step": 7320 + }, + { + "epoch": 2.767081917704794, + "grad_norm": 0.6049063205718994, + "learning_rate": 9.522549980416619e-05, + "loss": 0.5062, + "step": 7330 + }, + { + "epoch": 2.770856927142318, + "grad_norm": 0.5581164956092834, + "learning_rate": 9.520785619542196e-05, + "loss": 0.5072, + "step": 7340 + }, + { + "epoch": 2.7746319365798415, + "grad_norm": 0.5462380051612854, + "learning_rate": 9.519018168753428e-05, + "loss": 0.5026, + "step": 7350 + }, + { + "epoch": 2.7784069460173653, + "grad_norm": 0.5068135857582092, + "learning_rate": 9.517247629258357e-05, + "loss": 0.5122, + "step": 7360 + }, + { + "epoch": 2.7821819554548886, + "grad_norm": 0.5248288512229919, + "learning_rate": 9.515474002267124e-05, + "loss": 0.5, + "step": 7370 + }, + { + "epoch": 2.7859569648924123, + "grad_norm": 0.5446317791938782, + "learning_rate": 9.513697288991989e-05, + "loss": 0.5096, + "step": 7380 + }, + { + "epoch": 2.7897319743299356, + "grad_norm": 0.5539740324020386, + "learning_rate": 9.51191749064732e-05, + "loss": 0.5037, + "step": 7390 + }, + { + "epoch": 2.7935069837674593, + "grad_norm": 0.48345574736595154, + "learning_rate": 9.51013460844959e-05, + "loss": 0.4938, + "step": 7400 + }, + { + "epoch": 2.797281993204983, + "grad_norm": 0.538747251033783, + "learning_rate": 9.508348643617382e-05, + "loss": 0.501, + "step": 7410 + }, + { + "epoch": 2.801057002642507, + "grad_norm": 0.4825133681297302, + "learning_rate": 9.50655959737139e-05, + "loss": 0.4987, + "step": 7420 + }, + { + "epoch": 2.80483201208003, + "grad_norm": 0.5534849762916565, + "learning_rate": 9.504767470934405e-05, + "loss": 0.4992, + "step": 7430 + }, + { + "epoch": 2.808607021517554, + "grad_norm": 0.5716150999069214, + "learning_rate": 9.502972265531332e-05, + "loss": 0.5085, + "step": 7440 + }, + { + "epoch": 2.8123820309550775, + "grad_norm": 0.47607553005218506, + "learning_rate": 9.501173982389175e-05, + "loss": 0.501, + "step": 7450 + }, + { + "epoch": 2.816157040392601, + "grad_norm": 0.5296433568000793, + "learning_rate": 9.499372622737047e-05, + "loss": 0.5004, + "step": 7460 + }, + { + "epoch": 2.8199320498301246, + "grad_norm": 0.5382222533226013, + "learning_rate": 9.497568187806156e-05, + "loss": 0.5099, + "step": 7470 + }, + { + "epoch": 2.8237070592676483, + "grad_norm": 0.480925977230072, + "learning_rate": 9.495760678829819e-05, + "loss": 0.5031, + "step": 7480 + }, + { + "epoch": 2.8274820687051716, + "grad_norm": 0.5002157092094421, + "learning_rate": 9.49395009704345e-05, + "loss": 0.5087, + "step": 7490 + }, + { + "epoch": 2.8312570781426953, + "grad_norm": 0.517042338848114, + "learning_rate": 9.492136443684565e-05, + "loss": 0.5033, + "step": 7500 + }, + { + "epoch": 2.835032087580219, + "grad_norm": 0.5893968343734741, + "learning_rate": 9.49031971999278e-05, + "loss": 0.5025, + "step": 7510 + }, + { + "epoch": 2.8388070970177424, + "grad_norm": 0.47264131903648376, + "learning_rate": 9.488499927209806e-05, + "loss": 0.4947, + "step": 7520 + }, + { + "epoch": 2.842582106455266, + "grad_norm": 0.5126714706420898, + "learning_rate": 9.486677066579456e-05, + "loss": 0.4967, + "step": 7530 + }, + { + "epoch": 2.84635711589279, + "grad_norm": 0.5384365916252136, + "learning_rate": 9.484851139347639e-05, + "loss": 0.4961, + "step": 7540 + }, + { + "epoch": 2.8501321253303136, + "grad_norm": 0.5498970746994019, + "learning_rate": 9.483022146762358e-05, + "loss": 0.4979, + "step": 7550 + }, + { + "epoch": 2.853907134767837, + "grad_norm": 0.5383314490318298, + "learning_rate": 9.48119009007371e-05, + "loss": 0.4993, + "step": 7560 + }, + { + "epoch": 2.8576821442053606, + "grad_norm": 0.46135392785072327, + "learning_rate": 9.479354970533892e-05, + "loss": 0.4989, + "step": 7570 + }, + { + "epoch": 2.861457153642884, + "grad_norm": 0.5344179272651672, + "learning_rate": 9.477516789397189e-05, + "loss": 0.4975, + "step": 7580 + }, + { + "epoch": 2.8652321630804076, + "grad_norm": 0.47662675380706787, + "learning_rate": 9.475675547919983e-05, + "loss": 0.5002, + "step": 7590 + }, + { + "epoch": 2.8690071725179314, + "grad_norm": 0.5206663012504578, + "learning_rate": 9.47383124736074e-05, + "loss": 0.4969, + "step": 7600 + }, + { + "epoch": 2.872782181955455, + "grad_norm": 0.5600720643997192, + "learning_rate": 9.471983888980028e-05, + "loss": 0.4965, + "step": 7610 + }, + { + "epoch": 2.8765571913929784, + "grad_norm": 0.5266720056533813, + "learning_rate": 9.470133474040494e-05, + "loss": 0.5027, + "step": 7620 + }, + { + "epoch": 2.880332200830502, + "grad_norm": 2.0006062984466553, + "learning_rate": 9.468280003806882e-05, + "loss": 0.5084, + "step": 7630 + }, + { + "epoch": 2.8841072102680254, + "grad_norm": 0.5050598978996277, + "learning_rate": 9.46642347954602e-05, + "loss": 0.5037, + "step": 7640 + }, + { + "epoch": 2.887882219705549, + "grad_norm": 0.47348907589912415, + "learning_rate": 9.464563902526826e-05, + "loss": 0.494, + "step": 7650 + }, + { + "epoch": 2.891657229143073, + "grad_norm": 0.5870567560195923, + "learning_rate": 9.462701274020303e-05, + "loss": 0.4916, + "step": 7660 + }, + { + "epoch": 2.8954322385805966, + "grad_norm": 0.580295979976654, + "learning_rate": 9.46083559529954e-05, + "loss": 0.4945, + "step": 7670 + }, + { + "epoch": 2.89920724801812, + "grad_norm": 0.5126973390579224, + "learning_rate": 9.458966867639712e-05, + "loss": 0.4977, + "step": 7680 + }, + { + "epoch": 2.9029822574556436, + "grad_norm": 0.5632660984992981, + "learning_rate": 9.457095092318074e-05, + "loss": 0.5005, + "step": 7690 + }, + { + "epoch": 2.9067572668931674, + "grad_norm": 0.5285568833351135, + "learning_rate": 9.455220270613969e-05, + "loss": 0.4945, + "step": 7700 + }, + { + "epoch": 2.9105322763306907, + "grad_norm": 0.8250537514686584, + "learning_rate": 9.45334240380882e-05, + "loss": 0.49, + "step": 7710 + }, + { + "epoch": 2.9143072857682144, + "grad_norm": 0.5369483232498169, + "learning_rate": 9.451461493186129e-05, + "loss": 0.4984, + "step": 7720 + }, + { + "epoch": 2.918082295205738, + "grad_norm": 0.49857333302497864, + "learning_rate": 9.449577540031482e-05, + "loss": 0.4919, + "step": 7730 + }, + { + "epoch": 2.921857304643262, + "grad_norm": 0.5103145241737366, + "learning_rate": 9.447690545632544e-05, + "loss": 0.4908, + "step": 7740 + }, + { + "epoch": 2.925632314080785, + "grad_norm": 0.5372836589813232, + "learning_rate": 9.445800511279058e-05, + "loss": 0.496, + "step": 7750 + }, + { + "epoch": 2.929407323518309, + "grad_norm": 0.5468769669532776, + "learning_rate": 9.443907438262844e-05, + "loss": 0.5104, + "step": 7760 + }, + { + "epoch": 2.933182332955832, + "grad_norm": 0.6511932015419006, + "learning_rate": 9.442011327877798e-05, + "loss": 0.4951, + "step": 7770 + }, + { + "epoch": 2.936957342393356, + "grad_norm": 0.5281775593757629, + "learning_rate": 9.440112181419895e-05, + "loss": 0.498, + "step": 7780 + }, + { + "epoch": 2.9407323518308797, + "grad_norm": 0.6145192384719849, + "learning_rate": 9.438210000187186e-05, + "loss": 0.4998, + "step": 7790 + }, + { + "epoch": 2.9445073612684034, + "grad_norm": 0.5092405676841736, + "learning_rate": 9.43630478547979e-05, + "loss": 0.4879, + "step": 7800 + }, + { + "epoch": 2.9482823707059267, + "grad_norm": 0.5056966543197632, + "learning_rate": 9.434396538599906e-05, + "loss": 0.49, + "step": 7810 + }, + { + "epoch": 2.9520573801434504, + "grad_norm": 0.49092942476272583, + "learning_rate": 9.4324852608518e-05, + "loss": 0.4925, + "step": 7820 + }, + { + "epoch": 2.9558323895809737, + "grad_norm": 0.48439252376556396, + "learning_rate": 9.430570953541816e-05, + "loss": 0.4888, + "step": 7830 + }, + { + "epoch": 2.9596073990184975, + "grad_norm": 0.553947389125824, + "learning_rate": 9.428653617978362e-05, + "loss": 0.4942, + "step": 7840 + }, + { + "epoch": 2.963382408456021, + "grad_norm": 0.5610196590423584, + "learning_rate": 9.42673325547192e-05, + "loss": 0.4906, + "step": 7850 + }, + { + "epoch": 2.967157417893545, + "grad_norm": 0.6619260907173157, + "learning_rate": 9.42480986733504e-05, + "loss": 0.4923, + "step": 7860 + }, + { + "epoch": 2.970932427331068, + "grad_norm": 0.5648078918457031, + "learning_rate": 9.422883454882338e-05, + "loss": 0.4951, + "step": 7870 + }, + { + "epoch": 2.974707436768592, + "grad_norm": 0.505842387676239, + "learning_rate": 9.4209540194305e-05, + "loss": 0.5008, + "step": 7880 + }, + { + "epoch": 2.9784824462061152, + "grad_norm": 0.5517747402191162, + "learning_rate": 9.419021562298278e-05, + "loss": 0.4977, + "step": 7890 + }, + { + "epoch": 2.982257455643639, + "grad_norm": 0.5466206073760986, + "learning_rate": 9.417086084806486e-05, + "loss": 0.5011, + "step": 7900 + }, + { + "epoch": 2.9860324650811627, + "grad_norm": 0.5874969363212585, + "learning_rate": 9.415147588278005e-05, + "loss": 0.4834, + "step": 7910 + }, + { + "epoch": 2.9898074745186864, + "grad_norm": 0.6627646684646606, + "learning_rate": 9.413206074037781e-05, + "loss": 0.4895, + "step": 7920 + }, + { + "epoch": 2.9935824839562097, + "grad_norm": 0.5905627608299255, + "learning_rate": 9.411261543412819e-05, + "loss": 0.4971, + "step": 7930 + }, + { + "epoch": 2.9973574933937335, + "grad_norm": 0.5630371570587158, + "learning_rate": 9.409313997732188e-05, + "loss": 0.4933, + "step": 7940 + }, + { + "epoch": 3.001132502831257, + "grad_norm": 0.5096926093101501, + "learning_rate": 9.407363438327018e-05, + "loss": 0.5077, + "step": 7950 + }, + { + "epoch": 3.0049075122687805, + "grad_norm": 0.545479416847229, + "learning_rate": 9.405409866530496e-05, + "loss": 0.4925, + "step": 7960 + }, + { + "epoch": 3.0086825217063042, + "grad_norm": 0.4964505434036255, + "learning_rate": 9.40345328367787e-05, + "loss": 0.4872, + "step": 7970 + }, + { + "epoch": 3.012457531143828, + "grad_norm": 0.5932561159133911, + "learning_rate": 9.401493691106446e-05, + "loss": 0.4889, + "step": 7980 + }, + { + "epoch": 3.0162325405813513, + "grad_norm": 0.4858354926109314, + "learning_rate": 9.39953109015559e-05, + "loss": 0.4962, + "step": 7990 + }, + { + "epoch": 3.020007550018875, + "grad_norm": 0.4681301712989807, + "learning_rate": 9.397565482166718e-05, + "loss": 0.493, + "step": 8000 + }, + { + "epoch": 3.0237825594563987, + "grad_norm": 0.5294594168663025, + "learning_rate": 9.395596868483305e-05, + "loss": 0.4941, + "step": 8010 + }, + { + "epoch": 3.027557568893922, + "grad_norm": 0.5004336833953857, + "learning_rate": 9.393625250450881e-05, + "loss": 0.4886, + "step": 8020 + }, + { + "epoch": 3.0313325783314458, + "grad_norm": 0.5321143269538879, + "learning_rate": 9.391650629417028e-05, + "loss": 0.497, + "step": 8030 + }, + { + "epoch": 3.0351075877689695, + "grad_norm": 0.6064821481704712, + "learning_rate": 9.38967300673138e-05, + "loss": 0.4979, + "step": 8040 + }, + { + "epoch": 3.0388825972064932, + "grad_norm": 0.5963532328605652, + "learning_rate": 9.387692383745624e-05, + "loss": 0.4883, + "step": 8050 + }, + { + "epoch": 3.0426576066440165, + "grad_norm": 0.510930597782135, + "learning_rate": 9.385708761813496e-05, + "loss": 0.4906, + "step": 8060 + }, + { + "epoch": 3.0464326160815403, + "grad_norm": 0.627373456954956, + "learning_rate": 9.383722142290787e-05, + "loss": 0.4881, + "step": 8070 + }, + { + "epoch": 3.050207625519064, + "grad_norm": 0.5580819249153137, + "learning_rate": 9.381732526535328e-05, + "loss": 0.4941, + "step": 8080 + }, + { + "epoch": 3.0539826349565873, + "grad_norm": 0.4533594250679016, + "learning_rate": 9.379739915907004e-05, + "loss": 0.4924, + "step": 8090 + }, + { + "epoch": 3.057757644394111, + "grad_norm": 0.47666409611701965, + "learning_rate": 9.377744311767746e-05, + "loss": 0.4957, + "step": 8100 + }, + { + "epoch": 3.0615326538316348, + "grad_norm": 0.5064053535461426, + "learning_rate": 9.375745715481532e-05, + "loss": 0.4978, + "step": 8110 + }, + { + "epoch": 3.065307663269158, + "grad_norm": 0.5491259694099426, + "learning_rate": 9.373744128414382e-05, + "loss": 0.4879, + "step": 8120 + }, + { + "epoch": 3.069082672706682, + "grad_norm": 0.5816760063171387, + "learning_rate": 9.371739551934359e-05, + "loss": 0.4947, + "step": 8130 + }, + { + "epoch": 3.0728576821442055, + "grad_norm": 0.63339763879776, + "learning_rate": 9.36973198741158e-05, + "loss": 0.4921, + "step": 8140 + }, + { + "epoch": 3.076632691581729, + "grad_norm": 0.48095303773880005, + "learning_rate": 9.367721436218189e-05, + "loss": 0.5039, + "step": 8150 + }, + { + "epoch": 3.0804077010192525, + "grad_norm": 1.313503623008728, + "learning_rate": 9.365707899728383e-05, + "loss": 0.4874, + "step": 8160 + }, + { + "epoch": 3.0841827104567763, + "grad_norm": 0.5337556004524231, + "learning_rate": 9.363691379318391e-05, + "loss": 0.4949, + "step": 8170 + }, + { + "epoch": 3.0879577198942996, + "grad_norm": 0.5201692581176758, + "learning_rate": 9.361671876366491e-05, + "loss": 0.4903, + "step": 8180 + }, + { + "epoch": 3.0917327293318233, + "grad_norm": 0.48047223687171936, + "learning_rate": 9.35964939225299e-05, + "loss": 0.4895, + "step": 8190 + }, + { + "epoch": 3.095507738769347, + "grad_norm": 0.5520181655883789, + "learning_rate": 9.35762392836024e-05, + "loss": 0.5046, + "step": 8200 + }, + { + "epoch": 3.0992827482068703, + "grad_norm": 0.5537580847740173, + "learning_rate": 9.355595486072624e-05, + "loss": 0.4904, + "step": 8210 + }, + { + "epoch": 3.103057757644394, + "grad_norm": 0.5307925343513489, + "learning_rate": 9.353564066776563e-05, + "loss": 0.4851, + "step": 8220 + }, + { + "epoch": 3.106832767081918, + "grad_norm": 0.5735664367675781, + "learning_rate": 9.351529671860516e-05, + "loss": 0.4852, + "step": 8230 + }, + { + "epoch": 3.110607776519441, + "grad_norm": 0.553432047367096, + "learning_rate": 9.349492302714969e-05, + "loss": 0.5036, + "step": 8240 + }, + { + "epoch": 3.114382785956965, + "grad_norm": 0.6025165319442749, + "learning_rate": 9.347451960732447e-05, + "loss": 0.4889, + "step": 8250 + }, + { + "epoch": 3.1181577953944886, + "grad_norm": 10.570920944213867, + "learning_rate": 9.345408647307506e-05, + "loss": 0.5668, + "step": 8260 + }, + { + "epoch": 3.121932804832012, + "grad_norm": 0.6134909391403198, + "learning_rate": 9.343362363836726e-05, + "loss": 0.488, + "step": 8270 + }, + { + "epoch": 3.1257078142695356, + "grad_norm": 0.5113944411277771, + "learning_rate": 9.341313111718728e-05, + "loss": 0.4943, + "step": 8280 + }, + { + "epoch": 3.1294828237070593, + "grad_norm": 0.586234986782074, + "learning_rate": 9.339260892354153e-05, + "loss": 0.4888, + "step": 8290 + }, + { + "epoch": 3.133257833144583, + "grad_norm": 0.5822952389717102, + "learning_rate": 9.337205707145676e-05, + "loss": 0.5001, + "step": 8300 + }, + { + "epoch": 3.1370328425821064, + "grad_norm": 0.6024602651596069, + "learning_rate": 9.335147557497994e-05, + "loss": 0.4862, + "step": 8310 + }, + { + "epoch": 3.14080785201963, + "grad_norm": 0.5428383946418762, + "learning_rate": 9.333086444817835e-05, + "loss": 0.4889, + "step": 8320 + }, + { + "epoch": 3.144582861457154, + "grad_norm": 0.5743405818939209, + "learning_rate": 9.331022370513951e-05, + "loss": 0.4873, + "step": 8330 + }, + { + "epoch": 3.148357870894677, + "grad_norm": 0.497543066740036, + "learning_rate": 9.328955335997113e-05, + "loss": 0.4962, + "step": 8340 + }, + { + "epoch": 3.152132880332201, + "grad_norm": 0.4729820489883423, + "learning_rate": 9.326885342680125e-05, + "loss": 0.4873, + "step": 8350 + }, + { + "epoch": 3.1559078897697246, + "grad_norm": 0.5927625298500061, + "learning_rate": 9.324812391977806e-05, + "loss": 0.4924, + "step": 8360 + }, + { + "epoch": 3.159682899207248, + "grad_norm": 0.5837076902389526, + "learning_rate": 9.322736485306995e-05, + "loss": 0.4886, + "step": 8370 + }, + { + "epoch": 3.1634579086447716, + "grad_norm": 0.5531527996063232, + "learning_rate": 9.32065762408656e-05, + "loss": 0.4897, + "step": 8380 + }, + { + "epoch": 3.1672329180822953, + "grad_norm": 0.5851390361785889, + "learning_rate": 9.318575809737377e-05, + "loss": 0.4864, + "step": 8390 + }, + { + "epoch": 3.1710079275198186, + "grad_norm": 0.5682461261749268, + "learning_rate": 9.316491043682351e-05, + "loss": 0.4849, + "step": 8400 + }, + { + "epoch": 3.1747829369573424, + "grad_norm": 0.623104989528656, + "learning_rate": 9.3144033273464e-05, + "loss": 0.5003, + "step": 8410 + }, + { + "epoch": 3.178557946394866, + "grad_norm": 0.6800477504730225, + "learning_rate": 9.312312662156457e-05, + "loss": 0.4887, + "step": 8420 + }, + { + "epoch": 3.1823329558323894, + "grad_norm": 0.5521610975265503, + "learning_rate": 9.310219049541471e-05, + "loss": 0.5015, + "step": 8430 + }, + { + "epoch": 3.186107965269913, + "grad_norm": 0.619483232498169, + "learning_rate": 9.308122490932409e-05, + "loss": 0.4849, + "step": 8440 + }, + { + "epoch": 3.189882974707437, + "grad_norm": 0.5283095240592957, + "learning_rate": 9.306022987762246e-05, + "loss": 0.4877, + "step": 8450 + }, + { + "epoch": 3.19365798414496, + "grad_norm": 0.569065511226654, + "learning_rate": 9.303920541465974e-05, + "loss": 0.4844, + "step": 8460 + }, + { + "epoch": 3.197432993582484, + "grad_norm": 0.4644882082939148, + "learning_rate": 9.301815153480598e-05, + "loss": 0.4875, + "step": 8470 + }, + { + "epoch": 3.2012080030200076, + "grad_norm": 0.5540872812271118, + "learning_rate": 9.299706825245126e-05, + "loss": 0.4936, + "step": 8480 + }, + { + "epoch": 3.2049830124575314, + "grad_norm": 0.5254048705101013, + "learning_rate": 9.297595558200583e-05, + "loss": 0.4812, + "step": 8490 + }, + { + "epoch": 3.2087580218950547, + "grad_norm": 0.4865260422229767, + "learning_rate": 9.295481353790001e-05, + "loss": 0.4874, + "step": 8500 + }, + { + "epoch": 3.2125330313325784, + "grad_norm": 0.6707580089569092, + "learning_rate": 9.293364213458417e-05, + "loss": 0.4893, + "step": 8510 + }, + { + "epoch": 3.216308040770102, + "grad_norm": 0.5525599718093872, + "learning_rate": 9.291244138652877e-05, + "loss": 0.4847, + "step": 8520 + }, + { + "epoch": 3.2200830502076254, + "grad_norm": 0.5599222183227539, + "learning_rate": 9.289121130822431e-05, + "loss": 0.4823, + "step": 8530 + }, + { + "epoch": 3.223858059645149, + "grad_norm": 0.6073276996612549, + "learning_rate": 9.286995191418137e-05, + "loss": 0.481, + "step": 8540 + }, + { + "epoch": 3.227633069082673, + "grad_norm": 0.5141738057136536, + "learning_rate": 9.284866321893056e-05, + "loss": 0.4807, + "step": 8550 + }, + { + "epoch": 3.231408078520196, + "grad_norm": 0.5618724822998047, + "learning_rate": 9.282734523702244e-05, + "loss": 0.4911, + "step": 8560 + }, + { + "epoch": 3.23518308795772, + "grad_norm": 0.5364367961883545, + "learning_rate": 9.280599798302771e-05, + "loss": 0.4817, + "step": 8570 + }, + { + "epoch": 3.2389580973952437, + "grad_norm": 0.5237820744514465, + "learning_rate": 9.278462147153699e-05, + "loss": 0.4798, + "step": 8580 + }, + { + "epoch": 3.242733106832767, + "grad_norm": 0.5670002698898315, + "learning_rate": 9.276321571716094e-05, + "loss": 0.4839, + "step": 8590 + }, + { + "epoch": 3.2465081162702907, + "grad_norm": 0.5286555290222168, + "learning_rate": 9.274178073453017e-05, + "loss": 0.4827, + "step": 8600 + }, + { + "epoch": 3.2502831257078144, + "grad_norm": 0.7837060689926147, + "learning_rate": 9.27203165382953e-05, + "loss": 0.4848, + "step": 8610 + }, + { + "epoch": 3.2540581351453377, + "grad_norm": 0.5733458995819092, + "learning_rate": 9.26988231431269e-05, + "loss": 0.4874, + "step": 8620 + }, + { + "epoch": 3.2578331445828614, + "grad_norm": 0.5995279550552368, + "learning_rate": 9.267730056371551e-05, + "loss": 0.4856, + "step": 8630 + }, + { + "epoch": 3.261608154020385, + "grad_norm": 0.585941731929779, + "learning_rate": 9.265574881477161e-05, + "loss": 0.495, + "step": 8640 + }, + { + "epoch": 3.2653831634579085, + "grad_norm": 0.592987596988678, + "learning_rate": 9.263416791102561e-05, + "loss": 0.4823, + "step": 8650 + }, + { + "epoch": 3.269158172895432, + "grad_norm": 0.47990015149116516, + "learning_rate": 9.261255786722786e-05, + "loss": 0.487, + "step": 8660 + }, + { + "epoch": 3.272933182332956, + "grad_norm": 0.49177876114845276, + "learning_rate": 9.259091869814864e-05, + "loss": 0.4745, + "step": 8670 + }, + { + "epoch": 3.2767081917704797, + "grad_norm": 0.5345214009284973, + "learning_rate": 9.256925041857807e-05, + "loss": 0.4883, + "step": 8680 + }, + { + "epoch": 3.280483201208003, + "grad_norm": 0.5280712842941284, + "learning_rate": 9.254755304332626e-05, + "loss": 0.4837, + "step": 8690 + }, + { + "epoch": 3.2842582106455267, + "grad_norm": 0.5407823920249939, + "learning_rate": 9.252582658722313e-05, + "loss": 0.4824, + "step": 8700 + }, + { + "epoch": 3.28803322008305, + "grad_norm": 0.510508120059967, + "learning_rate": 9.250407106511853e-05, + "loss": 0.4763, + "step": 8710 + }, + { + "epoch": 3.2918082295205737, + "grad_norm": 0.5562644004821777, + "learning_rate": 9.248228649188215e-05, + "loss": 0.4756, + "step": 8720 + }, + { + "epoch": 3.2955832389580975, + "grad_norm": 0.5866154432296753, + "learning_rate": 9.246047288240354e-05, + "loss": 0.4792, + "step": 8730 + }, + { + "epoch": 3.299358248395621, + "grad_norm": 0.5227904319763184, + "learning_rate": 9.243863025159208e-05, + "loss": 0.48, + "step": 8740 + }, + { + "epoch": 3.3031332578331445, + "grad_norm": 0.6132093667984009, + "learning_rate": 9.241675861437704e-05, + "loss": 0.4762, + "step": 8750 + }, + { + "epoch": 3.3069082672706682, + "grad_norm": 0.5366108417510986, + "learning_rate": 9.239485798570742e-05, + "loss": 0.4813, + "step": 8760 + }, + { + "epoch": 3.310683276708192, + "grad_norm": 0.5920717716217041, + "learning_rate": 9.237292838055215e-05, + "loss": 0.4834, + "step": 8770 + }, + { + "epoch": 3.3144582861457152, + "grad_norm": 0.5619731545448303, + "learning_rate": 9.235096981389988e-05, + "loss": 0.4846, + "step": 8780 + }, + { + "epoch": 3.318233295583239, + "grad_norm": 0.45175281167030334, + "learning_rate": 9.232898230075909e-05, + "loss": 0.4807, + "step": 8790 + }, + { + "epoch": 3.3220083050207627, + "grad_norm": 0.613134503364563, + "learning_rate": 9.230696585615801e-05, + "loss": 0.4724, + "step": 8800 + }, + { + "epoch": 3.325783314458286, + "grad_norm": 0.5235288739204407, + "learning_rate": 9.228492049514473e-05, + "loss": 0.4767, + "step": 8810 + }, + { + "epoch": 3.3295583238958097, + "grad_norm": 0.5533393025398254, + "learning_rate": 9.2262846232787e-05, + "loss": 0.4836, + "step": 8820 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 0.5668758749961853, + "learning_rate": 9.224074308417238e-05, + "loss": 0.485, + "step": 8830 + }, + { + "epoch": 3.3371083427708568, + "grad_norm": 0.568915069103241, + "learning_rate": 9.221861106440818e-05, + "loss": 0.4767, + "step": 8840 + }, + { + "epoch": 3.3408833522083805, + "grad_norm": 0.6423974633216858, + "learning_rate": 9.219645018862142e-05, + "loss": 0.4766, + "step": 8850 + }, + { + "epoch": 3.3446583616459042, + "grad_norm": 0.8083536028862, + "learning_rate": 9.217426047195882e-05, + "loss": 0.4734, + "step": 8860 + }, + { + "epoch": 3.3484333710834275, + "grad_norm": 0.5543457865715027, + "learning_rate": 9.21520419295869e-05, + "loss": 0.4806, + "step": 8870 + }, + { + "epoch": 3.3522083805209513, + "grad_norm": 0.5718511939048767, + "learning_rate": 9.212979457669178e-05, + "loss": 0.4761, + "step": 8880 + }, + { + "epoch": 3.355983389958475, + "grad_norm": 0.6094539165496826, + "learning_rate": 9.210751842847931e-05, + "loss": 0.4789, + "step": 8890 + }, + { + "epoch": 3.3597583993959983, + "grad_norm": 0.5675216913223267, + "learning_rate": 9.208521350017506e-05, + "loss": 0.4796, + "step": 8900 + }, + { + "epoch": 3.363533408833522, + "grad_norm": 0.5290115475654602, + "learning_rate": 9.206287980702422e-05, + "loss": 0.477, + "step": 8910 + }, + { + "epoch": 3.3673084182710458, + "grad_norm": 0.5056905150413513, + "learning_rate": 9.204051736429166e-05, + "loss": 0.4738, + "step": 8920 + }, + { + "epoch": 3.3710834277085695, + "grad_norm": 0.8471300601959229, + "learning_rate": 9.20181261872619e-05, + "loss": 0.4825, + "step": 8930 + }, + { + "epoch": 3.374858437146093, + "grad_norm": 0.516025185585022, + "learning_rate": 9.19957062912391e-05, + "loss": 0.4708, + "step": 8940 + }, + { + "epoch": 3.3786334465836165, + "grad_norm": 0.5982151627540588, + "learning_rate": 9.197325769154704e-05, + "loss": 0.4698, + "step": 8950 + }, + { + "epoch": 3.38240845602114, + "grad_norm": 0.6391962766647339, + "learning_rate": 9.195078040352914e-05, + "loss": 0.477, + "step": 8960 + }, + { + "epoch": 3.3861834654586636, + "grad_norm": 0.6175360679626465, + "learning_rate": 9.192827444254839e-05, + "loss": 0.4771, + "step": 8970 + }, + { + "epoch": 3.3899584748961873, + "grad_norm": 0.5909090042114258, + "learning_rate": 9.190573982398744e-05, + "loss": 0.4789, + "step": 8980 + }, + { + "epoch": 3.393733484333711, + "grad_norm": 0.6406884789466858, + "learning_rate": 9.188317656324845e-05, + "loss": 0.4735, + "step": 8990 + }, + { + "epoch": 3.3975084937712343, + "grad_norm": 0.509844183921814, + "learning_rate": 9.186058467575322e-05, + "loss": 0.4786, + "step": 9000 + }, + { + "epoch": 3.401283503208758, + "grad_norm": 0.524660587310791, + "learning_rate": 9.183796417694309e-05, + "loss": 0.4789, + "step": 9010 + }, + { + "epoch": 3.405058512646282, + "grad_norm": 0.5184488892555237, + "learning_rate": 9.181531508227897e-05, + "loss": 0.4794, + "step": 9020 + }, + { + "epoch": 3.408833522083805, + "grad_norm": 0.4693617522716522, + "learning_rate": 9.179263740724125e-05, + "loss": 0.483, + "step": 9030 + }, + { + "epoch": 3.412608531521329, + "grad_norm": 0.5032320618629456, + "learning_rate": 9.176993116732996e-05, + "loss": 0.4748, + "step": 9040 + }, + { + "epoch": 3.4163835409588525, + "grad_norm": 0.5809365510940552, + "learning_rate": 9.17471963780646e-05, + "loss": 0.4671, + "step": 9050 + }, + { + "epoch": 3.420158550396376, + "grad_norm": 0.547820508480072, + "learning_rate": 9.172443305498414e-05, + "loss": 0.4729, + "step": 9060 + }, + { + "epoch": 3.4239335598338996, + "grad_norm": 0.48663732409477234, + "learning_rate": 9.170164121364714e-05, + "loss": 0.475, + "step": 9070 + }, + { + "epoch": 3.4277085692714233, + "grad_norm": 0.6125722527503967, + "learning_rate": 9.167882086963158e-05, + "loss": 0.4802, + "step": 9080 + }, + { + "epoch": 3.4314835787089466, + "grad_norm": 0.5475627183914185, + "learning_rate": 9.165597203853499e-05, + "loss": 0.4822, + "step": 9090 + }, + { + "epoch": 3.4352585881464703, + "grad_norm": 1.4172425270080566, + "learning_rate": 9.163309473597428e-05, + "loss": 0.4796, + "step": 9100 + }, + { + "epoch": 3.439033597583994, + "grad_norm": 0.5351138114929199, + "learning_rate": 9.161018897758591e-05, + "loss": 0.4766, + "step": 9110 + }, + { + "epoch": 3.4428086070215174, + "grad_norm": 0.6007091999053955, + "learning_rate": 9.158725477902572e-05, + "loss": 0.4755, + "step": 9120 + }, + { + "epoch": 3.446583616459041, + "grad_norm": 0.5067334175109863, + "learning_rate": 9.156429215596905e-05, + "loss": 0.4923, + "step": 9130 + }, + { + "epoch": 3.450358625896565, + "grad_norm": 0.7098965644836426, + "learning_rate": 9.154130112411061e-05, + "loss": 0.4741, + "step": 9140 + }, + { + "epoch": 3.454133635334088, + "grad_norm": 0.555362343788147, + "learning_rate": 9.151828169916455e-05, + "loss": 0.4741, + "step": 9150 + }, + { + "epoch": 3.457908644771612, + "grad_norm": 0.5457745790481567, + "learning_rate": 9.149523389686445e-05, + "loss": 0.4723, + "step": 9160 + }, + { + "epoch": 3.4616836542091356, + "grad_norm": 0.5627350807189941, + "learning_rate": 9.147215773296327e-05, + "loss": 0.4732, + "step": 9170 + }, + { + "epoch": 3.4654586636466593, + "grad_norm": 0.5835473537445068, + "learning_rate": 9.144905322323334e-05, + "loss": 0.4693, + "step": 9180 + }, + { + "epoch": 3.4692336730841826, + "grad_norm": 0.5265465974807739, + "learning_rate": 9.142592038346635e-05, + "loss": 0.477, + "step": 9190 + }, + { + "epoch": 3.4730086825217064, + "grad_norm": 0.5562682747840881, + "learning_rate": 9.140275922947343e-05, + "loss": 0.4757, + "step": 9200 + }, + { + "epoch": 3.4767836919592296, + "grad_norm": 0.6338258981704712, + "learning_rate": 9.137956977708496e-05, + "loss": 0.4843, + "step": 9210 + }, + { + "epoch": 3.4805587013967534, + "grad_norm": 0.5517134666442871, + "learning_rate": 9.135635204215075e-05, + "loss": 0.47, + "step": 9220 + }, + { + "epoch": 3.484333710834277, + "grad_norm": 0.6941575407981873, + "learning_rate": 9.133310604053986e-05, + "loss": 0.4773, + "step": 9230 + }, + { + "epoch": 3.488108720271801, + "grad_norm": 0.532227098941803, + "learning_rate": 9.130983178814077e-05, + "loss": 0.4677, + "step": 9240 + }, + { + "epoch": 3.491883729709324, + "grad_norm": 0.5691418051719666, + "learning_rate": 9.128652930086116e-05, + "loss": 0.476, + "step": 9250 + }, + { + "epoch": 3.495658739146848, + "grad_norm": 0.5184800028800964, + "learning_rate": 9.126319859462808e-05, + "loss": 0.4822, + "step": 9260 + }, + { + "epoch": 3.4994337485843716, + "grad_norm": 0.55794358253479, + "learning_rate": 9.123983968538787e-05, + "loss": 0.4786, + "step": 9270 + }, + { + "epoch": 3.503208758021895, + "grad_norm": 0.5698232650756836, + "learning_rate": 9.121645258910609e-05, + "loss": 0.4716, + "step": 9280 + }, + { + "epoch": 3.5069837674594186, + "grad_norm": 0.53619784116745, + "learning_rate": 9.119303732176762e-05, + "loss": 0.4761, + "step": 9290 + }, + { + "epoch": 3.5107587768969424, + "grad_norm": 0.5619392395019531, + "learning_rate": 9.116959389937656e-05, + "loss": 0.4741, + "step": 9300 + }, + { + "epoch": 3.5145337863344657, + "grad_norm": 0.5324605107307434, + "learning_rate": 9.114612233795628e-05, + "loss": 0.4683, + "step": 9310 + }, + { + "epoch": 3.5183087957719894, + "grad_norm": 0.5709056258201599, + "learning_rate": 9.112262265354937e-05, + "loss": 0.4778, + "step": 9320 + }, + { + "epoch": 3.522083805209513, + "grad_norm": 0.5106092691421509, + "learning_rate": 9.109909486221763e-05, + "loss": 0.4673, + "step": 9330 + }, + { + "epoch": 3.5258588146470364, + "grad_norm": 0.5639387369155884, + "learning_rate": 9.107553898004208e-05, + "loss": 0.4741, + "step": 9340 + }, + { + "epoch": 3.52963382408456, + "grad_norm": 0.5696660280227661, + "learning_rate": 9.105195502312298e-05, + "loss": 0.4716, + "step": 9350 + }, + { + "epoch": 3.533408833522084, + "grad_norm": 0.5708215236663818, + "learning_rate": 9.102834300757972e-05, + "loss": 0.4703, + "step": 9360 + }, + { + "epoch": 3.5371838429596076, + "grad_norm": 0.5708835124969482, + "learning_rate": 9.100470294955087e-05, + "loss": 0.4707, + "step": 9370 + }, + { + "epoch": 3.540958852397131, + "grad_norm": 0.49455785751342773, + "learning_rate": 9.098103486519423e-05, + "loss": 0.4693, + "step": 9380 + }, + { + "epoch": 3.5447338618346547, + "grad_norm": 0.5507346987724304, + "learning_rate": 9.095733877068666e-05, + "loss": 0.468, + "step": 9390 + }, + { + "epoch": 3.548508871272178, + "grad_norm": 0.5424711108207703, + "learning_rate": 9.093361468222426e-05, + "loss": 0.4784, + "step": 9400 + }, + { + "epoch": 3.5522838807097017, + "grad_norm": 0.5007194876670837, + "learning_rate": 9.090986261602221e-05, + "loss": 0.4659, + "step": 9410 + }, + { + "epoch": 3.5560588901472254, + "grad_norm": 0.6124007701873779, + "learning_rate": 9.088608258831482e-05, + "loss": 0.4768, + "step": 9420 + }, + { + "epoch": 3.559833899584749, + "grad_norm": 0.5798971056938171, + "learning_rate": 9.08622746153555e-05, + "loss": 0.4814, + "step": 9430 + }, + { + "epoch": 3.5636089090222725, + "grad_norm": 0.5349039435386658, + "learning_rate": 9.083843871341679e-05, + "loss": 0.4676, + "step": 9440 + }, + { + "epoch": 3.567383918459796, + "grad_norm": 0.557998776435852, + "learning_rate": 9.08145748987903e-05, + "loss": 0.4758, + "step": 9450 + }, + { + "epoch": 3.5711589278973195, + "grad_norm": 0.5525006055831909, + "learning_rate": 9.079068318778674e-05, + "loss": 0.4692, + "step": 9460 + }, + { + "epoch": 3.574933937334843, + "grad_norm": 0.6234112977981567, + "learning_rate": 9.076676359673586e-05, + "loss": 0.4658, + "step": 9470 + }, + { + "epoch": 3.578708946772367, + "grad_norm": 0.6095515489578247, + "learning_rate": 9.074281614198645e-05, + "loss": 0.4685, + "step": 9480 + }, + { + "epoch": 3.5824839562098907, + "grad_norm": 0.5489786267280579, + "learning_rate": 9.071884083990638e-05, + "loss": 0.4732, + "step": 9490 + }, + { + "epoch": 3.586258965647414, + "grad_norm": 0.5583869814872742, + "learning_rate": 9.069483770688259e-05, + "loss": 0.4704, + "step": 9500 + }, + { + "epoch": 3.5900339750849377, + "grad_norm": 0.5421421527862549, + "learning_rate": 9.067080675932093e-05, + "loss": 0.4735, + "step": 9510 + }, + { + "epoch": 3.5938089845224614, + "grad_norm": 0.6016372442245483, + "learning_rate": 9.064674801364634e-05, + "loss": 0.4794, + "step": 9520 + }, + { + "epoch": 3.5975839939599847, + "grad_norm": 0.5686151385307312, + "learning_rate": 9.062266148630279e-05, + "loss": 0.4758, + "step": 9530 + }, + { + "epoch": 3.6013590033975085, + "grad_norm": 0.5975391864776611, + "learning_rate": 9.059854719375313e-05, + "loss": 0.4748, + "step": 9540 + }, + { + "epoch": 3.605134012835032, + "grad_norm": 0.5515162944793701, + "learning_rate": 9.057440515247932e-05, + "loss": 0.4738, + "step": 9550 + }, + { + "epoch": 3.608909022272556, + "grad_norm": 0.5409767627716064, + "learning_rate": 9.055023537898215e-05, + "loss": 0.4786, + "step": 9560 + }, + { + "epoch": 3.6126840317100792, + "grad_norm": 0.4877788722515106, + "learning_rate": 9.052603788978149e-05, + "loss": 0.4849, + "step": 9570 + }, + { + "epoch": 3.616459041147603, + "grad_norm": 0.6089911460876465, + "learning_rate": 9.050181270141606e-05, + "loss": 0.4728, + "step": 9580 + }, + { + "epoch": 3.6202340505851263, + "grad_norm": 0.64964359998703, + "learning_rate": 9.047755983044355e-05, + "loss": 0.4684, + "step": 9590 + }, + { + "epoch": 3.62400906002265, + "grad_norm": 0.5478320121765137, + "learning_rate": 9.045327929344058e-05, + "loss": 0.4765, + "step": 9600 + }, + { + "epoch": 3.6277840694601737, + "grad_norm": 0.5902840495109558, + "learning_rate": 9.042897110700268e-05, + "loss": 0.4705, + "step": 9610 + }, + { + "epoch": 3.6315590788976975, + "grad_norm": 0.5538387894630432, + "learning_rate": 9.040463528774423e-05, + "loss": 0.473, + "step": 9620 + }, + { + "epoch": 3.6353340883352208, + "grad_norm": 0.5468869209289551, + "learning_rate": 9.038027185229856e-05, + "loss": 0.4745, + "step": 9630 + }, + { + "epoch": 3.6391090977727445, + "grad_norm": 0.6634562015533447, + "learning_rate": 9.035588081731784e-05, + "loss": 0.4749, + "step": 9640 + }, + { + "epoch": 3.642884107210268, + "grad_norm": 0.5892646908760071, + "learning_rate": 9.033146219947313e-05, + "loss": 0.4674, + "step": 9650 + }, + { + "epoch": 3.6466591166477915, + "grad_norm": 0.5538533926010132, + "learning_rate": 9.030701601545429e-05, + "loss": 0.4665, + "step": 9660 + }, + { + "epoch": 3.6504341260853153, + "grad_norm": 0.5980792045593262, + "learning_rate": 9.028254228197009e-05, + "loss": 0.4635, + "step": 9670 + }, + { + "epoch": 3.654209135522839, + "grad_norm": 0.5423790216445923, + "learning_rate": 9.025804101574808e-05, + "loss": 0.4637, + "step": 9680 + }, + { + "epoch": 3.6579841449603623, + "grad_norm": 0.5625879168510437, + "learning_rate": 9.023351223353463e-05, + "loss": 0.4744, + "step": 9690 + }, + { + "epoch": 3.661759154397886, + "grad_norm": 0.5283533930778503, + "learning_rate": 9.020895595209498e-05, + "loss": 0.4697, + "step": 9700 + }, + { + "epoch": 3.6655341638354098, + "grad_norm": 0.569664478302002, + "learning_rate": 9.018437218821305e-05, + "loss": 0.4591, + "step": 9710 + }, + { + "epoch": 3.669309173272933, + "grad_norm": 0.5049920678138733, + "learning_rate": 9.015976095869166e-05, + "loss": 0.4713, + "step": 9720 + }, + { + "epoch": 3.673084182710457, + "grad_norm": 0.5221184492111206, + "learning_rate": 9.013512228035235e-05, + "loss": 0.4634, + "step": 9730 + }, + { + "epoch": 3.6768591921479805, + "grad_norm": 0.6969385743141174, + "learning_rate": 9.01104561700354e-05, + "loss": 0.4678, + "step": 9740 + }, + { + "epoch": 3.680634201585504, + "grad_norm": 0.6279971599578857, + "learning_rate": 9.008576264459987e-05, + "loss": 0.4684, + "step": 9750 + }, + { + "epoch": 3.6844092110230275, + "grad_norm": 0.7082238793373108, + "learning_rate": 9.006104172092357e-05, + "loss": 0.4734, + "step": 9760 + }, + { + "epoch": 3.6881842204605513, + "grad_norm": 0.5783301591873169, + "learning_rate": 9.0036293415903e-05, + "loss": 0.494, + "step": 9770 + }, + { + "epoch": 3.6919592298980746, + "grad_norm": 0.5902776122093201, + "learning_rate": 9.00115177464534e-05, + "loss": 0.5231, + "step": 9780 + }, + { + "epoch": 3.6957342393355983, + "grad_norm": 0.6361207365989685, + "learning_rate": 8.998671472950872e-05, + "loss": 0.4692, + "step": 9790 + }, + { + "epoch": 3.699509248773122, + "grad_norm": 0.5117760300636292, + "learning_rate": 8.996188438202156e-05, + "loss": 0.4679, + "step": 9800 + }, + { + "epoch": 3.7032842582106458, + "grad_norm": 0.6711943745613098, + "learning_rate": 8.993702672096324e-05, + "loss": 0.4669, + "step": 9810 + }, + { + "epoch": 3.707059267648169, + "grad_norm": 0.5595587491989136, + "learning_rate": 8.991214176332377e-05, + "loss": 0.4706, + "step": 9820 + }, + { + "epoch": 3.710834277085693, + "grad_norm": 0.5421229004859924, + "learning_rate": 8.988722952611171e-05, + "loss": 0.47, + "step": 9830 + }, + { + "epoch": 3.714609286523216, + "grad_norm": 0.5640469789505005, + "learning_rate": 8.986229002635441e-05, + "loss": 0.4672, + "step": 9840 + }, + { + "epoch": 3.71838429596074, + "grad_norm": 0.5714800357818604, + "learning_rate": 8.983732328109775e-05, + "loss": 0.4649, + "step": 9850 + }, + { + "epoch": 3.7221593053982636, + "grad_norm": 0.5369188189506531, + "learning_rate": 8.981232930740627e-05, + "loss": 0.4677, + "step": 9860 + }, + { + "epoch": 3.7259343148357873, + "grad_norm": 0.8434441089630127, + "learning_rate": 8.978730812236313e-05, + "loss": 0.4664, + "step": 9870 + }, + { + "epoch": 3.7297093242733106, + "grad_norm": 0.5156511664390564, + "learning_rate": 8.976225974307004e-05, + "loss": 0.4658, + "step": 9880 + }, + { + "epoch": 3.7334843337108343, + "grad_norm": 0.5229775905609131, + "learning_rate": 8.973718418664736e-05, + "loss": 0.4692, + "step": 9890 + }, + { + "epoch": 3.7372593431483576, + "grad_norm": 0.5336432456970215, + "learning_rate": 8.971208147023397e-05, + "loss": 0.4629, + "step": 9900 + }, + { + "epoch": 3.7410343525858814, + "grad_norm": 0.5237134695053101, + "learning_rate": 8.968695161098739e-05, + "loss": 0.4709, + "step": 9910 + }, + { + "epoch": 3.744809362023405, + "grad_norm": 0.5551522970199585, + "learning_rate": 8.96617946260836e-05, + "loss": 0.4639, + "step": 9920 + }, + { + "epoch": 3.748584371460929, + "grad_norm": 0.5532174706459045, + "learning_rate": 8.963661053271716e-05, + "loss": 0.4654, + "step": 9930 + }, + { + "epoch": 3.752359380898452, + "grad_norm": 0.5492597818374634, + "learning_rate": 8.961139934810117e-05, + "loss": 0.4636, + "step": 9940 + }, + { + "epoch": 3.756134390335976, + "grad_norm": 0.7447717785835266, + "learning_rate": 8.958616108946725e-05, + "loss": 0.4703, + "step": 9950 + }, + { + "epoch": 3.7599093997734996, + "grad_norm": 0.5668697357177734, + "learning_rate": 8.95608957740655e-05, + "loss": 0.4647, + "step": 9960 + }, + { + "epoch": 3.763684409211023, + "grad_norm": 0.6193100214004517, + "learning_rate": 8.953560341916452e-05, + "loss": 0.4757, + "step": 9970 + }, + { + "epoch": 3.7674594186485466, + "grad_norm": 0.522419273853302, + "learning_rate": 8.95102840420514e-05, + "loss": 0.4662, + "step": 9980 + }, + { + "epoch": 3.7712344280860703, + "grad_norm": 0.5857412219047546, + "learning_rate": 8.94849376600317e-05, + "loss": 0.4657, + "step": 9990 + }, + { + "epoch": 3.775009437523594, + "grad_norm": 0.557212769985199, + "learning_rate": 8.945956429042943e-05, + "loss": 0.4557, + "step": 10000 + }, + { + "epoch": 3.7787844469611174, + "grad_norm": 0.600222647190094, + "learning_rate": 8.943416395058705e-05, + "loss": 0.4661, + "step": 10010 + }, + { + "epoch": 3.782559456398641, + "grad_norm": 0.6195318102836609, + "learning_rate": 8.940873665786544e-05, + "loss": 0.467, + "step": 10020 + }, + { + "epoch": 3.7863344658361644, + "grad_norm": 0.8778566122055054, + "learning_rate": 8.938328242964394e-05, + "loss": 0.4602, + "step": 10030 + }, + { + "epoch": 3.790109475273688, + "grad_norm": 0.6205072999000549, + "learning_rate": 8.935780128332026e-05, + "loss": 0.4604, + "step": 10040 + }, + { + "epoch": 3.793884484711212, + "grad_norm": 0.5959478616714478, + "learning_rate": 8.933229323631052e-05, + "loss": 0.4582, + "step": 10050 + }, + { + "epoch": 3.7976594941487356, + "grad_norm": 0.5634093880653381, + "learning_rate": 8.930675830604925e-05, + "loss": 0.4649, + "step": 10060 + }, + { + "epoch": 3.801434503586259, + "grad_norm": 0.5344751477241516, + "learning_rate": 8.92811965099893e-05, + "loss": 0.465, + "step": 10070 + }, + { + "epoch": 3.8052095130237826, + "grad_norm": 0.5962951183319092, + "learning_rate": 8.925560786560194e-05, + "loss": 0.4676, + "step": 10080 + }, + { + "epoch": 3.808984522461306, + "grad_norm": 0.5603546500205994, + "learning_rate": 8.922999239037677e-05, + "loss": 0.4597, + "step": 10090 + }, + { + "epoch": 3.8127595318988297, + "grad_norm": 0.5416293144226074, + "learning_rate": 8.920435010182171e-05, + "loss": 0.4627, + "step": 10100 + }, + { + "epoch": 3.8165345413363534, + "grad_norm": 0.5338233709335327, + "learning_rate": 8.917868101746302e-05, + "loss": 0.4745, + "step": 10110 + }, + { + "epoch": 3.820309550773877, + "grad_norm": 0.6542847752571106, + "learning_rate": 8.91529851548453e-05, + "loss": 0.4676, + "step": 10120 + }, + { + "epoch": 3.8240845602114004, + "grad_norm": 0.6008440256118774, + "learning_rate": 8.912726253153142e-05, + "loss": 0.4685, + "step": 10130 + }, + { + "epoch": 3.827859569648924, + "grad_norm": 0.6180557012557983, + "learning_rate": 8.910151316510255e-05, + "loss": 0.4589, + "step": 10140 + }, + { + "epoch": 3.8316345790864474, + "grad_norm": 0.6951040625572205, + "learning_rate": 8.907573707315813e-05, + "loss": 0.4624, + "step": 10150 + }, + { + "epoch": 3.835409588523971, + "grad_norm": 0.5660848021507263, + "learning_rate": 8.904993427331588e-05, + "loss": 0.4647, + "step": 10160 + }, + { + "epoch": 3.839184597961495, + "grad_norm": 0.5535010099411011, + "learning_rate": 8.902410478321176e-05, + "loss": 0.4602, + "step": 10170 + }, + { + "epoch": 3.8429596073990187, + "grad_norm": 0.5885663628578186, + "learning_rate": 8.899824862050002e-05, + "loss": 0.4686, + "step": 10180 + }, + { + "epoch": 3.846734616836542, + "grad_norm": 0.6347377300262451, + "learning_rate": 8.897236580285308e-05, + "loss": 0.4584, + "step": 10190 + }, + { + "epoch": 3.8505096262740657, + "grad_norm": 0.6041572093963623, + "learning_rate": 8.894645634796159e-05, + "loss": 0.4584, + "step": 10200 + }, + { + "epoch": 3.8542846357115894, + "grad_norm": 0.5319682359695435, + "learning_rate": 8.892052027353444e-05, + "loss": 0.4608, + "step": 10210 + }, + { + "epoch": 3.8580596451491127, + "grad_norm": 0.5556275248527527, + "learning_rate": 8.889455759729866e-05, + "loss": 0.465, + "step": 10220 + }, + { + "epoch": 3.8618346545866364, + "grad_norm": 0.6249091029167175, + "learning_rate": 8.886856833699955e-05, + "loss": 0.4611, + "step": 10230 + }, + { + "epoch": 3.86560966402416, + "grad_norm": 0.5192009806632996, + "learning_rate": 8.884255251040046e-05, + "loss": 0.4656, + "step": 10240 + }, + { + "epoch": 3.869384673461684, + "grad_norm": 0.5966506600379944, + "learning_rate": 8.8816510135283e-05, + "loss": 0.4614, + "step": 10250 + }, + { + "epoch": 3.873159682899207, + "grad_norm": 0.5522090196609497, + "learning_rate": 8.879044122944688e-05, + "loss": 0.4583, + "step": 10260 + }, + { + "epoch": 3.876934692336731, + "grad_norm": 0.8987534046173096, + "learning_rate": 8.876434581070996e-05, + "loss": 0.4639, + "step": 10270 + }, + { + "epoch": 3.8807097017742542, + "grad_norm": 0.5576398372650146, + "learning_rate": 8.87382238969082e-05, + "loss": 0.4597, + "step": 10280 + }, + { + "epoch": 3.884484711211778, + "grad_norm": 0.5064070820808411, + "learning_rate": 8.871207550589568e-05, + "loss": 0.4612, + "step": 10290 + }, + { + "epoch": 3.8882597206493017, + "grad_norm": 0.536137580871582, + "learning_rate": 8.868590065554458e-05, + "loss": 0.4681, + "step": 10300 + }, + { + "epoch": 3.8920347300868254, + "grad_norm": 0.5211617350578308, + "learning_rate": 8.865969936374519e-05, + "loss": 0.458, + "step": 10310 + }, + { + "epoch": 3.8958097395243487, + "grad_norm": 0.5458037853240967, + "learning_rate": 8.863347164840581e-05, + "loss": 0.469, + "step": 10320 + }, + { + "epoch": 3.8995847489618725, + "grad_norm": 0.5937955379486084, + "learning_rate": 8.860721752745285e-05, + "loss": 0.4657, + "step": 10330 + }, + { + "epoch": 3.9033597583993958, + "grad_norm": 0.5774049758911133, + "learning_rate": 8.858093701883077e-05, + "loss": 0.454, + "step": 10340 + }, + { + "epoch": 3.9071347678369195, + "grad_norm": 0.5954073667526245, + "learning_rate": 8.8554630140502e-05, + "loss": 0.4606, + "step": 10350 + }, + { + "epoch": 3.910909777274443, + "grad_norm": 0.568393349647522, + "learning_rate": 8.85282969104471e-05, + "loss": 0.4559, + "step": 10360 + }, + { + "epoch": 3.914684786711967, + "grad_norm": 0.7279831767082214, + "learning_rate": 8.850193734666456e-05, + "loss": 0.462, + "step": 10370 + }, + { + "epoch": 3.9184597961494902, + "grad_norm": 0.610234260559082, + "learning_rate": 8.84755514671709e-05, + "loss": 0.4536, + "step": 10380 + }, + { + "epoch": 3.922234805587014, + "grad_norm": 0.6353870630264282, + "learning_rate": 8.84491392900006e-05, + "loss": 0.456, + "step": 10390 + }, + { + "epoch": 3.9260098150245377, + "grad_norm": 0.6264570951461792, + "learning_rate": 8.842270083320617e-05, + "loss": 0.4607, + "step": 10400 + }, + { + "epoch": 3.929784824462061, + "grad_norm": 0.5441379547119141, + "learning_rate": 8.839623611485801e-05, + "loss": 0.4638, + "step": 10410 + }, + { + "epoch": 3.9335598338995847, + "grad_norm": 0.5710632801055908, + "learning_rate": 8.836974515304453e-05, + "loss": 0.4552, + "step": 10420 + }, + { + "epoch": 3.9373348433371085, + "grad_norm": 0.6466676592826843, + "learning_rate": 8.834322796587204e-05, + "loss": 0.4643, + "step": 10430 + }, + { + "epoch": 3.9411098527746318, + "grad_norm": 0.5665135383605957, + "learning_rate": 8.831668457146478e-05, + "loss": 0.4565, + "step": 10440 + }, + { + "epoch": 3.9448848622121555, + "grad_norm": 0.48944804072380066, + "learning_rate": 8.829011498796493e-05, + "loss": 0.4624, + "step": 10450 + }, + { + "epoch": 3.9486598716496792, + "grad_norm": 0.4873522222042084, + "learning_rate": 8.826351923353253e-05, + "loss": 0.4627, + "step": 10460 + }, + { + "epoch": 3.9524348810872025, + "grad_norm": 0.7126550674438477, + "learning_rate": 8.823689732634555e-05, + "loss": 0.4545, + "step": 10470 + }, + { + "epoch": 3.9562098905247263, + "grad_norm": 0.5510223507881165, + "learning_rate": 8.82102492845998e-05, + "loss": 0.466, + "step": 10480 + }, + { + "epoch": 3.95998489996225, + "grad_norm": 0.5616976022720337, + "learning_rate": 8.818357512650896e-05, + "loss": 0.4593, + "step": 10490 + }, + { + "epoch": 3.9637599093997737, + "grad_norm": 0.523526668548584, + "learning_rate": 8.815687487030458e-05, + "loss": 0.4557, + "step": 10500 + }, + { + "epoch": 3.967534918837297, + "grad_norm": 0.5704669952392578, + "learning_rate": 8.8130148534236e-05, + "loss": 0.4687, + "step": 10510 + }, + { + "epoch": 3.9713099282748208, + "grad_norm": 0.6883910894393921, + "learning_rate": 8.810339613657047e-05, + "loss": 0.4598, + "step": 10520 + }, + { + "epoch": 3.975084937712344, + "grad_norm": 0.9575225114822388, + "learning_rate": 8.807661769559295e-05, + "loss": 0.4557, + "step": 10530 + }, + { + "epoch": 3.978859947149868, + "grad_norm": 0.5738497376441956, + "learning_rate": 8.804981322960628e-05, + "loss": 0.4612, + "step": 10540 + }, + { + "epoch": 3.9826349565873915, + "grad_norm": 0.530934751033783, + "learning_rate": 8.802298275693106e-05, + "loss": 0.4532, + "step": 10550 + }, + { + "epoch": 3.9864099660249153, + "grad_norm": 0.5506056547164917, + "learning_rate": 8.799612629590568e-05, + "loss": 0.4542, + "step": 10560 + }, + { + "epoch": 3.9901849754624386, + "grad_norm": 0.6080657243728638, + "learning_rate": 8.796924386488624e-05, + "loss": 0.4601, + "step": 10570 + }, + { + "epoch": 3.9939599848999623, + "grad_norm": 0.5610547661781311, + "learning_rate": 8.794233548224666e-05, + "loss": 0.4545, + "step": 10580 + }, + { + "epoch": 3.9977349943374856, + "grad_norm": 0.5351642966270447, + "learning_rate": 8.791540116637853e-05, + "loss": 0.4598, + "step": 10590 + }, + { + "epoch": 4.001510003775009, + "grad_norm": 0.5863658785820007, + "learning_rate": 8.788844093569124e-05, + "loss": 0.4532, + "step": 10600 + }, + { + "epoch": 4.005285013212533, + "grad_norm": 0.530657172203064, + "learning_rate": 8.786145480861184e-05, + "loss": 0.4474, + "step": 10610 + }, + { + "epoch": 4.009060022650057, + "grad_norm": 0.6831036806106567, + "learning_rate": 8.783444280358507e-05, + "loss": 0.4602, + "step": 10620 + }, + { + "epoch": 4.0128350320875805, + "grad_norm": 0.579365074634552, + "learning_rate": 8.780740493907342e-05, + "loss": 0.4698, + "step": 10630 + }, + { + "epoch": 4.016610041525103, + "grad_norm": 0.5720619559288025, + "learning_rate": 8.778034123355698e-05, + "loss": 0.4594, + "step": 10640 + }, + { + "epoch": 4.020385050962627, + "grad_norm": 0.6572504043579102, + "learning_rate": 8.775325170553357e-05, + "loss": 0.4755, + "step": 10650 + }, + { + "epoch": 4.024160060400151, + "grad_norm": 0.584198534488678, + "learning_rate": 8.77261363735186e-05, + "loss": 0.4579, + "step": 10660 + }, + { + "epoch": 4.027935069837675, + "grad_norm": 0.5127620100975037, + "learning_rate": 8.769899525604517e-05, + "loss": 0.4559, + "step": 10670 + }, + { + "epoch": 4.031710079275198, + "grad_norm": 0.5613308548927307, + "learning_rate": 8.767182837166397e-05, + "loss": 0.4607, + "step": 10680 + }, + { + "epoch": 4.035485088712722, + "grad_norm": 0.49144065380096436, + "learning_rate": 8.764463573894328e-05, + "loss": 0.4517, + "step": 10690 + }, + { + "epoch": 4.039260098150246, + "grad_norm": 0.533999502658844, + "learning_rate": 8.761741737646902e-05, + "loss": 0.4584, + "step": 10700 + }, + { + "epoch": 4.043035107587769, + "grad_norm": 0.521372377872467, + "learning_rate": 8.759017330284471e-05, + "loss": 0.4653, + "step": 10710 + }, + { + "epoch": 4.046810117025292, + "grad_norm": 0.6814863681793213, + "learning_rate": 8.756290353669142e-05, + "loss": 0.4665, + "step": 10720 + }, + { + "epoch": 4.050585126462816, + "grad_norm": 0.5394108891487122, + "learning_rate": 8.753560809664774e-05, + "loss": 0.4549, + "step": 10730 + }, + { + "epoch": 4.05436013590034, + "grad_norm": 0.624742329120636, + "learning_rate": 8.750828700136986e-05, + "loss": 0.4544, + "step": 10740 + }, + { + "epoch": 4.058135145337864, + "grad_norm": 0.6260218620300293, + "learning_rate": 8.74809402695315e-05, + "loss": 0.4608, + "step": 10750 + }, + { + "epoch": 4.061910154775387, + "grad_norm": 0.5796566009521484, + "learning_rate": 8.745356791982391e-05, + "loss": 0.4623, + "step": 10760 + }, + { + "epoch": 4.06568516421291, + "grad_norm": 0.5631840825080872, + "learning_rate": 8.742616997095578e-05, + "loss": 0.4543, + "step": 10770 + }, + { + "epoch": 4.069460173650434, + "grad_norm": 0.6347076892852783, + "learning_rate": 8.739874644165341e-05, + "loss": 0.4613, + "step": 10780 + }, + { + "epoch": 4.073235183087958, + "grad_norm": 0.5708393454551697, + "learning_rate": 8.737129735066048e-05, + "loss": 0.4609, + "step": 10790 + }, + { + "epoch": 4.077010192525481, + "grad_norm": 0.5556269288063049, + "learning_rate": 8.734382271673821e-05, + "loss": 0.4542, + "step": 10800 + }, + { + "epoch": 4.080785201963005, + "grad_norm": 0.587650716304779, + "learning_rate": 8.731632255866525e-05, + "loss": 0.4546, + "step": 10810 + }, + { + "epoch": 4.084560211400529, + "grad_norm": 0.5899844765663147, + "learning_rate": 8.728879689523767e-05, + "loss": 0.4493, + "step": 10820 + }, + { + "epoch": 4.088335220838052, + "grad_norm": 0.5399772524833679, + "learning_rate": 8.726124574526905e-05, + "loss": 0.4597, + "step": 10830 + }, + { + "epoch": 4.092110230275575, + "grad_norm": 0.534979522228241, + "learning_rate": 8.72336691275903e-05, + "loss": 0.4519, + "step": 10840 + }, + { + "epoch": 4.095885239713099, + "grad_norm": 0.5515267848968506, + "learning_rate": 8.720606706104979e-05, + "loss": 0.453, + "step": 10850 + }, + { + "epoch": 4.099660249150623, + "grad_norm": 0.5918541550636292, + "learning_rate": 8.71784395645133e-05, + "loss": 0.4586, + "step": 10860 + }, + { + "epoch": 4.103435258588147, + "grad_norm": 0.5579758286476135, + "learning_rate": 8.715078665686392e-05, + "loss": 0.4613, + "step": 10870 + }, + { + "epoch": 4.10721026802567, + "grad_norm": 0.5700817108154297, + "learning_rate": 8.712310835700218e-05, + "loss": 0.4591, + "step": 10880 + }, + { + "epoch": 4.110985277463194, + "grad_norm": 0.5911275148391724, + "learning_rate": 8.709540468384591e-05, + "loss": 0.4544, + "step": 10890 + }, + { + "epoch": 4.114760286900717, + "grad_norm": 0.6004135012626648, + "learning_rate": 8.706767565633033e-05, + "loss": 0.4484, + "step": 10900 + }, + { + "epoch": 4.118535296338241, + "grad_norm": 0.6015802621841431, + "learning_rate": 8.7039921293408e-05, + "loss": 0.4672, + "step": 10910 + }, + { + "epoch": 4.122310305775764, + "grad_norm": 0.6620537638664246, + "learning_rate": 8.70121416140487e-05, + "loss": 0.4506, + "step": 10920 + }, + { + "epoch": 4.126085315213288, + "grad_norm": 0.5017173886299133, + "learning_rate": 8.698433663723962e-05, + "loss": 0.4554, + "step": 10930 + }, + { + "epoch": 4.129860324650812, + "grad_norm": 0.5419957637786865, + "learning_rate": 8.695650638198518e-05, + "loss": 0.456, + "step": 10940 + }, + { + "epoch": 4.133635334088336, + "grad_norm": 0.6066375374794006, + "learning_rate": 8.692865086730713e-05, + "loss": 0.4519, + "step": 10950 + }, + { + "epoch": 4.1374103435258585, + "grad_norm": 0.5887925624847412, + "learning_rate": 8.69007701122444e-05, + "loss": 0.4557, + "step": 10960 + }, + { + "epoch": 4.141185352963382, + "grad_norm": 0.5485467910766602, + "learning_rate": 8.687286413585328e-05, + "loss": 0.4606, + "step": 10970 + }, + { + "epoch": 4.144960362400906, + "grad_norm": 0.5847924947738647, + "learning_rate": 8.684493295720719e-05, + "loss": 0.4481, + "step": 10980 + }, + { + "epoch": 4.14873537183843, + "grad_norm": 0.6698928475379944, + "learning_rate": 8.681697659539685e-05, + "loss": 0.4543, + "step": 10990 + }, + { + "epoch": 4.152510381275953, + "grad_norm": 0.578162431716919, + "learning_rate": 8.678899506953019e-05, + "loss": 0.4537, + "step": 11000 + }, + { + "epoch": 4.156285390713477, + "grad_norm": 0.5858516097068787, + "learning_rate": 8.676098839873227e-05, + "loss": 0.4531, + "step": 11010 + }, + { + "epoch": 4.160060400151, + "grad_norm": 0.6587435603141785, + "learning_rate": 8.673295660214545e-05, + "loss": 0.4573, + "step": 11020 + }, + { + "epoch": 4.163835409588524, + "grad_norm": 0.5828598141670227, + "learning_rate": 8.670489969892914e-05, + "loss": 0.4543, + "step": 11030 + }, + { + "epoch": 4.1676104190260475, + "grad_norm": 0.6038519740104675, + "learning_rate": 8.667681770826e-05, + "loss": 0.448, + "step": 11040 + }, + { + "epoch": 4.171385428463571, + "grad_norm": 0.5508007407188416, + "learning_rate": 8.66487106493318e-05, + "loss": 0.4513, + "step": 11050 + }, + { + "epoch": 4.175160437901095, + "grad_norm": 0.5896838903427124, + "learning_rate": 8.662057854135544e-05, + "loss": 0.4543, + "step": 11060 + }, + { + "epoch": 4.178935447338619, + "grad_norm": 0.5867491960525513, + "learning_rate": 8.659242140355897e-05, + "loss": 0.449, + "step": 11070 + }, + { + "epoch": 4.182710456776142, + "grad_norm": 0.6184877157211304, + "learning_rate": 8.65642392551875e-05, + "loss": 0.4619, + "step": 11080 + }, + { + "epoch": 4.186485466213665, + "grad_norm": 0.6027196049690247, + "learning_rate": 8.65360321155033e-05, + "loss": 0.4591, + "step": 11090 + }, + { + "epoch": 4.190260475651189, + "grad_norm": 0.588248610496521, + "learning_rate": 8.650780000378566e-05, + "loss": 0.4567, + "step": 11100 + }, + { + "epoch": 4.194035485088713, + "grad_norm": 0.5748762488365173, + "learning_rate": 8.647954293933096e-05, + "loss": 0.4568, + "step": 11110 + }, + { + "epoch": 4.1978104945262364, + "grad_norm": 0.5934643149375916, + "learning_rate": 8.645126094145264e-05, + "loss": 0.4538, + "step": 11120 + }, + { + "epoch": 4.20158550396376, + "grad_norm": 0.6221973896026611, + "learning_rate": 8.642295402948117e-05, + "loss": 0.4602, + "step": 11130 + }, + { + "epoch": 4.205360513401284, + "grad_norm": 0.5034570097923279, + "learning_rate": 8.639462222276409e-05, + "loss": 0.4543, + "step": 11140 + }, + { + "epoch": 4.209135522838807, + "grad_norm": 0.6960034966468811, + "learning_rate": 8.636626554066589e-05, + "loss": 0.4548, + "step": 11150 + }, + { + "epoch": 4.2129105322763305, + "grad_norm": 0.5671536326408386, + "learning_rate": 8.633788400256811e-05, + "loss": 0.4484, + "step": 11160 + }, + { + "epoch": 4.216685541713854, + "grad_norm": 0.48511791229248047, + "learning_rate": 8.630947762786927e-05, + "loss": 0.4558, + "step": 11170 + }, + { + "epoch": 4.220460551151378, + "grad_norm": 0.6003875732421875, + "learning_rate": 8.628104643598483e-05, + "loss": 0.4512, + "step": 11180 + }, + { + "epoch": 4.224235560588902, + "grad_norm": 0.544647753238678, + "learning_rate": 8.625259044634726e-05, + "loss": 0.4552, + "step": 11190 + }, + { + "epoch": 4.228010570026425, + "grad_norm": 0.5275722146034241, + "learning_rate": 8.622410967840597e-05, + "loss": 0.4498, + "step": 11200 + }, + { + "epoch": 4.231785579463948, + "grad_norm": 0.5396122932434082, + "learning_rate": 8.619560415162731e-05, + "loss": 0.4526, + "step": 11210 + }, + { + "epoch": 4.235560588901472, + "grad_norm": 0.5510291457176208, + "learning_rate": 8.616707388549447e-05, + "loss": 0.4465, + "step": 11220 + }, + { + "epoch": 4.239335598338996, + "grad_norm": 0.6948357820510864, + "learning_rate": 8.613851889950771e-05, + "loss": 0.4498, + "step": 11230 + }, + { + "epoch": 4.2431106077765195, + "grad_norm": 0.6219735145568848, + "learning_rate": 8.610993921318402e-05, + "loss": 0.4567, + "step": 11240 + }, + { + "epoch": 4.246885617214043, + "grad_norm": 0.5520322322845459, + "learning_rate": 8.608133484605738e-05, + "loss": 0.4523, + "step": 11250 + }, + { + "epoch": 4.250660626651567, + "grad_norm": 0.5849624872207642, + "learning_rate": 8.605270581767859e-05, + "loss": 0.4526, + "step": 11260 + }, + { + "epoch": 4.25443563608909, + "grad_norm": 0.5052843689918518, + "learning_rate": 8.602405214761536e-05, + "loss": 0.4525, + "step": 11270 + }, + { + "epoch": 4.2582106455266135, + "grad_norm": 0.518661618232727, + "learning_rate": 8.599537385545215e-05, + "loss": 0.4503, + "step": 11280 + }, + { + "epoch": 4.261985654964137, + "grad_norm": 0.5502333045005798, + "learning_rate": 8.596667096079032e-05, + "loss": 0.4477, + "step": 11290 + }, + { + "epoch": 4.265760664401661, + "grad_norm": 0.6355045437812805, + "learning_rate": 8.593794348324806e-05, + "loss": 0.4492, + "step": 11300 + }, + { + "epoch": 4.269535673839185, + "grad_norm": 0.5576472878456116, + "learning_rate": 8.590919144246028e-05, + "loss": 0.4517, + "step": 11310 + }, + { + "epoch": 4.2733106832767085, + "grad_norm": 0.5880599021911621, + "learning_rate": 8.588041485807876e-05, + "loss": 0.4478, + "step": 11320 + }, + { + "epoch": 4.277085692714232, + "grad_norm": 0.5388513207435608, + "learning_rate": 8.585161374977202e-05, + "loss": 0.4523, + "step": 11330 + }, + { + "epoch": 4.280860702151755, + "grad_norm": 0.5540359616279602, + "learning_rate": 8.582278813722533e-05, + "loss": 0.4484, + "step": 11340 + }, + { + "epoch": 4.284635711589279, + "grad_norm": 0.6409743428230286, + "learning_rate": 8.579393804014076e-05, + "loss": 0.4489, + "step": 11350 + }, + { + "epoch": 4.2884107210268025, + "grad_norm": 0.6246625185012817, + "learning_rate": 8.576506347823703e-05, + "loss": 0.4488, + "step": 11360 + }, + { + "epoch": 4.292185730464326, + "grad_norm": 0.7034008502960205, + "learning_rate": 8.573616447124968e-05, + "loss": 0.4492, + "step": 11370 + }, + { + "epoch": 4.29596073990185, + "grad_norm": 0.6897478103637695, + "learning_rate": 8.570724103893086e-05, + "loss": 0.4477, + "step": 11380 + }, + { + "epoch": 4.299735749339374, + "grad_norm": 0.5276645421981812, + "learning_rate": 8.567829320104951e-05, + "loss": 0.4424, + "step": 11390 + }, + { + "epoch": 4.303510758776897, + "grad_norm": 0.5786343812942505, + "learning_rate": 8.564932097739118e-05, + "loss": 0.4503, + "step": 11400 + }, + { + "epoch": 4.30728576821442, + "grad_norm": 0.5901648998260498, + "learning_rate": 8.562032438775811e-05, + "loss": 0.4426, + "step": 11410 + }, + { + "epoch": 4.311060777651944, + "grad_norm": 0.60638427734375, + "learning_rate": 8.559130345196921e-05, + "loss": 0.4519, + "step": 11420 + }, + { + "epoch": 4.314835787089468, + "grad_norm": 0.6317540407180786, + "learning_rate": 8.556225818986e-05, + "loss": 0.4567, + "step": 11430 + }, + { + "epoch": 4.3186107965269915, + "grad_norm": 0.5992687344551086, + "learning_rate": 8.553318862128265e-05, + "loss": 0.4547, + "step": 11440 + }, + { + "epoch": 4.322385805964515, + "grad_norm": 0.6526298522949219, + "learning_rate": 8.550409476610593e-05, + "loss": 0.4417, + "step": 11450 + }, + { + "epoch": 4.326160815402038, + "grad_norm": 0.5577118992805481, + "learning_rate": 8.547497664421522e-05, + "loss": 0.4517, + "step": 11460 + }, + { + "epoch": 4.329935824839562, + "grad_norm": 0.6398622989654541, + "learning_rate": 8.544583427551249e-05, + "loss": 0.4474, + "step": 11470 + }, + { + "epoch": 4.333710834277086, + "grad_norm": 0.5577737092971802, + "learning_rate": 8.541666767991628e-05, + "loss": 0.4508, + "step": 11480 + }, + { + "epoch": 4.337485843714609, + "grad_norm": 0.5385574698448181, + "learning_rate": 8.538747687736166e-05, + "loss": 0.4417, + "step": 11490 + }, + { + "epoch": 4.341260853152133, + "grad_norm": 0.595480740070343, + "learning_rate": 8.535826188780026e-05, + "loss": 0.4463, + "step": 11500 + }, + { + "epoch": 4.345035862589657, + "grad_norm": 0.5585368871688843, + "learning_rate": 8.532902273120029e-05, + "loss": 0.4421, + "step": 11510 + }, + { + "epoch": 4.34881087202718, + "grad_norm": 0.6641288995742798, + "learning_rate": 8.52997594275464e-05, + "loss": 0.444, + "step": 11520 + }, + { + "epoch": 4.352585881464703, + "grad_norm": 0.5479051470756531, + "learning_rate": 8.52704719968398e-05, + "loss": 0.4611, + "step": 11530 + }, + { + "epoch": 4.356360890902227, + "grad_norm": 0.5787330269813538, + "learning_rate": 8.524116045909818e-05, + "loss": 0.4524, + "step": 11540 + }, + { + "epoch": 4.360135900339751, + "grad_norm": 0.5680203437805176, + "learning_rate": 8.521182483435569e-05, + "loss": 0.4497, + "step": 11550 + }, + { + "epoch": 4.363910909777275, + "grad_norm": 0.6969168186187744, + "learning_rate": 8.518246514266295e-05, + "loss": 0.4511, + "step": 11560 + }, + { + "epoch": 4.367685919214798, + "grad_norm": 0.6396414041519165, + "learning_rate": 8.515308140408703e-05, + "loss": 0.4435, + "step": 11570 + }, + { + "epoch": 4.371460928652322, + "grad_norm": 0.6805716753005981, + "learning_rate": 8.512367363871145e-05, + "loss": 0.4603, + "step": 11580 + }, + { + "epoch": 4.375235938089845, + "grad_norm": 0.712642252445221, + "learning_rate": 8.509424186663614e-05, + "loss": 0.4488, + "step": 11590 + }, + { + "epoch": 4.379010947527369, + "grad_norm": 0.6053329110145569, + "learning_rate": 8.506478610797743e-05, + "loss": 0.4522, + "step": 11600 + }, + { + "epoch": 4.382785956964892, + "grad_norm": 0.5076550245285034, + "learning_rate": 8.503530638286805e-05, + "loss": 0.4405, + "step": 11610 + }, + { + "epoch": 4.386560966402416, + "grad_norm": 0.6560049057006836, + "learning_rate": 8.500580271145712e-05, + "loss": 0.4475, + "step": 11620 + }, + { + "epoch": 4.39033597583994, + "grad_norm": 0.5975435972213745, + "learning_rate": 8.497627511391014e-05, + "loss": 0.4492, + "step": 11630 + }, + { + "epoch": 4.394110985277464, + "grad_norm": 0.6190789341926575, + "learning_rate": 8.494672361040891e-05, + "loss": 0.4587, + "step": 11640 + }, + { + "epoch": 4.397885994714986, + "grad_norm": 0.6516814827919006, + "learning_rate": 8.491714822115162e-05, + "loss": 0.4436, + "step": 11650 + }, + { + "epoch": 4.40166100415251, + "grad_norm": 0.5875669121742249, + "learning_rate": 8.488754896635277e-05, + "loss": 0.445, + "step": 11660 + }, + { + "epoch": 4.405436013590034, + "grad_norm": 0.49789679050445557, + "learning_rate": 8.485792586624317e-05, + "loss": 0.4551, + "step": 11670 + }, + { + "epoch": 4.409211023027558, + "grad_norm": 0.6263242959976196, + "learning_rate": 8.482827894106993e-05, + "loss": 0.4457, + "step": 11680 + }, + { + "epoch": 4.412986032465081, + "grad_norm": 0.5868949294090271, + "learning_rate": 8.479860821109646e-05, + "loss": 0.4527, + "step": 11690 + }, + { + "epoch": 4.416761041902605, + "grad_norm": 0.6368350386619568, + "learning_rate": 8.476891369660239e-05, + "loss": 0.4464, + "step": 11700 + }, + { + "epoch": 4.420536051340128, + "grad_norm": 0.5921139717102051, + "learning_rate": 8.473919541788366e-05, + "loss": 0.4423, + "step": 11710 + }, + { + "epoch": 4.424311060777652, + "grad_norm": 0.5540608763694763, + "learning_rate": 8.470945339525245e-05, + "loss": 0.4407, + "step": 11720 + }, + { + "epoch": 4.428086070215175, + "grad_norm": 0.5866498351097107, + "learning_rate": 8.467968764903713e-05, + "loss": 0.4408, + "step": 11730 + }, + { + "epoch": 4.431861079652699, + "grad_norm": 0.6014940738677979, + "learning_rate": 8.46498981995823e-05, + "loss": 0.4541, + "step": 11740 + }, + { + "epoch": 4.435636089090223, + "grad_norm": 0.5633965134620667, + "learning_rate": 8.462008506724879e-05, + "loss": 0.4428, + "step": 11750 + }, + { + "epoch": 4.439411098527747, + "grad_norm": 0.5375277400016785, + "learning_rate": 8.459024827241359e-05, + "loss": 0.4449, + "step": 11760 + }, + { + "epoch": 4.4431861079652695, + "grad_norm": 0.566908597946167, + "learning_rate": 8.456038783546985e-05, + "loss": 0.4425, + "step": 11770 + }, + { + "epoch": 4.446961117402793, + "grad_norm": 0.5475563406944275, + "learning_rate": 8.453050377682691e-05, + "loss": 0.4469, + "step": 11780 + }, + { + "epoch": 4.450736126840317, + "grad_norm": 0.5368761420249939, + "learning_rate": 8.450059611691026e-05, + "loss": 0.441, + "step": 11790 + }, + { + "epoch": 4.454511136277841, + "grad_norm": 0.561191976070404, + "learning_rate": 8.447066487616146e-05, + "loss": 0.4444, + "step": 11800 + }, + { + "epoch": 4.458286145715364, + "grad_norm": 0.6211469173431396, + "learning_rate": 8.444071007503826e-05, + "loss": 0.443, + "step": 11810 + }, + { + "epoch": 4.462061155152888, + "grad_norm": 0.6266103982925415, + "learning_rate": 8.441073173401449e-05, + "loss": 0.4392, + "step": 11820 + }, + { + "epoch": 4.465836164590412, + "grad_norm": 0.5911449193954468, + "learning_rate": 8.438072987358006e-05, + "loss": 0.4502, + "step": 11830 + }, + { + "epoch": 4.469611174027935, + "grad_norm": 0.7175887227058411, + "learning_rate": 8.435070451424094e-05, + "loss": 0.4447, + "step": 11840 + }, + { + "epoch": 4.4733861834654585, + "grad_norm": 0.5663455724716187, + "learning_rate": 8.432065567651919e-05, + "loss": 0.4387, + "step": 11850 + }, + { + "epoch": 4.477161192902982, + "grad_norm": 0.8334063291549683, + "learning_rate": 8.429058338095291e-05, + "loss": 0.4453, + "step": 11860 + }, + { + "epoch": 4.480936202340506, + "grad_norm": 0.5345563888549805, + "learning_rate": 8.426048764809624e-05, + "loss": 0.446, + "step": 11870 + }, + { + "epoch": 4.48471121177803, + "grad_norm": 0.5361315608024597, + "learning_rate": 8.423036849851932e-05, + "loss": 0.4468, + "step": 11880 + }, + { + "epoch": 4.488486221215553, + "grad_norm": 0.624729573726654, + "learning_rate": 8.42002259528083e-05, + "loss": 0.4496, + "step": 11890 + }, + { + "epoch": 4.492261230653076, + "grad_norm": 0.5892227292060852, + "learning_rate": 8.417006003156532e-05, + "loss": 0.4537, + "step": 11900 + }, + { + "epoch": 4.4960362400906, + "grad_norm": 0.5822799205780029, + "learning_rate": 8.413987075540852e-05, + "loss": 0.4537, + "step": 11910 + }, + { + "epoch": 4.499811249528124, + "grad_norm": 0.656360924243927, + "learning_rate": 8.4109658144972e-05, + "loss": 0.4475, + "step": 11920 + }, + { + "epoch": 4.5035862589656475, + "grad_norm": 0.6090147495269775, + "learning_rate": 8.407942222090573e-05, + "loss": 0.4528, + "step": 11930 + }, + { + "epoch": 4.507361268403171, + "grad_norm": 0.6417983174324036, + "learning_rate": 8.404916300387576e-05, + "loss": 0.4479, + "step": 11940 + }, + { + "epoch": 4.511136277840695, + "grad_norm": 0.6105488538742065, + "learning_rate": 8.401888051456391e-05, + "loss": 0.4563, + "step": 11950 + }, + { + "epoch": 4.514911287278219, + "grad_norm": 0.6055814027786255, + "learning_rate": 8.398857477366803e-05, + "loss": 0.4442, + "step": 11960 + }, + { + "epoch": 4.5186862967157415, + "grad_norm": 0.6824854612350464, + "learning_rate": 8.395824580190178e-05, + "loss": 0.4473, + "step": 11970 + }, + { + "epoch": 4.522461306153265, + "grad_norm": 0.5766848921775818, + "learning_rate": 8.392789361999473e-05, + "loss": 0.4444, + "step": 11980 + }, + { + "epoch": 4.526236315590789, + "grad_norm": 0.5585546493530273, + "learning_rate": 8.38975182486923e-05, + "loss": 0.4419, + "step": 11990 + }, + { + "epoch": 4.530011325028313, + "grad_norm": 0.5684893727302551, + "learning_rate": 8.386711970875581e-05, + "loss": 0.4381, + "step": 12000 + }, + { + "epoch": 4.5337863344658365, + "grad_norm": 0.5678910613059998, + "learning_rate": 8.383669802096232e-05, + "loss": 0.4394, + "step": 12010 + }, + { + "epoch": 4.537561343903359, + "grad_norm": 0.5642674565315247, + "learning_rate": 8.38062532061048e-05, + "loss": 0.4401, + "step": 12020 + }, + { + "epoch": 4.541336353340883, + "grad_norm": 0.5675539374351501, + "learning_rate": 8.3775785284992e-05, + "loss": 0.4403, + "step": 12030 + }, + { + "epoch": 4.545111362778407, + "grad_norm": 0.6496412754058838, + "learning_rate": 8.374529427844843e-05, + "loss": 0.4428, + "step": 12040 + }, + { + "epoch": 4.5488863722159305, + "grad_norm": 0.6716994643211365, + "learning_rate": 8.371478020731442e-05, + "loss": 0.4419, + "step": 12050 + }, + { + "epoch": 4.552661381653454, + "grad_norm": 0.5777934193611145, + "learning_rate": 8.368424309244607e-05, + "loss": 0.4414, + "step": 12060 + }, + { + "epoch": 4.556436391090978, + "grad_norm": 0.6052347421646118, + "learning_rate": 8.365368295471517e-05, + "loss": 0.4444, + "step": 12070 + }, + { + "epoch": 4.560211400528502, + "grad_norm": 0.6728206276893616, + "learning_rate": 8.362309981500931e-05, + "loss": 0.4366, + "step": 12080 + }, + { + "epoch": 4.563986409966025, + "grad_norm": 0.576066792011261, + "learning_rate": 8.359249369423177e-05, + "loss": 0.4433, + "step": 12090 + }, + { + "epoch": 4.567761419403548, + "grad_norm": 0.566498875617981, + "learning_rate": 8.356186461330155e-05, + "loss": 0.444, + "step": 12100 + }, + { + "epoch": 4.571536428841072, + "grad_norm": 0.5168943405151367, + "learning_rate": 8.353121259315334e-05, + "loss": 0.4345, + "step": 12110 + }, + { + "epoch": 4.575311438278596, + "grad_norm": 0.6019729375839233, + "learning_rate": 8.350053765473751e-05, + "loss": 0.4342, + "step": 12120 + }, + { + "epoch": 4.5790864477161195, + "grad_norm": 0.6395158767700195, + "learning_rate": 8.346983981902005e-05, + "loss": 0.4388, + "step": 12130 + }, + { + "epoch": 4.582861457153643, + "grad_norm": 0.6709539890289307, + "learning_rate": 8.343911910698271e-05, + "loss": 0.4448, + "step": 12140 + }, + { + "epoch": 4.586636466591166, + "grad_norm": 0.6900480389595032, + "learning_rate": 8.340837553962278e-05, + "loss": 0.4453, + "step": 12150 + }, + { + "epoch": 4.59041147602869, + "grad_norm": 0.6444860696792603, + "learning_rate": 8.337760913795316e-05, + "loss": 0.4453, + "step": 12160 + }, + { + "epoch": 4.5941864854662136, + "grad_norm": 0.5633404850959778, + "learning_rate": 8.334681992300244e-05, + "loss": 0.4447, + "step": 12170 + }, + { + "epoch": 4.597961494903737, + "grad_norm": 0.5962954163551331, + "learning_rate": 8.331600791581475e-05, + "loss": 0.4448, + "step": 12180 + }, + { + "epoch": 4.601736504341261, + "grad_norm": 0.5366525053977966, + "learning_rate": 8.328517313744978e-05, + "loss": 0.4433, + "step": 12190 + }, + { + "epoch": 4.605511513778785, + "grad_norm": 0.5605337619781494, + "learning_rate": 8.325431560898286e-05, + "loss": 0.4383, + "step": 12200 + }, + { + "epoch": 4.6092865232163085, + "grad_norm": 0.7788808941841125, + "learning_rate": 8.322343535150478e-05, + "loss": 0.4409, + "step": 12210 + }, + { + "epoch": 4.613061532653831, + "grad_norm": 0.6166380643844604, + "learning_rate": 8.319253238612191e-05, + "loss": 0.4415, + "step": 12220 + }, + { + "epoch": 4.616836542091355, + "grad_norm": 0.5688962936401367, + "learning_rate": 8.316160673395614e-05, + "loss": 0.4401, + "step": 12230 + }, + { + "epoch": 4.620611551528879, + "grad_norm": 0.6172645688056946, + "learning_rate": 8.313065841614487e-05, + "loss": 0.4439, + "step": 12240 + }, + { + "epoch": 4.6243865609664025, + "grad_norm": 0.7282325029373169, + "learning_rate": 8.309968745384096e-05, + "loss": 0.4383, + "step": 12250 + }, + { + "epoch": 4.628161570403926, + "grad_norm": 0.519595742225647, + "learning_rate": 8.306869386821282e-05, + "loss": 0.4384, + "step": 12260 + }, + { + "epoch": 4.631936579841449, + "grad_norm": 0.5448688268661499, + "learning_rate": 8.30376776804442e-05, + "loss": 0.4418, + "step": 12270 + }, + { + "epoch": 4.635711589278973, + "grad_norm": 0.5774092674255371, + "learning_rate": 8.300663891173443e-05, + "loss": 0.4396, + "step": 12280 + }, + { + "epoch": 4.639486598716497, + "grad_norm": 0.6279200315475464, + "learning_rate": 8.297557758329822e-05, + "loss": 0.4434, + "step": 12290 + }, + { + "epoch": 4.64326160815402, + "grad_norm": 0.6159646511077881, + "learning_rate": 8.294449371636564e-05, + "loss": 0.4406, + "step": 12300 + }, + { + "epoch": 4.647036617591544, + "grad_norm": 0.6391906142234802, + "learning_rate": 8.291338733218226e-05, + "loss": 0.4379, + "step": 12310 + }, + { + "epoch": 4.650811627029068, + "grad_norm": 0.6429669260978699, + "learning_rate": 8.2882258452009e-05, + "loss": 0.4328, + "step": 12320 + }, + { + "epoch": 4.6545866364665915, + "grad_norm": 0.5893004536628723, + "learning_rate": 8.285110709712214e-05, + "loss": 0.4385, + "step": 12330 + }, + { + "epoch": 4.658361645904114, + "grad_norm": 0.5372092127799988, + "learning_rate": 8.281993328881337e-05, + "loss": 0.434, + "step": 12340 + }, + { + "epoch": 4.662136655341638, + "grad_norm": 0.5648928284645081, + "learning_rate": 8.278873704838964e-05, + "loss": 0.4438, + "step": 12350 + }, + { + "epoch": 4.665911664779162, + "grad_norm": 0.5447239875793457, + "learning_rate": 8.275751839717334e-05, + "loss": 0.4314, + "step": 12360 + }, + { + "epoch": 4.669686674216686, + "grad_norm": 0.5694062113761902, + "learning_rate": 8.272627735650208e-05, + "loss": 0.443, + "step": 12370 + }, + { + "epoch": 4.673461683654209, + "grad_norm": 0.6510103940963745, + "learning_rate": 8.269501394772884e-05, + "loss": 0.4352, + "step": 12380 + }, + { + "epoch": 4.677236693091733, + "grad_norm": 0.5314978361129761, + "learning_rate": 8.266372819222189e-05, + "loss": 0.4385, + "step": 12390 + }, + { + "epoch": 4.681011702529256, + "grad_norm": 0.6842690110206604, + "learning_rate": 8.26324201113647e-05, + "loss": 0.4378, + "step": 12400 + }, + { + "epoch": 4.68478671196678, + "grad_norm": 0.6124461889266968, + "learning_rate": 8.260108972655606e-05, + "loss": 0.4416, + "step": 12410 + }, + { + "epoch": 4.688561721404303, + "grad_norm": 0.568488359451294, + "learning_rate": 8.256973705921e-05, + "loss": 0.4378, + "step": 12420 + }, + { + "epoch": 4.692336730841827, + "grad_norm": 0.5795332193374634, + "learning_rate": 8.25383621307558e-05, + "loss": 0.4411, + "step": 12430 + }, + { + "epoch": 4.696111740279351, + "grad_norm": 0.5848597884178162, + "learning_rate": 8.25069649626379e-05, + "loss": 0.4318, + "step": 12440 + }, + { + "epoch": 4.699886749716875, + "grad_norm": 0.5752859115600586, + "learning_rate": 8.247554557631596e-05, + "loss": 0.4373, + "step": 12450 + }, + { + "epoch": 4.703661759154398, + "grad_norm": 0.5697815418243408, + "learning_rate": 8.244410399326483e-05, + "loss": 0.4434, + "step": 12460 + }, + { + "epoch": 4.707436768591921, + "grad_norm": 0.5662826895713806, + "learning_rate": 8.241264023497457e-05, + "loss": 0.4412, + "step": 12470 + }, + { + "epoch": 4.711211778029445, + "grad_norm": 0.5779763460159302, + "learning_rate": 8.238115432295034e-05, + "loss": 0.4347, + "step": 12480 + }, + { + "epoch": 4.714986787466969, + "grad_norm": 0.5694817304611206, + "learning_rate": 8.234964627871247e-05, + "loss": 0.44, + "step": 12490 + }, + { + "epoch": 4.718761796904492, + "grad_norm": 0.6271215677261353, + "learning_rate": 8.231811612379639e-05, + "loss": 0.4388, + "step": 12500 + }, + { + "epoch": 4.722536806342016, + "grad_norm": 0.647670567035675, + "learning_rate": 8.228656387975268e-05, + "loss": 0.4424, + "step": 12510 + }, + { + "epoch": 4.726311815779539, + "grad_norm": 0.58333820104599, + "learning_rate": 8.225498956814702e-05, + "loss": 0.4432, + "step": 12520 + }, + { + "epoch": 4.730086825217063, + "grad_norm": 0.5770929455757141, + "learning_rate": 8.222339321056014e-05, + "loss": 0.4366, + "step": 12530 + }, + { + "epoch": 4.733861834654586, + "grad_norm": 0.5523325800895691, + "learning_rate": 8.219177482858785e-05, + "loss": 0.4332, + "step": 12540 + }, + { + "epoch": 4.73763684409211, + "grad_norm": 0.8121219873428345, + "learning_rate": 8.216013444384099e-05, + "loss": 0.4383, + "step": 12550 + }, + { + "epoch": 4.741411853529634, + "grad_norm": 0.5820472240447998, + "learning_rate": 8.21284720779455e-05, + "loss": 0.4418, + "step": 12560 + }, + { + "epoch": 4.745186862967158, + "grad_norm": 0.6363267302513123, + "learning_rate": 8.209678775254231e-05, + "loss": 0.4436, + "step": 12570 + }, + { + "epoch": 4.748961872404681, + "grad_norm": 0.5754694938659668, + "learning_rate": 8.206508148928733e-05, + "loss": 0.448, + "step": 12580 + }, + { + "epoch": 4.752736881842204, + "grad_norm": 0.6008169054985046, + "learning_rate": 8.203335330985151e-05, + "loss": 0.4448, + "step": 12590 + }, + { + "epoch": 4.756511891279728, + "grad_norm": 0.6524280905723572, + "learning_rate": 8.200160323592076e-05, + "loss": 0.4572, + "step": 12600 + }, + { + "epoch": 4.760286900717252, + "grad_norm": 0.5874181389808655, + "learning_rate": 8.196983128919598e-05, + "loss": 0.4404, + "step": 12610 + }, + { + "epoch": 4.764061910154775, + "grad_norm": 0.8897712826728821, + "learning_rate": 8.193803749139295e-05, + "loss": 0.4319, + "step": 12620 + }, + { + "epoch": 4.767836919592299, + "grad_norm": 0.5824182033538818, + "learning_rate": 8.190622186424244e-05, + "loss": 0.4436, + "step": 12630 + }, + { + "epoch": 4.771611929029823, + "grad_norm": 0.6077965497970581, + "learning_rate": 8.187438442949016e-05, + "loss": 0.4422, + "step": 12640 + }, + { + "epoch": 4.775386938467346, + "grad_norm": 0.7708337903022766, + "learning_rate": 8.184252520889668e-05, + "loss": 0.4312, + "step": 12650 + }, + { + "epoch": 4.7791619479048695, + "grad_norm": 0.5734137296676636, + "learning_rate": 8.181064422423748e-05, + "loss": 0.4382, + "step": 12660 + }, + { + "epoch": 4.782936957342393, + "grad_norm": 0.5938032865524292, + "learning_rate": 8.177874149730289e-05, + "loss": 0.4417, + "step": 12670 + }, + { + "epoch": 4.786711966779917, + "grad_norm": 0.6791886687278748, + "learning_rate": 8.174681704989816e-05, + "loss": 0.4365, + "step": 12680 + }, + { + "epoch": 4.790486976217441, + "grad_norm": 0.6398690342903137, + "learning_rate": 8.171487090384333e-05, + "loss": 0.4407, + "step": 12690 + }, + { + "epoch": 4.794261985654964, + "grad_norm": 0.6101489067077637, + "learning_rate": 8.168290308097328e-05, + "loss": 0.4296, + "step": 12700 + }, + { + "epoch": 4.798036995092488, + "grad_norm": 0.6080114841461182, + "learning_rate": 8.165091360313774e-05, + "loss": 0.4463, + "step": 12710 + }, + { + "epoch": 4.801812004530011, + "grad_norm": 0.5545832514762878, + "learning_rate": 8.161890249220119e-05, + "loss": 0.4354, + "step": 12720 + }, + { + "epoch": 4.805587013967535, + "grad_norm": 0.6806719899177551, + "learning_rate": 8.158686977004295e-05, + "loss": 0.4329, + "step": 12730 + }, + { + "epoch": 4.8093620234050585, + "grad_norm": 0.6922208070755005, + "learning_rate": 8.155481545855706e-05, + "loss": 0.4361, + "step": 12740 + }, + { + "epoch": 4.813137032842582, + "grad_norm": 0.6132349371910095, + "learning_rate": 8.152273957965233e-05, + "loss": 0.4392, + "step": 12750 + }, + { + "epoch": 4.816912042280106, + "grad_norm": 0.5297583341598511, + "learning_rate": 8.149064215525237e-05, + "loss": 0.4371, + "step": 12760 + }, + { + "epoch": 4.82068705171763, + "grad_norm": 0.5206873416900635, + "learning_rate": 8.14585232072954e-05, + "loss": 0.4358, + "step": 12770 + }, + { + "epoch": 4.8244620611551525, + "grad_norm": 0.5716885328292847, + "learning_rate": 8.142638275773449e-05, + "loss": 0.4391, + "step": 12780 + }, + { + "epoch": 4.828237070592676, + "grad_norm": 0.5915391445159912, + "learning_rate": 8.139422082853729e-05, + "loss": 0.4369, + "step": 12790 + }, + { + "epoch": 4.8320120800302, + "grad_norm": 0.5491467118263245, + "learning_rate": 8.136203744168618e-05, + "loss": 0.4377, + "step": 12800 + }, + { + "epoch": 4.835787089467724, + "grad_norm": 0.5877269506454468, + "learning_rate": 8.132983261917819e-05, + "loss": 0.4327, + "step": 12810 + }, + { + "epoch": 4.8395620989052475, + "grad_norm": 0.8957927227020264, + "learning_rate": 8.129760638302504e-05, + "loss": 0.4363, + "step": 12820 + }, + { + "epoch": 4.843337108342771, + "grad_norm": 0.6469703912734985, + "learning_rate": 8.126535875525305e-05, + "loss": 0.4354, + "step": 12830 + }, + { + "epoch": 4.847112117780295, + "grad_norm": 0.5585578680038452, + "learning_rate": 8.123308975790316e-05, + "loss": 0.4417, + "step": 12840 + }, + { + "epoch": 4.850887127217818, + "grad_norm": 0.858641505241394, + "learning_rate": 8.120079941303094e-05, + "loss": 0.4365, + "step": 12850 + }, + { + "epoch": 4.8546621366553415, + "grad_norm": 0.6256632804870605, + "learning_rate": 8.116848774270651e-05, + "loss": 0.44, + "step": 12860 + }, + { + "epoch": 4.858437146092865, + "grad_norm": 0.6245275735855103, + "learning_rate": 8.113615476901461e-05, + "loss": 0.4331, + "step": 12870 + }, + { + "epoch": 4.862212155530389, + "grad_norm": 0.6392505168914795, + "learning_rate": 8.110380051405454e-05, + "loss": 0.4316, + "step": 12880 + }, + { + "epoch": 4.865987164967913, + "grad_norm": 0.6834761500358582, + "learning_rate": 8.107142499994009e-05, + "loss": 0.4369, + "step": 12890 + }, + { + "epoch": 4.869762174405436, + "grad_norm": 0.6249173879623413, + "learning_rate": 8.103902824879966e-05, + "loss": 0.4378, + "step": 12900 + }, + { + "epoch": 4.873537183842959, + "grad_norm": 0.7045850157737732, + "learning_rate": 8.10066102827761e-05, + "loss": 0.4355, + "step": 12910 + }, + { + "epoch": 4.877312193280483, + "grad_norm": 0.6210693120956421, + "learning_rate": 8.097417112402676e-05, + "loss": 0.4387, + "step": 12920 + }, + { + "epoch": 4.881087202718007, + "grad_norm": 0.6081416010856628, + "learning_rate": 8.094171079472355e-05, + "loss": 0.4362, + "step": 12930 + }, + { + "epoch": 4.8848622121555305, + "grad_norm": 0.651164710521698, + "learning_rate": 8.090922931705277e-05, + "loss": 0.4349, + "step": 12940 + }, + { + "epoch": 4.888637221593054, + "grad_norm": 0.6259034276008606, + "learning_rate": 8.08767267132152e-05, + "loss": 0.4379, + "step": 12950 + }, + { + "epoch": 4.892412231030578, + "grad_norm": 0.5768272876739502, + "learning_rate": 8.084420300542608e-05, + "loss": 0.4296, + "step": 12960 + }, + { + "epoch": 4.896187240468101, + "grad_norm": 0.6301991939544678, + "learning_rate": 8.081165821591505e-05, + "loss": 0.434, + "step": 12970 + }, + { + "epoch": 4.899962249905625, + "grad_norm": 0.6101955771446228, + "learning_rate": 8.077909236692615e-05, + "loss": 0.4301, + "step": 12980 + }, + { + "epoch": 4.903737259343148, + "grad_norm": 0.6158044338226318, + "learning_rate": 8.074650548071787e-05, + "loss": 0.4378, + "step": 12990 + }, + { + "epoch": 4.907512268780672, + "grad_norm": 0.5665403604507446, + "learning_rate": 8.071389757956301e-05, + "loss": 0.436, + "step": 13000 + }, + { + "epoch": 4.911287278218196, + "grad_norm": 0.6046426892280579, + "learning_rate": 8.068126868574876e-05, + "loss": 0.4314, + "step": 13010 + }, + { + "epoch": 4.9150622876557195, + "grad_norm": 0.5915471315383911, + "learning_rate": 8.064861882157668e-05, + "loss": 0.435, + "step": 13020 + }, + { + "epoch": 4.918837297093242, + "grad_norm": 0.6827958226203918, + "learning_rate": 8.061594800936263e-05, + "loss": 0.4365, + "step": 13030 + }, + { + "epoch": 4.922612306530766, + "grad_norm": 0.5622991323471069, + "learning_rate": 8.058325627143681e-05, + "loss": 0.4344, + "step": 13040 + }, + { + "epoch": 4.92638731596829, + "grad_norm": 0.5483256578445435, + "learning_rate": 8.055054363014372e-05, + "loss": 0.4383, + "step": 13050 + }, + { + "epoch": 4.930162325405814, + "grad_norm": 0.6701416969299316, + "learning_rate": 8.051781010784211e-05, + "loss": 0.4397, + "step": 13060 + }, + { + "epoch": 4.933937334843337, + "grad_norm": 0.5472255945205688, + "learning_rate": 8.048505572690506e-05, + "loss": 0.4302, + "step": 13070 + }, + { + "epoch": 4.937712344280861, + "grad_norm": 0.655945897102356, + "learning_rate": 8.045228050971988e-05, + "loss": 0.4367, + "step": 13080 + }, + { + "epoch": 4.941487353718385, + "grad_norm": 0.5672075748443604, + "learning_rate": 8.041948447868814e-05, + "loss": 0.4292, + "step": 13090 + }, + { + "epoch": 4.945262363155908, + "grad_norm": 0.7679628133773804, + "learning_rate": 8.038666765622558e-05, + "loss": 0.4333, + "step": 13100 + }, + { + "epoch": 4.949037372593431, + "grad_norm": 0.5666245818138123, + "learning_rate": 8.03538300647622e-05, + "loss": 0.443, + "step": 13110 + }, + { + "epoch": 4.952812382030955, + "grad_norm": 0.6425677537918091, + "learning_rate": 8.03209717267422e-05, + "loss": 0.4288, + "step": 13120 + }, + { + "epoch": 4.956587391468479, + "grad_norm": 0.585444450378418, + "learning_rate": 8.028809266462395e-05, + "loss": 0.4317, + "step": 13130 + }, + { + "epoch": 4.9603624009060026, + "grad_norm": 0.6069588661193848, + "learning_rate": 8.025519290087994e-05, + "loss": 0.4352, + "step": 13140 + }, + { + "epoch": 4.964137410343525, + "grad_norm": 0.5473458170890808, + "learning_rate": 8.022227245799688e-05, + "loss": 0.4407, + "step": 13150 + }, + { + "epoch": 4.967912419781049, + "grad_norm": 0.5702902674674988, + "learning_rate": 8.018933135847557e-05, + "loss": 0.4293, + "step": 13160 + }, + { + "epoch": 4.971687429218573, + "grad_norm": 0.6425644755363464, + "learning_rate": 8.015636962483096e-05, + "loss": 0.4394, + "step": 13170 + }, + { + "epoch": 4.975462438656097, + "grad_norm": 0.5478054285049438, + "learning_rate": 8.012338727959205e-05, + "loss": 0.4267, + "step": 13180 + }, + { + "epoch": 4.97923744809362, + "grad_norm": 0.6049672961235046, + "learning_rate": 8.009038434530198e-05, + "loss": 0.4381, + "step": 13190 + }, + { + "epoch": 4.983012457531144, + "grad_norm": 0.5757096409797668, + "learning_rate": 8.005736084451796e-05, + "loss": 0.4292, + "step": 13200 + }, + { + "epoch": 4.986787466968668, + "grad_norm": 0.6315653920173645, + "learning_rate": 8.002431679981122e-05, + "loss": 0.4249, + "step": 13210 + }, + { + "epoch": 4.990562476406191, + "grad_norm": 0.6188434958457947, + "learning_rate": 7.999125223376706e-05, + "loss": 0.4312, + "step": 13220 + }, + { + "epoch": 4.994337485843714, + "grad_norm": 0.5495557188987732, + "learning_rate": 7.99581671689848e-05, + "loss": 0.4272, + "step": 13230 + }, + { + "epoch": 4.998112495281238, + "grad_norm": 0.6032848954200745, + "learning_rate": 7.992506162807775e-05, + "loss": 0.4378, + "step": 13240 + }, + { + "epoch": 5.001887504718762, + "grad_norm": 0.563420832157135, + "learning_rate": 7.989193563367328e-05, + "loss": 0.4333, + "step": 13250 + }, + { + "epoch": 5.005662514156286, + "grad_norm": 0.5446894764900208, + "learning_rate": 7.985878920841266e-05, + "loss": 0.4252, + "step": 13260 + }, + { + "epoch": 5.009437523593809, + "grad_norm": 0.6635614633560181, + "learning_rate": 7.982562237495117e-05, + "loss": 0.4289, + "step": 13270 + }, + { + "epoch": 5.013212533031332, + "grad_norm": 0.6846327781677246, + "learning_rate": 7.979243515595802e-05, + "loss": 0.4408, + "step": 13280 + }, + { + "epoch": 5.016987542468856, + "grad_norm": 0.6627950668334961, + "learning_rate": 7.975922757411636e-05, + "loss": 0.4443, + "step": 13290 + }, + { + "epoch": 5.02076255190638, + "grad_norm": 0.5397953987121582, + "learning_rate": 7.972599965212329e-05, + "loss": 0.4393, + "step": 13300 + }, + { + "epoch": 5.024537561343903, + "grad_norm": 0.5803020000457764, + "learning_rate": 7.969275141268973e-05, + "loss": 0.4316, + "step": 13310 + }, + { + "epoch": 5.028312570781427, + "grad_norm": 0.625580370426178, + "learning_rate": 7.96594828785406e-05, + "loss": 0.4343, + "step": 13320 + }, + { + "epoch": 5.032087580218951, + "grad_norm": 0.5824500918388367, + "learning_rate": 7.962619407241456e-05, + "loss": 0.4314, + "step": 13330 + }, + { + "epoch": 5.035862589656474, + "grad_norm": 0.6776422262191772, + "learning_rate": 7.959288501706424e-05, + "loss": 0.444, + "step": 13340 + }, + { + "epoch": 5.0396375990939974, + "grad_norm": 0.5621914267539978, + "learning_rate": 7.955955573525605e-05, + "loss": 0.435, + "step": 13350 + }, + { + "epoch": 5.043412608531521, + "grad_norm": 0.6005207300186157, + "learning_rate": 7.952620624977026e-05, + "loss": 0.4282, + "step": 13360 + }, + { + "epoch": 5.047187617969045, + "grad_norm": 0.5816413760185242, + "learning_rate": 7.949283658340089e-05, + "loss": 0.4396, + "step": 13370 + }, + { + "epoch": 5.050962627406569, + "grad_norm": 0.6071826219558716, + "learning_rate": 7.945944675895585e-05, + "loss": 0.4344, + "step": 13380 + }, + { + "epoch": 5.054737636844092, + "grad_norm": 0.5671345591545105, + "learning_rate": 7.942603679925671e-05, + "loss": 0.4225, + "step": 13390 + }, + { + "epoch": 5.058512646281616, + "grad_norm": 0.5134578943252563, + "learning_rate": 7.93926067271389e-05, + "loss": 0.4412, + "step": 13400 + }, + { + "epoch": 5.062287655719139, + "grad_norm": 0.5748825669288635, + "learning_rate": 7.935915656545155e-05, + "loss": 0.4246, + "step": 13410 + }, + { + "epoch": 5.066062665156663, + "grad_norm": 0.6098251938819885, + "learning_rate": 7.932568633705752e-05, + "loss": 0.4316, + "step": 13420 + }, + { + "epoch": 5.069837674594186, + "grad_norm": 0.5781272649765015, + "learning_rate": 7.929219606483341e-05, + "loss": 0.4375, + "step": 13430 + }, + { + "epoch": 5.07361268403171, + "grad_norm": 0.5721918940544128, + "learning_rate": 7.925868577166948e-05, + "loss": 0.4328, + "step": 13440 + }, + { + "epoch": 5.077387693469234, + "grad_norm": 0.6408784985542297, + "learning_rate": 7.922515548046974e-05, + "loss": 0.4351, + "step": 13450 + }, + { + "epoch": 5.081162702906758, + "grad_norm": 0.5710509419441223, + "learning_rate": 7.919160521415179e-05, + "loss": 0.4305, + "step": 13460 + }, + { + "epoch": 5.0849377123442805, + "grad_norm": 0.6249564290046692, + "learning_rate": 7.915803499564694e-05, + "loss": 0.4246, + "step": 13470 + }, + { + "epoch": 5.088712721781804, + "grad_norm": 0.6229961514472961, + "learning_rate": 7.912444484790013e-05, + "loss": 0.4307, + "step": 13480 + }, + { + "epoch": 5.092487731219328, + "grad_norm": 0.5835480093955994, + "learning_rate": 7.909083479386987e-05, + "loss": 0.4262, + "step": 13490 + }, + { + "epoch": 5.096262740656852, + "grad_norm": 0.6598435044288635, + "learning_rate": 7.905720485652836e-05, + "loss": 0.4326, + "step": 13500 + }, + { + "epoch": 5.100037750094375, + "grad_norm": 0.6232707500457764, + "learning_rate": 7.902355505886132e-05, + "loss": 0.4312, + "step": 13510 + }, + { + "epoch": 5.103812759531899, + "grad_norm": 0.5751008987426758, + "learning_rate": 7.898988542386805e-05, + "loss": 0.4318, + "step": 13520 + }, + { + "epoch": 5.107587768969422, + "grad_norm": 0.7185839414596558, + "learning_rate": 7.895619597456147e-05, + "loss": 0.431, + "step": 13530 + }, + { + "epoch": 5.111362778406946, + "grad_norm": 0.6064713597297668, + "learning_rate": 7.892248673396798e-05, + "loss": 0.4428, + "step": 13540 + }, + { + "epoch": 5.1151377878444695, + "grad_norm": 0.5585970282554626, + "learning_rate": 7.888875772512754e-05, + "loss": 0.4285, + "step": 13550 + }, + { + "epoch": 5.118912797281993, + "grad_norm": 0.590541660785675, + "learning_rate": 7.885500897109359e-05, + "loss": 0.4327, + "step": 13560 + }, + { + "epoch": 5.122687806719517, + "grad_norm": 0.6079216003417969, + "learning_rate": 7.882124049493309e-05, + "loss": 0.4219, + "step": 13570 + }, + { + "epoch": 5.126462816157041, + "grad_norm": 0.6529505848884583, + "learning_rate": 7.878745231972649e-05, + "loss": 0.4195, + "step": 13580 + }, + { + "epoch": 5.130237825594564, + "grad_norm": 0.6022513508796692, + "learning_rate": 7.875364446856766e-05, + "loss": 0.4318, + "step": 13590 + }, + { + "epoch": 5.134012835032087, + "grad_norm": 0.6363440752029419, + "learning_rate": 7.871981696456398e-05, + "loss": 0.4323, + "step": 13600 + }, + { + "epoch": 5.137787844469611, + "grad_norm": 0.6193153262138367, + "learning_rate": 7.868596983083623e-05, + "loss": 0.4285, + "step": 13610 + }, + { + "epoch": 5.141562853907135, + "grad_norm": 0.7710440158843994, + "learning_rate": 7.865210309051858e-05, + "loss": 0.4266, + "step": 13620 + }, + { + "epoch": 5.1453378633446585, + "grad_norm": 0.5872613787651062, + "learning_rate": 7.861821676675863e-05, + "loss": 0.4347, + "step": 13630 + }, + { + "epoch": 5.149112872782182, + "grad_norm": 0.536893367767334, + "learning_rate": 7.858431088271739e-05, + "loss": 0.4304, + "step": 13640 + }, + { + "epoch": 5.152887882219706, + "grad_norm": 0.5925413966178894, + "learning_rate": 7.855038546156918e-05, + "loss": 0.4284, + "step": 13650 + }, + { + "epoch": 5.156662891657229, + "grad_norm": 0.562650203704834, + "learning_rate": 7.851644052650173e-05, + "loss": 0.4268, + "step": 13660 + }, + { + "epoch": 5.1604379010947525, + "grad_norm": 0.5925480723381042, + "learning_rate": 7.848247610071609e-05, + "loss": 0.4251, + "step": 13670 + }, + { + "epoch": 5.164212910532276, + "grad_norm": 0.6896264553070068, + "learning_rate": 7.844849220742658e-05, + "loss": 0.43, + "step": 13680 + }, + { + "epoch": 5.1679879199698, + "grad_norm": 0.643406867980957, + "learning_rate": 7.841448886986092e-05, + "loss": 0.4298, + "step": 13690 + }, + { + "epoch": 5.171762929407324, + "grad_norm": 0.5790614485740662, + "learning_rate": 7.838046611126004e-05, + "loss": 0.4333, + "step": 13700 + }, + { + "epoch": 5.1755379388448475, + "grad_norm": 0.6230219602584839, + "learning_rate": 7.834642395487819e-05, + "loss": 0.432, + "step": 13710 + }, + { + "epoch": 5.17931294828237, + "grad_norm": 0.6495296955108643, + "learning_rate": 7.831236242398285e-05, + "loss": 0.4336, + "step": 13720 + }, + { + "epoch": 5.183087957719894, + "grad_norm": 0.5876701474189758, + "learning_rate": 7.827828154185477e-05, + "loss": 0.4362, + "step": 13730 + }, + { + "epoch": 5.186862967157418, + "grad_norm": 0.5457055568695068, + "learning_rate": 7.82441813317879e-05, + "loss": 0.43, + "step": 13740 + }, + { + "epoch": 5.1906379765949415, + "grad_norm": 0.618160605430603, + "learning_rate": 7.821006181708944e-05, + "loss": 0.4281, + "step": 13750 + }, + { + "epoch": 5.194412986032465, + "grad_norm": 0.5848748087882996, + "learning_rate": 7.81759230210797e-05, + "loss": 0.4397, + "step": 13760 + }, + { + "epoch": 5.198187995469989, + "grad_norm": 0.5916726589202881, + "learning_rate": 7.814176496709227e-05, + "loss": 0.4268, + "step": 13770 + }, + { + "epoch": 5.201963004907512, + "grad_norm": 0.5741629004478455, + "learning_rate": 7.810758767847385e-05, + "loss": 0.4286, + "step": 13780 + }, + { + "epoch": 5.205738014345036, + "grad_norm": 0.548261284828186, + "learning_rate": 7.807339117858427e-05, + "loss": 0.4242, + "step": 13790 + }, + { + "epoch": 5.209513023782559, + "grad_norm": 0.5949342846870422, + "learning_rate": 7.803917549079655e-05, + "loss": 0.4291, + "step": 13800 + }, + { + "epoch": 5.213288033220083, + "grad_norm": 0.611446738243103, + "learning_rate": 7.800494063849679e-05, + "loss": 0.4322, + "step": 13810 + }, + { + "epoch": 5.217063042657607, + "grad_norm": 0.5855134129524231, + "learning_rate": 7.797068664508416e-05, + "loss": 0.4329, + "step": 13820 + }, + { + "epoch": 5.2208380520951305, + "grad_norm": 0.615392804145813, + "learning_rate": 7.793641353397096e-05, + "loss": 0.4321, + "step": 13830 + }, + { + "epoch": 5.224613061532654, + "grad_norm": 0.6607218980789185, + "learning_rate": 7.790212132858253e-05, + "loss": 0.4314, + "step": 13840 + }, + { + "epoch": 5.228388070970177, + "grad_norm": 0.6020424365997314, + "learning_rate": 7.786781005235728e-05, + "loss": 0.4282, + "step": 13850 + }, + { + "epoch": 5.232163080407701, + "grad_norm": 0.5580222606658936, + "learning_rate": 7.783347972874662e-05, + "loss": 0.4303, + "step": 13860 + }, + { + "epoch": 5.235938089845225, + "grad_norm": 0.5938263535499573, + "learning_rate": 7.779913038121504e-05, + "loss": 0.4318, + "step": 13870 + }, + { + "epoch": 5.239713099282748, + "grad_norm": 0.5820572376251221, + "learning_rate": 7.776476203323997e-05, + "loss": 0.4244, + "step": 13880 + }, + { + "epoch": 5.243488108720272, + "grad_norm": 0.5716726779937744, + "learning_rate": 7.773037470831185e-05, + "loss": 0.4268, + "step": 13890 + }, + { + "epoch": 5.247263118157796, + "grad_norm": 0.6699029803276062, + "learning_rate": 7.76959684299341e-05, + "loss": 0.4448, + "step": 13900 + }, + { + "epoch": 5.251038127595319, + "grad_norm": 0.6544981002807617, + "learning_rate": 7.76615432216231e-05, + "loss": 0.4324, + "step": 13910 + }, + { + "epoch": 5.254813137032842, + "grad_norm": 0.6434966325759888, + "learning_rate": 7.762709910690811e-05, + "loss": 0.431, + "step": 13920 + }, + { + "epoch": 5.258588146470366, + "grad_norm": 0.6706520915031433, + "learning_rate": 7.759263610933141e-05, + "loss": 0.4239, + "step": 13930 + }, + { + "epoch": 5.26236315590789, + "grad_norm": 0.7404385209083557, + "learning_rate": 7.755815425244811e-05, + "loss": 0.4211, + "step": 13940 + }, + { + "epoch": 5.266138165345414, + "grad_norm": 0.750700831413269, + "learning_rate": 7.752365355982624e-05, + "loss": 0.4297, + "step": 13950 + }, + { + "epoch": 5.269913174782937, + "grad_norm": 0.6595298051834106, + "learning_rate": 7.748913405504668e-05, + "loss": 0.4395, + "step": 13960 + }, + { + "epoch": 5.27368818422046, + "grad_norm": 0.7262568473815918, + "learning_rate": 7.745459576170322e-05, + "loss": 0.4189, + "step": 13970 + }, + { + "epoch": 5.277463193657984, + "grad_norm": 0.6093210577964783, + "learning_rate": 7.742003870340242e-05, + "loss": 0.4403, + "step": 13980 + }, + { + "epoch": 5.281238203095508, + "grad_norm": 0.6888540387153625, + "learning_rate": 7.738546290376373e-05, + "loss": 0.4357, + "step": 13990 + }, + { + "epoch": 5.285013212533031, + "grad_norm": 0.9906258583068848, + "learning_rate": 7.735086838641937e-05, + "loss": 0.4655, + "step": 14000 + }, + { + "epoch": 5.288788221970555, + "grad_norm": 0.6373776197433472, + "learning_rate": 7.731625517501437e-05, + "loss": 0.4296, + "step": 14010 + }, + { + "epoch": 5.292563231408079, + "grad_norm": 0.5870838761329651, + "learning_rate": 7.728162329320655e-05, + "loss": 0.4318, + "step": 14020 + }, + { + "epoch": 5.296338240845602, + "grad_norm": 0.5722277164459229, + "learning_rate": 7.724697276466645e-05, + "loss": 0.426, + "step": 14030 + }, + { + "epoch": 5.300113250283125, + "grad_norm": 0.595984697341919, + "learning_rate": 7.721230361307738e-05, + "loss": 0.4287, + "step": 14040 + }, + { + "epoch": 5.303888259720649, + "grad_norm": 2.9559850692749023, + "learning_rate": 7.71776158621354e-05, + "loss": 0.4298, + "step": 14050 + }, + { + "epoch": 5.307663269158173, + "grad_norm": 0.5847383737564087, + "learning_rate": 7.714290953554925e-05, + "loss": 0.4304, + "step": 14060 + }, + { + "epoch": 5.311438278595697, + "grad_norm": 0.5809390544891357, + "learning_rate": 7.710818465704037e-05, + "loss": 0.4305, + "step": 14070 + }, + { + "epoch": 5.31521328803322, + "grad_norm": 0.5706303119659424, + "learning_rate": 7.707344125034288e-05, + "loss": 0.4256, + "step": 14080 + }, + { + "epoch": 5.318988297470744, + "grad_norm": 0.5360699892044067, + "learning_rate": 7.703867933920359e-05, + "loss": 0.4197, + "step": 14090 + }, + { + "epoch": 5.322763306908267, + "grad_norm": 0.5971339344978333, + "learning_rate": 7.700389894738194e-05, + "loss": 0.4181, + "step": 14100 + }, + { + "epoch": 5.326538316345791, + "grad_norm": 0.6393430233001709, + "learning_rate": 7.696910009864999e-05, + "loss": 0.4214, + "step": 14110 + }, + { + "epoch": 5.330313325783314, + "grad_norm": 0.5459288358688354, + "learning_rate": 7.693428281679241e-05, + "loss": 0.422, + "step": 14120 + }, + { + "epoch": 5.334088335220838, + "grad_norm": 0.516596794128418, + "learning_rate": 7.689944712560652e-05, + "loss": 0.4297, + "step": 14130 + }, + { + "epoch": 5.337863344658362, + "grad_norm": 0.5618916153907776, + "learning_rate": 7.686459304890214e-05, + "loss": 0.426, + "step": 14140 + }, + { + "epoch": 5.341638354095886, + "grad_norm": 0.6064931750297546, + "learning_rate": 7.682972061050175e-05, + "loss": 0.424, + "step": 14150 + }, + { + "epoch": 5.3454133635334085, + "grad_norm": 0.6051250696182251, + "learning_rate": 7.679482983424032e-05, + "loss": 0.4242, + "step": 14160 + }, + { + "epoch": 5.349188372970932, + "grad_norm": 0.6006349921226501, + "learning_rate": 7.675992074396534e-05, + "loss": 0.4295, + "step": 14170 + }, + { + "epoch": 5.352963382408456, + "grad_norm": 1.0653759241104126, + "learning_rate": 7.672499336353687e-05, + "loss": 0.4233, + "step": 14180 + }, + { + "epoch": 5.35673839184598, + "grad_norm": 0.6030228734016418, + "learning_rate": 7.669004771682744e-05, + "loss": 0.4259, + "step": 14190 + }, + { + "epoch": 5.360513401283503, + "grad_norm": 0.6684322357177734, + "learning_rate": 7.665508382772206e-05, + "loss": 0.4259, + "step": 14200 + }, + { + "epoch": 5.364288410721027, + "grad_norm": 0.6103203892707825, + "learning_rate": 7.662010172011824e-05, + "loss": 0.4218, + "step": 14210 + }, + { + "epoch": 5.36806342015855, + "grad_norm": 0.5854779481887817, + "learning_rate": 7.658510141792588e-05, + "loss": 0.4223, + "step": 14220 + }, + { + "epoch": 5.371838429596074, + "grad_norm": 0.6130976676940918, + "learning_rate": 7.65500829450674e-05, + "loss": 0.4235, + "step": 14230 + }, + { + "epoch": 5.3756134390335975, + "grad_norm": 0.5876962542533875, + "learning_rate": 7.651504632547759e-05, + "loss": 0.4262, + "step": 14240 + }, + { + "epoch": 5.379388448471121, + "grad_norm": 0.5816875100135803, + "learning_rate": 7.647999158310364e-05, + "loss": 0.4238, + "step": 14250 + }, + { + "epoch": 5.383163457908645, + "grad_norm": 0.6344910860061646, + "learning_rate": 7.644491874190512e-05, + "loss": 0.4277, + "step": 14260 + }, + { + "epoch": 5.386938467346169, + "grad_norm": 0.5912450551986694, + "learning_rate": 7.6409827825854e-05, + "loss": 0.4217, + "step": 14270 + }, + { + "epoch": 5.3907134767836915, + "grad_norm": 0.5558596849441528, + "learning_rate": 7.637471885893459e-05, + "loss": 0.4291, + "step": 14280 + }, + { + "epoch": 5.394488486221215, + "grad_norm": 0.6248513460159302, + "learning_rate": 7.633959186514354e-05, + "loss": 0.4281, + "step": 14290 + }, + { + "epoch": 5.398263495658739, + "grad_norm": 0.623848557472229, + "learning_rate": 7.630444686848984e-05, + "loss": 0.4189, + "step": 14300 + }, + { + "epoch": 5.402038505096263, + "grad_norm": 0.5933666825294495, + "learning_rate": 7.626928389299471e-05, + "loss": 0.4238, + "step": 14310 + }, + { + "epoch": 5.4058135145337864, + "grad_norm": 0.6335635781288147, + "learning_rate": 7.623410296269175e-05, + "loss": 0.4352, + "step": 14320 + }, + { + "epoch": 5.40958852397131, + "grad_norm": 0.5906968116760254, + "learning_rate": 7.61989041016268e-05, + "loss": 0.4217, + "step": 14330 + }, + { + "epoch": 5.413363533408834, + "grad_norm": 1.1786829233169556, + "learning_rate": 7.616368733385793e-05, + "loss": 0.4292, + "step": 14340 + }, + { + "epoch": 5.417138542846357, + "grad_norm": 0.6028378009796143, + "learning_rate": 7.612845268345547e-05, + "loss": 0.4245, + "step": 14350 + }, + { + "epoch": 5.4209135522838805, + "grad_norm": 1.1031107902526855, + "learning_rate": 7.609320017450199e-05, + "loss": 0.4235, + "step": 14360 + }, + { + "epoch": 5.424688561721404, + "grad_norm": 0.6411410570144653, + "learning_rate": 7.605792983109222e-05, + "loss": 0.4187, + "step": 14370 + }, + { + "epoch": 5.428463571158928, + "grad_norm": 0.547097384929657, + "learning_rate": 7.602264167733313e-05, + "loss": 0.4282, + "step": 14380 + }, + { + "epoch": 5.432238580596452, + "grad_norm": 0.6103999018669128, + "learning_rate": 7.598733573734384e-05, + "loss": 0.4256, + "step": 14390 + }, + { + "epoch": 5.436013590033975, + "grad_norm": 0.5741267800331116, + "learning_rate": 7.595201203525561e-05, + "loss": 0.4254, + "step": 14400 + }, + { + "epoch": 5.439788599471498, + "grad_norm": 0.5826230645179749, + "learning_rate": 7.591667059521187e-05, + "loss": 0.4279, + "step": 14410 + }, + { + "epoch": 5.443563608909022, + "grad_norm": 0.5484496355056763, + "learning_rate": 7.588131144136815e-05, + "loss": 0.429, + "step": 14420 + }, + { + "epoch": 5.447338618346546, + "grad_norm": 0.566040575504303, + "learning_rate": 7.584593459789212e-05, + "loss": 0.4213, + "step": 14430 + }, + { + "epoch": 5.4511136277840695, + "grad_norm": 0.5821641087532043, + "learning_rate": 7.58105400889635e-05, + "loss": 0.4327, + "step": 14440 + }, + { + "epoch": 5.454888637221593, + "grad_norm": 0.6450483798980713, + "learning_rate": 7.57751279387741e-05, + "loss": 0.4227, + "step": 14450 + }, + { + "epoch": 5.458663646659117, + "grad_norm": 0.5825949907302856, + "learning_rate": 7.573969817152782e-05, + "loss": 0.423, + "step": 14460 + }, + { + "epoch": 5.462438656096641, + "grad_norm": 0.5413886308670044, + "learning_rate": 7.570425081144052e-05, + "loss": 0.425, + "step": 14470 + }, + { + "epoch": 5.4662136655341635, + "grad_norm": 0.5597252249717712, + "learning_rate": 7.56687858827402e-05, + "loss": 0.4192, + "step": 14480 + }, + { + "epoch": 5.469988674971687, + "grad_norm": 0.5918856263160706, + "learning_rate": 7.563330340966675e-05, + "loss": 0.421, + "step": 14490 + }, + { + "epoch": 5.473763684409211, + "grad_norm": 0.574697732925415, + "learning_rate": 7.559780341647212e-05, + "loss": 0.4286, + "step": 14500 + }, + { + "epoch": 5.477538693846735, + "grad_norm": 0.7420601844787598, + "learning_rate": 7.556228592742026e-05, + "loss": 0.4226, + "step": 14510 + }, + { + "epoch": 5.4813137032842585, + "grad_norm": 0.8182733058929443, + "learning_rate": 7.552675096678696e-05, + "loss": 0.4199, + "step": 14520 + }, + { + "epoch": 5.485088712721781, + "grad_norm": 0.6243227124214172, + "learning_rate": 7.549119855886012e-05, + "loss": 0.4234, + "step": 14530 + }, + { + "epoch": 5.488863722159305, + "grad_norm": 0.5414767265319824, + "learning_rate": 7.545562872793941e-05, + "loss": 0.4243, + "step": 14540 + }, + { + "epoch": 5.492638731596829, + "grad_norm": 0.6131302714347839, + "learning_rate": 7.542004149833648e-05, + "loss": 0.4277, + "step": 14550 + }, + { + "epoch": 5.4964137410343525, + "grad_norm": 0.6857340335845947, + "learning_rate": 7.538443689437492e-05, + "loss": 0.4259, + "step": 14560 + }, + { + "epoch": 5.500188750471876, + "grad_norm": 0.5744028091430664, + "learning_rate": 7.53488149403901e-05, + "loss": 0.4208, + "step": 14570 + }, + { + "epoch": 5.5039637599094, + "grad_norm": 0.602278470993042, + "learning_rate": 7.531317566072929e-05, + "loss": 0.4211, + "step": 14580 + }, + { + "epoch": 5.507738769346924, + "grad_norm": 0.6976267695426941, + "learning_rate": 7.527751907975158e-05, + "loss": 0.4207, + "step": 14590 + }, + { + "epoch": 5.511513778784447, + "grad_norm": 0.5509871244430542, + "learning_rate": 7.524184522182793e-05, + "loss": 0.4165, + "step": 14600 + }, + { + "epoch": 5.51528878822197, + "grad_norm": 0.6053850650787354, + "learning_rate": 7.520615411134112e-05, + "loss": 0.4168, + "step": 14610 + }, + { + "epoch": 5.519063797659494, + "grad_norm": 0.592635989189148, + "learning_rate": 7.517044577268564e-05, + "loss": 0.4263, + "step": 14620 + }, + { + "epoch": 5.522838807097018, + "grad_norm": 0.6511247158050537, + "learning_rate": 7.513472023026782e-05, + "loss": 0.4141, + "step": 14630 + }, + { + "epoch": 5.5266138165345415, + "grad_norm": 0.5748429298400879, + "learning_rate": 7.509897750850572e-05, + "loss": 0.4176, + "step": 14640 + }, + { + "epoch": 5.530388825972065, + "grad_norm": 0.6025387048721313, + "learning_rate": 7.506321763182918e-05, + "loss": 0.4144, + "step": 14650 + }, + { + "epoch": 5.534163835409588, + "grad_norm": 0.6291730999946594, + "learning_rate": 7.50274406246797e-05, + "loss": 0.4273, + "step": 14660 + }, + { + "epoch": 5.537938844847112, + "grad_norm": 0.8842790722846985, + "learning_rate": 7.499164651151056e-05, + "loss": 0.4211, + "step": 14670 + }, + { + "epoch": 5.541713854284636, + "grad_norm": 0.6468051075935364, + "learning_rate": 7.495583531678669e-05, + "loss": 0.4204, + "step": 14680 + }, + { + "epoch": 5.545488863722159, + "grad_norm": 0.649259090423584, + "learning_rate": 7.492000706498469e-05, + "loss": 0.4179, + "step": 14690 + }, + { + "epoch": 5.549263873159683, + "grad_norm": 0.5695837140083313, + "learning_rate": 7.488416178059284e-05, + "loss": 0.4255, + "step": 14700 + }, + { + "epoch": 5.553038882597207, + "grad_norm": 0.66133713722229, + "learning_rate": 7.484829948811107e-05, + "loss": 0.4263, + "step": 14710 + }, + { + "epoch": 5.5568138920347305, + "grad_norm": 0.6037610173225403, + "learning_rate": 7.48124202120509e-05, + "loss": 0.4221, + "step": 14720 + }, + { + "epoch": 5.560588901472253, + "grad_norm": 0.6156513094902039, + "learning_rate": 7.477652397693549e-05, + "loss": 0.4173, + "step": 14730 + }, + { + "epoch": 5.564363910909777, + "grad_norm": 0.5722167491912842, + "learning_rate": 7.474061080729955e-05, + "loss": 0.4218, + "step": 14740 + }, + { + "epoch": 5.568138920347301, + "grad_norm": 1.3760823011398315, + "learning_rate": 7.470468072768941e-05, + "loss": 0.4296, + "step": 14750 + }, + { + "epoch": 5.571913929784825, + "grad_norm": 0.7008207440376282, + "learning_rate": 7.466873376266297e-05, + "loss": 0.4214, + "step": 14760 + }, + { + "epoch": 5.575688939222348, + "grad_norm": 0.5777148604393005, + "learning_rate": 7.46327699367896e-05, + "loss": 0.4208, + "step": 14770 + }, + { + "epoch": 5.579463948659871, + "grad_norm": 0.6137166023254395, + "learning_rate": 7.459678927465026e-05, + "loss": 0.4207, + "step": 14780 + }, + { + "epoch": 5.583238958097395, + "grad_norm": 0.6016892194747925, + "learning_rate": 7.456079180083737e-05, + "loss": 0.42, + "step": 14790 + }, + { + "epoch": 5.587013967534919, + "grad_norm": 0.6223365068435669, + "learning_rate": 7.452477753995489e-05, + "loss": 0.4188, + "step": 14800 + }, + { + "epoch": 5.590788976972442, + "grad_norm": 0.5763020515441895, + "learning_rate": 7.448874651661823e-05, + "loss": 0.4192, + "step": 14810 + }, + { + "epoch": 5.594563986409966, + "grad_norm": 0.5202614665031433, + "learning_rate": 7.445269875545423e-05, + "loss": 0.4213, + "step": 14820 + }, + { + "epoch": 5.59833899584749, + "grad_norm": 0.6435422301292419, + "learning_rate": 7.44166342811012e-05, + "loss": 0.4166, + "step": 14830 + }, + { + "epoch": 5.602114005285014, + "grad_norm": 0.6565425992012024, + "learning_rate": 7.438055311820886e-05, + "loss": 0.42, + "step": 14840 + }, + { + "epoch": 5.605889014722536, + "grad_norm": 0.5886382460594177, + "learning_rate": 7.434445529143837e-05, + "loss": 0.4143, + "step": 14850 + }, + { + "epoch": 5.60966402416006, + "grad_norm": 0.5876275897026062, + "learning_rate": 7.430834082546225e-05, + "loss": 0.4178, + "step": 14860 + }, + { + "epoch": 5.613439033597584, + "grad_norm": 0.5644301772117615, + "learning_rate": 7.427220974496438e-05, + "loss": 0.4271, + "step": 14870 + }, + { + "epoch": 5.617214043035108, + "grad_norm": 0.6446945667266846, + "learning_rate": 7.423606207464005e-05, + "loss": 0.4173, + "step": 14880 + }, + { + "epoch": 5.620989052472631, + "grad_norm": 0.6406450271606445, + "learning_rate": 7.419989783919578e-05, + "loss": 0.4194, + "step": 14890 + }, + { + "epoch": 5.624764061910155, + "grad_norm": 0.6164763569831848, + "learning_rate": 7.416371706334956e-05, + "loss": 0.4221, + "step": 14900 + }, + { + "epoch": 5.628539071347678, + "grad_norm": 0.6049771904945374, + "learning_rate": 7.412751977183056e-05, + "loss": 0.4181, + "step": 14910 + }, + { + "epoch": 5.632314080785202, + "grad_norm": 0.6096076965332031, + "learning_rate": 7.409130598937932e-05, + "loss": 0.4198, + "step": 14920 + }, + { + "epoch": 5.636089090222725, + "grad_norm": 0.5442625880241394, + "learning_rate": 7.40550757407476e-05, + "loss": 0.4373, + "step": 14930 + }, + { + "epoch": 5.639864099660249, + "grad_norm": 0.6001285910606384, + "learning_rate": 7.401882905069843e-05, + "loss": 0.4248, + "step": 14940 + }, + { + "epoch": 5.643639109097773, + "grad_norm": 0.60331791639328, + "learning_rate": 7.39825659440061e-05, + "loss": 0.4246, + "step": 14950 + }, + { + "epoch": 5.647414118535297, + "grad_norm": 0.5998172760009766, + "learning_rate": 7.394628644545609e-05, + "loss": 0.4225, + "step": 14960 + }, + { + "epoch": 5.65118912797282, + "grad_norm": 0.6505115628242493, + "learning_rate": 7.390999057984507e-05, + "loss": 0.4245, + "step": 14970 + }, + { + "epoch": 5.654964137410343, + "grad_norm": 0.6152286529541016, + "learning_rate": 7.387367837198097e-05, + "loss": 0.4314, + "step": 14980 + }, + { + "epoch": 5.658739146847867, + "grad_norm": 0.595940113067627, + "learning_rate": 7.383734984668281e-05, + "loss": 0.4192, + "step": 14990 + }, + { + "epoch": 5.662514156285391, + "grad_norm": 0.6145548820495605, + "learning_rate": 7.38010050287808e-05, + "loss": 0.414, + "step": 15000 + }, + { + "epoch": 5.666289165722914, + "grad_norm": 1.0223385095596313, + "learning_rate": 7.376464394311628e-05, + "loss": 0.4289, + "step": 15010 + }, + { + "epoch": 5.670064175160438, + "grad_norm": 0.6325558423995972, + "learning_rate": 7.372826661454172e-05, + "loss": 0.4227, + "step": 15020 + }, + { + "epoch": 5.673839184597962, + "grad_norm": 0.6463314294815063, + "learning_rate": 7.369187306792068e-05, + "loss": 0.4183, + "step": 15030 + }, + { + "epoch": 5.677614194035485, + "grad_norm": 0.6034806370735168, + "learning_rate": 7.365546332812779e-05, + "loss": 0.4133, + "step": 15040 + }, + { + "epoch": 5.6813892034730085, + "grad_norm": 0.5699142813682556, + "learning_rate": 7.361903742004876e-05, + "loss": 0.4185, + "step": 15050 + }, + { + "epoch": 5.685164212910532, + "grad_norm": 0.6562982797622681, + "learning_rate": 7.358259536858039e-05, + "loss": 0.4136, + "step": 15060 + }, + { + "epoch": 5.688939222348056, + "grad_norm": 0.691800594329834, + "learning_rate": 7.354613719863044e-05, + "loss": 0.4263, + "step": 15070 + }, + { + "epoch": 5.69271423178558, + "grad_norm": 0.6193991303443909, + "learning_rate": 7.350966293511776e-05, + "loss": 0.422, + "step": 15080 + }, + { + "epoch": 5.696489241223103, + "grad_norm": 0.6109247803688049, + "learning_rate": 7.347317260297212e-05, + "loss": 0.4272, + "step": 15090 + }, + { + "epoch": 5.700264250660627, + "grad_norm": 0.5521445274353027, + "learning_rate": 7.343666622713437e-05, + "loss": 0.4142, + "step": 15100 + }, + { + "epoch": 5.70403926009815, + "grad_norm": 0.5884120464324951, + "learning_rate": 7.340014383255624e-05, + "loss": 0.4103, + "step": 15110 + }, + { + "epoch": 5.707814269535674, + "grad_norm": 0.7118320465087891, + "learning_rate": 7.336360544420044e-05, + "loss": 0.4234, + "step": 15120 + }, + { + "epoch": 5.7115892789731975, + "grad_norm": 0.6134358644485474, + "learning_rate": 7.332705108704064e-05, + "loss": 0.4077, + "step": 15130 + }, + { + "epoch": 5.715364288410721, + "grad_norm": 0.5904216170310974, + "learning_rate": 7.329048078606138e-05, + "loss": 0.4202, + "step": 15140 + }, + { + "epoch": 5.719139297848245, + "grad_norm": 0.6160265803337097, + "learning_rate": 7.32538945662581e-05, + "loss": 0.4229, + "step": 15150 + }, + { + "epoch": 5.722914307285768, + "grad_norm": 0.6015536189079285, + "learning_rate": 7.321729245263718e-05, + "loss": 0.458, + "step": 15160 + }, + { + "epoch": 5.7266893167232915, + "grad_norm": 0.5921651721000671, + "learning_rate": 7.318067447021578e-05, + "loss": 0.42, + "step": 15170 + }, + { + "epoch": 5.730464326160815, + "grad_norm": 0.5869306921958923, + "learning_rate": 7.314404064402198e-05, + "loss": 0.4119, + "step": 15180 + }, + { + "epoch": 5.734239335598339, + "grad_norm": 0.5936274528503418, + "learning_rate": 7.310739099909461e-05, + "loss": 0.4105, + "step": 15190 + }, + { + "epoch": 5.738014345035863, + "grad_norm": 0.6002879738807678, + "learning_rate": 7.307072556048339e-05, + "loss": 0.4178, + "step": 15200 + }, + { + "epoch": 5.7417893544733865, + "grad_norm": 0.6081198453903198, + "learning_rate": 7.30340443532488e-05, + "loss": 0.4202, + "step": 15210 + }, + { + "epoch": 5.74556436391091, + "grad_norm": 0.6480271816253662, + "learning_rate": 7.299734740246208e-05, + "loss": 0.4105, + "step": 15220 + }, + { + "epoch": 5.749339373348433, + "grad_norm": 0.5739256739616394, + "learning_rate": 7.296063473320528e-05, + "loss": 0.4184, + "step": 15230 + }, + { + "epoch": 5.753114382785957, + "grad_norm": 0.749381422996521, + "learning_rate": 7.292390637057113e-05, + "loss": 0.4158, + "step": 15240 + }, + { + "epoch": 5.7568893922234805, + "grad_norm": 0.6204841732978821, + "learning_rate": 7.288716233966314e-05, + "loss": 0.4133, + "step": 15250 + }, + { + "epoch": 5.760664401661004, + "grad_norm": 1.603737235069275, + "learning_rate": 7.285040266559551e-05, + "loss": 0.4199, + "step": 15260 + }, + { + "epoch": 5.764439411098528, + "grad_norm": 0.6340755820274353, + "learning_rate": 7.281362737349312e-05, + "loss": 0.4139, + "step": 15270 + }, + { + "epoch": 5.768214420536052, + "grad_norm": 0.6750447154045105, + "learning_rate": 7.277683648849153e-05, + "loss": 0.4176, + "step": 15280 + }, + { + "epoch": 5.771989429973575, + "grad_norm": 0.5785810351371765, + "learning_rate": 7.2740030035737e-05, + "loss": 0.4137, + "step": 15290 + }, + { + "epoch": 5.775764439411098, + "grad_norm": 0.5631486773490906, + "learning_rate": 7.270320804038634e-05, + "loss": 0.4183, + "step": 15300 + }, + { + "epoch": 5.779539448848622, + "grad_norm": 0.6204792857170105, + "learning_rate": 7.266637052760708e-05, + "loss": 0.4185, + "step": 15310 + }, + { + "epoch": 5.783314458286146, + "grad_norm": 0.6543823480606079, + "learning_rate": 7.262951752257728e-05, + "loss": 0.4132, + "step": 15320 + }, + { + "epoch": 5.7870894677236695, + "grad_norm": 0.6588866114616394, + "learning_rate": 7.259264905048564e-05, + "loss": 0.4141, + "step": 15330 + }, + { + "epoch": 5.790864477161193, + "grad_norm": 0.5547715425491333, + "learning_rate": 7.255576513653142e-05, + "loss": 0.4132, + "step": 15340 + }, + { + "epoch": 5.794639486598717, + "grad_norm": 0.642467200756073, + "learning_rate": 7.251886580592439e-05, + "loss": 0.4144, + "step": 15350 + }, + { + "epoch": 5.79841449603624, + "grad_norm": 0.6495168209075928, + "learning_rate": 7.248195108388496e-05, + "loss": 0.4233, + "step": 15360 + }, + { + "epoch": 5.8021895054737636, + "grad_norm": 0.6559742093086243, + "learning_rate": 7.244502099564395e-05, + "loss": 0.415, + "step": 15370 + }, + { + "epoch": 5.805964514911287, + "grad_norm": 0.6655861139297485, + "learning_rate": 7.240807556644271e-05, + "loss": 0.4143, + "step": 15380 + }, + { + "epoch": 5.809739524348811, + "grad_norm": 0.5971380472183228, + "learning_rate": 7.237111482153314e-05, + "loss": 0.4144, + "step": 15390 + }, + { + "epoch": 5.813514533786335, + "grad_norm": 0.6410373449325562, + "learning_rate": 7.233413878617751e-05, + "loss": 0.4212, + "step": 15400 + }, + { + "epoch": 5.817289543223858, + "grad_norm": 0.5639101266860962, + "learning_rate": 7.229714748564864e-05, + "loss": 0.4118, + "step": 15410 + }, + { + "epoch": 5.821064552661381, + "grad_norm": 0.631568193435669, + "learning_rate": 7.22601409452297e-05, + "loss": 0.4152, + "step": 15420 + }, + { + "epoch": 5.824839562098905, + "grad_norm": 0.6407223343849182, + "learning_rate": 7.222311919021433e-05, + "loss": 0.4209, + "step": 15430 + }, + { + "epoch": 5.828614571536429, + "grad_norm": 0.8518263101577759, + "learning_rate": 7.218608224590655e-05, + "loss": 0.4134, + "step": 15440 + }, + { + "epoch": 5.8323895809739525, + "grad_norm": 0.6463886499404907, + "learning_rate": 7.214903013762074e-05, + "loss": 0.4219, + "step": 15450 + }, + { + "epoch": 5.836164590411476, + "grad_norm": 0.6677048206329346, + "learning_rate": 7.21119628906817e-05, + "loss": 0.4104, + "step": 15460 + }, + { + "epoch": 5.839939599849, + "grad_norm": 0.6562634706497192, + "learning_rate": 7.207488053042454e-05, + "loss": 0.4146, + "step": 15470 + }, + { + "epoch": 5.843714609286523, + "grad_norm": 0.6001513600349426, + "learning_rate": 7.203778308219467e-05, + "loss": 0.4196, + "step": 15480 + }, + { + "epoch": 5.847489618724047, + "grad_norm": 0.571530818939209, + "learning_rate": 7.200067057134787e-05, + "loss": 0.4142, + "step": 15490 + }, + { + "epoch": 5.85126462816157, + "grad_norm": 0.6734705567359924, + "learning_rate": 7.196354302325019e-05, + "loss": 0.4137, + "step": 15500 + }, + { + "epoch": 5.855039637599094, + "grad_norm": 0.7626957297325134, + "learning_rate": 7.192640046327795e-05, + "loss": 0.4193, + "step": 15510 + }, + { + "epoch": 5.858814647036618, + "grad_norm": 0.589971125125885, + "learning_rate": 7.188924291681777e-05, + "loss": 0.4144, + "step": 15520 + }, + { + "epoch": 5.8625896564741415, + "grad_norm": 0.6109906435012817, + "learning_rate": 7.185207040926643e-05, + "loss": 0.4154, + "step": 15530 + }, + { + "epoch": 5.866364665911664, + "grad_norm": 0.6139832139015198, + "learning_rate": 7.181488296603103e-05, + "loss": 0.415, + "step": 15540 + }, + { + "epoch": 5.870139675349188, + "grad_norm": 0.5642750263214111, + "learning_rate": 7.177768061252885e-05, + "loss": 0.4155, + "step": 15550 + }, + { + "epoch": 5.873914684786712, + "grad_norm": 0.6058228015899658, + "learning_rate": 7.174046337418729e-05, + "loss": 0.4193, + "step": 15560 + }, + { + "epoch": 5.877689694224236, + "grad_norm": 0.6035988926887512, + "learning_rate": 7.170323127644403e-05, + "loss": 0.4152, + "step": 15570 + }, + { + "epoch": 5.881464703661759, + "grad_norm": 0.5951765179634094, + "learning_rate": 7.166598434474683e-05, + "loss": 0.4125, + "step": 15580 + }, + { + "epoch": 5.885239713099283, + "grad_norm": 0.6099646091461182, + "learning_rate": 7.162872260455364e-05, + "loss": 0.4121, + "step": 15590 + }, + { + "epoch": 5.889014722536807, + "grad_norm": 0.6386427879333496, + "learning_rate": 7.159144608133248e-05, + "loss": 0.4079, + "step": 15600 + }, + { + "epoch": 5.89278973197433, + "grad_norm": 0.6007016897201538, + "learning_rate": 7.155415480056153e-05, + "loss": 0.411, + "step": 15610 + }, + { + "epoch": 5.896564741411853, + "grad_norm": 0.6118659973144531, + "learning_rate": 7.151684878772902e-05, + "loss": 0.4271, + "step": 15620 + }, + { + "epoch": 5.900339750849377, + "grad_norm": 0.6127933859825134, + "learning_rate": 7.147952806833324e-05, + "loss": 0.4097, + "step": 15630 + }, + { + "epoch": 5.904114760286901, + "grad_norm": 0.6609582304954529, + "learning_rate": 7.14421926678826e-05, + "loss": 0.4102, + "step": 15640 + }, + { + "epoch": 5.907889769724425, + "grad_norm": 0.6080712080001831, + "learning_rate": 7.140484261189543e-05, + "loss": 0.4141, + "step": 15650 + }, + { + "epoch": 5.911664779161947, + "grad_norm": 0.621368408203125, + "learning_rate": 7.136747792590017e-05, + "loss": 0.4098, + "step": 15660 + }, + { + "epoch": 5.915439788599471, + "grad_norm": 0.6989858746528625, + "learning_rate": 7.133009863543524e-05, + "loss": 0.4273, + "step": 15670 + }, + { + "epoch": 5.919214798036995, + "grad_norm": 0.5902137160301208, + "learning_rate": 7.129270476604901e-05, + "loss": 0.4121, + "step": 15680 + }, + { + "epoch": 5.922989807474519, + "grad_norm": 0.6309996843338013, + "learning_rate": 7.125529634329988e-05, + "loss": 0.4136, + "step": 15690 + }, + { + "epoch": 5.926764816912042, + "grad_norm": 0.6389601230621338, + "learning_rate": 7.12178733927561e-05, + "loss": 0.4177, + "step": 15700 + }, + { + "epoch": 5.930539826349566, + "grad_norm": 0.5543851852416992, + "learning_rate": 7.118043593999593e-05, + "loss": 0.4172, + "step": 15710 + }, + { + "epoch": 5.93431483578709, + "grad_norm": 0.5611990094184875, + "learning_rate": 7.114298401060752e-05, + "loss": 0.4153, + "step": 15720 + }, + { + "epoch": 5.938089845224613, + "grad_norm": 0.6465316414833069, + "learning_rate": 7.11055176301889e-05, + "loss": 0.4102, + "step": 15730 + }, + { + "epoch": 5.941864854662136, + "grad_norm": 0.6097255349159241, + "learning_rate": 7.1068036824348e-05, + "loss": 0.4169, + "step": 15740 + }, + { + "epoch": 5.94563986409966, + "grad_norm": 0.5505424737930298, + "learning_rate": 7.10305416187026e-05, + "loss": 0.4134, + "step": 15750 + }, + { + "epoch": 5.949414873537184, + "grad_norm": 0.554263710975647, + "learning_rate": 7.099303203888029e-05, + "loss": 0.4127, + "step": 15760 + }, + { + "epoch": 5.953189882974708, + "grad_norm": 0.6733128428459167, + "learning_rate": 7.095550811051855e-05, + "loss": 0.4059, + "step": 15770 + }, + { + "epoch": 5.956964892412231, + "grad_norm": 0.6691094040870667, + "learning_rate": 7.09179698592646e-05, + "loss": 0.416, + "step": 15780 + }, + { + "epoch": 5.960739901849754, + "grad_norm": 0.581087052822113, + "learning_rate": 7.088041731077551e-05, + "loss": 0.4081, + "step": 15790 + }, + { + "epoch": 5.964514911287278, + "grad_norm": 0.6719875931739807, + "learning_rate": 7.084285049071806e-05, + "loss": 0.4086, + "step": 15800 + }, + { + "epoch": 5.968289920724802, + "grad_norm": 0.6516507267951965, + "learning_rate": 7.080526942476886e-05, + "loss": 0.4166, + "step": 15810 + }, + { + "epoch": 5.972064930162325, + "grad_norm": 0.6685953140258789, + "learning_rate": 7.076767413861418e-05, + "loss": 0.4253, + "step": 15820 + }, + { + "epoch": 5.975839939599849, + "grad_norm": 0.6204284429550171, + "learning_rate": 7.073006465795005e-05, + "loss": 0.4134, + "step": 15830 + }, + { + "epoch": 5.979614949037373, + "grad_norm": 0.5573307275772095, + "learning_rate": 7.06924410084822e-05, + "loss": 0.4112, + "step": 15840 + }, + { + "epoch": 5.983389958474897, + "grad_norm": 0.5689889192581177, + "learning_rate": 7.065480321592604e-05, + "loss": 0.4107, + "step": 15850 + }, + { + "epoch": 5.9871649679124195, + "grad_norm": 0.6300421953201294, + "learning_rate": 7.061715130600663e-05, + "loss": 0.4019, + "step": 15860 + }, + { + "epoch": 5.990939977349943, + "grad_norm": 0.5939746499061584, + "learning_rate": 7.057948530445873e-05, + "loss": 0.4093, + "step": 15870 + }, + { + "epoch": 5.994714986787467, + "grad_norm": 0.576476514339447, + "learning_rate": 7.054180523702668e-05, + "loss": 0.4084, + "step": 15880 + }, + { + "epoch": 5.998489996224991, + "grad_norm": 0.6696039438247681, + "learning_rate": 7.050411112946442e-05, + "loss": 0.412, + "step": 15890 + }, + { + "epoch": 6.002265005662514, + "grad_norm": 0.6292182803153992, + "learning_rate": 7.046640300753557e-05, + "loss": 0.4151, + "step": 15900 + }, + { + "epoch": 6.006040015100038, + "grad_norm": 0.5818023085594177, + "learning_rate": 7.042868089701325e-05, + "loss": 0.409, + "step": 15910 + }, + { + "epoch": 6.009815024537561, + "grad_norm": 0.601872444152832, + "learning_rate": 7.039094482368016e-05, + "loss": 0.4074, + "step": 15920 + }, + { + "epoch": 6.013590033975085, + "grad_norm": 0.629568874835968, + "learning_rate": 7.035319481332858e-05, + "loss": 0.4241, + "step": 15930 + }, + { + "epoch": 6.0173650434126085, + "grad_norm": 0.6180824637413025, + "learning_rate": 7.031543089176023e-05, + "loss": 0.413, + "step": 15940 + }, + { + "epoch": 6.021140052850132, + "grad_norm": 0.7090234160423279, + "learning_rate": 7.027765308478644e-05, + "loss": 0.4159, + "step": 15950 + }, + { + "epoch": 6.024915062287656, + "grad_norm": 0.8287419676780701, + "learning_rate": 7.023986141822798e-05, + "loss": 0.4117, + "step": 15960 + }, + { + "epoch": 6.02869007172518, + "grad_norm": 0.6538956761360168, + "learning_rate": 7.02020559179151e-05, + "loss": 0.4134, + "step": 15970 + }, + { + "epoch": 6.0324650811627025, + "grad_norm": 0.6877662539482117, + "learning_rate": 7.016423660968748e-05, + "loss": 0.4132, + "step": 15980 + }, + { + "epoch": 6.036240090600226, + "grad_norm": 0.9112296104431152, + "learning_rate": 7.012640351939428e-05, + "loss": 0.4225, + "step": 15990 + }, + { + "epoch": 6.04001510003775, + "grad_norm": 0.6331153512001038, + "learning_rate": 7.008855667289404e-05, + "loss": 0.4109, + "step": 16000 + }, + { + "epoch": 6.043790109475274, + "grad_norm": 0.6433661580085754, + "learning_rate": 7.005069609605476e-05, + "loss": 0.4129, + "step": 16010 + }, + { + "epoch": 6.0475651189127975, + "grad_norm": 0.5703040361404419, + "learning_rate": 7.001282181475377e-05, + "loss": 0.422, + "step": 16020 + }, + { + "epoch": 6.051340128350321, + "grad_norm": 0.5761269330978394, + "learning_rate": 6.997493385487775e-05, + "loss": 0.4079, + "step": 16030 + }, + { + "epoch": 6.055115137787844, + "grad_norm": 0.6320202350616455, + "learning_rate": 6.99370322423228e-05, + "loss": 0.405, + "step": 16040 + }, + { + "epoch": 6.058890147225368, + "grad_norm": 0.5953516960144043, + "learning_rate": 6.989911700299433e-05, + "loss": 0.4178, + "step": 16050 + }, + { + "epoch": 6.0626651566628915, + "grad_norm": 0.6841408014297485, + "learning_rate": 6.9861188162807e-05, + "loss": 0.4176, + "step": 16060 + }, + { + "epoch": 6.066440166100415, + "grad_norm": 0.6144746541976929, + "learning_rate": 6.982324574768487e-05, + "loss": 0.4184, + "step": 16070 + }, + { + "epoch": 6.070215175537939, + "grad_norm": 0.6368954181671143, + "learning_rate": 6.978528978356117e-05, + "loss": 0.4169, + "step": 16080 + }, + { + "epoch": 6.073990184975463, + "grad_norm": 0.6184800863265991, + "learning_rate": 6.974732029637846e-05, + "loss": 0.4123, + "step": 16090 + }, + { + "epoch": 6.0777651944129865, + "grad_norm": 0.6910690665245056, + "learning_rate": 6.970933731208855e-05, + "loss": 0.413, + "step": 16100 + }, + { + "epoch": 6.081540203850509, + "grad_norm": 0.5969312787055969, + "learning_rate": 6.967134085665244e-05, + "loss": 0.4145, + "step": 16110 + }, + { + "epoch": 6.085315213288033, + "grad_norm": 0.5405552983283997, + "learning_rate": 6.963333095604034e-05, + "loss": 0.4136, + "step": 16120 + }, + { + "epoch": 6.089090222725557, + "grad_norm": 0.5933461785316467, + "learning_rate": 6.959530763623166e-05, + "loss": 0.4144, + "step": 16130 + }, + { + "epoch": 6.0928652321630805, + "grad_norm": 0.6700628995895386, + "learning_rate": 6.955727092321497e-05, + "loss": 0.4092, + "step": 16140 + }, + { + "epoch": 6.096640241600604, + "grad_norm": 0.6101819276809692, + "learning_rate": 6.951922084298803e-05, + "loss": 0.418, + "step": 16150 + }, + { + "epoch": 6.100415251038128, + "grad_norm": 0.5824779272079468, + "learning_rate": 6.948115742155769e-05, + "loss": 0.4169, + "step": 16160 + }, + { + "epoch": 6.104190260475651, + "grad_norm": 0.5644822716712952, + "learning_rate": 6.944308068493996e-05, + "loss": 0.42, + "step": 16170 + }, + { + "epoch": 6.107965269913175, + "grad_norm": 0.6036456823348999, + "learning_rate": 6.940499065915992e-05, + "loss": 0.4083, + "step": 16180 + }, + { + "epoch": 6.111740279350698, + "grad_norm": 0.5669860243797302, + "learning_rate": 6.936688737025173e-05, + "loss": 0.4195, + "step": 16190 + }, + { + "epoch": 6.115515288788222, + "grad_norm": 0.5517399311065674, + "learning_rate": 6.932877084425867e-05, + "loss": 0.414, + "step": 16200 + }, + { + "epoch": 6.119290298225746, + "grad_norm": 0.5780723094940186, + "learning_rate": 6.929064110723297e-05, + "loss": 0.4133, + "step": 16210 + }, + { + "epoch": 6.1230653076632695, + "grad_norm": 0.5744929909706116, + "learning_rate": 6.925249818523598e-05, + "loss": 0.4055, + "step": 16220 + }, + { + "epoch": 6.126840317100792, + "grad_norm": 0.5747778415679932, + "learning_rate": 6.921434210433801e-05, + "loss": 0.4068, + "step": 16230 + }, + { + "epoch": 6.130615326538316, + "grad_norm": 0.6638509631156921, + "learning_rate": 6.917617289061841e-05, + "loss": 0.4097, + "step": 16240 + }, + { + "epoch": 6.13439033597584, + "grad_norm": 0.6439339518547058, + "learning_rate": 6.913799057016547e-05, + "loss": 0.4181, + "step": 16250 + }, + { + "epoch": 6.138165345413364, + "grad_norm": 1.5434315204620361, + "learning_rate": 6.909979516907641e-05, + "loss": 0.4165, + "step": 16260 + }, + { + "epoch": 6.141940354850887, + "grad_norm": 0.7095974683761597, + "learning_rate": 6.906158671345746e-05, + "loss": 0.4082, + "step": 16270 + }, + { + "epoch": 6.145715364288411, + "grad_norm": 0.5394503474235535, + "learning_rate": 6.902336522942374e-05, + "loss": 0.4141, + "step": 16280 + }, + { + "epoch": 6.149490373725934, + "grad_norm": 1.054152011871338, + "learning_rate": 6.898513074309924e-05, + "loss": 0.4109, + "step": 16290 + }, + { + "epoch": 6.153265383163458, + "grad_norm": 0.6516214609146118, + "learning_rate": 6.894688328061693e-05, + "loss": 0.4187, + "step": 16300 + }, + { + "epoch": 6.157040392600981, + "grad_norm": 0.5728663802146912, + "learning_rate": 6.890862286811853e-05, + "loss": 0.4231, + "step": 16310 + }, + { + "epoch": 6.160815402038505, + "grad_norm": 0.6336321830749512, + "learning_rate": 6.88703495317547e-05, + "loss": 0.4066, + "step": 16320 + }, + { + "epoch": 6.164590411476029, + "grad_norm": 0.6610395908355713, + "learning_rate": 6.883206329768492e-05, + "loss": 0.4141, + "step": 16330 + }, + { + "epoch": 6.1683654209135526, + "grad_norm": 0.6659107208251953, + "learning_rate": 6.879376419207743e-05, + "loss": 0.4055, + "step": 16340 + }, + { + "epoch": 6.172140430351076, + "grad_norm": 0.6790131330490112, + "learning_rate": 6.875545224110935e-05, + "loss": 0.4178, + "step": 16350 + }, + { + "epoch": 6.175915439788599, + "grad_norm": 0.5676288604736328, + "learning_rate": 6.871712747096651e-05, + "loss": 0.4282, + "step": 16360 + }, + { + "epoch": 6.179690449226123, + "grad_norm": 0.6402314901351929, + "learning_rate": 6.867878990784353e-05, + "loss": 0.4146, + "step": 16370 + }, + { + "epoch": 6.183465458663647, + "grad_norm": 0.6174314022064209, + "learning_rate": 6.864043957794377e-05, + "loss": 0.414, + "step": 16380 + }, + { + "epoch": 6.18724046810117, + "grad_norm": 0.6669917702674866, + "learning_rate": 6.860207650747934e-05, + "loss": 0.4154, + "step": 16390 + }, + { + "epoch": 6.191015477538694, + "grad_norm": 0.6218419671058655, + "learning_rate": 6.856370072267104e-05, + "loss": 0.4125, + "step": 16400 + }, + { + "epoch": 6.194790486976218, + "grad_norm": 0.5821741223335266, + "learning_rate": 6.852531224974831e-05, + "loss": 0.4081, + "step": 16410 + }, + { + "epoch": 6.198565496413741, + "grad_norm": 0.6421241164207458, + "learning_rate": 6.848691111494936e-05, + "loss": 0.4067, + "step": 16420 + }, + { + "epoch": 6.202340505851264, + "grad_norm": 0.7413673996925354, + "learning_rate": 6.844849734452097e-05, + "loss": 0.4059, + "step": 16430 + }, + { + "epoch": 6.206115515288788, + "grad_norm": 0.715653657913208, + "learning_rate": 6.841007096471862e-05, + "loss": 0.414, + "step": 16440 + }, + { + "epoch": 6.209890524726312, + "grad_norm": 0.9158740639686584, + "learning_rate": 6.837163200180636e-05, + "loss": 0.4127, + "step": 16450 + }, + { + "epoch": 6.213665534163836, + "grad_norm": 0.6491902470588684, + "learning_rate": 6.833318048205684e-05, + "loss": 0.4046, + "step": 16460 + }, + { + "epoch": 6.217440543601359, + "grad_norm": 0.5672116875648499, + "learning_rate": 6.829471643175136e-05, + "loss": 0.4076, + "step": 16470 + }, + { + "epoch": 6.221215553038882, + "grad_norm": 0.7057968378067017, + "learning_rate": 6.825623987717969e-05, + "loss": 0.4077, + "step": 16480 + }, + { + "epoch": 6.224990562476406, + "grad_norm": 0.6522260308265686, + "learning_rate": 6.821775084464022e-05, + "loss": 0.4096, + "step": 16490 + }, + { + "epoch": 6.22876557191393, + "grad_norm": 0.6316922903060913, + "learning_rate": 6.817924936043982e-05, + "loss": 0.4026, + "step": 16500 + }, + { + "epoch": 6.232540581351453, + "grad_norm": 0.6496508717536926, + "learning_rate": 6.81407354508939e-05, + "loss": 0.4087, + "step": 16510 + }, + { + "epoch": 6.236315590788977, + "grad_norm": 0.5875167846679688, + "learning_rate": 6.810220914232636e-05, + "loss": 0.4075, + "step": 16520 + }, + { + "epoch": 6.240090600226501, + "grad_norm": 0.6535125970840454, + "learning_rate": 6.806367046106959e-05, + "loss": 0.4127, + "step": 16530 + }, + { + "epoch": 6.243865609664024, + "grad_norm": 0.6569979786872864, + "learning_rate": 6.802511943346435e-05, + "loss": 0.4154, + "step": 16540 + }, + { + "epoch": 6.247640619101547, + "grad_norm": 0.6199260354042053, + "learning_rate": 6.798655608585997e-05, + "loss": 0.4116, + "step": 16550 + }, + { + "epoch": 6.251415628539071, + "grad_norm": 0.6537362337112427, + "learning_rate": 6.79479804446141e-05, + "loss": 0.4057, + "step": 16560 + }, + { + "epoch": 6.255190637976595, + "grad_norm": 0.6324570775032043, + "learning_rate": 6.790939253609284e-05, + "loss": 0.4028, + "step": 16570 + }, + { + "epoch": 6.258965647414119, + "grad_norm": 0.6022115349769592, + "learning_rate": 6.787079238667065e-05, + "loss": 0.4087, + "step": 16580 + }, + { + "epoch": 6.262740656851642, + "grad_norm": 0.6702650189399719, + "learning_rate": 6.783218002273039e-05, + "loss": 0.4071, + "step": 16590 + }, + { + "epoch": 6.266515666289166, + "grad_norm": 0.6910784244537354, + "learning_rate": 6.779355547066322e-05, + "loss": 0.4038, + "step": 16600 + }, + { + "epoch": 6.270290675726689, + "grad_norm": 0.6253206729888916, + "learning_rate": 6.775491875686865e-05, + "loss": 0.4126, + "step": 16610 + }, + { + "epoch": 6.274065685164213, + "grad_norm": 0.5710755586624146, + "learning_rate": 6.771626990775457e-05, + "loss": 0.4108, + "step": 16620 + }, + { + "epoch": 6.277840694601736, + "grad_norm": 0.6698145866394043, + "learning_rate": 6.767760894973704e-05, + "loss": 0.4424, + "step": 16630 + }, + { + "epoch": 6.28161570403926, + "grad_norm": 1.365607500076294, + "learning_rate": 6.763893590924048e-05, + "loss": 0.4173, + "step": 16640 + }, + { + "epoch": 6.285390713476784, + "grad_norm": 0.6131840944290161, + "learning_rate": 6.760025081269756e-05, + "loss": 0.4056, + "step": 16650 + }, + { + "epoch": 6.289165722914308, + "grad_norm": 0.5999181866645813, + "learning_rate": 6.756155368654915e-05, + "loss": 0.4133, + "step": 16660 + }, + { + "epoch": 6.2929407323518305, + "grad_norm": 0.6875386238098145, + "learning_rate": 6.752284455724442e-05, + "loss": 0.4055, + "step": 16670 + }, + { + "epoch": 6.296715741789354, + "grad_norm": 0.5545741319656372, + "learning_rate": 6.748412345124065e-05, + "loss": 0.4132, + "step": 16680 + }, + { + "epoch": 6.300490751226878, + "grad_norm": 0.623226523399353, + "learning_rate": 6.744539039500335e-05, + "loss": 0.4087, + "step": 16690 + }, + { + "epoch": 6.304265760664402, + "grad_norm": 0.6600329279899597, + "learning_rate": 6.740664541500625e-05, + "loss": 0.4003, + "step": 16700 + }, + { + "epoch": 6.308040770101925, + "grad_norm": 0.5790383219718933, + "learning_rate": 6.736788853773112e-05, + "loss": 0.4076, + "step": 16710 + }, + { + "epoch": 6.311815779539449, + "grad_norm": 0.6672446131706238, + "learning_rate": 6.732911978966796e-05, + "loss": 0.4089, + "step": 16720 + }, + { + "epoch": 6.315590788976973, + "grad_norm": 0.5762278437614441, + "learning_rate": 6.729033919731482e-05, + "loss": 0.4116, + "step": 16730 + }, + { + "epoch": 6.319365798414496, + "grad_norm": 0.5791032314300537, + "learning_rate": 6.725154678717787e-05, + "loss": 0.407, + "step": 16740 + }, + { + "epoch": 6.3231408078520195, + "grad_norm": 0.6549326777458191, + "learning_rate": 6.721274258577138e-05, + "loss": 0.4114, + "step": 16750 + }, + { + "epoch": 6.326915817289543, + "grad_norm": 0.6120374202728271, + "learning_rate": 6.717392661961763e-05, + "loss": 0.4089, + "step": 16760 + }, + { + "epoch": 6.330690826727067, + "grad_norm": 0.6165977716445923, + "learning_rate": 6.713509891524697e-05, + "loss": 0.4128, + "step": 16770 + }, + { + "epoch": 6.334465836164591, + "grad_norm": 0.6072388887405396, + "learning_rate": 6.709625949919777e-05, + "loss": 0.4091, + "step": 16780 + }, + { + "epoch": 6.3382408456021135, + "grad_norm": 0.6050112843513489, + "learning_rate": 6.705740839801642e-05, + "loss": 0.4031, + "step": 16790 + }, + { + "epoch": 6.342015855039637, + "grad_norm": 0.7295091152191162, + "learning_rate": 6.701854563825727e-05, + "loss": 0.4105, + "step": 16800 + }, + { + "epoch": 6.345790864477161, + "grad_norm": 0.6450115442276001, + "learning_rate": 6.697967124648266e-05, + "loss": 0.4026, + "step": 16810 + }, + { + "epoch": 6.349565873914685, + "grad_norm": 0.6749234199523926, + "learning_rate": 6.694078524926285e-05, + "loss": 0.4106, + "step": 16820 + }, + { + "epoch": 6.3533408833522085, + "grad_norm": 0.6108365058898926, + "learning_rate": 6.690188767317607e-05, + "loss": 0.403, + "step": 16830 + }, + { + "epoch": 6.357115892789732, + "grad_norm": 0.6435592174530029, + "learning_rate": 6.686297854480843e-05, + "loss": 0.4091, + "step": 16840 + }, + { + "epoch": 6.360890902227256, + "grad_norm": 0.5970647931098938, + "learning_rate": 6.682405789075398e-05, + "loss": 0.404, + "step": 16850 + }, + { + "epoch": 6.364665911664779, + "grad_norm": 0.5717617869377136, + "learning_rate": 6.67851257376146e-05, + "loss": 0.3988, + "step": 16860 + }, + { + "epoch": 6.3684409211023025, + "grad_norm": 0.6971402168273926, + "learning_rate": 6.674618211200004e-05, + "loss": 0.4016, + "step": 16870 + }, + { + "epoch": 6.372215930539826, + "grad_norm": 0.5539186000823975, + "learning_rate": 6.670722704052792e-05, + "loss": 0.4109, + "step": 16880 + }, + { + "epoch": 6.37599093997735, + "grad_norm": 0.5876161456108093, + "learning_rate": 6.666826054982365e-05, + "loss": 0.406, + "step": 16890 + }, + { + "epoch": 6.379765949414874, + "grad_norm": 0.6357724666595459, + "learning_rate": 6.662928266652048e-05, + "loss": 0.4059, + "step": 16900 + }, + { + "epoch": 6.3835409588523975, + "grad_norm": 0.7432243227958679, + "learning_rate": 6.659029341725941e-05, + "loss": 0.4126, + "step": 16910 + }, + { + "epoch": 6.38731596828992, + "grad_norm": 0.5893529057502747, + "learning_rate": 6.655129282868923e-05, + "loss": 0.4053, + "step": 16920 + }, + { + "epoch": 6.391090977727444, + "grad_norm": 0.5809387564659119, + "learning_rate": 6.651228092746646e-05, + "loss": 0.41, + "step": 16930 + }, + { + "epoch": 6.394865987164968, + "grad_norm": 0.6010847687721252, + "learning_rate": 6.647325774025539e-05, + "loss": 0.42, + "step": 16940 + }, + { + "epoch": 6.3986409966024915, + "grad_norm": 0.6078194975852966, + "learning_rate": 6.643422329372798e-05, + "loss": 0.4069, + "step": 16950 + }, + { + "epoch": 6.402416006040015, + "grad_norm": 0.637675940990448, + "learning_rate": 6.639517761456392e-05, + "loss": 0.407, + "step": 16960 + }, + { + "epoch": 6.406191015477539, + "grad_norm": 0.5883272886276245, + "learning_rate": 6.635612072945054e-05, + "loss": 0.4043, + "step": 16970 + }, + { + "epoch": 6.409966024915063, + "grad_norm": 0.5617126226425171, + "learning_rate": 6.631705266508289e-05, + "loss": 0.4079, + "step": 16980 + }, + { + "epoch": 6.413741034352586, + "grad_norm": 0.6175063252449036, + "learning_rate": 6.62779734481636e-05, + "loss": 0.4003, + "step": 16990 + }, + { + "epoch": 6.417516043790109, + "grad_norm": 0.6425361037254333, + "learning_rate": 6.623888310540294e-05, + "loss": 0.4033, + "step": 17000 + }, + { + "epoch": 6.421291053227633, + "grad_norm": 0.5959002375602722, + "learning_rate": 6.619978166351882e-05, + "loss": 0.4037, + "step": 17010 + }, + { + "epoch": 6.425066062665157, + "grad_norm": 0.6390359997749329, + "learning_rate": 6.616066914923666e-05, + "loss": 0.4058, + "step": 17020 + }, + { + "epoch": 6.4288410721026805, + "grad_norm": 0.5752782225608826, + "learning_rate": 6.612154558928955e-05, + "loss": 0.4081, + "step": 17030 + }, + { + "epoch": 6.432616081540204, + "grad_norm": 0.684485137462616, + "learning_rate": 6.608241101041804e-05, + "loss": 0.4, + "step": 17040 + }, + { + "epoch": 6.436391090977727, + "grad_norm": 0.6156774759292603, + "learning_rate": 6.604326543937025e-05, + "loss": 0.4071, + "step": 17050 + }, + { + "epoch": 6.440166100415251, + "grad_norm": 0.6333221197128296, + "learning_rate": 6.60041089029018e-05, + "loss": 0.4074, + "step": 17060 + }, + { + "epoch": 6.443941109852775, + "grad_norm": 0.6324211955070496, + "learning_rate": 6.596494142777583e-05, + "loss": 0.4126, + "step": 17070 + }, + { + "epoch": 6.447716119290298, + "grad_norm": 0.6180362701416016, + "learning_rate": 6.592576304076294e-05, + "loss": 0.4113, + "step": 17080 + }, + { + "epoch": 6.451491128727822, + "grad_norm": 0.5131887197494507, + "learning_rate": 6.588657376864119e-05, + "loss": 0.4074, + "step": 17090 + }, + { + "epoch": 6.455266138165346, + "grad_norm": 0.570022702217102, + "learning_rate": 6.584737363819605e-05, + "loss": 0.4051, + "step": 17100 + }, + { + "epoch": 6.459041147602869, + "grad_norm": 0.5947955846786499, + "learning_rate": 6.580816267622048e-05, + "loss": 0.3996, + "step": 17110 + }, + { + "epoch": 6.462816157040392, + "grad_norm": 0.5993560552597046, + "learning_rate": 6.576894090951478e-05, + "loss": 0.4017, + "step": 17120 + }, + { + "epoch": 6.466591166477916, + "grad_norm": 0.5770085453987122, + "learning_rate": 6.572970836488665e-05, + "loss": 0.4095, + "step": 17130 + }, + { + "epoch": 6.47036617591544, + "grad_norm": 0.5673043727874756, + "learning_rate": 6.569046506915119e-05, + "loss": 0.3987, + "step": 17140 + }, + { + "epoch": 6.474141185352964, + "grad_norm": 0.6231826543807983, + "learning_rate": 6.56512110491308e-05, + "loss": 0.3943, + "step": 17150 + }, + { + "epoch": 6.477916194790487, + "grad_norm": 0.6213942766189575, + "learning_rate": 6.561194633165523e-05, + "loss": 0.4093, + "step": 17160 + }, + { + "epoch": 6.48169120422801, + "grad_norm": 0.6126642823219299, + "learning_rate": 6.557267094356155e-05, + "loss": 0.4164, + "step": 17170 + }, + { + "epoch": 6.485466213665534, + "grad_norm": 0.5821068286895752, + "learning_rate": 6.553338491169414e-05, + "loss": 0.4137, + "step": 17180 + }, + { + "epoch": 6.489241223103058, + "grad_norm": 0.5541254878044128, + "learning_rate": 6.54940882629046e-05, + "loss": 0.4062, + "step": 17190 + }, + { + "epoch": 6.493016232540581, + "grad_norm": 6.614664554595947, + "learning_rate": 6.545478102405184e-05, + "loss": 0.4263, + "step": 17200 + }, + { + "epoch": 6.496791241978105, + "grad_norm": 0.6379892230033875, + "learning_rate": 6.541546322200199e-05, + "loss": 0.4085, + "step": 17210 + }, + { + "epoch": 6.500566251415629, + "grad_norm": 0.5639785528182983, + "learning_rate": 6.537613488362837e-05, + "loss": 0.3997, + "step": 17220 + }, + { + "epoch": 6.504341260853153, + "grad_norm": 0.6088283061981201, + "learning_rate": 6.533679603581155e-05, + "loss": 0.4141, + "step": 17230 + }, + { + "epoch": 6.508116270290675, + "grad_norm": 0.7054754495620728, + "learning_rate": 6.529744670543926e-05, + "loss": 0.4144, + "step": 17240 + }, + { + "epoch": 6.511891279728199, + "grad_norm": 0.6340017914772034, + "learning_rate": 6.52580869194064e-05, + "loss": 0.4131, + "step": 17250 + }, + { + "epoch": 6.515666289165723, + "grad_norm": 0.6918211579322815, + "learning_rate": 6.521871670461499e-05, + "loss": 0.411, + "step": 17260 + }, + { + "epoch": 6.519441298603247, + "grad_norm": 0.6365815997123718, + "learning_rate": 6.517933608797422e-05, + "loss": 0.4104, + "step": 17270 + }, + { + "epoch": 6.52321630804077, + "grad_norm": 0.599440336227417, + "learning_rate": 6.513994509640038e-05, + "loss": 0.4075, + "step": 17280 + }, + { + "epoch": 6.526991317478293, + "grad_norm": 0.6258647441864014, + "learning_rate": 6.510054375681682e-05, + "loss": 0.4004, + "step": 17290 + }, + { + "epoch": 6.530766326915817, + "grad_norm": 0.7459936141967773, + "learning_rate": 6.506113209615398e-05, + "loss": 0.4078, + "step": 17300 + }, + { + "epoch": 6.534541336353341, + "grad_norm": 0.5855075120925903, + "learning_rate": 6.502171014134938e-05, + "loss": 0.4061, + "step": 17310 + }, + { + "epoch": 6.538316345790864, + "grad_norm": 0.6313618421554565, + "learning_rate": 6.498227791934755e-05, + "loss": 0.4085, + "step": 17320 + }, + { + "epoch": 6.542091355228388, + "grad_norm": 0.714318573474884, + "learning_rate": 6.494283545710003e-05, + "loss": 0.4015, + "step": 17330 + }, + { + "epoch": 6.545866364665912, + "grad_norm": 0.6481465697288513, + "learning_rate": 6.490338278156538e-05, + "loss": 0.4065, + "step": 17340 + }, + { + "epoch": 6.549641374103436, + "grad_norm": 0.7039319276809692, + "learning_rate": 6.486391991970913e-05, + "loss": 0.4136, + "step": 17350 + }, + { + "epoch": 6.553416383540959, + "grad_norm": 0.642465353012085, + "learning_rate": 6.482444689850377e-05, + "loss": 0.4026, + "step": 17360 + }, + { + "epoch": 6.557191392978482, + "grad_norm": 0.5987950563430786, + "learning_rate": 6.478496374492875e-05, + "loss": 0.4091, + "step": 17370 + }, + { + "epoch": 6.560966402416006, + "grad_norm": 0.6216861009597778, + "learning_rate": 6.474547048597042e-05, + "loss": 0.4006, + "step": 17380 + }, + { + "epoch": 6.56474141185353, + "grad_norm": 0.6449426412582397, + "learning_rate": 6.470596714862205e-05, + "loss": 0.4062, + "step": 17390 + }, + { + "epoch": 6.568516421291053, + "grad_norm": 0.5876918435096741, + "learning_rate": 6.46664537598838e-05, + "loss": 0.4061, + "step": 17400 + }, + { + "epoch": 6.572291430728577, + "grad_norm": 0.5914142727851868, + "learning_rate": 6.462693034676271e-05, + "loss": 0.4012, + "step": 17410 + }, + { + "epoch": 6.5760664401661, + "grad_norm": 0.561425507068634, + "learning_rate": 6.458739693627265e-05, + "loss": 0.3989, + "step": 17420 + }, + { + "epoch": 6.579841449603624, + "grad_norm": 0.6401806473731995, + "learning_rate": 6.454785355543432e-05, + "loss": 0.3995, + "step": 17430 + }, + { + "epoch": 6.5836164590411475, + "grad_norm": 0.660929262638092, + "learning_rate": 6.450830023127528e-05, + "loss": 0.4131, + "step": 17440 + }, + { + "epoch": 6.587391468478671, + "grad_norm": 0.6242609620094299, + "learning_rate": 6.446873699082982e-05, + "loss": 0.403, + "step": 17450 + }, + { + "epoch": 6.591166477916195, + "grad_norm": 0.5753269195556641, + "learning_rate": 6.44291638611391e-05, + "loss": 0.4042, + "step": 17460 + }, + { + "epoch": 6.594941487353719, + "grad_norm": 0.6394173502922058, + "learning_rate": 6.43895808692509e-05, + "loss": 0.4081, + "step": 17470 + }, + { + "epoch": 6.598716496791242, + "grad_norm": 0.6524138450622559, + "learning_rate": 6.434998804221986e-05, + "loss": 0.4026, + "step": 17480 + }, + { + "epoch": 6.602491506228765, + "grad_norm": 0.5924917459487915, + "learning_rate": 6.431038540710732e-05, + "loss": 0.4033, + "step": 17490 + }, + { + "epoch": 6.606266515666289, + "grad_norm": 2.0175888538360596, + "learning_rate": 6.427077299098129e-05, + "loss": 0.4034, + "step": 17500 + }, + { + "epoch": 6.610041525103813, + "grad_norm": 0.6277138590812683, + "learning_rate": 6.423115082091651e-05, + "loss": 0.3949, + "step": 17510 + }, + { + "epoch": 6.6138165345413364, + "grad_norm": 0.5728242993354797, + "learning_rate": 6.419151892399429e-05, + "loss": 0.3917, + "step": 17520 + }, + { + "epoch": 6.61759154397886, + "grad_norm": 0.6258454918861389, + "learning_rate": 6.415187732730273e-05, + "loss": 0.4026, + "step": 17530 + }, + { + "epoch": 6.621366553416384, + "grad_norm": 0.574672281742096, + "learning_rate": 6.411222605793645e-05, + "loss": 0.4017, + "step": 17540 + }, + { + "epoch": 6.625141562853907, + "grad_norm": 0.6237361431121826, + "learning_rate": 6.407256514299674e-05, + "loss": 0.3992, + "step": 17550 + }, + { + "epoch": 6.6289165722914305, + "grad_norm": 0.6078923940658569, + "learning_rate": 6.403289460959147e-05, + "loss": 0.4027, + "step": 17560 + }, + { + "epoch": 6.632691581728954, + "grad_norm": 0.6238769888877869, + "learning_rate": 6.399321448483501e-05, + "loss": 0.3993, + "step": 17570 + }, + { + "epoch": 6.636466591166478, + "grad_norm": 0.732795238494873, + "learning_rate": 6.395352479584844e-05, + "loss": 0.4017, + "step": 17580 + }, + { + "epoch": 6.640241600604002, + "grad_norm": 0.6137060523033142, + "learning_rate": 6.391382556975923e-05, + "loss": 0.3989, + "step": 17590 + }, + { + "epoch": 6.644016610041525, + "grad_norm": 0.6174196004867554, + "learning_rate": 6.387411683370144e-05, + "loss": 0.4061, + "step": 17600 + }, + { + "epoch": 6.647791619479049, + "grad_norm": 0.6093177199363708, + "learning_rate": 6.383439861481562e-05, + "loss": 0.3993, + "step": 17610 + }, + { + "epoch": 6.651566628916572, + "grad_norm": 0.624611496925354, + "learning_rate": 6.379467094024879e-05, + "loss": 0.4061, + "step": 17620 + }, + { + "epoch": 6.655341638354096, + "grad_norm": 0.7298497557640076, + "learning_rate": 6.375493383715445e-05, + "loss": 0.3959, + "step": 17630 + }, + { + "epoch": 6.6591166477916195, + "grad_norm": 0.5991964936256409, + "learning_rate": 6.371518733269254e-05, + "loss": 0.3992, + "step": 17640 + }, + { + "epoch": 6.662891657229143, + "grad_norm": 0.6233243942260742, + "learning_rate": 6.367543145402942e-05, + "loss": 0.4132, + "step": 17650 + }, + { + "epoch": 6.666666666666667, + "grad_norm": 0.6464060544967651, + "learning_rate": 6.363566622833785e-05, + "loss": 0.399, + "step": 17660 + }, + { + "epoch": 6.67044167610419, + "grad_norm": 0.7238352298736572, + "learning_rate": 6.359589168279698e-05, + "loss": 0.3989, + "step": 17670 + }, + { + "epoch": 6.6742166855417135, + "grad_norm": 0.7250881791114807, + "learning_rate": 6.355610784459235e-05, + "loss": 0.4039, + "step": 17680 + }, + { + "epoch": 6.677991694979237, + "grad_norm": 0.5912888646125793, + "learning_rate": 6.351631474091585e-05, + "loss": 0.4017, + "step": 17690 + }, + { + "epoch": 6.681766704416761, + "grad_norm": 0.625900149345398, + "learning_rate": 6.347651239896566e-05, + "loss": 0.4034, + "step": 17700 + }, + { + "epoch": 6.685541713854285, + "grad_norm": 0.6302257180213928, + "learning_rate": 6.343670084594633e-05, + "loss": 0.4045, + "step": 17710 + }, + { + "epoch": 6.6893167232918085, + "grad_norm": 0.6661359667778015, + "learning_rate": 6.339688010906866e-05, + "loss": 0.4092, + "step": 17720 + }, + { + "epoch": 6.693091732729332, + "grad_norm": 0.5921412706375122, + "learning_rate": 6.335705021554975e-05, + "loss": 0.3951, + "step": 17730 + }, + { + "epoch": 6.696866742166855, + "grad_norm": 0.6364487409591675, + "learning_rate": 6.3317211192613e-05, + "loss": 0.4093, + "step": 17740 + }, + { + "epoch": 6.700641751604379, + "grad_norm": 0.6267909407615662, + "learning_rate": 6.327736306748795e-05, + "loss": 0.4035, + "step": 17750 + }, + { + "epoch": 6.7044167610419025, + "grad_norm": 0.5861899852752686, + "learning_rate": 6.323750586741047e-05, + "loss": 0.4027, + "step": 17760 + }, + { + "epoch": 6.708191770479426, + "grad_norm": 0.647732138633728, + "learning_rate": 6.319763961962252e-05, + "loss": 0.4064, + "step": 17770 + }, + { + "epoch": 6.71196677991695, + "grad_norm": 0.618825376033783, + "learning_rate": 6.315776435137233e-05, + "loss": 0.4008, + "step": 17780 + }, + { + "epoch": 6.715741789354474, + "grad_norm": 0.7474274039268494, + "learning_rate": 6.311788008991432e-05, + "loss": 0.4089, + "step": 17790 + }, + { + "epoch": 6.719516798791997, + "grad_norm": 0.6482340693473816, + "learning_rate": 6.307798686250891e-05, + "loss": 0.4037, + "step": 17800 + }, + { + "epoch": 6.72329180822952, + "grad_norm": 0.6014053821563721, + "learning_rate": 6.303808469642284e-05, + "loss": 0.4031, + "step": 17810 + }, + { + "epoch": 6.727066817667044, + "grad_norm": 0.6458423137664795, + "learning_rate": 6.29981736189288e-05, + "loss": 0.3998, + "step": 17820 + }, + { + "epoch": 6.730841827104568, + "grad_norm": 0.6253558397293091, + "learning_rate": 6.295825365730567e-05, + "loss": 0.4044, + "step": 17830 + }, + { + "epoch": 6.7346168365420915, + "grad_norm": 0.6395803093910217, + "learning_rate": 6.291832483883835e-05, + "loss": 0.4026, + "step": 17840 + }, + { + "epoch": 6.738391845979615, + "grad_norm": 0.6076302528381348, + "learning_rate": 6.28783871908178e-05, + "loss": 0.3979, + "step": 17850 + }, + { + "epoch": 6.742166855417139, + "grad_norm": 0.591390073299408, + "learning_rate": 6.283844074054107e-05, + "loss": 0.3967, + "step": 17860 + }, + { + "epoch": 6.745941864854662, + "grad_norm": 0.5707950592041016, + "learning_rate": 6.279848551531112e-05, + "loss": 0.3941, + "step": 17870 + }, + { + "epoch": 6.749716874292186, + "grad_norm": 0.6005119681358337, + "learning_rate": 6.275852154243702e-05, + "loss": 0.407, + "step": 17880 + }, + { + "epoch": 6.753491883729709, + "grad_norm": 0.5978226065635681, + "learning_rate": 6.271854884923377e-05, + "loss": 0.3962, + "step": 17890 + }, + { + "epoch": 6.757266893167233, + "grad_norm": 0.637346088886261, + "learning_rate": 6.267856746302228e-05, + "loss": 0.3962, + "step": 17900 + }, + { + "epoch": 6.761041902604757, + "grad_norm": 0.7648394107818604, + "learning_rate": 6.263857741112948e-05, + "loss": 0.4004, + "step": 17910 + }, + { + "epoch": 6.76481691204228, + "grad_norm": 0.6177964806556702, + "learning_rate": 6.259857872088821e-05, + "loss": 0.3901, + "step": 17920 + }, + { + "epoch": 6.768591921479803, + "grad_norm": 0.6134040355682373, + "learning_rate": 6.255857141963719e-05, + "loss": 0.3985, + "step": 17930 + }, + { + "epoch": 6.772366930917327, + "grad_norm": 0.5989758968353271, + "learning_rate": 6.251855553472101e-05, + "loss": 0.398, + "step": 17940 + }, + { + "epoch": 6.776141940354851, + "grad_norm": 0.6749169230461121, + "learning_rate": 6.247853109349016e-05, + "loss": 0.3976, + "step": 17950 + }, + { + "epoch": 6.779916949792375, + "grad_norm": 0.6461236476898193, + "learning_rate": 6.243849812330098e-05, + "loss": 0.3982, + "step": 17960 + }, + { + "epoch": 6.783691959229898, + "grad_norm": 1.9516565799713135, + "learning_rate": 6.239845665151563e-05, + "loss": 0.3939, + "step": 17970 + }, + { + "epoch": 6.787466968667422, + "grad_norm": 0.6713792681694031, + "learning_rate": 6.235840670550204e-05, + "loss": 0.4018, + "step": 17980 + }, + { + "epoch": 6.791241978104945, + "grad_norm": 0.6273426413536072, + "learning_rate": 6.231834831263403e-05, + "loss": 0.401, + "step": 17990 + }, + { + "epoch": 6.795016987542469, + "grad_norm": 0.6532526612281799, + "learning_rate": 6.22782815002911e-05, + "loss": 0.4034, + "step": 18000 + }, + { + "epoch": 6.798791996979992, + "grad_norm": 0.6998550891876221, + "learning_rate": 6.223820629585852e-05, + "loss": 0.4019, + "step": 18010 + }, + { + "epoch": 6.802567006417516, + "grad_norm": 0.6373661756515503, + "learning_rate": 6.219812272672737e-05, + "loss": 0.4007, + "step": 18020 + }, + { + "epoch": 6.80634201585504, + "grad_norm": 0.6695569157600403, + "learning_rate": 6.215803082029434e-05, + "loss": 0.3977, + "step": 18030 + }, + { + "epoch": 6.810117025292564, + "grad_norm": 0.7880366444587708, + "learning_rate": 6.211793060396188e-05, + "loss": 0.3985, + "step": 18040 + }, + { + "epoch": 6.813892034730086, + "grad_norm": 0.6441872119903564, + "learning_rate": 6.207782210513811e-05, + "loss": 0.4013, + "step": 18050 + }, + { + "epoch": 6.81766704416761, + "grad_norm": 0.6251243948936462, + "learning_rate": 6.203770535123683e-05, + "loss": 0.4084, + "step": 18060 + }, + { + "epoch": 6.821442053605134, + "grad_norm": 0.710367739200592, + "learning_rate": 6.199758036967747e-05, + "loss": 0.4, + "step": 18070 + }, + { + "epoch": 6.825217063042658, + "grad_norm": 0.568421483039856, + "learning_rate": 6.195744718788503e-05, + "loss": 0.3971, + "step": 18080 + }, + { + "epoch": 6.828992072480181, + "grad_norm": 0.6238227486610413, + "learning_rate": 6.191730583329021e-05, + "loss": 0.3973, + "step": 18090 + }, + { + "epoch": 6.832767081917705, + "grad_norm": 0.6105861067771912, + "learning_rate": 6.187715633332921e-05, + "loss": 0.3994, + "step": 18100 + }, + { + "epoch": 6.836542091355229, + "grad_norm": 0.5954229235649109, + "learning_rate": 6.183699871544386e-05, + "loss": 0.4027, + "step": 18110 + }, + { + "epoch": 6.840317100792752, + "grad_norm": 0.5816265344619751, + "learning_rate": 6.179683300708152e-05, + "loss": 0.3998, + "step": 18120 + }, + { + "epoch": 6.844092110230275, + "grad_norm": 0.609210729598999, + "learning_rate": 6.175665923569503e-05, + "loss": 0.3993, + "step": 18130 + }, + { + "epoch": 6.847867119667799, + "grad_norm": 0.6947378516197205, + "learning_rate": 6.171647742874281e-05, + "loss": 0.3964, + "step": 18140 + }, + { + "epoch": 6.851642129105323, + "grad_norm": 0.6516598463058472, + "learning_rate": 6.167628761368875e-05, + "loss": 0.4027, + "step": 18150 + }, + { + "epoch": 6.855417138542847, + "grad_norm": 0.5862056612968445, + "learning_rate": 6.163608981800222e-05, + "loss": 0.4079, + "step": 18160 + }, + { + "epoch": 6.8591921479803695, + "grad_norm": 0.626376211643219, + "learning_rate": 6.159588406915803e-05, + "loss": 0.4003, + "step": 18170 + }, + { + "epoch": 6.862967157417893, + "grad_norm": 0.6769831776618958, + "learning_rate": 6.155567039463639e-05, + "loss": 0.4041, + "step": 18180 + }, + { + "epoch": 6.866742166855417, + "grad_norm": 0.6997120380401611, + "learning_rate": 6.151544882192302e-05, + "loss": 0.4028, + "step": 18190 + }, + { + "epoch": 6.870517176292941, + "grad_norm": 0.5860679745674133, + "learning_rate": 6.147521937850895e-05, + "loss": 0.4132, + "step": 18200 + }, + { + "epoch": 6.874292185730464, + "grad_norm": 0.5797774791717529, + "learning_rate": 6.143498209189066e-05, + "loss": 0.4019, + "step": 18210 + }, + { + "epoch": 6.878067195167988, + "grad_norm": 0.6230421662330627, + "learning_rate": 6.139473698956993e-05, + "loss": 0.4041, + "step": 18220 + }, + { + "epoch": 6.881842204605512, + "grad_norm": 0.5917826890945435, + "learning_rate": 6.13544840990539e-05, + "loss": 0.402, + "step": 18230 + }, + { + "epoch": 6.885617214043035, + "grad_norm": 1.265276312828064, + "learning_rate": 6.131422344785507e-05, + "loss": 0.3984, + "step": 18240 + }, + { + "epoch": 6.8893922234805585, + "grad_norm": 0.5542703866958618, + "learning_rate": 6.127395506349119e-05, + "loss": 0.3988, + "step": 18250 + }, + { + "epoch": 6.893167232918082, + "grad_norm": 0.736250102519989, + "learning_rate": 6.123367897348533e-05, + "loss": 0.3968, + "step": 18260 + }, + { + "epoch": 6.896942242355606, + "grad_norm": 0.8266544342041016, + "learning_rate": 6.119339520536584e-05, + "loss": 0.4013, + "step": 18270 + }, + { + "epoch": 6.90071725179313, + "grad_norm": 0.6568412780761719, + "learning_rate": 6.115310378666625e-05, + "loss": 0.4019, + "step": 18280 + }, + { + "epoch": 6.904492261230653, + "grad_norm": 0.5549454689025879, + "learning_rate": 6.11128047449254e-05, + "loss": 0.3982, + "step": 18290 + }, + { + "epoch": 6.908267270668176, + "grad_norm": 0.6285567283630371, + "learning_rate": 6.107249810768729e-05, + "loss": 0.4012, + "step": 18300 + }, + { + "epoch": 6.9120422801057, + "grad_norm": 0.6208999156951904, + "learning_rate": 6.1032183902501125e-05, + "loss": 0.3999, + "step": 18310 + }, + { + "epoch": 6.915817289543224, + "grad_norm": 0.5961741805076599, + "learning_rate": 6.099186215692131e-05, + "loss": 0.4093, + "step": 18320 + }, + { + "epoch": 6.9195922989807475, + "grad_norm": 0.5774527192115784, + "learning_rate": 6.095153289850734e-05, + "loss": 0.401, + "step": 18330 + }, + { + "epoch": 6.923367308418271, + "grad_norm": 0.6928045749664307, + "learning_rate": 6.0911196154823904e-05, + "loss": 0.3907, + "step": 18340 + }, + { + "epoch": 6.927142317855795, + "grad_norm": 0.7921097874641418, + "learning_rate": 6.087085195344079e-05, + "loss": 0.395, + "step": 18350 + }, + { + "epoch": 6.930917327293319, + "grad_norm": 0.6023475527763367, + "learning_rate": 6.083050032193286e-05, + "loss": 0.4031, + "step": 18360 + }, + { + "epoch": 6.9346923367308415, + "grad_norm": 1.3727141618728638, + "learning_rate": 6.0790141287880097e-05, + "loss": 0.4059, + "step": 18370 + }, + { + "epoch": 6.938467346168365, + "grad_norm": 0.606326699256897, + "learning_rate": 6.0749774878867496e-05, + "loss": 0.3957, + "step": 18380 + }, + { + "epoch": 6.942242355605889, + "grad_norm": 0.6082478165626526, + "learning_rate": 6.0709401122485146e-05, + "loss": 0.3975, + "step": 18390 + }, + { + "epoch": 6.946017365043413, + "grad_norm": 0.7824208736419678, + "learning_rate": 6.066902004632811e-05, + "loss": 0.3967, + "step": 18400 + }, + { + "epoch": 6.9497923744809365, + "grad_norm": 0.6232104301452637, + "learning_rate": 6.062863167799646e-05, + "loss": 0.4024, + "step": 18410 + }, + { + "epoch": 6.953567383918459, + "grad_norm": 0.6417713165283203, + "learning_rate": 6.058823604509529e-05, + "loss": 0.3967, + "step": 18420 + }, + { + "epoch": 6.957342393355983, + "grad_norm": 0.6295148730278015, + "learning_rate": 6.054783317523462e-05, + "loss": 0.3991, + "step": 18430 + }, + { + "epoch": 6.961117402793507, + "grad_norm": 0.6527547836303711, + "learning_rate": 6.050742309602944e-05, + "loss": 0.3988, + "step": 18440 + }, + { + "epoch": 6.9648924122310305, + "grad_norm": 0.6303055882453918, + "learning_rate": 6.046700583509965e-05, + "loss": 0.3933, + "step": 18450 + }, + { + "epoch": 6.968667421668554, + "grad_norm": 0.637154757976532, + "learning_rate": 6.042658142007007e-05, + "loss": 0.4024, + "step": 18460 + }, + { + "epoch": 6.972442431106078, + "grad_norm": 0.6554228663444519, + "learning_rate": 6.038614987857041e-05, + "loss": 0.3999, + "step": 18470 + }, + { + "epoch": 6.976217440543602, + "grad_norm": 0.6131272315979004, + "learning_rate": 6.0345711238235224e-05, + "loss": 0.3946, + "step": 18480 + }, + { + "epoch": 6.9799924499811254, + "grad_norm": 0.5949199199676514, + "learning_rate": 6.030526552670399e-05, + "loss": 0.4006, + "step": 18490 + }, + { + "epoch": 6.983767459418648, + "grad_norm": 0.5946569442749023, + "learning_rate": 6.0264812771620925e-05, + "loss": 0.3971, + "step": 18500 + }, + { + "epoch": 6.987542468856172, + "grad_norm": 0.7052402496337891, + "learning_rate": 6.022435300063512e-05, + "loss": 0.3864, + "step": 18510 + }, + { + "epoch": 6.991317478293696, + "grad_norm": 0.5748345255851746, + "learning_rate": 6.0183886241400466e-05, + "loss": 0.4049, + "step": 18520 + }, + { + "epoch": 6.9950924877312195, + "grad_norm": 0.6450919508934021, + "learning_rate": 6.0143412521575584e-05, + "loss": 0.4042, + "step": 18530 + }, + { + "epoch": 6.998867497168743, + "grad_norm": 0.7220432162284851, + "learning_rate": 6.010293186882389e-05, + "loss": 0.4025, + "step": 18540 + }, + { + "epoch": 7.002642506606266, + "grad_norm": 0.6318272352218628, + "learning_rate": 6.0062444310813525e-05, + "loss": 0.4026, + "step": 18550 + }, + { + "epoch": 7.00641751604379, + "grad_norm": 0.5739586353302002, + "learning_rate": 6.0021949875217355e-05, + "loss": 0.3905, + "step": 18560 + }, + { + "epoch": 7.0101925254813136, + "grad_norm": 0.6182407140731812, + "learning_rate": 5.998144858971295e-05, + "loss": 0.3936, + "step": 18570 + }, + { + "epoch": 7.013967534918837, + "grad_norm": 0.6267920732498169, + "learning_rate": 5.994094048198257e-05, + "loss": 0.4063, + "step": 18580 + }, + { + "epoch": 7.017742544356361, + "grad_norm": 0.7207063436508179, + "learning_rate": 5.990042557971307e-05, + "loss": 0.3992, + "step": 18590 + }, + { + "epoch": 7.021517553793885, + "grad_norm": 0.6681888103485107, + "learning_rate": 5.985990391059607e-05, + "loss": 0.3976, + "step": 18600 + }, + { + "epoch": 7.0252925632314085, + "grad_norm": 0.6886135935783386, + "learning_rate": 5.981937550232771e-05, + "loss": 0.3898, + "step": 18610 + }, + { + "epoch": 7.029067572668931, + "grad_norm": 0.7061811685562134, + "learning_rate": 5.9778840382608794e-05, + "loss": 0.4119, + "step": 18620 + }, + { + "epoch": 7.032842582106455, + "grad_norm": 0.5984750986099243, + "learning_rate": 5.9738298579144695e-05, + "loss": 0.3965, + "step": 18630 + }, + { + "epoch": 7.036617591543979, + "grad_norm": 0.6792024374008179, + "learning_rate": 5.9697750119645314e-05, + "loss": 0.4087, + "step": 18640 + }, + { + "epoch": 7.0403926009815025, + "grad_norm": 0.5802191495895386, + "learning_rate": 5.96571950318252e-05, + "loss": 0.4003, + "step": 18650 + }, + { + "epoch": 7.044167610419026, + "grad_norm": 0.6558259129524231, + "learning_rate": 5.9616633343403316e-05, + "loss": 0.408, + "step": 18660 + }, + { + "epoch": 7.04794261985655, + "grad_norm": 0.7390403151512146, + "learning_rate": 5.957606508210324e-05, + "loss": 0.404, + "step": 18670 + }, + { + "epoch": 7.051717629294073, + "grad_norm": 0.6561256051063538, + "learning_rate": 5.953549027565297e-05, + "loss": 0.3936, + "step": 18680 + }, + { + "epoch": 7.055492638731597, + "grad_norm": 0.6152408123016357, + "learning_rate": 5.949490895178501e-05, + "loss": 0.3959, + "step": 18690 + }, + { + "epoch": 7.05926764816912, + "grad_norm": 1.134881854057312, + "learning_rate": 5.945432113823632e-05, + "loss": 0.3955, + "step": 18700 + }, + { + "epoch": 7.063042657606644, + "grad_norm": 0.611450731754303, + "learning_rate": 5.9413726862748276e-05, + "loss": 0.3999, + "step": 18710 + }, + { + "epoch": 7.066817667044168, + "grad_norm": 1.1153512001037598, + "learning_rate": 5.9373126153066694e-05, + "loss": 0.3965, + "step": 18720 + }, + { + "epoch": 7.0705926764816915, + "grad_norm": 0.6631627678871155, + "learning_rate": 5.933251903694177e-05, + "loss": 0.4053, + "step": 18730 + }, + { + "epoch": 7.074367685919214, + "grad_norm": 0.6014063954353333, + "learning_rate": 5.929190554212807e-05, + "loss": 0.3936, + "step": 18740 + }, + { + "epoch": 7.078142695356738, + "grad_norm": 0.6215736269950867, + "learning_rate": 5.9251285696384565e-05, + "loss": 0.404, + "step": 18750 + }, + { + "epoch": 7.081917704794262, + "grad_norm": 0.791002094745636, + "learning_rate": 5.921065952747451e-05, + "loss": 0.3992, + "step": 18760 + }, + { + "epoch": 7.085692714231786, + "grad_norm": 0.9129757881164551, + "learning_rate": 5.917002706316552e-05, + "loss": 0.3984, + "step": 18770 + }, + { + "epoch": 7.089467723669309, + "grad_norm": 0.5969563722610474, + "learning_rate": 5.912938833122952e-05, + "loss": 0.3988, + "step": 18780 + }, + { + "epoch": 7.093242733106833, + "grad_norm": 1.1693058013916016, + "learning_rate": 5.908874335944265e-05, + "loss": 0.3969, + "step": 18790 + }, + { + "epoch": 7.097017742544356, + "grad_norm": 0.6244138479232788, + "learning_rate": 5.904809217558542e-05, + "loss": 0.3992, + "step": 18800 + }, + { + "epoch": 7.10079275198188, + "grad_norm": 0.6120061874389648, + "learning_rate": 5.90074348074425e-05, + "loss": 0.3984, + "step": 18810 + }, + { + "epoch": 7.104567761419403, + "grad_norm": 0.6857101917266846, + "learning_rate": 5.8966771282802814e-05, + "loss": 0.3949, + "step": 18820 + }, + { + "epoch": 7.108342770856927, + "grad_norm": 0.6439679265022278, + "learning_rate": 5.892610162945952e-05, + "loss": 0.401, + "step": 18830 + }, + { + "epoch": 7.112117780294451, + "grad_norm": 0.686291515827179, + "learning_rate": 5.8885425875209924e-05, + "loss": 0.3991, + "step": 18840 + }, + { + "epoch": 7.115892789731975, + "grad_norm": 0.893528938293457, + "learning_rate": 5.884474404785553e-05, + "loss": 0.3903, + "step": 18850 + }, + { + "epoch": 7.119667799169498, + "grad_norm": 0.6380524635314941, + "learning_rate": 5.8804056175201983e-05, + "loss": 0.3936, + "step": 18860 + }, + { + "epoch": 7.123442808607021, + "grad_norm": 0.5932635068893433, + "learning_rate": 5.876336228505904e-05, + "loss": 0.3988, + "step": 18870 + }, + { + "epoch": 7.127217818044545, + "grad_norm": 0.6427797079086304, + "learning_rate": 5.872266240524062e-05, + "loss": 0.4023, + "step": 18880 + }, + { + "epoch": 7.130992827482069, + "grad_norm": 0.625927209854126, + "learning_rate": 5.86819565635647e-05, + "loss": 0.3963, + "step": 18890 + }, + { + "epoch": 7.134767836919592, + "grad_norm": 0.5989848971366882, + "learning_rate": 5.8641244787853334e-05, + "loss": 0.3958, + "step": 18900 + }, + { + "epoch": 7.138542846357116, + "grad_norm": 0.6157662868499756, + "learning_rate": 5.860052710593265e-05, + "loss": 0.3974, + "step": 18910 + }, + { + "epoch": 7.14231785579464, + "grad_norm": 0.6159812211990356, + "learning_rate": 5.855980354563276e-05, + "loss": 0.3868, + "step": 18920 + }, + { + "epoch": 7.146092865232163, + "grad_norm": 0.5764108300209045, + "learning_rate": 5.8519074134787874e-05, + "loss": 0.3937, + "step": 18930 + }, + { + "epoch": 7.149867874669686, + "grad_norm": 0.657748281955719, + "learning_rate": 5.847833890123614e-05, + "loss": 0.3937, + "step": 18940 + }, + { + "epoch": 7.15364288410721, + "grad_norm": 0.6909071207046509, + "learning_rate": 5.8437597872819737e-05, + "loss": 0.397, + "step": 18950 + }, + { + "epoch": 7.157417893544734, + "grad_norm": 0.6924566030502319, + "learning_rate": 5.839685107738473e-05, + "loss": 0.397, + "step": 18960 + }, + { + "epoch": 7.161192902982258, + "grad_norm": 0.6301664710044861, + "learning_rate": 5.835609854278118e-05, + "loss": 0.3921, + "step": 18970 + }, + { + "epoch": 7.164967912419781, + "grad_norm": 0.6503030061721802, + "learning_rate": 5.831534029686308e-05, + "loss": 0.3958, + "step": 18980 + }, + { + "epoch": 7.168742921857304, + "grad_norm": 0.6070290207862854, + "learning_rate": 5.82745763674883e-05, + "loss": 0.3943, + "step": 18990 + }, + { + "epoch": 7.172517931294828, + "grad_norm": 0.6799952387809753, + "learning_rate": 5.823380678251861e-05, + "loss": 0.3899, + "step": 19000 + }, + { + "epoch": 7.176292940732352, + "grad_norm": 0.6328525543212891, + "learning_rate": 5.81930315698196e-05, + "loss": 0.3935, + "step": 19010 + }, + { + "epoch": 7.180067950169875, + "grad_norm": 0.6666650176048279, + "learning_rate": 5.815225075726076e-05, + "loss": 0.3931, + "step": 19020 + }, + { + "epoch": 7.183842959607399, + "grad_norm": 0.6060150861740112, + "learning_rate": 5.811146437271543e-05, + "loss": 0.3976, + "step": 19030 + }, + { + "epoch": 7.187617969044923, + "grad_norm": 0.6627577543258667, + "learning_rate": 5.807067244406066e-05, + "loss": 0.4025, + "step": 19040 + }, + { + "epoch": 7.191392978482447, + "grad_norm": 0.5842311978340149, + "learning_rate": 5.8029874999177405e-05, + "loss": 0.3941, + "step": 19050 + }, + { + "epoch": 7.1951679879199695, + "grad_norm": 1.3617955446243286, + "learning_rate": 5.798907206595029e-05, + "loss": 0.3934, + "step": 19060 + }, + { + "epoch": 7.198942997357493, + "grad_norm": 0.5990311503410339, + "learning_rate": 5.794826367226773e-05, + "loss": 0.3947, + "step": 19070 + }, + { + "epoch": 7.202718006795017, + "grad_norm": 0.5806323885917664, + "learning_rate": 5.790744984602193e-05, + "loss": 0.4021, + "step": 19080 + }, + { + "epoch": 7.206493016232541, + "grad_norm": 0.6318025588989258, + "learning_rate": 5.786663061510872e-05, + "loss": 0.3978, + "step": 19090 + }, + { + "epoch": 7.210268025670064, + "grad_norm": 0.6155949234962463, + "learning_rate": 5.782580600742765e-05, + "loss": 0.395, + "step": 19100 + }, + { + "epoch": 7.214043035107588, + "grad_norm": 0.6419726014137268, + "learning_rate": 5.7784976050881965e-05, + "loss": 0.395, + "step": 19110 + }, + { + "epoch": 7.217818044545111, + "grad_norm": 0.6743911504745483, + "learning_rate": 5.774414077337855e-05, + "loss": 0.3907, + "step": 19120 + }, + { + "epoch": 7.221593053982635, + "grad_norm": 0.6767613887786865, + "learning_rate": 5.770330020282796e-05, + "loss": 0.3911, + "step": 19130 + }, + { + "epoch": 7.2253680634201585, + "grad_norm": 0.6418374180793762, + "learning_rate": 5.7662454367144317e-05, + "loss": 0.3978, + "step": 19140 + }, + { + "epoch": 7.229143072857682, + "grad_norm": 0.5886910557746887, + "learning_rate": 5.762160329424536e-05, + "loss": 0.398, + "step": 19150 + }, + { + "epoch": 7.232918082295206, + "grad_norm": 0.56149822473526, + "learning_rate": 5.7580747012052416e-05, + "loss": 0.3917, + "step": 19160 + }, + { + "epoch": 7.23669309173273, + "grad_norm": 0.8316841125488281, + "learning_rate": 5.753988554849037e-05, + "loss": 0.3967, + "step": 19170 + }, + { + "epoch": 7.2404681011702525, + "grad_norm": 0.5947608351707458, + "learning_rate": 5.749901893148766e-05, + "loss": 0.395, + "step": 19180 + }, + { + "epoch": 7.244243110607776, + "grad_norm": 0.6548729538917542, + "learning_rate": 5.745814718897621e-05, + "loss": 0.3925, + "step": 19190 + }, + { + "epoch": 7.2480181200453, + "grad_norm": 0.5887550115585327, + "learning_rate": 5.74172703488915e-05, + "loss": 0.3922, + "step": 19200 + }, + { + "epoch": 7.251793129482824, + "grad_norm": 0.6358749270439148, + "learning_rate": 5.737638843917242e-05, + "loss": 0.3876, + "step": 19210 + }, + { + "epoch": 7.2555681389203475, + "grad_norm": 0.7427713871002197, + "learning_rate": 5.73355014877614e-05, + "loss": 0.3955, + "step": 19220 + }, + { + "epoch": 7.259343148357871, + "grad_norm": 0.6359094977378845, + "learning_rate": 5.7294609522604316e-05, + "loss": 0.3964, + "step": 19230 + }, + { + "epoch": 7.263118157795395, + "grad_norm": 0.5932873487472534, + "learning_rate": 5.7253712571650376e-05, + "loss": 0.3966, + "step": 19240 + }, + { + "epoch": 7.266893167232918, + "grad_norm": 0.6501133441925049, + "learning_rate": 5.721281066285229e-05, + "loss": 0.3909, + "step": 19250 + }, + { + "epoch": 7.2706681766704415, + "grad_norm": 0.6515145301818848, + "learning_rate": 5.717190382416615e-05, + "loss": 0.3967, + "step": 19260 + }, + { + "epoch": 7.274443186107965, + "grad_norm": 0.6237747073173523, + "learning_rate": 5.713099208355135e-05, + "loss": 0.3959, + "step": 19270 + }, + { + "epoch": 7.278218195545489, + "grad_norm": 0.5545824766159058, + "learning_rate": 5.709007546897074e-05, + "loss": 0.396, + "step": 19280 + }, + { + "epoch": 7.281993204983013, + "grad_norm": 0.6651487350463867, + "learning_rate": 5.704915400839037e-05, + "loss": 0.3994, + "step": 19290 + }, + { + "epoch": 7.2857682144205365, + "grad_norm": 0.701859176158905, + "learning_rate": 5.700822772977971e-05, + "loss": 0.4039, + "step": 19300 + }, + { + "epoch": 7.289543223858059, + "grad_norm": 0.9425735473632812, + "learning_rate": 5.696729666111148e-05, + "loss": 0.4047, + "step": 19310 + }, + { + "epoch": 7.293318233295583, + "grad_norm": 0.6198208928108215, + "learning_rate": 5.692636083036168e-05, + "loss": 0.4123, + "step": 19320 + }, + { + "epoch": 7.297093242733107, + "grad_norm": 0.7142148613929749, + "learning_rate": 5.688542026550958e-05, + "loss": 0.3903, + "step": 19330 + }, + { + "epoch": 7.3008682521706305, + "grad_norm": 0.6616962552070618, + "learning_rate": 5.684447499453763e-05, + "loss": 0.3973, + "step": 19340 + }, + { + "epoch": 7.304643261608154, + "grad_norm": 0.6994644403457642, + "learning_rate": 5.680352504543156e-05, + "loss": 0.3936, + "step": 19350 + }, + { + "epoch": 7.308418271045678, + "grad_norm": 0.6935665011405945, + "learning_rate": 5.67625704461803e-05, + "loss": 0.3882, + "step": 19360 + }, + { + "epoch": 7.312193280483201, + "grad_norm": 0.6593263745307922, + "learning_rate": 5.672161122477589e-05, + "loss": 0.4025, + "step": 19370 + }, + { + "epoch": 7.315968289920725, + "grad_norm": 1.2950445413589478, + "learning_rate": 5.668064740921359e-05, + "loss": 0.3984, + "step": 19380 + }, + { + "epoch": 7.319743299358248, + "grad_norm": 0.6627549529075623, + "learning_rate": 5.663967902749179e-05, + "loss": 0.3958, + "step": 19390 + }, + { + "epoch": 7.323518308795772, + "grad_norm": 0.7319017648696899, + "learning_rate": 5.6598706107611965e-05, + "loss": 0.3883, + "step": 19400 + }, + { + "epoch": 7.327293318233296, + "grad_norm": 0.6958197951316833, + "learning_rate": 5.655772867757876e-05, + "loss": 0.3962, + "step": 19410 + }, + { + "epoch": 7.3310683276708195, + "grad_norm": 0.6247055530548096, + "learning_rate": 5.651674676539982e-05, + "loss": 0.3911, + "step": 19420 + }, + { + "epoch": 7.334843337108342, + "grad_norm": 0.6021857857704163, + "learning_rate": 5.647576039908593e-05, + "loss": 0.3909, + "step": 19430 + }, + { + "epoch": 7.338618346545866, + "grad_norm": 0.6767669916152954, + "learning_rate": 5.6434769606650864e-05, + "loss": 0.3908, + "step": 19440 + }, + { + "epoch": 7.34239335598339, + "grad_norm": 0.8901523351669312, + "learning_rate": 5.639377441611143e-05, + "loss": 0.3988, + "step": 19450 + }, + { + "epoch": 7.346168365420914, + "grad_norm": 0.6947163939476013, + "learning_rate": 5.635277485548751e-05, + "loss": 0.3917, + "step": 19460 + }, + { + "epoch": 7.349943374858437, + "grad_norm": 0.7375092506408691, + "learning_rate": 5.631177095280186e-05, + "loss": 0.3965, + "step": 19470 + }, + { + "epoch": 7.353718384295961, + "grad_norm": 1.315040111541748, + "learning_rate": 5.627076273608027e-05, + "loss": 0.3952, + "step": 19480 + }, + { + "epoch": 7.357493393733485, + "grad_norm": 0.6428912878036499, + "learning_rate": 5.622975023335148e-05, + "loss": 0.3936, + "step": 19490 + }, + { + "epoch": 7.361268403171008, + "grad_norm": 0.7309068441390991, + "learning_rate": 5.618873347264716e-05, + "loss": 0.3972, + "step": 19500 + }, + { + "epoch": 7.365043412608531, + "grad_norm": 0.6258814930915833, + "learning_rate": 5.614771248200188e-05, + "loss": 0.3858, + "step": 19510 + }, + { + "epoch": 7.368818422046055, + "grad_norm": 0.64792400598526, + "learning_rate": 5.6106687289453066e-05, + "loss": 0.398, + "step": 19520 + }, + { + "epoch": 7.372593431483579, + "grad_norm": 0.6821390986442566, + "learning_rate": 5.606565792304108e-05, + "loss": 0.4027, + "step": 19530 + }, + { + "epoch": 7.3763684409211026, + "grad_norm": 0.7018711566925049, + "learning_rate": 5.602462441080909e-05, + "loss": 0.3907, + "step": 19540 + }, + { + "epoch": 7.380143450358626, + "grad_norm": 0.6835519671440125, + "learning_rate": 5.5983586780803135e-05, + "loss": 0.3912, + "step": 19550 + }, + { + "epoch": 7.383918459796149, + "grad_norm": 0.6249481439590454, + "learning_rate": 5.594254506107205e-05, + "loss": 0.3903, + "step": 19560 + }, + { + "epoch": 7.387693469233673, + "grad_norm": 0.6832369565963745, + "learning_rate": 5.590149927966743e-05, + "loss": 0.3882, + "step": 19570 + }, + { + "epoch": 7.391468478671197, + "grad_norm": 0.6031800508499146, + "learning_rate": 5.58604494646437e-05, + "loss": 0.3954, + "step": 19580 + }, + { + "epoch": 7.39524348810872, + "grad_norm": 0.6349557638168335, + "learning_rate": 5.5819395644058025e-05, + "loss": 0.3894, + "step": 19590 + }, + { + "epoch": 7.399018497546244, + "grad_norm": 0.6155382990837097, + "learning_rate": 5.577833784597031e-05, + "loss": 0.3928, + "step": 19600 + }, + { + "epoch": 7.402793506983768, + "grad_norm": 0.6620678305625916, + "learning_rate": 5.573727609844316e-05, + "loss": 0.3906, + "step": 19610 + }, + { + "epoch": 7.406568516421291, + "grad_norm": 0.6608960628509521, + "learning_rate": 5.5696210429541884e-05, + "loss": 0.3975, + "step": 19620 + }, + { + "epoch": 7.410343525858814, + "grad_norm": 0.6048488020896912, + "learning_rate": 5.565514086733451e-05, + "loss": 0.3969, + "step": 19630 + }, + { + "epoch": 7.414118535296338, + "grad_norm": 0.7584668397903442, + "learning_rate": 5.5614067439891657e-05, + "loss": 0.3942, + "step": 19640 + }, + { + "epoch": 7.417893544733862, + "grad_norm": 0.6703759431838989, + "learning_rate": 5.557299017528666e-05, + "loss": 0.3879, + "step": 19650 + }, + { + "epoch": 7.421668554171386, + "grad_norm": 0.6314090490341187, + "learning_rate": 5.5531909101595436e-05, + "loss": 0.391, + "step": 19660 + }, + { + "epoch": 7.425443563608909, + "grad_norm": 0.6210800409317017, + "learning_rate": 5.549082424689649e-05, + "loss": 0.4007, + "step": 19670 + }, + { + "epoch": 7.429218573046432, + "grad_norm": 0.6025400161743164, + "learning_rate": 5.544973563927095e-05, + "loss": 0.3935, + "step": 19680 + }, + { + "epoch": 7.432993582483956, + "grad_norm": 0.7302088141441345, + "learning_rate": 5.540864330680249e-05, + "loss": 0.4046, + "step": 19690 + }, + { + "epoch": 7.43676859192148, + "grad_norm": 0.6146891117095947, + "learning_rate": 5.536754727757733e-05, + "loss": 0.396, + "step": 19700 + }, + { + "epoch": 7.440543601359003, + "grad_norm": 0.6504469513893127, + "learning_rate": 5.532644757968422e-05, + "loss": 0.3957, + "step": 19710 + }, + { + "epoch": 7.444318610796527, + "grad_norm": 0.6892192363739014, + "learning_rate": 5.528534424121441e-05, + "loss": 0.5172, + "step": 19720 + }, + { + "epoch": 7.448093620234051, + "grad_norm": 0.673173189163208, + "learning_rate": 5.524423729026165e-05, + "loss": 0.3959, + "step": 19730 + }, + { + "epoch": 7.451868629671575, + "grad_norm": 0.689117431640625, + "learning_rate": 5.5203126754922164e-05, + "loss": 0.3885, + "step": 19740 + }, + { + "epoch": 7.455643639109097, + "grad_norm": 0.6626030802726746, + "learning_rate": 5.5162012663294585e-05, + "loss": 0.3877, + "step": 19750 + }, + { + "epoch": 7.459418648546621, + "grad_norm": 0.7295408248901367, + "learning_rate": 5.512089504348003e-05, + "loss": 0.392, + "step": 19760 + }, + { + "epoch": 7.463193657984145, + "grad_norm": 0.6033636331558228, + "learning_rate": 5.5079773923582e-05, + "loss": 0.3939, + "step": 19770 + }, + { + "epoch": 7.466968667421669, + "grad_norm": 0.6206973791122437, + "learning_rate": 5.50386493317064e-05, + "loss": 0.3915, + "step": 19780 + }, + { + "epoch": 7.470743676859192, + "grad_norm": 0.6662564277648926, + "learning_rate": 5.49975212959615e-05, + "loss": 0.3923, + "step": 19790 + }, + { + "epoch": 7.474518686296716, + "grad_norm": 0.6757304072380066, + "learning_rate": 5.4956389844457904e-05, + "loss": 0.3936, + "step": 19800 + }, + { + "epoch": 7.478293695734239, + "grad_norm": 0.717755913734436, + "learning_rate": 5.491525500530859e-05, + "loss": 0.3932, + "step": 19810 + }, + { + "epoch": 7.482068705171763, + "grad_norm": 0.6577935218811035, + "learning_rate": 5.487411680662882e-05, + "loss": 0.3985, + "step": 19820 + }, + { + "epoch": 7.485843714609286, + "grad_norm": 0.6445197463035583, + "learning_rate": 5.483297527653618e-05, + "loss": 0.3979, + "step": 19830 + }, + { + "epoch": 7.48961872404681, + "grad_norm": 0.6646354794502258, + "learning_rate": 5.4791830443150516e-05, + "loss": 0.3896, + "step": 19840 + }, + { + "epoch": 7.493393733484334, + "grad_norm": 0.6168741583824158, + "learning_rate": 5.475068233459392e-05, + "loss": 0.398, + "step": 19850 + }, + { + "epoch": 7.497168742921858, + "grad_norm": 0.7372630834579468, + "learning_rate": 5.470953097899075e-05, + "loss": 0.3972, + "step": 19860 + }, + { + "epoch": 7.500943752359381, + "grad_norm": 0.7275758981704712, + "learning_rate": 5.466837640446756e-05, + "loss": 0.3923, + "step": 19870 + }, + { + "epoch": 7.504718761796904, + "grad_norm": 0.6699514389038086, + "learning_rate": 5.462721863915312e-05, + "loss": 0.3958, + "step": 19880 + }, + { + "epoch": 7.508493771234428, + "grad_norm": 0.6782812476158142, + "learning_rate": 5.4586057711178374e-05, + "loss": 0.3939, + "step": 19890 + }, + { + "epoch": 7.512268780671952, + "grad_norm": 0.6210011839866638, + "learning_rate": 5.454489364867642e-05, + "loss": 0.3935, + "step": 19900 + }, + { + "epoch": 7.516043790109475, + "grad_norm": 0.6767253279685974, + "learning_rate": 5.4503726479782523e-05, + "loss": 0.3922, + "step": 19910 + }, + { + "epoch": 7.519818799546999, + "grad_norm": 0.5947399735450745, + "learning_rate": 5.446255623263403e-05, + "loss": 0.3867, + "step": 19920 + }, + { + "epoch": 7.523593808984522, + "grad_norm": 0.6279717683792114, + "learning_rate": 5.4421382935370445e-05, + "loss": 0.3933, + "step": 19930 + }, + { + "epoch": 7.527368818422046, + "grad_norm": 0.6436152458190918, + "learning_rate": 5.438020661613331e-05, + "loss": 0.3879, + "step": 19940 + }, + { + "epoch": 7.5311438278595695, + "grad_norm": 0.6227508187294006, + "learning_rate": 5.433902730306625e-05, + "loss": 0.3899, + "step": 19950 + }, + { + "epoch": 7.534918837297093, + "grad_norm": 0.7720093131065369, + "learning_rate": 5.429784502431495e-05, + "loss": 0.3946, + "step": 19960 + }, + { + "epoch": 7.538693846734617, + "grad_norm": 0.7744880318641663, + "learning_rate": 5.42566598080271e-05, + "loss": 0.3946, + "step": 19970 + }, + { + "epoch": 7.542468856172141, + "grad_norm": 1.112042784690857, + "learning_rate": 5.421547168235241e-05, + "loss": 0.392, + "step": 19980 + }, + { + "epoch": 7.546243865609664, + "grad_norm": 0.6036012172698975, + "learning_rate": 5.417428067544258e-05, + "loss": 0.3993, + "step": 19990 + }, + { + "epoch": 7.550018875047187, + "grad_norm": 0.6224648952484131, + "learning_rate": 5.413308681545126e-05, + "loss": 0.3947, + "step": 20000 + }, + { + "epoch": 7.553793884484711, + "grad_norm": 0.6193528771400452, + "learning_rate": 5.409189013053408e-05, + "loss": 0.3918, + "step": 20010 + }, + { + "epoch": 7.557568893922235, + "grad_norm": 0.5908026099205017, + "learning_rate": 5.4050690648848576e-05, + "loss": 0.3911, + "step": 20020 + }, + { + "epoch": 7.5613439033597585, + "grad_norm": 0.6350475549697876, + "learning_rate": 5.400948839855421e-05, + "loss": 0.3936, + "step": 20030 + }, + { + "epoch": 7.565118912797282, + "grad_norm": 0.6277138590812683, + "learning_rate": 5.396828340781234e-05, + "loss": 0.3937, + "step": 20040 + }, + { + "epoch": 7.568893922234806, + "grad_norm": 0.8242809176445007, + "learning_rate": 5.392707570478617e-05, + "loss": 0.407, + "step": 20050 + }, + { + "epoch": 7.572668931672329, + "grad_norm": 0.7098134756088257, + "learning_rate": 5.388586531764078e-05, + "loss": 0.3781, + "step": 20060 + }, + { + "epoch": 7.5764439411098525, + "grad_norm": 0.9367830753326416, + "learning_rate": 5.384465227454311e-05, + "loss": 0.3971, + "step": 20070 + }, + { + "epoch": 7.580218950547376, + "grad_norm": 0.5797680616378784, + "learning_rate": 5.380343660366184e-05, + "loss": 0.3923, + "step": 20080 + }, + { + "epoch": 7.5839939599849, + "grad_norm": 0.6200445890426636, + "learning_rate": 5.376221833316752e-05, + "loss": 0.3958, + "step": 20090 + }, + { + "epoch": 7.587768969422424, + "grad_norm": 0.651992678642273, + "learning_rate": 5.3720997491232436e-05, + "loss": 0.3893, + "step": 20100 + }, + { + "epoch": 7.5915439788599475, + "grad_norm": 0.6731122732162476, + "learning_rate": 5.367977410603068e-05, + "loss": 0.3918, + "step": 20110 + }, + { + "epoch": 7.595318988297471, + "grad_norm": 0.5849369764328003, + "learning_rate": 5.3638548205738004e-05, + "loss": 0.3909, + "step": 20120 + }, + { + "epoch": 7.599093997734994, + "grad_norm": 0.6614940762519836, + "learning_rate": 5.359731981853194e-05, + "loss": 0.3949, + "step": 20130 + }, + { + "epoch": 7.602869007172518, + "grad_norm": 3.908480644226074, + "learning_rate": 5.35560889725917e-05, + "loss": 0.3875, + "step": 20140 + }, + { + "epoch": 7.6066440166100415, + "grad_norm": 0.6389253735542297, + "learning_rate": 5.3514855696098176e-05, + "loss": 0.3883, + "step": 20150 + }, + { + "epoch": 7.610419026047565, + "grad_norm": 0.6248882412910461, + "learning_rate": 5.347362001723394e-05, + "loss": 0.3848, + "step": 20160 + }, + { + "epoch": 7.614194035485089, + "grad_norm": 0.6055360436439514, + "learning_rate": 5.3432381964183176e-05, + "loss": 0.3936, + "step": 20170 + }, + { + "epoch": 7.617969044922612, + "grad_norm": 0.7338094115257263, + "learning_rate": 5.3391141565131685e-05, + "loss": 0.3904, + "step": 20180 + }, + { + "epoch": 7.621744054360136, + "grad_norm": 0.8716061115264893, + "learning_rate": 5.3349898848266935e-05, + "loss": 0.3902, + "step": 20190 + }, + { + "epoch": 7.625519063797659, + "grad_norm": 1.380915641784668, + "learning_rate": 5.330865384177789e-05, + "loss": 0.3927, + "step": 20200 + }, + { + "epoch": 7.629294073235183, + "grad_norm": 0.6246406435966492, + "learning_rate": 5.326740657385515e-05, + "loss": 0.389, + "step": 20210 + }, + { + "epoch": 7.633069082672707, + "grad_norm": 0.6647712588310242, + "learning_rate": 5.322615707269083e-05, + "loss": 0.3956, + "step": 20220 + }, + { + "epoch": 7.6368440921102305, + "grad_norm": 0.6063628792762756, + "learning_rate": 5.318490536647856e-05, + "loss": 0.3887, + "step": 20230 + }, + { + "epoch": 7.640619101547754, + "grad_norm": 0.625741720199585, + "learning_rate": 5.3143651483413524e-05, + "loss": 0.3936, + "step": 20240 + }, + { + "epoch": 7.644394110985277, + "grad_norm": 0.676576554775238, + "learning_rate": 5.310239545169232e-05, + "loss": 0.3916, + "step": 20250 + }, + { + "epoch": 7.648169120422801, + "grad_norm": 0.5982394218444824, + "learning_rate": 5.30611372995131e-05, + "loss": 0.3868, + "step": 20260 + }, + { + "epoch": 7.651944129860325, + "grad_norm": 0.6712155342102051, + "learning_rate": 5.30198770550754e-05, + "loss": 0.3903, + "step": 20270 + }, + { + "epoch": 7.655719139297848, + "grad_norm": 0.5862340331077576, + "learning_rate": 5.297861474658019e-05, + "loss": 0.389, + "step": 20280 + }, + { + "epoch": 7.659494148735372, + "grad_norm": 0.6452902555465698, + "learning_rate": 5.29373504022299e-05, + "loss": 0.3847, + "step": 20290 + }, + { + "epoch": 7.663269158172896, + "grad_norm": 0.6566249132156372, + "learning_rate": 5.28960840502283e-05, + "loss": 0.3969, + "step": 20300 + }, + { + "epoch": 7.667044167610419, + "grad_norm": 0.6458132863044739, + "learning_rate": 5.285481571878056e-05, + "loss": 0.3942, + "step": 20310 + }, + { + "epoch": 7.670819177047942, + "grad_norm": 0.6279521584510803, + "learning_rate": 5.281354543609321e-05, + "loss": 0.3903, + "step": 20320 + }, + { + "epoch": 7.674594186485466, + "grad_norm": 0.5799058079719543, + "learning_rate": 5.277227323037406e-05, + "loss": 0.3871, + "step": 20330 + }, + { + "epoch": 7.67836919592299, + "grad_norm": 0.5981761813163757, + "learning_rate": 5.273099912983233e-05, + "loss": 0.3854, + "step": 20340 + }, + { + "epoch": 7.682144205360514, + "grad_norm": 0.6565678119659424, + "learning_rate": 5.268972316267843e-05, + "loss": 0.3878, + "step": 20350 + }, + { + "epoch": 7.685919214798037, + "grad_norm": 0.6143926382064819, + "learning_rate": 5.26484453571241e-05, + "loss": 0.3997, + "step": 20360 + }, + { + "epoch": 7.689694224235561, + "grad_norm": 0.630825936794281, + "learning_rate": 5.260716574138235e-05, + "loss": 0.3865, + "step": 20370 + }, + { + "epoch": 7.693469233673084, + "grad_norm": 0.636688768863678, + "learning_rate": 5.256588434366739e-05, + "loss": 0.3895, + "step": 20380 + }, + { + "epoch": 7.697244243110608, + "grad_norm": 0.675250768661499, + "learning_rate": 5.25246011921947e-05, + "loss": 0.3901, + "step": 20390 + }, + { + "epoch": 7.701019252548131, + "grad_norm": 0.6246991157531738, + "learning_rate": 5.248331631518089e-05, + "loss": 0.385, + "step": 20400 + }, + { + "epoch": 7.704794261985655, + "grad_norm": 0.6083090901374817, + "learning_rate": 5.244202974084379e-05, + "loss": 0.3906, + "step": 20410 + }, + { + "epoch": 7.708569271423179, + "grad_norm": 0.63591068983078, + "learning_rate": 5.240074149740239e-05, + "loss": 0.3842, + "step": 20420 + }, + { + "epoch": 7.712344280860702, + "grad_norm": 0.7097567319869995, + "learning_rate": 5.2359451613076814e-05, + "loss": 0.3777, + "step": 20430 + }, + { + "epoch": 7.716119290298225, + "grad_norm": 0.6984173059463501, + "learning_rate": 5.231816011608832e-05, + "loss": 0.3889, + "step": 20440 + }, + { + "epoch": 7.719894299735749, + "grad_norm": 0.6347232460975647, + "learning_rate": 5.227686703465924e-05, + "loss": 0.3888, + "step": 20450 + }, + { + "epoch": 7.723669309173273, + "grad_norm": 0.5677266120910645, + "learning_rate": 5.2235572397013e-05, + "loss": 0.3899, + "step": 20460 + }, + { + "epoch": 7.727444318610797, + "grad_norm": 0.579582929611206, + "learning_rate": 5.2194276231374114e-05, + "loss": 0.3896, + "step": 20470 + }, + { + "epoch": 7.73121932804832, + "grad_norm": 0.5820940732955933, + "learning_rate": 5.21529785659681e-05, + "loss": 0.3921, + "step": 20480 + }, + { + "epoch": 7.734994337485844, + "grad_norm": 0.6141514778137207, + "learning_rate": 5.2111679429021565e-05, + "loss": 0.3986, + "step": 20490 + }, + { + "epoch": 7.738769346923367, + "grad_norm": 0.6320546269416809, + "learning_rate": 5.207037884876205e-05, + "loss": 0.3882, + "step": 20500 + }, + { + "epoch": 7.742544356360891, + "grad_norm": 0.6534587740898132, + "learning_rate": 5.202907685341809e-05, + "loss": 0.3885, + "step": 20510 + }, + { + "epoch": 7.746319365798414, + "grad_norm": 0.8131640553474426, + "learning_rate": 5.198777347121926e-05, + "loss": 0.3791, + "step": 20520 + }, + { + "epoch": 7.750094375235938, + "grad_norm": 0.5980712175369263, + "learning_rate": 5.194646873039598e-05, + "loss": 0.3823, + "step": 20530 + }, + { + "epoch": 7.753869384673462, + "grad_norm": 0.6374219059944153, + "learning_rate": 5.1905162659179696e-05, + "loss": 0.3967, + "step": 20540 + }, + { + "epoch": 7.757644394110986, + "grad_norm": 0.6131422519683838, + "learning_rate": 5.18638552858027e-05, + "loss": 0.3865, + "step": 20550 + }, + { + "epoch": 7.7614194035485085, + "grad_norm": 1.034074306488037, + "learning_rate": 5.182254663849818e-05, + "loss": 0.3907, + "step": 20560 + }, + { + "epoch": 7.765194412986032, + "grad_norm": 0.6826948523521423, + "learning_rate": 5.178123674550023e-05, + "loss": 0.3857, + "step": 20570 + }, + { + "epoch": 7.768969422423556, + "grad_norm": 0.5983694195747375, + "learning_rate": 5.173992563504375e-05, + "loss": 0.3914, + "step": 20580 + }, + { + "epoch": 7.77274443186108, + "grad_norm": 0.5786053538322449, + "learning_rate": 5.169861333536451e-05, + "loss": 0.3863, + "step": 20590 + }, + { + "epoch": 7.776519441298603, + "grad_norm": 0.607843816280365, + "learning_rate": 5.165729987469907e-05, + "loss": 0.3928, + "step": 20600 + }, + { + "epoch": 7.780294450736127, + "grad_norm": 0.597486138343811, + "learning_rate": 5.161598528128478e-05, + "loss": 0.3845, + "step": 20610 + }, + { + "epoch": 7.784069460173651, + "grad_norm": 0.6441140174865723, + "learning_rate": 5.157466958335981e-05, + "loss": 0.3856, + "step": 20620 + }, + { + "epoch": 7.787844469611174, + "grad_norm": 0.8996423482894897, + "learning_rate": 5.1533352809163025e-05, + "loss": 0.3818, + "step": 20630 + }, + { + "epoch": 7.7916194790486974, + "grad_norm": 0.6202008724212646, + "learning_rate": 5.1492034986934046e-05, + "loss": 0.3865, + "step": 20640 + }, + { + "epoch": 7.795394488486221, + "grad_norm": 0.6499298810958862, + "learning_rate": 5.1450716144913225e-05, + "loss": 0.3887, + "step": 20650 + }, + { + "epoch": 7.799169497923745, + "grad_norm": 0.6905611753463745, + "learning_rate": 5.1409396311341595e-05, + "loss": 0.3841, + "step": 20660 + }, + { + "epoch": 7.802944507361269, + "grad_norm": 0.6405202150344849, + "learning_rate": 5.136807551446089e-05, + "loss": 0.3883, + "step": 20670 + }, + { + "epoch": 7.8067195167987915, + "grad_norm": 0.6478602290153503, + "learning_rate": 5.132675378251346e-05, + "loss": 0.3856, + "step": 20680 + }, + { + "epoch": 7.810494526236315, + "grad_norm": 0.6184169054031372, + "learning_rate": 5.1285431143742325e-05, + "loss": 0.3901, + "step": 20690 + }, + { + "epoch": 7.814269535673839, + "grad_norm": 0.6250163316726685, + "learning_rate": 5.1244107626391136e-05, + "loss": 0.4039, + "step": 20700 + }, + { + "epoch": 7.818044545111363, + "grad_norm": 0.6539636254310608, + "learning_rate": 5.12027832587041e-05, + "loss": 0.3866, + "step": 20710 + }, + { + "epoch": 7.821819554548886, + "grad_norm": 0.6491449475288391, + "learning_rate": 5.116145806892607e-05, + "loss": 0.3965, + "step": 20720 + }, + { + "epoch": 7.82559456398641, + "grad_norm": 0.653968870639801, + "learning_rate": 5.1120132085302384e-05, + "loss": 0.3825, + "step": 20730 + }, + { + "epoch": 7.829369573423934, + "grad_norm": 0.7389346957206726, + "learning_rate": 5.107880533607898e-05, + "loss": 0.3935, + "step": 20740 + }, + { + "epoch": 7.833144582861458, + "grad_norm": 0.6781933903694153, + "learning_rate": 5.103747784950231e-05, + "loss": 0.3852, + "step": 20750 + }, + { + "epoch": 7.8369195922989805, + "grad_norm": 0.6142136454582214, + "learning_rate": 5.09961496538193e-05, + "loss": 0.3956, + "step": 20760 + }, + { + "epoch": 7.840694601736504, + "grad_norm": 0.5901447534561157, + "learning_rate": 5.095482077727742e-05, + "loss": 0.3849, + "step": 20770 + }, + { + "epoch": 7.844469611174028, + "grad_norm": 0.6106001138687134, + "learning_rate": 5.091349124812452e-05, + "loss": 0.3852, + "step": 20780 + }, + { + "epoch": 7.848244620611552, + "grad_norm": 0.7389256954193115, + "learning_rate": 5.087216109460897e-05, + "loss": 0.3912, + "step": 20790 + }, + { + "epoch": 7.852019630049075, + "grad_norm": 0.7008233070373535, + "learning_rate": 5.083083034497954e-05, + "loss": 0.3856, + "step": 20800 + }, + { + "epoch": 7.855794639486598, + "grad_norm": 0.5946593880653381, + "learning_rate": 5.07894990274854e-05, + "loss": 0.3878, + "step": 20810 + }, + { + "epoch": 7.859569648924122, + "grad_norm": 0.5986919403076172, + "learning_rate": 5.074816717037614e-05, + "loss": 0.3842, + "step": 20820 + }, + { + "epoch": 7.863344658361646, + "grad_norm": 0.7096017003059387, + "learning_rate": 5.070683480190165e-05, + "loss": 0.3923, + "step": 20830 + }, + { + "epoch": 7.8671196677991695, + "grad_norm": 0.6921922564506531, + "learning_rate": 5.066550195031223e-05, + "loss": 0.3864, + "step": 20840 + }, + { + "epoch": 7.870894677236693, + "grad_norm": 0.5950612425804138, + "learning_rate": 5.062416864385852e-05, + "loss": 0.384, + "step": 20850 + }, + { + "epoch": 7.874669686674217, + "grad_norm": 0.6273109912872314, + "learning_rate": 5.058283491079142e-05, + "loss": 0.3778, + "step": 20860 + }, + { + "epoch": 7.878444696111741, + "grad_norm": 0.7356699109077454, + "learning_rate": 5.054150077936216e-05, + "loss": 0.3905, + "step": 20870 + }, + { + "epoch": 7.8822197055492635, + "grad_norm": 0.6300787329673767, + "learning_rate": 5.0500166277822214e-05, + "loss": 0.387, + "step": 20880 + }, + { + "epoch": 7.885994714986787, + "grad_norm": 0.6487736105918884, + "learning_rate": 5.0458831434423334e-05, + "loss": 0.3935, + "step": 20890 + }, + { + "epoch": 7.889769724424311, + "grad_norm": 0.6694430708885193, + "learning_rate": 5.0417496277417506e-05, + "loss": 0.3982, + "step": 20900 + }, + { + "epoch": 7.893544733861835, + "grad_norm": 0.6112816333770752, + "learning_rate": 5.037616083505691e-05, + "loss": 0.3769, + "step": 20910 + }, + { + "epoch": 7.8973197432993585, + "grad_norm": 0.5769428014755249, + "learning_rate": 5.0334825135593935e-05, + "loss": 0.387, + "step": 20920 + }, + { + "epoch": 7.901094752736882, + "grad_norm": 0.623289167881012, + "learning_rate": 5.029348920728111e-05, + "loss": 0.3874, + "step": 20930 + }, + { + "epoch": 7.904869762174405, + "grad_norm": 0.6469810605049133, + "learning_rate": 5.0252153078371186e-05, + "loss": 0.3863, + "step": 20940 + }, + { + "epoch": 7.908644771611929, + "grad_norm": 0.6396733522415161, + "learning_rate": 5.021081677711704e-05, + "loss": 0.3933, + "step": 20950 + }, + { + "epoch": 7.9124197810494525, + "grad_norm": 0.6472081542015076, + "learning_rate": 5.016948033177159e-05, + "loss": 0.3804, + "step": 20960 + }, + { + "epoch": 7.916194790486976, + "grad_norm": 0.670315146446228, + "learning_rate": 5.012814377058793e-05, + "loss": 0.3922, + "step": 20970 + }, + { + "epoch": 7.9199697999245, + "grad_norm": 0.6452091932296753, + "learning_rate": 5.008680712181921e-05, + "loss": 0.3906, + "step": 20980 + }, + { + "epoch": 7.923744809362024, + "grad_norm": 0.6105216145515442, + "learning_rate": 5.0045470413718645e-05, + "loss": 0.3864, + "step": 20990 + }, + { + "epoch": 7.9275198187995475, + "grad_norm": 0.6681371331214905, + "learning_rate": 5.00041336745395e-05, + "loss": 0.3806, + "step": 21000 + }, + { + "epoch": 7.93129482823707, + "grad_norm": 0.6200874447822571, + "learning_rate": 4.996279693253499e-05, + "loss": 0.3895, + "step": 21010 + }, + { + "epoch": 7.935069837674594, + "grad_norm": 0.620955765247345, + "learning_rate": 4.992146021595847e-05, + "loss": 0.3872, + "step": 21020 + }, + { + "epoch": 7.938844847112118, + "grad_norm": 0.6408780217170715, + "learning_rate": 4.988012355306313e-05, + "loss": 0.3821, + "step": 21030 + }, + { + "epoch": 7.9426198565496415, + "grad_norm": 0.6567278504371643, + "learning_rate": 4.98387869721022e-05, + "loss": 0.3987, + "step": 21040 + }, + { + "epoch": 7.946394865987165, + "grad_norm": 0.6321324706077576, + "learning_rate": 4.9797450501328866e-05, + "loss": 0.382, + "step": 21050 + }, + { + "epoch": 7.950169875424688, + "grad_norm": 0.6207996606826782, + "learning_rate": 4.97561141689962e-05, + "loss": 0.3921, + "step": 21060 + }, + { + "epoch": 7.953944884862212, + "grad_norm": 0.6274701952934265, + "learning_rate": 4.971477800335721e-05, + "loss": 0.3803, + "step": 21070 + }, + { + "epoch": 7.957719894299736, + "grad_norm": 0.6380013227462769, + "learning_rate": 4.967344203266475e-05, + "loss": 0.3857, + "step": 21080 + }, + { + "epoch": 7.961494903737259, + "grad_norm": 0.6619207859039307, + "learning_rate": 4.9632106285171584e-05, + "loss": 0.3828, + "step": 21090 + }, + { + "epoch": 7.965269913174783, + "grad_norm": 0.6286654472351074, + "learning_rate": 4.959077078913031e-05, + "loss": 0.3857, + "step": 21100 + }, + { + "epoch": 7.969044922612307, + "grad_norm": 0.7368075847625732, + "learning_rate": 4.954943557279333e-05, + "loss": 0.3843, + "step": 21110 + }, + { + "epoch": 7.9728199320498305, + "grad_norm": 0.6597092747688293, + "learning_rate": 4.9508100664412916e-05, + "loss": 0.3883, + "step": 21120 + }, + { + "epoch": 7.976594941487353, + "grad_norm": 0.673981249332428, + "learning_rate": 4.946676609224105e-05, + "loss": 0.3855, + "step": 21130 + }, + { + "epoch": 7.980369950924877, + "grad_norm": 0.6356333494186401, + "learning_rate": 4.942543188452952e-05, + "loss": 0.3922, + "step": 21140 + }, + { + "epoch": 7.984144960362401, + "grad_norm": 0.6984882354736328, + "learning_rate": 4.938409806952988e-05, + "loss": 0.3834, + "step": 21150 + }, + { + "epoch": 7.987919969799925, + "grad_norm": 0.6634735465049744, + "learning_rate": 4.93427646754934e-05, + "loss": 0.3838, + "step": 21160 + }, + { + "epoch": 7.991694979237448, + "grad_norm": 0.6481113433837891, + "learning_rate": 4.930143173067108e-05, + "loss": 0.3863, + "step": 21170 + }, + { + "epoch": 7.995469988674972, + "grad_norm": 0.684773325920105, + "learning_rate": 4.9260099263313565e-05, + "loss": 0.379, + "step": 21180 + }, + { + "epoch": 7.999244998112495, + "grad_norm": 0.6397860646247864, + "learning_rate": 4.921876730167123e-05, + "loss": 0.3962, + "step": 21190 + }, + { + "epoch": 8.003020007550019, + "grad_norm": 0.6320487260818481, + "learning_rate": 4.917743587399409e-05, + "loss": 0.3839, + "step": 21200 + }, + { + "epoch": 8.006795016987542, + "grad_norm": 0.6578797101974487, + "learning_rate": 4.913610500853178e-05, + "loss": 0.3841, + "step": 21210 + }, + { + "epoch": 8.010570026425066, + "grad_norm": 0.6178431510925293, + "learning_rate": 4.909477473353354e-05, + "loss": 0.3848, + "step": 21220 + }, + { + "epoch": 8.01434503586259, + "grad_norm": 0.6298217177391052, + "learning_rate": 4.9053445077248236e-05, + "loss": 0.3876, + "step": 21230 + }, + { + "epoch": 8.018120045300114, + "grad_norm": 0.6131578087806702, + "learning_rate": 4.901211606792429e-05, + "loss": 0.3914, + "step": 21240 + }, + { + "epoch": 8.021895054737637, + "grad_norm": 0.6626631021499634, + "learning_rate": 4.89707877338097e-05, + "loss": 0.3864, + "step": 21250 + }, + { + "epoch": 8.025670064175161, + "grad_norm": 0.6320914626121521, + "learning_rate": 4.892946010315199e-05, + "loss": 0.3853, + "step": 21260 + }, + { + "epoch": 8.029445073612685, + "grad_norm": 0.657503068447113, + "learning_rate": 4.8888133204198204e-05, + "loss": 0.3864, + "step": 21270 + }, + { + "epoch": 8.033220083050207, + "grad_norm": 0.6045080423355103, + "learning_rate": 4.8846807065194886e-05, + "loss": 0.3859, + "step": 21280 + }, + { + "epoch": 8.03699509248773, + "grad_norm": 0.6179243326187134, + "learning_rate": 4.880548171438806e-05, + "loss": 0.39, + "step": 21290 + }, + { + "epoch": 8.040770101925254, + "grad_norm": 0.7276541590690613, + "learning_rate": 4.8764157180023245e-05, + "loss": 0.3869, + "step": 21300 + }, + { + "epoch": 8.044545111362778, + "grad_norm": 0.5852580666542053, + "learning_rate": 4.872283349034533e-05, + "loss": 0.3847, + "step": 21310 + }, + { + "epoch": 8.048320120800302, + "grad_norm": 0.8796269297599792, + "learning_rate": 4.8681510673598674e-05, + "loss": 0.3854, + "step": 21320 + }, + { + "epoch": 8.052095130237825, + "grad_norm": 0.5888720750808716, + "learning_rate": 4.8640188758027046e-05, + "loss": 0.3767, + "step": 21330 + }, + { + "epoch": 8.05587013967535, + "grad_norm": 0.6211332678794861, + "learning_rate": 4.859886777187357e-05, + "loss": 0.392, + "step": 21340 + }, + { + "epoch": 8.059645149112873, + "grad_norm": 0.6416304111480713, + "learning_rate": 4.855754774338077e-05, + "loss": 0.3906, + "step": 21350 + }, + { + "epoch": 8.063420158550397, + "grad_norm": 0.6397213339805603, + "learning_rate": 4.851622870079048e-05, + "loss": 0.3865, + "step": 21360 + }, + { + "epoch": 8.06719516798792, + "grad_norm": 0.655972957611084, + "learning_rate": 4.847491067234389e-05, + "loss": 0.3845, + "step": 21370 + }, + { + "epoch": 8.070970177425444, + "grad_norm": 0.7151166796684265, + "learning_rate": 4.843359368628146e-05, + "loss": 0.379, + "step": 21380 + }, + { + "epoch": 8.074745186862968, + "grad_norm": 0.7403479814529419, + "learning_rate": 4.8392277770842975e-05, + "loss": 0.3857, + "step": 21390 + }, + { + "epoch": 8.078520196300492, + "grad_norm": 0.8259317278862, + "learning_rate": 4.83509629542675e-05, + "loss": 0.3833, + "step": 21400 + }, + { + "epoch": 8.082295205738014, + "grad_norm": 0.5347198843955994, + "learning_rate": 4.830964926479329e-05, + "loss": 0.3884, + "step": 21410 + }, + { + "epoch": 8.086070215175537, + "grad_norm": 0.6426698565483093, + "learning_rate": 4.826833673065785e-05, + "loss": 0.3822, + "step": 21420 + }, + { + "epoch": 8.089845224613061, + "grad_norm": 0.6192202568054199, + "learning_rate": 4.822702538009794e-05, + "loss": 0.386, + "step": 21430 + }, + { + "epoch": 8.093620234050585, + "grad_norm": 0.622089684009552, + "learning_rate": 4.818571524134945e-05, + "loss": 0.3854, + "step": 21440 + }, + { + "epoch": 8.097395243488108, + "grad_norm": 0.6114779114723206, + "learning_rate": 4.8144406342647496e-05, + "loss": 0.391, + "step": 21450 + }, + { + "epoch": 8.101170252925632, + "grad_norm": 0.6789860129356384, + "learning_rate": 4.81030987122263e-05, + "loss": 0.3818, + "step": 21460 + }, + { + "epoch": 8.104945262363156, + "grad_norm": 0.6163697242736816, + "learning_rate": 4.806179237831926e-05, + "loss": 0.3791, + "step": 21470 + }, + { + "epoch": 8.10872027180068, + "grad_norm": 0.7106571793556213, + "learning_rate": 4.802048736915884e-05, + "loss": 0.3811, + "step": 21480 + }, + { + "epoch": 8.112495281238203, + "grad_norm": 0.6782732009887695, + "learning_rate": 4.797918371297666e-05, + "loss": 0.3845, + "step": 21490 + }, + { + "epoch": 8.116270290675727, + "grad_norm": 0.662436842918396, + "learning_rate": 4.793788143800334e-05, + "loss": 0.39, + "step": 21500 + }, + { + "epoch": 8.12004530011325, + "grad_norm": 0.6009316444396973, + "learning_rate": 4.789658057246862e-05, + "loss": 0.3929, + "step": 21510 + }, + { + "epoch": 8.123820309550775, + "grad_norm": 0.628048300743103, + "learning_rate": 4.7855281144601227e-05, + "loss": 0.386, + "step": 21520 + }, + { + "epoch": 8.127595318988298, + "grad_norm": 0.6932970881462097, + "learning_rate": 4.781398318262897e-05, + "loss": 0.3889, + "step": 21530 + }, + { + "epoch": 8.13137032842582, + "grad_norm": 0.6603420376777649, + "learning_rate": 4.777268671477858e-05, + "loss": 0.3847, + "step": 21540 + }, + { + "epoch": 8.135145337863344, + "grad_norm": 0.7748346328735352, + "learning_rate": 4.773139176927582e-05, + "loss": 0.3916, + "step": 21550 + }, + { + "epoch": 8.138920347300868, + "grad_norm": 0.7710997462272644, + "learning_rate": 4.769009837434539e-05, + "loss": 0.3834, + "step": 21560 + }, + { + "epoch": 8.142695356738392, + "grad_norm": 0.6665760278701782, + "learning_rate": 4.764880655821095e-05, + "loss": 0.3885, + "step": 21570 + }, + { + "epoch": 8.146470366175915, + "grad_norm": 0.6791707873344421, + "learning_rate": 4.760751634909508e-05, + "loss": 0.3798, + "step": 21580 + }, + { + "epoch": 8.150245375613439, + "grad_norm": 0.6167095899581909, + "learning_rate": 4.756622777521919e-05, + "loss": 0.3799, + "step": 21590 + }, + { + "epoch": 8.154020385050963, + "grad_norm": 0.7474955320358276, + "learning_rate": 4.752494086480368e-05, + "loss": 0.3873, + "step": 21600 + }, + { + "epoch": 8.157795394488486, + "grad_norm": 0.6586377024650574, + "learning_rate": 4.7483655646067744e-05, + "loss": 0.386, + "step": 21610 + }, + { + "epoch": 8.16157040392601, + "grad_norm": 0.6137120127677917, + "learning_rate": 4.744237214722944e-05, + "loss": 0.3842, + "step": 21620 + }, + { + "epoch": 8.165345413363534, + "grad_norm": 0.613304615020752, + "learning_rate": 4.740109039650567e-05, + "loss": 0.3886, + "step": 21630 + }, + { + "epoch": 8.169120422801058, + "grad_norm": 0.6361426115036011, + "learning_rate": 4.73598104221121e-05, + "loss": 0.3856, + "step": 21640 + }, + { + "epoch": 8.172895432238581, + "grad_norm": 1.1138379573822021, + "learning_rate": 4.731853225226322e-05, + "loss": 0.3922, + "step": 21650 + }, + { + "epoch": 8.176670441676103, + "grad_norm": 0.666292667388916, + "learning_rate": 4.727725591517225e-05, + "loss": 0.3847, + "step": 21660 + }, + { + "epoch": 8.180445451113627, + "grad_norm": 0.6527111530303955, + "learning_rate": 4.723598143905119e-05, + "loss": 0.3905, + "step": 21670 + }, + { + "epoch": 8.18422046055115, + "grad_norm": 0.634685754776001, + "learning_rate": 4.719470885211077e-05, + "loss": 0.3734, + "step": 21680 + }, + { + "epoch": 8.187995469988675, + "grad_norm": 0.7132300734519958, + "learning_rate": 4.7153438182560387e-05, + "loss": 0.3891, + "step": 21690 + }, + { + "epoch": 8.191770479426198, + "grad_norm": 0.6142542958259583, + "learning_rate": 4.711216945860815e-05, + "loss": 0.3857, + "step": 21700 + }, + { + "epoch": 8.195545488863722, + "grad_norm": 0.6071975827217102, + "learning_rate": 4.707090270846088e-05, + "loss": 0.3857, + "step": 21710 + }, + { + "epoch": 8.199320498301246, + "grad_norm": 0.7059184312820435, + "learning_rate": 4.702963796032397e-05, + "loss": 0.3856, + "step": 21720 + }, + { + "epoch": 8.20309550773877, + "grad_norm": 0.6421445608139038, + "learning_rate": 4.6988375242401514e-05, + "loss": 0.3778, + "step": 21730 + }, + { + "epoch": 8.206870517176293, + "grad_norm": 0.7011498808860779, + "learning_rate": 4.694711458289618e-05, + "loss": 0.3898, + "step": 21740 + }, + { + "epoch": 8.210645526613817, + "grad_norm": 0.7598146796226501, + "learning_rate": 4.690585601000925e-05, + "loss": 0.3778, + "step": 21750 + }, + { + "epoch": 8.21442053605134, + "grad_norm": 0.6215041875839233, + "learning_rate": 4.686459955194055e-05, + "loss": 0.3854, + "step": 21760 + }, + { + "epoch": 8.218195545488864, + "grad_norm": 0.7370675206184387, + "learning_rate": 4.6823345236888504e-05, + "loss": 0.3823, + "step": 21770 + }, + { + "epoch": 8.221970554926388, + "grad_norm": 0.6150240898132324, + "learning_rate": 4.678209309305002e-05, + "loss": 0.3955, + "step": 21780 + }, + { + "epoch": 8.22574556436391, + "grad_norm": 0.5834974050521851, + "learning_rate": 4.674084314862057e-05, + "loss": 0.3852, + "step": 21790 + }, + { + "epoch": 8.229520573801434, + "grad_norm": 0.6348870992660522, + "learning_rate": 4.669959543179409e-05, + "loss": 0.3903, + "step": 21800 + }, + { + "epoch": 8.233295583238958, + "grad_norm": 0.7599389553070068, + "learning_rate": 4.665834997076303e-05, + "loss": 0.3849, + "step": 21810 + }, + { + "epoch": 8.237070592676481, + "grad_norm": 0.7290443181991577, + "learning_rate": 4.661710679371823e-05, + "loss": 0.3846, + "step": 21820 + }, + { + "epoch": 8.240845602114005, + "grad_norm": 0.6612056493759155, + "learning_rate": 4.657586592884905e-05, + "loss": 0.3849, + "step": 21830 + }, + { + "epoch": 8.244620611551529, + "grad_norm": 0.5885748863220215, + "learning_rate": 4.653462740434322e-05, + "loss": 0.3801, + "step": 21840 + }, + { + "epoch": 8.248395620989053, + "grad_norm": 0.7608574032783508, + "learning_rate": 4.649339124838689e-05, + "loss": 0.3873, + "step": 21850 + }, + { + "epoch": 8.252170630426576, + "grad_norm": 0.5708411335945129, + "learning_rate": 4.6452157489164574e-05, + "loss": 0.3858, + "step": 21860 + }, + { + "epoch": 8.2559456398641, + "grad_norm": 0.6522685289382935, + "learning_rate": 4.6410926154859155e-05, + "loss": 0.3869, + "step": 21870 + }, + { + "epoch": 8.259720649301624, + "grad_norm": 0.688291609287262, + "learning_rate": 4.636969727365186e-05, + "loss": 0.3839, + "step": 21880 + }, + { + "epoch": 8.263495658739147, + "grad_norm": 0.7088874578475952, + "learning_rate": 4.632847087372226e-05, + "loss": 0.3872, + "step": 21890 + }, + { + "epoch": 8.267270668176671, + "grad_norm": 0.6327393054962158, + "learning_rate": 4.628724698324818e-05, + "loss": 0.3819, + "step": 21900 + }, + { + "epoch": 8.271045677614193, + "grad_norm": 0.6789700388908386, + "learning_rate": 4.6246025630405795e-05, + "loss": 0.384, + "step": 21910 + }, + { + "epoch": 8.274820687051717, + "grad_norm": 0.6304154396057129, + "learning_rate": 4.6204806843369474e-05, + "loss": 0.3883, + "step": 21920 + }, + { + "epoch": 8.27859569648924, + "grad_norm": 0.6723567843437195, + "learning_rate": 4.616359065031191e-05, + "loss": 0.386, + "step": 21930 + }, + { + "epoch": 8.282370705926764, + "grad_norm": 0.801613450050354, + "learning_rate": 4.6122377079403946e-05, + "loss": 0.3893, + "step": 21940 + }, + { + "epoch": 8.286145715364288, + "grad_norm": 0.6835589408874512, + "learning_rate": 4.6081166158814695e-05, + "loss": 0.3801, + "step": 21950 + }, + { + "epoch": 8.289920724801812, + "grad_norm": 0.6579275131225586, + "learning_rate": 4.603995791671144e-05, + "loss": 0.3877, + "step": 21960 + }, + { + "epoch": 8.293695734239336, + "grad_norm": 0.6721914410591125, + "learning_rate": 4.599875238125957e-05, + "loss": 0.382, + "step": 21970 + }, + { + "epoch": 8.29747074367686, + "grad_norm": 0.6746684312820435, + "learning_rate": 4.595754958062273e-05, + "loss": 0.3833, + "step": 21980 + }, + { + "epoch": 8.301245753114383, + "grad_norm": 0.6919302940368652, + "learning_rate": 4.591634954296265e-05, + "loss": 0.3862, + "step": 21990 + }, + { + "epoch": 8.305020762551907, + "grad_norm": 0.6918922662734985, + "learning_rate": 4.587515229643913e-05, + "loss": 0.3834, + "step": 22000 + }, + { + "epoch": 8.30879577198943, + "grad_norm": 0.7155516743659973, + "learning_rate": 4.583395786921013e-05, + "loss": 0.3745, + "step": 22010 + }, + { + "epoch": 8.312570781426954, + "grad_norm": 0.607598602771759, + "learning_rate": 4.579276628943164e-05, + "loss": 0.3794, + "step": 22020 + }, + { + "epoch": 8.316345790864478, + "grad_norm": 0.673291027545929, + "learning_rate": 4.575157758525772e-05, + "loss": 0.3824, + "step": 22030 + }, + { + "epoch": 8.320120800302, + "grad_norm": 0.7992842197418213, + "learning_rate": 4.571039178484046e-05, + "loss": 0.3756, + "step": 22040 + }, + { + "epoch": 8.323895809739524, + "grad_norm": 0.7000048756599426, + "learning_rate": 4.566920891632998e-05, + "loss": 0.38, + "step": 22050 + }, + { + "epoch": 8.327670819177047, + "grad_norm": 0.6638155579566956, + "learning_rate": 4.562802900787436e-05, + "loss": 0.3841, + "step": 22060 + }, + { + "epoch": 8.331445828614571, + "grad_norm": 0.6273589730262756, + "learning_rate": 4.558685208761968e-05, + "loss": 0.3826, + "step": 22070 + }, + { + "epoch": 8.335220838052095, + "grad_norm": 0.6618062853813171, + "learning_rate": 4.554567818370998e-05, + "loss": 0.3874, + "step": 22080 + }, + { + "epoch": 8.338995847489619, + "grad_norm": 0.6083613038063049, + "learning_rate": 4.550450732428726e-05, + "loss": 0.3755, + "step": 22090 + }, + { + "epoch": 8.342770856927142, + "grad_norm": 0.6382494568824768, + "learning_rate": 4.546333953749137e-05, + "loss": 0.3794, + "step": 22100 + }, + { + "epoch": 8.346545866364666, + "grad_norm": 0.6135038733482361, + "learning_rate": 4.5422174851460154e-05, + "loss": 0.3803, + "step": 22110 + }, + { + "epoch": 8.35032087580219, + "grad_norm": 0.6180480718612671, + "learning_rate": 4.538101329432924e-05, + "loss": 0.3844, + "step": 22120 + }, + { + "epoch": 8.354095885239714, + "grad_norm": 0.6306928396224976, + "learning_rate": 4.5339854894232195e-05, + "loss": 0.3843, + "step": 22130 + }, + { + "epoch": 8.357870894677237, + "grad_norm": 0.594245433807373, + "learning_rate": 4.52986996793004e-05, + "loss": 0.3936, + "step": 22140 + }, + { + "epoch": 8.361645904114761, + "grad_norm": 0.6511313319206238, + "learning_rate": 4.5257547677663024e-05, + "loss": 0.3795, + "step": 22150 + }, + { + "epoch": 8.365420913552285, + "grad_norm": 0.6905748248100281, + "learning_rate": 4.52163989174471e-05, + "loss": 0.378, + "step": 22160 + }, + { + "epoch": 8.369195922989807, + "grad_norm": 0.6889414191246033, + "learning_rate": 4.51752534267774e-05, + "loss": 0.3779, + "step": 22170 + }, + { + "epoch": 8.37297093242733, + "grad_norm": 0.7038361430168152, + "learning_rate": 4.513411123377649e-05, + "loss": 0.3789, + "step": 22180 + }, + { + "epoch": 8.376745941864854, + "grad_norm": 0.8612080216407776, + "learning_rate": 4.5092972366564675e-05, + "loss": 0.3814, + "step": 22190 + }, + { + "epoch": 8.380520951302378, + "grad_norm": 0.6390385627746582, + "learning_rate": 4.505183685325997e-05, + "loss": 0.3787, + "step": 22200 + }, + { + "epoch": 8.384295960739902, + "grad_norm": 0.7001070976257324, + "learning_rate": 4.5010704721978125e-05, + "loss": 0.3821, + "step": 22210 + }, + { + "epoch": 8.388070970177425, + "grad_norm": 0.6782472729682922, + "learning_rate": 4.496957600083255e-05, + "loss": 0.3777, + "step": 22220 + }, + { + "epoch": 8.39184597961495, + "grad_norm": 0.6302990317344666, + "learning_rate": 4.4928450717934343e-05, + "loss": 0.379, + "step": 22230 + }, + { + "epoch": 8.395620989052473, + "grad_norm": 0.6895720958709717, + "learning_rate": 4.488732890139227e-05, + "loss": 0.3809, + "step": 22240 + }, + { + "epoch": 8.399395998489997, + "grad_norm": 0.6838264465332031, + "learning_rate": 4.4846210579312665e-05, + "loss": 0.3814, + "step": 22250 + }, + { + "epoch": 8.40317100792752, + "grad_norm": 0.6539376974105835, + "learning_rate": 4.480509577979953e-05, + "loss": 0.3828, + "step": 22260 + }, + { + "epoch": 8.406946017365044, + "grad_norm": 0.6490859985351562, + "learning_rate": 4.476398453095445e-05, + "loss": 0.387, + "step": 22270 + }, + { + "epoch": 8.410721026802568, + "grad_norm": 0.6515709161758423, + "learning_rate": 4.472287686087656e-05, + "loss": 0.389, + "step": 22280 + }, + { + "epoch": 8.41449603624009, + "grad_norm": 0.6394408941268921, + "learning_rate": 4.468177279766259e-05, + "loss": 0.381, + "step": 22290 + }, + { + "epoch": 8.418271045677614, + "grad_norm": 0.6975342631340027, + "learning_rate": 4.4640672369406746e-05, + "loss": 0.373, + "step": 22300 + }, + { + "epoch": 8.422046055115137, + "grad_norm": 0.6658486723899841, + "learning_rate": 4.459957560420082e-05, + "loss": 0.3773, + "step": 22310 + }, + { + "epoch": 8.425821064552661, + "grad_norm": 0.649869441986084, + "learning_rate": 4.455848253013403e-05, + "loss": 0.3816, + "step": 22320 + }, + { + "epoch": 8.429596073990185, + "grad_norm": 0.7822151184082031, + "learning_rate": 4.4517393175293146e-05, + "loss": 0.3872, + "step": 22330 + }, + { + "epoch": 8.433371083427708, + "grad_norm": 0.6239002346992493, + "learning_rate": 4.447630756776232e-05, + "loss": 0.3834, + "step": 22340 + }, + { + "epoch": 8.437146092865232, + "grad_norm": 0.6562296748161316, + "learning_rate": 4.443522573562318e-05, + "loss": 0.3933, + "step": 22350 + }, + { + "epoch": 8.440921102302756, + "grad_norm": 0.6250916123390198, + "learning_rate": 4.4394147706954776e-05, + "loss": 0.383, + "step": 22360 + }, + { + "epoch": 8.44469611174028, + "grad_norm": 0.6783263683319092, + "learning_rate": 4.435307350983355e-05, + "loss": 0.382, + "step": 22370 + }, + { + "epoch": 8.448471121177803, + "grad_norm": 0.5855938792228699, + "learning_rate": 4.4312003172333326e-05, + "loss": 0.3753, + "step": 22380 + }, + { + "epoch": 8.452246130615327, + "grad_norm": 0.6959243416786194, + "learning_rate": 4.427093672252531e-05, + "loss": 0.3769, + "step": 22390 + }, + { + "epoch": 8.45602114005285, + "grad_norm": 0.750425398349762, + "learning_rate": 4.422987418847802e-05, + "loss": 0.3819, + "step": 22400 + }, + { + "epoch": 8.459796149490373, + "grad_norm": 0.6782823204994202, + "learning_rate": 4.4188815598257325e-05, + "loss": 0.3751, + "step": 22410 + }, + { + "epoch": 8.463571158927897, + "grad_norm": 0.7516863346099854, + "learning_rate": 4.414776097992638e-05, + "loss": 0.3794, + "step": 22420 + }, + { + "epoch": 8.46734616836542, + "grad_norm": 0.694094717502594, + "learning_rate": 4.4106710361545595e-05, + "loss": 0.3835, + "step": 22430 + }, + { + "epoch": 8.471121177802944, + "grad_norm": 0.7622425556182861, + "learning_rate": 4.406566377117272e-05, + "loss": 0.3817, + "step": 22440 + }, + { + "epoch": 8.474896187240468, + "grad_norm": 0.6150171160697937, + "learning_rate": 4.40246212368627e-05, + "loss": 0.3821, + "step": 22450 + }, + { + "epoch": 8.478671196677992, + "grad_norm": 0.5998985171318054, + "learning_rate": 4.3983582786667715e-05, + "loss": 0.3769, + "step": 22460 + }, + { + "epoch": 8.482446206115515, + "grad_norm": 0.6367303133010864, + "learning_rate": 4.394254844863716e-05, + "loss": 0.3853, + "step": 22470 + }, + { + "epoch": 8.486221215553039, + "grad_norm": 0.7072235941886902, + "learning_rate": 4.390151825081762e-05, + "loss": 0.3844, + "step": 22480 + }, + { + "epoch": 8.489996224990563, + "grad_norm": 0.678936779499054, + "learning_rate": 4.386049222125286e-05, + "loss": 0.3695, + "step": 22490 + }, + { + "epoch": 8.493771234428086, + "grad_norm": 0.6695368885993958, + "learning_rate": 4.3819470387983774e-05, + "loss": 0.3838, + "step": 22500 + }, + { + "epoch": 8.49754624386561, + "grad_norm": 0.6475124359130859, + "learning_rate": 4.377845277904841e-05, + "loss": 0.3847, + "step": 22510 + }, + { + "epoch": 8.501321253303134, + "grad_norm": 0.6073307991027832, + "learning_rate": 4.37374394224819e-05, + "loss": 0.3805, + "step": 22520 + }, + { + "epoch": 8.505096262740658, + "grad_norm": 0.6134095191955566, + "learning_rate": 4.369643034631648e-05, + "loss": 0.388, + "step": 22530 + }, + { + "epoch": 8.50887127217818, + "grad_norm": 1.0386673212051392, + "learning_rate": 4.365542557858149e-05, + "loss": 0.3853, + "step": 22540 + }, + { + "epoch": 8.512646281615703, + "grad_norm": 0.6191480755805969, + "learning_rate": 4.361442514730329e-05, + "loss": 0.3866, + "step": 22550 + }, + { + "epoch": 8.516421291053227, + "grad_norm": 0.6350696682929993, + "learning_rate": 4.357342908050528e-05, + "loss": 0.3817, + "step": 22560 + }, + { + "epoch": 8.52019630049075, + "grad_norm": 0.6684324145317078, + "learning_rate": 4.3532437406207895e-05, + "loss": 0.382, + "step": 22570 + }, + { + "epoch": 8.523971309928275, + "grad_norm": 0.6637814044952393, + "learning_rate": 4.349145015242856e-05, + "loss": 0.3775, + "step": 22580 + }, + { + "epoch": 8.527746319365798, + "grad_norm": 0.7314398884773254, + "learning_rate": 4.345046734718168e-05, + "loss": 0.3755, + "step": 22590 + }, + { + "epoch": 8.531521328803322, + "grad_norm": 0.6954400539398193, + "learning_rate": 4.34094890184786e-05, + "loss": 0.3816, + "step": 22600 + }, + { + "epoch": 8.535296338240846, + "grad_norm": 0.6581496596336365, + "learning_rate": 4.336851519432765e-05, + "loss": 0.3869, + "step": 22610 + }, + { + "epoch": 8.53907134767837, + "grad_norm": 0.6038832068443298, + "learning_rate": 4.332754590273403e-05, + "loss": 0.3821, + "step": 22620 + }, + { + "epoch": 8.542846357115893, + "grad_norm": 0.6651731133460999, + "learning_rate": 4.3286581171699855e-05, + "loss": 0.3916, + "step": 22630 + }, + { + "epoch": 8.546621366553417, + "grad_norm": 0.6608867645263672, + "learning_rate": 4.324562102922416e-05, + "loss": 0.3764, + "step": 22640 + }, + { + "epoch": 8.55039637599094, + "grad_norm": 0.6170617938041687, + "learning_rate": 4.320466550330278e-05, + "loss": 0.3805, + "step": 22650 + }, + { + "epoch": 8.554171385428464, + "grad_norm": 0.6135375499725342, + "learning_rate": 4.3163714621928466e-05, + "loss": 0.3792, + "step": 22660 + }, + { + "epoch": 8.557946394865986, + "grad_norm": 0.6618521809577942, + "learning_rate": 4.312276841309074e-05, + "loss": 0.3768, + "step": 22670 + }, + { + "epoch": 8.56172140430351, + "grad_norm": 0.6322848200798035, + "learning_rate": 4.3081826904775945e-05, + "loss": 0.3733, + "step": 22680 + }, + { + "epoch": 8.565496413741034, + "grad_norm": 0.6854828000068665, + "learning_rate": 4.3040890124967246e-05, + "loss": 0.3895, + "step": 22690 + }, + { + "epoch": 8.569271423178558, + "grad_norm": 0.6372854113578796, + "learning_rate": 4.2999958101644537e-05, + "loss": 0.3766, + "step": 22700 + }, + { + "epoch": 8.573046432616081, + "grad_norm": 0.6355828642845154, + "learning_rate": 4.2959030862784435e-05, + "loss": 0.3769, + "step": 22710 + }, + { + "epoch": 8.576821442053605, + "grad_norm": 0.6458565592765808, + "learning_rate": 4.291810843636036e-05, + "loss": 0.3925, + "step": 22720 + }, + { + "epoch": 8.580596451491129, + "grad_norm": 0.6839278936386108, + "learning_rate": 4.2877190850342375e-05, + "loss": 0.388, + "step": 22730 + }, + { + "epoch": 8.584371460928653, + "grad_norm": 0.6284090876579285, + "learning_rate": 4.2836278132697294e-05, + "loss": 0.3827, + "step": 22740 + }, + { + "epoch": 8.588146470366176, + "grad_norm": 0.6752241253852844, + "learning_rate": 4.279537031138855e-05, + "loss": 0.3969, + "step": 22750 + }, + { + "epoch": 8.5919214798037, + "grad_norm": 0.6706361770629883, + "learning_rate": 4.275446741437625e-05, + "loss": 0.3845, + "step": 22760 + }, + { + "epoch": 8.595696489241224, + "grad_norm": 0.6637896299362183, + "learning_rate": 4.2713569469617176e-05, + "loss": 0.3749, + "step": 22770 + }, + { + "epoch": 8.599471498678747, + "grad_norm": 0.7166405916213989, + "learning_rate": 4.267267650506465e-05, + "loss": 0.3803, + "step": 22780 + }, + { + "epoch": 8.603246508116271, + "grad_norm": 1.1399887800216675, + "learning_rate": 4.263178854866866e-05, + "loss": 0.3818, + "step": 22790 + }, + { + "epoch": 8.607021517553793, + "grad_norm": 0.5608764886856079, + "learning_rate": 4.259090562837571e-05, + "loss": 0.3855, + "step": 22800 + }, + { + "epoch": 8.610796526991317, + "grad_norm": 0.6282104253768921, + "learning_rate": 4.255002777212888e-05, + "loss": 0.3775, + "step": 22810 + }, + { + "epoch": 8.61457153642884, + "grad_norm": 0.7265238165855408, + "learning_rate": 4.250915500786783e-05, + "loss": 0.38, + "step": 22820 + }, + { + "epoch": 8.618346545866364, + "grad_norm": 0.6828112006187439, + "learning_rate": 4.24682873635287e-05, + "loss": 0.3821, + "step": 22830 + }, + { + "epoch": 8.622121555303888, + "grad_norm": 0.6247879862785339, + "learning_rate": 4.242742486704414e-05, + "loss": 0.388, + "step": 22840 + }, + { + "epoch": 8.625896564741412, + "grad_norm": 0.6014358997344971, + "learning_rate": 4.238656754634327e-05, + "loss": 0.3832, + "step": 22850 + }, + { + "epoch": 8.629671574178936, + "grad_norm": 0.6423870921134949, + "learning_rate": 4.234571542935168e-05, + "loss": 0.3951, + "step": 22860 + }, + { + "epoch": 8.63344658361646, + "grad_norm": 3.2498388290405273, + "learning_rate": 4.230486854399144e-05, + "loss": 0.3945, + "step": 22870 + }, + { + "epoch": 8.637221593053983, + "grad_norm": 0.9906812906265259, + "learning_rate": 4.226402691818098e-05, + "loss": 0.3884, + "step": 22880 + }, + { + "epoch": 8.640996602491507, + "grad_norm": 0.6815230250358582, + "learning_rate": 4.2223190579835196e-05, + "loss": 0.3842, + "step": 22890 + }, + { + "epoch": 8.64477161192903, + "grad_norm": 0.6938421726226807, + "learning_rate": 4.218235955686531e-05, + "loss": 0.3797, + "step": 22900 + }, + { + "epoch": 8.648546621366553, + "grad_norm": 1.4999902248382568, + "learning_rate": 4.214153387717894e-05, + "loss": 0.3921, + "step": 22910 + }, + { + "epoch": 8.652321630804076, + "grad_norm": 1.1413220167160034, + "learning_rate": 4.210071356868007e-05, + "loss": 0.3825, + "step": 22920 + }, + { + "epoch": 8.6560966402416, + "grad_norm": 0.633385419845581, + "learning_rate": 4.205989865926898e-05, + "loss": 0.3819, + "step": 22930 + }, + { + "epoch": 8.659871649679124, + "grad_norm": 0.680814802646637, + "learning_rate": 4.2019089176842294e-05, + "loss": 0.3818, + "step": 22940 + }, + { + "epoch": 8.663646659116647, + "grad_norm": 0.5858637094497681, + "learning_rate": 4.1978285149292894e-05, + "loss": 0.3804, + "step": 22950 + }, + { + "epoch": 8.667421668554171, + "grad_norm": 0.6605408191680908, + "learning_rate": 4.193748660450996e-05, + "loss": 0.389, + "step": 22960 + }, + { + "epoch": 8.671196677991695, + "grad_norm": 0.6238226890563965, + "learning_rate": 4.189669357037891e-05, + "loss": 0.378, + "step": 22970 + }, + { + "epoch": 8.674971687429219, + "grad_norm": 0.7179723978042603, + "learning_rate": 4.1855906074781405e-05, + "loss": 0.3778, + "step": 22980 + }, + { + "epoch": 8.678746696866742, + "grad_norm": 0.6245715618133545, + "learning_rate": 4.1815124145595285e-05, + "loss": 0.3823, + "step": 22990 + }, + { + "epoch": 8.682521706304266, + "grad_norm": 0.6498119831085205, + "learning_rate": 4.1774347810694644e-05, + "loss": 0.3801, + "step": 23000 + }, + { + "epoch": 8.68629671574179, + "grad_norm": 0.8200873136520386, + "learning_rate": 4.17335770979497e-05, + "loss": 0.3793, + "step": 23010 + }, + { + "epoch": 8.690071725179314, + "grad_norm": 0.6485256552696228, + "learning_rate": 4.169281203522687e-05, + "loss": 0.3811, + "step": 23020 + }, + { + "epoch": 8.693846734616837, + "grad_norm": 0.5570192933082581, + "learning_rate": 4.1652052650388674e-05, + "loss": 0.3783, + "step": 23030 + }, + { + "epoch": 8.69762174405436, + "grad_norm": 0.7071910500526428, + "learning_rate": 4.1611298971293786e-05, + "loss": 0.3839, + "step": 23040 + }, + { + "epoch": 8.701396753491883, + "grad_norm": 0.6622625589370728, + "learning_rate": 4.1570551025796935e-05, + "loss": 0.3754, + "step": 23050 + }, + { + "epoch": 8.705171762929407, + "grad_norm": 0.680098831653595, + "learning_rate": 4.152980884174897e-05, + "loss": 0.3802, + "step": 23060 + }, + { + "epoch": 8.70894677236693, + "grad_norm": 0.6106149554252625, + "learning_rate": 4.148907244699682e-05, + "loss": 0.3755, + "step": 23070 + }, + { + "epoch": 8.712721781804454, + "grad_norm": 0.5729265809059143, + "learning_rate": 4.1448341869383395e-05, + "loss": 0.38, + "step": 23080 + }, + { + "epoch": 8.716496791241978, + "grad_norm": 0.7992091774940491, + "learning_rate": 4.140761713674765e-05, + "loss": 0.3833, + "step": 23090 + }, + { + "epoch": 8.720271800679502, + "grad_norm": 0.6829302906990051, + "learning_rate": 4.1366898276924574e-05, + "loss": 0.3739, + "step": 23100 + }, + { + "epoch": 8.724046810117025, + "grad_norm": 0.6297142505645752, + "learning_rate": 4.132618531774512e-05, + "loss": 0.3754, + "step": 23110 + }, + { + "epoch": 8.72782181955455, + "grad_norm": 0.6362572908401489, + "learning_rate": 4.128547828703622e-05, + "loss": 0.3764, + "step": 23120 + }, + { + "epoch": 8.731596828992073, + "grad_norm": 0.7810057997703552, + "learning_rate": 4.1244777212620725e-05, + "loss": 0.3822, + "step": 23130 + }, + { + "epoch": 8.735371838429597, + "grad_norm": 0.6120405197143555, + "learning_rate": 4.120408212231746e-05, + "loss": 0.3755, + "step": 23140 + }, + { + "epoch": 8.73914684786712, + "grad_norm": 0.6071982979774475, + "learning_rate": 4.116339304394111e-05, + "loss": 0.3795, + "step": 23150 + }, + { + "epoch": 8.742921857304644, + "grad_norm": 0.7162799835205078, + "learning_rate": 4.112271000530229e-05, + "loss": 0.3815, + "step": 23160 + }, + { + "epoch": 8.746696866742166, + "grad_norm": 0.5881337523460388, + "learning_rate": 4.10820330342075e-05, + "loss": 0.3798, + "step": 23170 + }, + { + "epoch": 8.75047187617969, + "grad_norm": 0.6906888484954834, + "learning_rate": 4.1041362158459027e-05, + "loss": 0.3732, + "step": 23180 + }, + { + "epoch": 8.754246885617214, + "grad_norm": 0.6561579704284668, + "learning_rate": 4.1000697405855024e-05, + "loss": 0.3825, + "step": 23190 + }, + { + "epoch": 8.758021895054737, + "grad_norm": 0.5920860171318054, + "learning_rate": 4.096003880418951e-05, + "loss": 0.3819, + "step": 23200 + }, + { + "epoch": 8.761796904492261, + "grad_norm": 1.4881322383880615, + "learning_rate": 4.0919386381252215e-05, + "loss": 0.3763, + "step": 23210 + }, + { + "epoch": 8.765571913929785, + "grad_norm": 0.7468271255493164, + "learning_rate": 4.087874016482872e-05, + "loss": 0.3806, + "step": 23220 + }, + { + "epoch": 8.769346923367308, + "grad_norm": 0.5837022066116333, + "learning_rate": 4.0838100182700295e-05, + "loss": 0.382, + "step": 23230 + }, + { + "epoch": 8.773121932804832, + "grad_norm": 0.590907633304596, + "learning_rate": 4.079746646264402e-05, + "loss": 0.3886, + "step": 23240 + }, + { + "epoch": 8.776896942242356, + "grad_norm": 0.9434071779251099, + "learning_rate": 4.075683903243262e-05, + "loss": 0.3737, + "step": 23250 + }, + { + "epoch": 8.78067195167988, + "grad_norm": 0.7620961666107178, + "learning_rate": 4.071621791983462e-05, + "loss": 0.3776, + "step": 23260 + }, + { + "epoch": 8.784446961117403, + "grad_norm": 0.5980284214019775, + "learning_rate": 4.06756031526141e-05, + "loss": 0.3804, + "step": 23270 + }, + { + "epoch": 8.788221970554927, + "grad_norm": 0.6051507592201233, + "learning_rate": 4.063499475853092e-05, + "loss": 0.3804, + "step": 23280 + }, + { + "epoch": 8.791996979992451, + "grad_norm": 0.6261473894119263, + "learning_rate": 4.0594392765340506e-05, + "loss": 0.3658, + "step": 23290 + }, + { + "epoch": 8.795771989429973, + "grad_norm": 0.6253399848937988, + "learning_rate": 4.0553797200793954e-05, + "loss": 0.3819, + "step": 23300 + }, + { + "epoch": 8.799546998867497, + "grad_norm": 0.6369916200637817, + "learning_rate": 4.0513208092637926e-05, + "loss": 0.3783, + "step": 23310 + }, + { + "epoch": 8.80332200830502, + "grad_norm": 0.6382907032966614, + "learning_rate": 4.0472625468614735e-05, + "loss": 0.3774, + "step": 23320 + }, + { + "epoch": 8.807097017742544, + "grad_norm": 1.013975977897644, + "learning_rate": 4.043204935646218e-05, + "loss": 0.3976, + "step": 23330 + }, + { + "epoch": 8.810872027180068, + "grad_norm": 0.7413721084594727, + "learning_rate": 4.0391479783913675e-05, + "loss": 0.3813, + "step": 23340 + }, + { + "epoch": 8.814647036617592, + "grad_norm": 0.7551056146621704, + "learning_rate": 4.0350916778698155e-05, + "loss": 0.3754, + "step": 23350 + }, + { + "epoch": 8.818422046055115, + "grad_norm": 0.6768625378608704, + "learning_rate": 4.031036036854001e-05, + "loss": 0.3672, + "step": 23360 + }, + { + "epoch": 8.822197055492639, + "grad_norm": 0.7512798309326172, + "learning_rate": 4.026981058115918e-05, + "loss": 0.3806, + "step": 23370 + }, + { + "epoch": 8.825972064930163, + "grad_norm": 0.6166073679924011, + "learning_rate": 4.022926744427108e-05, + "loss": 0.3818, + "step": 23380 + }, + { + "epoch": 8.829747074367686, + "grad_norm": 0.6651231050491333, + "learning_rate": 4.018873098558654e-05, + "loss": 0.379, + "step": 23390 + }, + { + "epoch": 8.83352208380521, + "grad_norm": 0.6436861157417297, + "learning_rate": 4.014820123281186e-05, + "loss": 0.3816, + "step": 23400 + }, + { + "epoch": 8.837297093242734, + "grad_norm": 0.6689978837966919, + "learning_rate": 4.0107678213648735e-05, + "loss": 0.3756, + "step": 23410 + }, + { + "epoch": 8.841072102680256, + "grad_norm": 0.6367429494857788, + "learning_rate": 4.006716195579428e-05, + "loss": 0.3759, + "step": 23420 + }, + { + "epoch": 8.84484711211778, + "grad_norm": 0.649011492729187, + "learning_rate": 4.002665248694096e-05, + "loss": 0.3727, + "step": 23430 + }, + { + "epoch": 8.848622121555303, + "grad_norm": 0.7108133435249329, + "learning_rate": 3.998614983477664e-05, + "loss": 0.3741, + "step": 23440 + }, + { + "epoch": 8.852397130992827, + "grad_norm": 0.6738606691360474, + "learning_rate": 3.994565402698448e-05, + "loss": 0.3754, + "step": 23450 + }, + { + "epoch": 8.85617214043035, + "grad_norm": 0.6250699758529663, + "learning_rate": 3.9905165091242975e-05, + "loss": 0.3832, + "step": 23460 + }, + { + "epoch": 8.859947149867875, + "grad_norm": 0.6175537705421448, + "learning_rate": 3.9864683055225936e-05, + "loss": 0.3705, + "step": 23470 + }, + { + "epoch": 8.863722159305398, + "grad_norm": 0.6735917925834656, + "learning_rate": 3.982420794660247e-05, + "loss": 0.3799, + "step": 23480 + }, + { + "epoch": 8.867497168742922, + "grad_norm": 0.7157145738601685, + "learning_rate": 3.978373979303691e-05, + "loss": 0.3749, + "step": 23490 + }, + { + "epoch": 8.871272178180446, + "grad_norm": 0.6760987639427185, + "learning_rate": 3.974327862218888e-05, + "loss": 0.3851, + "step": 23500 + }, + { + "epoch": 8.87504718761797, + "grad_norm": 0.6828089356422424, + "learning_rate": 3.970282446171318e-05, + "loss": 0.3787, + "step": 23510 + }, + { + "epoch": 8.878822197055493, + "grad_norm": 0.6380763053894043, + "learning_rate": 3.966237733925988e-05, + "loss": 0.3748, + "step": 23520 + }, + { + "epoch": 8.882597206493017, + "grad_norm": 0.6100180745124817, + "learning_rate": 3.962193728247418e-05, + "loss": 0.3767, + "step": 23530 + }, + { + "epoch": 8.886372215930539, + "grad_norm": 0.6232542991638184, + "learning_rate": 3.958150431899651e-05, + "loss": 0.381, + "step": 23540 + }, + { + "epoch": 8.890147225368063, + "grad_norm": 0.636439323425293, + "learning_rate": 3.954107847646238e-05, + "loss": 0.3755, + "step": 23550 + }, + { + "epoch": 8.893922234805586, + "grad_norm": 0.611513078212738, + "learning_rate": 3.950065978250249e-05, + "loss": 0.3761, + "step": 23560 + }, + { + "epoch": 8.89769724424311, + "grad_norm": 0.6682894825935364, + "learning_rate": 3.9460248264742624e-05, + "loss": 0.3789, + "step": 23570 + }, + { + "epoch": 8.901472253680634, + "grad_norm": 0.6221024990081787, + "learning_rate": 3.941984395080371e-05, + "loss": 0.3734, + "step": 23580 + }, + { + "epoch": 8.905247263118158, + "grad_norm": 0.6244845390319824, + "learning_rate": 3.937944686830167e-05, + "loss": 0.3826, + "step": 23590 + }, + { + "epoch": 8.909022272555681, + "grad_norm": 0.6587332487106323, + "learning_rate": 3.933905704484756e-05, + "loss": 0.3759, + "step": 23600 + }, + { + "epoch": 8.912797281993205, + "grad_norm": 0.6589406728744507, + "learning_rate": 3.929867450804743e-05, + "loss": 0.3756, + "step": 23610 + }, + { + "epoch": 8.916572291430729, + "grad_norm": 0.6549832224845886, + "learning_rate": 3.925829928550237e-05, + "loss": 0.3883, + "step": 23620 + }, + { + "epoch": 8.920347300868253, + "grad_norm": 0.7647480964660645, + "learning_rate": 3.921793140480847e-05, + "loss": 0.3709, + "step": 23630 + }, + { + "epoch": 8.924122310305776, + "grad_norm": 0.6407207250595093, + "learning_rate": 3.917757089355677e-05, + "loss": 0.3736, + "step": 23640 + }, + { + "epoch": 8.9278973197433, + "grad_norm": 1.7529853582382202, + "learning_rate": 3.9137217779333326e-05, + "loss": 0.3909, + "step": 23650 + }, + { + "epoch": 8.931672329180824, + "grad_norm": 0.7098379135131836, + "learning_rate": 3.9096872089719083e-05, + "loss": 0.3768, + "step": 23660 + }, + { + "epoch": 8.935447338618346, + "grad_norm": 0.5911573767662048, + "learning_rate": 3.905653385228996e-05, + "loss": 0.3764, + "step": 23670 + }, + { + "epoch": 8.93922234805587, + "grad_norm": 0.652247965335846, + "learning_rate": 3.901620309461677e-05, + "loss": 0.3774, + "step": 23680 + }, + { + "epoch": 8.942997357493393, + "grad_norm": 0.6572662591934204, + "learning_rate": 3.897587984426518e-05, + "loss": 0.3836, + "step": 23690 + }, + { + "epoch": 8.946772366930917, + "grad_norm": 0.6210899949073792, + "learning_rate": 3.893556412879577e-05, + "loss": 0.3724, + "step": 23700 + }, + { + "epoch": 8.95054737636844, + "grad_norm": 0.6767255067825317, + "learning_rate": 3.889525597576395e-05, + "loss": 0.3752, + "step": 23710 + }, + { + "epoch": 8.954322385805964, + "grad_norm": 0.6062052249908447, + "learning_rate": 3.8854955412719965e-05, + "loss": 0.3703, + "step": 23720 + }, + { + "epoch": 8.958097395243488, + "grad_norm": 0.7351146936416626, + "learning_rate": 3.881466246720887e-05, + "loss": 0.3779, + "step": 23730 + }, + { + "epoch": 8.961872404681012, + "grad_norm": 0.6535313725471497, + "learning_rate": 3.8774377166770484e-05, + "loss": 0.3723, + "step": 23740 + }, + { + "epoch": 8.965647414118536, + "grad_norm": 0.6908565163612366, + "learning_rate": 3.8734099538939474e-05, + "loss": 0.374, + "step": 23750 + }, + { + "epoch": 8.96942242355606, + "grad_norm": 0.6240267157554626, + "learning_rate": 3.869382961124518e-05, + "loss": 0.3791, + "step": 23760 + }, + { + "epoch": 8.973197432993583, + "grad_norm": 0.6270217895507812, + "learning_rate": 3.8653567411211736e-05, + "loss": 0.3766, + "step": 23770 + }, + { + "epoch": 8.976972442431107, + "grad_norm": 0.6301257014274597, + "learning_rate": 3.8613312966357987e-05, + "loss": 0.3816, + "step": 23780 + }, + { + "epoch": 8.98074745186863, + "grad_norm": 0.6038647890090942, + "learning_rate": 3.857306630419745e-05, + "loss": 0.3798, + "step": 23790 + }, + { + "epoch": 8.984522461306153, + "grad_norm": 0.6066597104072571, + "learning_rate": 3.853282745223834e-05, + "loss": 0.3679, + "step": 23800 + }, + { + "epoch": 8.988297470743676, + "grad_norm": 0.7371004819869995, + "learning_rate": 3.8492596437983546e-05, + "loss": 0.3713, + "step": 23810 + }, + { + "epoch": 8.9920724801812, + "grad_norm": 0.7040576934814453, + "learning_rate": 3.8452373288930586e-05, + "loss": 0.3827, + "step": 23820 + }, + { + "epoch": 8.995847489618724, + "grad_norm": 0.6456139087677002, + "learning_rate": 3.841215803257159e-05, + "loss": 0.3755, + "step": 23830 + }, + { + "epoch": 8.999622499056247, + "grad_norm": 0.7062108516693115, + "learning_rate": 3.83719506963933e-05, + "loss": 0.374, + "step": 23840 + }, + { + "epoch": 9.003397508493771, + "grad_norm": 0.6744314432144165, + "learning_rate": 3.8331751307877087e-05, + "loss": 0.372, + "step": 23850 + }, + { + "epoch": 9.007172517931295, + "grad_norm": 0.6453410983085632, + "learning_rate": 3.82915598944988e-05, + "loss": 0.3676, + "step": 23860 + }, + { + "epoch": 9.010947527368819, + "grad_norm": 0.7047562599182129, + "learning_rate": 3.825137648372893e-05, + "loss": 0.3786, + "step": 23870 + }, + { + "epoch": 9.014722536806342, + "grad_norm": 0.6693297624588013, + "learning_rate": 3.8211201103032465e-05, + "loss": 0.3704, + "step": 23880 + }, + { + "epoch": 9.018497546243866, + "grad_norm": 0.6609267592430115, + "learning_rate": 3.817103377986887e-05, + "loss": 0.3851, + "step": 23890 + }, + { + "epoch": 9.02227255568139, + "grad_norm": 0.6702944040298462, + "learning_rate": 3.813087454169215e-05, + "loss": 0.3761, + "step": 23900 + }, + { + "epoch": 9.026047565118914, + "grad_norm": 0.61748206615448, + "learning_rate": 3.809072341595078e-05, + "loss": 0.3738, + "step": 23910 + }, + { + "epoch": 9.029822574556436, + "grad_norm": 0.6437124609947205, + "learning_rate": 3.8050580430087636e-05, + "loss": 0.3763, + "step": 23920 + }, + { + "epoch": 9.03359758399396, + "grad_norm": 0.7444436550140381, + "learning_rate": 3.8010445611540096e-05, + "loss": 0.3817, + "step": 23930 + }, + { + "epoch": 9.037372593431483, + "grad_norm": 0.6082374453544617, + "learning_rate": 3.797031898773992e-05, + "loss": 0.3766, + "step": 23940 + }, + { + "epoch": 9.041147602869007, + "grad_norm": 0.6564713716506958, + "learning_rate": 3.793020058611329e-05, + "loss": 0.3716, + "step": 23950 + }, + { + "epoch": 9.04492261230653, + "grad_norm": 0.5503236651420593, + "learning_rate": 3.789009043408074e-05, + "loss": 0.3771, + "step": 23960 + }, + { + "epoch": 9.048697621744054, + "grad_norm": 0.6116209030151367, + "learning_rate": 3.7849988559057194e-05, + "loss": 0.3796, + "step": 23970 + }, + { + "epoch": 9.052472631181578, + "grad_norm": 0.6737613081932068, + "learning_rate": 3.78098949884519e-05, + "loss": 0.3765, + "step": 23980 + }, + { + "epoch": 9.056247640619102, + "grad_norm": 0.6291811466217041, + "learning_rate": 3.776980974966843e-05, + "loss": 0.3685, + "step": 23990 + }, + { + "epoch": 9.060022650056625, + "grad_norm": 0.6767562031745911, + "learning_rate": 3.772973287010468e-05, + "loss": 0.3759, + "step": 24000 + }, + { + "epoch": 9.06379765949415, + "grad_norm": 0.6803905963897705, + "learning_rate": 3.768966437715283e-05, + "loss": 0.372, + "step": 24010 + }, + { + "epoch": 9.067572668931673, + "grad_norm": 0.6342119574546814, + "learning_rate": 3.7649604298199274e-05, + "loss": 0.3688, + "step": 24020 + }, + { + "epoch": 9.071347678369197, + "grad_norm": 0.6667430996894836, + "learning_rate": 3.760955266062473e-05, + "loss": 0.3792, + "step": 24030 + }, + { + "epoch": 9.07512268780672, + "grad_norm": 0.6684991717338562, + "learning_rate": 3.75695094918041e-05, + "loss": 0.3715, + "step": 24040 + }, + { + "epoch": 9.078897697244242, + "grad_norm": 0.6000959277153015, + "learning_rate": 3.752947481910652e-05, + "loss": 0.3751, + "step": 24050 + }, + { + "epoch": 9.082672706681766, + "grad_norm": 0.6405602693557739, + "learning_rate": 3.7489448669895324e-05, + "loss": 0.3863, + "step": 24060 + }, + { + "epoch": 9.08644771611929, + "grad_norm": 0.6103823781013489, + "learning_rate": 3.744943107152798e-05, + "loss": 0.3703, + "step": 24070 + }, + { + "epoch": 9.090222725556814, + "grad_norm": 0.6388285756111145, + "learning_rate": 3.7409422051356165e-05, + "loss": 0.3784, + "step": 24080 + }, + { + "epoch": 9.093997734994337, + "grad_norm": 0.7293974161148071, + "learning_rate": 3.736942163672564e-05, + "loss": 0.393, + "step": 24090 + }, + { + "epoch": 9.097772744431861, + "grad_norm": 0.7579730749130249, + "learning_rate": 3.732942985497636e-05, + "loss": 0.3702, + "step": 24100 + }, + { + "epoch": 9.101547753869385, + "grad_norm": 0.6169874668121338, + "learning_rate": 3.728944673344228e-05, + "loss": 0.3748, + "step": 24110 + }, + { + "epoch": 9.105322763306908, + "grad_norm": 0.6163791418075562, + "learning_rate": 3.72494722994515e-05, + "loss": 0.3786, + "step": 24120 + }, + { + "epoch": 9.109097772744432, + "grad_norm": 0.7307747602462769, + "learning_rate": 3.720950658032617e-05, + "loss": 0.3765, + "step": 24130 + }, + { + "epoch": 9.112872782181956, + "grad_norm": 0.6685255765914917, + "learning_rate": 3.716954960338249e-05, + "loss": 0.3714, + "step": 24140 + }, + { + "epoch": 9.11664779161948, + "grad_norm": 0.6877933144569397, + "learning_rate": 3.712960139593066e-05, + "loss": 0.3769, + "step": 24150 + }, + { + "epoch": 9.120422801057003, + "grad_norm": 0.6554901599884033, + "learning_rate": 3.708966198527493e-05, + "loss": 0.3784, + "step": 24160 + }, + { + "epoch": 9.124197810494525, + "grad_norm": 0.7071563005447388, + "learning_rate": 3.704973139871349e-05, + "loss": 0.3673, + "step": 24170 + }, + { + "epoch": 9.12797281993205, + "grad_norm": 0.5478484630584717, + "learning_rate": 3.700980966353853e-05, + "loss": 0.3713, + "step": 24180 + }, + { + "epoch": 9.131747829369573, + "grad_norm": 0.6506239175796509, + "learning_rate": 3.696989680703619e-05, + "loss": 0.3786, + "step": 24190 + }, + { + "epoch": 9.135522838807097, + "grad_norm": 0.6592217087745667, + "learning_rate": 3.69299928564865e-05, + "loss": 0.3782, + "step": 24200 + }, + { + "epoch": 9.13929784824462, + "grad_norm": 0.6553889513015747, + "learning_rate": 3.689009783916345e-05, + "loss": 0.3705, + "step": 24210 + }, + { + "epoch": 9.143072857682144, + "grad_norm": 0.6616779565811157, + "learning_rate": 3.6850211782334895e-05, + "loss": 0.3685, + "step": 24220 + }, + { + "epoch": 9.146847867119668, + "grad_norm": 0.6655992269515991, + "learning_rate": 3.681033471326261e-05, + "loss": 0.3741, + "step": 24230 + }, + { + "epoch": 9.150622876557192, + "grad_norm": 0.6539639830589294, + "learning_rate": 3.677046665920216e-05, + "loss": 0.3741, + "step": 24240 + }, + { + "epoch": 9.154397885994715, + "grad_norm": 0.6297611594200134, + "learning_rate": 3.6730607647403005e-05, + "loss": 0.3766, + "step": 24250 + }, + { + "epoch": 9.158172895432239, + "grad_norm": 0.6319004893302917, + "learning_rate": 3.6690757705108416e-05, + "loss": 0.3773, + "step": 24260 + }, + { + "epoch": 9.161947904869763, + "grad_norm": 1.1053649187088013, + "learning_rate": 3.665091685955542e-05, + "loss": 0.3735, + "step": 24270 + }, + { + "epoch": 9.165722914307286, + "grad_norm": 0.6567557454109192, + "learning_rate": 3.6611085137974896e-05, + "loss": 0.3817, + "step": 24280 + }, + { + "epoch": 9.16949792374481, + "grad_norm": 0.6980771422386169, + "learning_rate": 3.657126256759143e-05, + "loss": 0.3696, + "step": 24290 + }, + { + "epoch": 9.173272933182332, + "grad_norm": 0.6510254144668579, + "learning_rate": 3.653144917562335e-05, + "loss": 0.3991, + "step": 24300 + }, + { + "epoch": 9.177047942619856, + "grad_norm": 0.5957704782485962, + "learning_rate": 3.649164498928277e-05, + "loss": 0.3752, + "step": 24310 + }, + { + "epoch": 9.18082295205738, + "grad_norm": 0.7079484462738037, + "learning_rate": 3.645185003577546e-05, + "loss": 0.3757, + "step": 24320 + }, + { + "epoch": 9.184597961494903, + "grad_norm": 0.6439237594604492, + "learning_rate": 3.6412064342300906e-05, + "loss": 0.3793, + "step": 24330 + }, + { + "epoch": 9.188372970932427, + "grad_norm": 0.6560051441192627, + "learning_rate": 3.637228793605224e-05, + "loss": 0.3781, + "step": 24340 + }, + { + "epoch": 9.19214798036995, + "grad_norm": 0.6205071210861206, + "learning_rate": 3.6332520844216264e-05, + "loss": 0.373, + "step": 24350 + }, + { + "epoch": 9.195922989807475, + "grad_norm": 0.8601105213165283, + "learning_rate": 3.6292763093973425e-05, + "loss": 0.3697, + "step": 24360 + }, + { + "epoch": 9.199697999244998, + "grad_norm": 0.6811267733573914, + "learning_rate": 3.6253014712497754e-05, + "loss": 0.3826, + "step": 24370 + }, + { + "epoch": 9.203473008682522, + "grad_norm": 0.6935433745384216, + "learning_rate": 3.621327572695692e-05, + "loss": 0.3782, + "step": 24380 + }, + { + "epoch": 9.207248018120046, + "grad_norm": 0.5961577892303467, + "learning_rate": 3.617354616451211e-05, + "loss": 0.375, + "step": 24390 + }, + { + "epoch": 9.21102302755757, + "grad_norm": 0.5627548694610596, + "learning_rate": 3.6133826052318116e-05, + "loss": 0.3731, + "step": 24400 + }, + { + "epoch": 9.214798036995093, + "grad_norm": 0.6557216048240662, + "learning_rate": 3.609411541752327e-05, + "loss": 0.3741, + "step": 24410 + }, + { + "epoch": 9.218573046432615, + "grad_norm": 0.7413145899772644, + "learning_rate": 3.6054414287269405e-05, + "loss": 0.3712, + "step": 24420 + }, + { + "epoch": 9.222348055870139, + "grad_norm": 0.6858183145523071, + "learning_rate": 3.601472268869188e-05, + "loss": 0.3814, + "step": 24430 + }, + { + "epoch": 9.226123065307663, + "grad_norm": 0.6973282098770142, + "learning_rate": 3.597504064891952e-05, + "loss": 0.377, + "step": 24440 + }, + { + "epoch": 9.229898074745186, + "grad_norm": 0.6426398158073425, + "learning_rate": 3.5935368195074636e-05, + "loss": 0.3757, + "step": 24450 + }, + { + "epoch": 9.23367308418271, + "grad_norm": 0.6414214372634888, + "learning_rate": 3.589570535427297e-05, + "loss": 0.3642, + "step": 24460 + }, + { + "epoch": 9.237448093620234, + "grad_norm": 0.6278491020202637, + "learning_rate": 3.585605215362371e-05, + "loss": 0.3778, + "step": 24470 + }, + { + "epoch": 9.241223103057758, + "grad_norm": 0.695353090763092, + "learning_rate": 3.581640862022941e-05, + "loss": 0.3763, + "step": 24480 + }, + { + "epoch": 9.244998112495281, + "grad_norm": 0.6020937561988831, + "learning_rate": 3.57767747811861e-05, + "loss": 0.3761, + "step": 24490 + }, + { + "epoch": 9.248773121932805, + "grad_norm": 0.7220026254653931, + "learning_rate": 3.573715066358308e-05, + "loss": 0.3751, + "step": 24500 + }, + { + "epoch": 9.252548131370329, + "grad_norm": 0.6463865041732788, + "learning_rate": 3.569753629450311e-05, + "loss": 0.3783, + "step": 24510 + }, + { + "epoch": 9.256323140807853, + "grad_norm": 0.6139450073242188, + "learning_rate": 3.565793170102221e-05, + "loss": 0.3754, + "step": 24520 + }, + { + "epoch": 9.260098150245376, + "grad_norm": 0.6605015993118286, + "learning_rate": 3.561833691020976e-05, + "loss": 0.3685, + "step": 24530 + }, + { + "epoch": 9.2638731596829, + "grad_norm": 0.6926522254943848, + "learning_rate": 3.5578751949128415e-05, + "loss": 0.374, + "step": 24540 + }, + { + "epoch": 9.267648169120422, + "grad_norm": 0.6264640092849731, + "learning_rate": 3.5539176844834125e-05, + "loss": 0.3783, + "step": 24550 + }, + { + "epoch": 9.271423178557946, + "grad_norm": 0.9688851237297058, + "learning_rate": 3.5499611624376125e-05, + "loss": 0.3795, + "step": 24560 + }, + { + "epoch": 9.27519818799547, + "grad_norm": 0.5947738885879517, + "learning_rate": 3.546005631479684e-05, + "loss": 0.3662, + "step": 24570 + }, + { + "epoch": 9.278973197432993, + "grad_norm": 0.6355003714561462, + "learning_rate": 3.542051094313196e-05, + "loss": 0.3785, + "step": 24580 + }, + { + "epoch": 9.282748206870517, + "grad_norm": 0.6188831329345703, + "learning_rate": 3.5380975536410364e-05, + "loss": 0.3804, + "step": 24590 + }, + { + "epoch": 9.28652321630804, + "grad_norm": 0.6349472403526306, + "learning_rate": 3.534145012165415e-05, + "loss": 0.3757, + "step": 24600 + }, + { + "epoch": 9.290298225745564, + "grad_norm": 0.6269373297691345, + "learning_rate": 3.5301934725878546e-05, + "loss": 0.3698, + "step": 24610 + }, + { + "epoch": 9.294073235183088, + "grad_norm": 0.6995264291763306, + "learning_rate": 3.526242937609197e-05, + "loss": 0.3798, + "step": 24620 + }, + { + "epoch": 9.297848244620612, + "grad_norm": 0.6309302449226379, + "learning_rate": 3.522293409929595e-05, + "loss": 0.3652, + "step": 24630 + }, + { + "epoch": 9.301623254058136, + "grad_norm": 0.8019225597381592, + "learning_rate": 3.518344892248513e-05, + "loss": 0.3808, + "step": 24640 + }, + { + "epoch": 9.30539826349566, + "grad_norm": 0.6818584203720093, + "learning_rate": 3.514397387264725e-05, + "loss": 0.3726, + "step": 24650 + }, + { + "epoch": 9.309173272933183, + "grad_norm": 0.6697105169296265, + "learning_rate": 3.5104508976763176e-05, + "loss": 0.3678, + "step": 24660 + }, + { + "epoch": 9.312948282370705, + "grad_norm": 0.5741595029830933, + "learning_rate": 3.506505426180674e-05, + "loss": 0.38, + "step": 24670 + }, + { + "epoch": 9.316723291808229, + "grad_norm": 0.6436684131622314, + "learning_rate": 3.502560975474488e-05, + "loss": 0.3747, + "step": 24680 + }, + { + "epoch": 9.320498301245753, + "grad_norm": 0.6101210117340088, + "learning_rate": 3.4986175482537566e-05, + "loss": 0.3678, + "step": 24690 + }, + { + "epoch": 9.324273310683276, + "grad_norm": 0.8269137144088745, + "learning_rate": 3.4946751472137725e-05, + "loss": 0.3721, + "step": 24700 + }, + { + "epoch": 9.3280483201208, + "grad_norm": 0.6154971718788147, + "learning_rate": 3.490733775049132e-05, + "loss": 0.379, + "step": 24710 + }, + { + "epoch": 9.331823329558324, + "grad_norm": 0.642996072769165, + "learning_rate": 3.4867934344537236e-05, + "loss": 0.3823, + "step": 24720 + }, + { + "epoch": 9.335598338995847, + "grad_norm": 0.6304564476013184, + "learning_rate": 3.482854128120735e-05, + "loss": 0.3742, + "step": 24730 + }, + { + "epoch": 9.339373348433371, + "grad_norm": 0.6240081787109375, + "learning_rate": 3.478915858742643e-05, + "loss": 0.3698, + "step": 24740 + }, + { + "epoch": 9.343148357870895, + "grad_norm": 0.6386808156967163, + "learning_rate": 3.4749786290112205e-05, + "loss": 0.3709, + "step": 24750 + }, + { + "epoch": 9.346923367308419, + "grad_norm": 0.6619960069656372, + "learning_rate": 3.471042441617524e-05, + "loss": 0.3689, + "step": 24760 + }, + { + "epoch": 9.350698376745942, + "grad_norm": 0.8115395903587341, + "learning_rate": 3.467107299251902e-05, + "loss": 0.3721, + "step": 24770 + }, + { + "epoch": 9.354473386183466, + "grad_norm": 0.6916590929031372, + "learning_rate": 3.463173204603984e-05, + "loss": 0.3736, + "step": 24780 + }, + { + "epoch": 9.35824839562099, + "grad_norm": 0.6382922530174255, + "learning_rate": 3.4592401603626924e-05, + "loss": 0.3715, + "step": 24790 + }, + { + "epoch": 9.362023405058512, + "grad_norm": 0.6073492765426636, + "learning_rate": 3.45530816921622e-05, + "loss": 0.3714, + "step": 24800 + }, + { + "epoch": 9.365798414496036, + "grad_norm": 0.6381345987319946, + "learning_rate": 3.451377233852051e-05, + "loss": 0.3758, + "step": 24810 + }, + { + "epoch": 9.36957342393356, + "grad_norm": 0.5805862545967102, + "learning_rate": 3.4474473569569385e-05, + "loss": 0.37, + "step": 24820 + }, + { + "epoch": 9.373348433371083, + "grad_norm": 0.7061961889266968, + "learning_rate": 3.443518541216918e-05, + "loss": 0.3721, + "step": 24830 + }, + { + "epoch": 9.377123442808607, + "grad_norm": 0.6575196981430054, + "learning_rate": 3.439590789317299e-05, + "loss": 0.3731, + "step": 24840 + }, + { + "epoch": 9.38089845224613, + "grad_norm": 0.6737225651741028, + "learning_rate": 3.4356641039426607e-05, + "loss": 0.3712, + "step": 24850 + }, + { + "epoch": 9.384673461683654, + "grad_norm": 0.7008658051490784, + "learning_rate": 3.431738487776857e-05, + "loss": 0.3713, + "step": 24860 + }, + { + "epoch": 9.388448471121178, + "grad_norm": 0.6766116619110107, + "learning_rate": 3.4278139435030084e-05, + "loss": 0.3772, + "step": 24870 + }, + { + "epoch": 9.392223480558702, + "grad_norm": 0.6456305384635925, + "learning_rate": 3.423890473803504e-05, + "loss": 0.3804, + "step": 24880 + }, + { + "epoch": 9.395998489996225, + "grad_norm": 0.7182512879371643, + "learning_rate": 3.41996808136e-05, + "loss": 0.37, + "step": 24890 + }, + { + "epoch": 9.39977349943375, + "grad_norm": 0.6886579990386963, + "learning_rate": 3.416046768853413e-05, + "loss": 0.3696, + "step": 24900 + }, + { + "epoch": 9.403548508871273, + "grad_norm": 0.6319290399551392, + "learning_rate": 3.412126538963925e-05, + "loss": 0.3771, + "step": 24910 + }, + { + "epoch": 9.407323518308797, + "grad_norm": 0.5859635472297668, + "learning_rate": 3.4082073943709727e-05, + "loss": 0.3729, + "step": 24920 + }, + { + "epoch": 9.411098527746319, + "grad_norm": 0.7094352841377258, + "learning_rate": 3.404289337753258e-05, + "loss": 0.374, + "step": 24930 + }, + { + "epoch": 9.414873537183842, + "grad_norm": 0.6192387938499451, + "learning_rate": 3.400372371788736e-05, + "loss": 0.3749, + "step": 24940 + }, + { + "epoch": 9.418648546621366, + "grad_norm": 0.5935202836990356, + "learning_rate": 3.3964564991546124e-05, + "loss": 0.3707, + "step": 24950 + }, + { + "epoch": 9.42242355605889, + "grad_norm": 0.6603958606719971, + "learning_rate": 3.392541722527351e-05, + "loss": 0.3825, + "step": 24960 + }, + { + "epoch": 9.426198565496414, + "grad_norm": 0.6225345134735107, + "learning_rate": 3.3886280445826644e-05, + "loss": 0.3669, + "step": 24970 + }, + { + "epoch": 9.429973574933937, + "grad_norm": 0.6600792407989502, + "learning_rate": 3.3847154679955154e-05, + "loss": 0.373, + "step": 24980 + }, + { + "epoch": 9.433748584371461, + "grad_norm": 0.6622351408004761, + "learning_rate": 3.380803995440113e-05, + "loss": 0.3696, + "step": 24990 + }, + { + "epoch": 9.437523593808985, + "grad_norm": 0.6164969801902771, + "learning_rate": 3.3768936295899115e-05, + "loss": 0.3682, + "step": 25000 + }, + { + "epoch": 9.441298603246508, + "grad_norm": 0.704637348651886, + "learning_rate": 3.3729843731176094e-05, + "loss": 0.3734, + "step": 25010 + }, + { + "epoch": 9.445073612684032, + "grad_norm": 0.6271415948867798, + "learning_rate": 3.369076228695146e-05, + "loss": 0.3691, + "step": 25020 + }, + { + "epoch": 9.448848622121556, + "grad_norm": 0.6947619318962097, + "learning_rate": 3.365169198993703e-05, + "loss": 0.3781, + "step": 25030 + }, + { + "epoch": 9.45262363155908, + "grad_norm": 0.6423031091690063, + "learning_rate": 3.361263286683697e-05, + "loss": 0.3713, + "step": 25040 + }, + { + "epoch": 9.456398640996602, + "grad_norm": 0.6936047673225403, + "learning_rate": 3.35735849443478e-05, + "loss": 0.3714, + "step": 25050 + }, + { + "epoch": 9.460173650434125, + "grad_norm": 0.6276016235351562, + "learning_rate": 3.3534548249158435e-05, + "loss": 0.3695, + "step": 25060 + }, + { + "epoch": 9.46394865987165, + "grad_norm": 0.6831814646720886, + "learning_rate": 3.3495522807950086e-05, + "loss": 0.3706, + "step": 25070 + }, + { + "epoch": 9.467723669309173, + "grad_norm": 0.6282784342765808, + "learning_rate": 3.345650864739627e-05, + "loss": 0.3717, + "step": 25080 + }, + { + "epoch": 9.471498678746697, + "grad_norm": 0.6343705654144287, + "learning_rate": 3.3417505794162794e-05, + "loss": 0.3685, + "step": 25090 + }, + { + "epoch": 9.47527368818422, + "grad_norm": 0.6878871917724609, + "learning_rate": 3.3378514274907745e-05, + "loss": 0.3773, + "step": 25100 + }, + { + "epoch": 9.479048697621744, + "grad_norm": 0.7008819580078125, + "learning_rate": 3.333953411628147e-05, + "loss": 0.3681, + "step": 25110 + }, + { + "epoch": 9.482823707059268, + "grad_norm": 0.653729259967804, + "learning_rate": 3.330056534492653e-05, + "loss": 0.3794, + "step": 25120 + }, + { + "epoch": 9.486598716496792, + "grad_norm": 0.7341142296791077, + "learning_rate": 3.32616079874777e-05, + "loss": 0.3691, + "step": 25130 + }, + { + "epoch": 9.490373725934315, + "grad_norm": 0.9911676049232483, + "learning_rate": 3.322266207056197e-05, + "loss": 0.37, + "step": 25140 + }, + { + "epoch": 9.494148735371839, + "grad_norm": 0.6838816404342651, + "learning_rate": 3.318372762079852e-05, + "loss": 0.3757, + "step": 25150 + }, + { + "epoch": 9.497923744809363, + "grad_norm": 0.7315078973770142, + "learning_rate": 3.3144804664798666e-05, + "loss": 0.3732, + "step": 25160 + }, + { + "epoch": 9.501698754246885, + "grad_norm": 0.6944950222969055, + "learning_rate": 3.3105893229165894e-05, + "loss": 0.3777, + "step": 25170 + }, + { + "epoch": 9.505473763684408, + "grad_norm": 0.6807419657707214, + "learning_rate": 3.30669933404958e-05, + "loss": 0.3688, + "step": 25180 + }, + { + "epoch": 9.509248773121932, + "grad_norm": 0.7115993499755859, + "learning_rate": 3.302810502537609e-05, + "loss": 0.3691, + "step": 25190 + }, + { + "epoch": 9.513023782559456, + "grad_norm": 0.8097192645072937, + "learning_rate": 3.298922831038655e-05, + "loss": 0.3776, + "step": 25200 + }, + { + "epoch": 9.51679879199698, + "grad_norm": 0.7089084386825562, + "learning_rate": 3.2950363222099073e-05, + "loss": 0.3708, + "step": 25210 + }, + { + "epoch": 9.520573801434503, + "grad_norm": 0.6845264434814453, + "learning_rate": 3.291150978707758e-05, + "loss": 0.3715, + "step": 25220 + }, + { + "epoch": 9.524348810872027, + "grad_norm": 0.6236247420310974, + "learning_rate": 3.287266803187798e-05, + "loss": 0.3671, + "step": 25230 + }, + { + "epoch": 9.52812382030955, + "grad_norm": 0.7404671907424927, + "learning_rate": 3.283383798304829e-05, + "loss": 0.3735, + "step": 25240 + }, + { + "epoch": 9.531898829747075, + "grad_norm": 0.6793319582939148, + "learning_rate": 3.279501966712847e-05, + "loss": 0.3674, + "step": 25250 + }, + { + "epoch": 9.535673839184598, + "grad_norm": 0.7164490222930908, + "learning_rate": 3.275621311065047e-05, + "loss": 0.3726, + "step": 25260 + }, + { + "epoch": 9.539448848622122, + "grad_norm": 0.8006794452667236, + "learning_rate": 3.271741834013822e-05, + "loss": 0.3671, + "step": 25270 + }, + { + "epoch": 9.543223858059646, + "grad_norm": 0.6206616759300232, + "learning_rate": 3.267863538210756e-05, + "loss": 0.3736, + "step": 25280 + }, + { + "epoch": 9.54699886749717, + "grad_norm": 0.6397393941879272, + "learning_rate": 3.2639864263066296e-05, + "loss": 0.3697, + "step": 25290 + }, + { + "epoch": 9.550773876934691, + "grad_norm": 1.0796784162521362, + "learning_rate": 3.26011050095141e-05, + "loss": 0.3907, + "step": 25300 + }, + { + "epoch": 9.554548886372215, + "grad_norm": 2.355830192565918, + "learning_rate": 3.256235764794259e-05, + "loss": 0.3707, + "step": 25310 + }, + { + "epoch": 9.558323895809739, + "grad_norm": 0.6870741844177246, + "learning_rate": 3.2523622204835194e-05, + "loss": 0.3682, + "step": 25320 + }, + { + "epoch": 9.562098905247263, + "grad_norm": 0.7428625822067261, + "learning_rate": 3.2484898706667214e-05, + "loss": 0.369, + "step": 25330 + }, + { + "epoch": 9.565873914684786, + "grad_norm": 0.7070319652557373, + "learning_rate": 3.2446187179905806e-05, + "loss": 0.3685, + "step": 25340 + }, + { + "epoch": 9.56964892412231, + "grad_norm": 0.6212445497512817, + "learning_rate": 3.240748765100995e-05, + "loss": 0.3802, + "step": 25350 + }, + { + "epoch": 9.573423933559834, + "grad_norm": 0.6794653534889221, + "learning_rate": 3.236880014643039e-05, + "loss": 0.3665, + "step": 25360 + }, + { + "epoch": 9.577198942997358, + "grad_norm": 0.6495829224586487, + "learning_rate": 3.233012469260969e-05, + "loss": 0.3645, + "step": 25370 + }, + { + "epoch": 9.580973952434881, + "grad_norm": 0.6042360067367554, + "learning_rate": 3.229146131598213e-05, + "loss": 0.3745, + "step": 25380 + }, + { + "epoch": 9.584748961872405, + "grad_norm": 2.035146713256836, + "learning_rate": 3.2252810042973794e-05, + "loss": 0.3827, + "step": 25390 + }, + { + "epoch": 9.588523971309929, + "grad_norm": 0.5512405633926392, + "learning_rate": 3.2214170900002456e-05, + "loss": 0.3873, + "step": 25400 + }, + { + "epoch": 9.592298980747453, + "grad_norm": 0.7587074041366577, + "learning_rate": 3.217554391347758e-05, + "loss": 0.3671, + "step": 25410 + }, + { + "epoch": 9.596073990184976, + "grad_norm": 0.7429296970367432, + "learning_rate": 3.213692910980037e-05, + "loss": 0.3708, + "step": 25420 + }, + { + "epoch": 9.599848999622498, + "grad_norm": 0.6271414160728455, + "learning_rate": 3.2098326515363666e-05, + "loss": 0.3715, + "step": 25430 + }, + { + "epoch": 9.603624009060022, + "grad_norm": 0.6165705919265747, + "learning_rate": 3.205973615655199e-05, + "loss": 0.3669, + "step": 25440 + }, + { + "epoch": 9.607399018497546, + "grad_norm": 0.5890395641326904, + "learning_rate": 3.202115805974149e-05, + "loss": 0.3681, + "step": 25450 + }, + { + "epoch": 9.61117402793507, + "grad_norm": 0.6870768070220947, + "learning_rate": 3.1982592251299916e-05, + "loss": 0.3736, + "step": 25460 + }, + { + "epoch": 9.614949037372593, + "grad_norm": 0.7155221700668335, + "learning_rate": 3.1944038757586656e-05, + "loss": 0.3716, + "step": 25470 + }, + { + "epoch": 9.618724046810117, + "grad_norm": 0.6171225905418396, + "learning_rate": 3.190549760495263e-05, + "loss": 0.3712, + "step": 25480 + }, + { + "epoch": 9.62249905624764, + "grad_norm": 0.7019498944282532, + "learning_rate": 3.186696881974039e-05, + "loss": 0.3641, + "step": 25490 + }, + { + "epoch": 9.626274065685164, + "grad_norm": 0.7143846154212952, + "learning_rate": 3.1828452428283986e-05, + "loss": 0.3769, + "step": 25500 + }, + { + "epoch": 9.630049075122688, + "grad_norm": 0.6226781606674194, + "learning_rate": 3.178994845690898e-05, + "loss": 0.3636, + "step": 25510 + }, + { + "epoch": 9.633824084560212, + "grad_norm": 0.6864441633224487, + "learning_rate": 3.17514569319325e-05, + "loss": 0.3741, + "step": 25520 + }, + { + "epoch": 9.637599093997736, + "grad_norm": 0.6476948261260986, + "learning_rate": 3.171297787966312e-05, + "loss": 0.3763, + "step": 25530 + }, + { + "epoch": 9.64137410343526, + "grad_norm": 0.6583799123764038, + "learning_rate": 3.167451132640093e-05, + "loss": 0.3715, + "step": 25540 + }, + { + "epoch": 9.645149112872783, + "grad_norm": 0.8483378291130066, + "learning_rate": 3.163605729843746e-05, + "loss": 0.3742, + "step": 25550 + }, + { + "epoch": 9.648924122310305, + "grad_norm": 0.7014634609222412, + "learning_rate": 3.159761582205565e-05, + "loss": 0.3693, + "step": 25560 + }, + { + "epoch": 9.652699131747829, + "grad_norm": 0.8034195899963379, + "learning_rate": 3.155918692352992e-05, + "loss": 0.3697, + "step": 25570 + }, + { + "epoch": 9.656474141185353, + "grad_norm": 0.7062366008758545, + "learning_rate": 3.152077062912602e-05, + "loss": 0.3776, + "step": 25580 + }, + { + "epoch": 9.660249150622876, + "grad_norm": 0.6510414481163025, + "learning_rate": 3.148236696510117e-05, + "loss": 0.3764, + "step": 25590 + }, + { + "epoch": 9.6640241600604, + "grad_norm": 0.7279618978500366, + "learning_rate": 3.144397595770388e-05, + "loss": 0.3773, + "step": 25600 + }, + { + "epoch": 9.667799169497924, + "grad_norm": 0.6099146008491516, + "learning_rate": 3.1405597633174036e-05, + "loss": 0.376, + "step": 25610 + }, + { + "epoch": 9.671574178935447, + "grad_norm": 0.6417626142501831, + "learning_rate": 3.136723201774289e-05, + "loss": 0.3717, + "step": 25620 + }, + { + "epoch": 9.675349188372971, + "grad_norm": 0.7016961574554443, + "learning_rate": 3.132887913763295e-05, + "loss": 0.3698, + "step": 25630 + }, + { + "epoch": 9.679124197810495, + "grad_norm": 0.7685357928276062, + "learning_rate": 3.129053901905806e-05, + "loss": 0.3648, + "step": 25640 + }, + { + "epoch": 9.682899207248019, + "grad_norm": 0.7496814131736755, + "learning_rate": 3.125221168822335e-05, + "loss": 0.3781, + "step": 25650 + }, + { + "epoch": 9.686674216685542, + "grad_norm": 0.6221609711647034, + "learning_rate": 3.1213897171325154e-05, + "loss": 0.3694, + "step": 25660 + }, + { + "epoch": 9.690449226123066, + "grad_norm": 0.6706365942955017, + "learning_rate": 3.1175595494551116e-05, + "loss": 0.3709, + "step": 25670 + }, + { + "epoch": 9.694224235560588, + "grad_norm": 0.5938177704811096, + "learning_rate": 3.1137306684080045e-05, + "loss": 0.3659, + "step": 25680 + }, + { + "epoch": 9.697999244998112, + "grad_norm": 0.69608074426651, + "learning_rate": 3.1099030766081985e-05, + "loss": 0.3712, + "step": 25690 + }, + { + "epoch": 9.701774254435636, + "grad_norm": 0.6536232829093933, + "learning_rate": 3.106076776671818e-05, + "loss": 0.3678, + "step": 25700 + }, + { + "epoch": 9.70554926387316, + "grad_norm": 0.6710280179977417, + "learning_rate": 3.102251771214101e-05, + "loss": 0.3667, + "step": 25710 + }, + { + "epoch": 9.709324273310683, + "grad_norm": 0.6235374808311462, + "learning_rate": 3.098428062849404e-05, + "loss": 0.3684, + "step": 25720 + }, + { + "epoch": 9.713099282748207, + "grad_norm": 0.6490816473960876, + "learning_rate": 3.094605654191195e-05, + "loss": 0.3715, + "step": 25730 + }, + { + "epoch": 9.71687429218573, + "grad_norm": 0.6581624746322632, + "learning_rate": 3.090784547852055e-05, + "loss": 0.3714, + "step": 25740 + }, + { + "epoch": 9.720649301623254, + "grad_norm": 0.7839778065681458, + "learning_rate": 3.0869647464436746e-05, + "loss": 0.373, + "step": 25750 + }, + { + "epoch": 9.724424311060778, + "grad_norm": 0.6403361558914185, + "learning_rate": 3.0831462525768496e-05, + "loss": 0.3684, + "step": 25760 + }, + { + "epoch": 9.728199320498302, + "grad_norm": 0.6090444326400757, + "learning_rate": 3.079329068861488e-05, + "loss": 0.3671, + "step": 25770 + }, + { + "epoch": 9.731974329935825, + "grad_norm": 0.6840675473213196, + "learning_rate": 3.075513197906597e-05, + "loss": 0.3682, + "step": 25780 + }, + { + "epoch": 9.73574933937335, + "grad_norm": 0.6021028757095337, + "learning_rate": 3.071698642320286e-05, + "loss": 0.3723, + "step": 25790 + }, + { + "epoch": 9.739524348810871, + "grad_norm": 0.6243388652801514, + "learning_rate": 3.067885404709772e-05, + "loss": 0.371, + "step": 25800 + }, + { + "epoch": 9.743299358248395, + "grad_norm": 0.6200324296951294, + "learning_rate": 3.0640734876813636e-05, + "loss": 0.3729, + "step": 25810 + }, + { + "epoch": 9.747074367685919, + "grad_norm": 1.0078961849212646, + "learning_rate": 3.060262893840473e-05, + "loss": 0.3695, + "step": 25820 + }, + { + "epoch": 9.750849377123442, + "grad_norm": 0.6697998642921448, + "learning_rate": 3.056453625791603e-05, + "loss": 0.3641, + "step": 25830 + }, + { + "epoch": 9.754624386560966, + "grad_norm": 0.8702282905578613, + "learning_rate": 3.052645686138353e-05, + "loss": 0.3734, + "step": 25840 + }, + { + "epoch": 9.75839939599849, + "grad_norm": 0.6085755825042725, + "learning_rate": 3.0488390774834153e-05, + "loss": 0.3702, + "step": 25850 + }, + { + "epoch": 9.762174405436014, + "grad_norm": 0.7779916524887085, + "learning_rate": 3.0450338024285684e-05, + "loss": 0.3718, + "step": 25860 + }, + { + "epoch": 9.765949414873537, + "grad_norm": 0.6143949627876282, + "learning_rate": 3.0412298635746855e-05, + "loss": 0.3663, + "step": 25870 + }, + { + "epoch": 9.769724424311061, + "grad_norm": 0.7032211422920227, + "learning_rate": 3.03742726352172e-05, + "loss": 0.3688, + "step": 25880 + }, + { + "epoch": 9.773499433748585, + "grad_norm": 0.6979837417602539, + "learning_rate": 3.0336260048687125e-05, + "loss": 0.3648, + "step": 25890 + }, + { + "epoch": 9.777274443186108, + "grad_norm": 0.6320093274116516, + "learning_rate": 3.0298260902137897e-05, + "loss": 0.3666, + "step": 25900 + }, + { + "epoch": 9.781049452623632, + "grad_norm": 0.6914851665496826, + "learning_rate": 3.0260275221541566e-05, + "loss": 0.3743, + "step": 25910 + }, + { + "epoch": 9.784824462061156, + "grad_norm": 0.6460880041122437, + "learning_rate": 3.0222303032860987e-05, + "loss": 0.376, + "step": 25920 + }, + { + "epoch": 9.788599471498678, + "grad_norm": 0.7050313353538513, + "learning_rate": 3.018434436204979e-05, + "loss": 0.3655, + "step": 25930 + }, + { + "epoch": 9.792374480936202, + "grad_norm": 0.7031101584434509, + "learning_rate": 3.014639923505237e-05, + "loss": 0.367, + "step": 25940 + }, + { + "epoch": 9.796149490373725, + "grad_norm": 0.6252807378768921, + "learning_rate": 3.0108467677803863e-05, + "loss": 0.3669, + "step": 25950 + }, + { + "epoch": 9.79992449981125, + "grad_norm": 0.6728451251983643, + "learning_rate": 3.0070549716230156e-05, + "loss": 0.3685, + "step": 25960 + }, + { + "epoch": 9.803699509248773, + "grad_norm": 0.6584720015525818, + "learning_rate": 3.003264537624777e-05, + "loss": 0.3704, + "step": 25970 + }, + { + "epoch": 9.807474518686297, + "grad_norm": 0.5711626410484314, + "learning_rate": 2.9994754683764e-05, + "loss": 0.3833, + "step": 25980 + }, + { + "epoch": 9.81124952812382, + "grad_norm": 0.7549001574516296, + "learning_rate": 2.9956877664676754e-05, + "loss": 0.3661, + "step": 25990 + }, + { + "epoch": 9.815024537561344, + "grad_norm": 0.6147065162658691, + "learning_rate": 2.9919014344874636e-05, + "loss": 0.3647, + "step": 26000 + }, + { + "epoch": 9.818799546998868, + "grad_norm": 0.7019761800765991, + "learning_rate": 2.9881164750236857e-05, + "loss": 0.3668, + "step": 26010 + }, + { + "epoch": 9.822574556436392, + "grad_norm": 0.6753397583961487, + "learning_rate": 2.984332890663326e-05, + "loss": 0.3729, + "step": 26020 + }, + { + "epoch": 9.826349565873915, + "grad_norm": 0.6772192120552063, + "learning_rate": 2.9805506839924292e-05, + "loss": 0.3735, + "step": 26030 + }, + { + "epoch": 9.830124575311439, + "grad_norm": 0.6899954080581665, + "learning_rate": 2.9767698575960968e-05, + "loss": 0.3628, + "step": 26040 + }, + { + "epoch": 9.833899584748963, + "grad_norm": 0.6952143907546997, + "learning_rate": 2.9729904140584913e-05, + "loss": 0.3724, + "step": 26050 + }, + { + "epoch": 9.837674594186485, + "grad_norm": 0.6387107968330383, + "learning_rate": 2.9692123559628234e-05, + "loss": 0.368, + "step": 26060 + }, + { + "epoch": 9.841449603624008, + "grad_norm": 0.6607224345207214, + "learning_rate": 2.9654356858913596e-05, + "loss": 0.3642, + "step": 26070 + }, + { + "epoch": 9.845224613061532, + "grad_norm": 0.6394532322883606, + "learning_rate": 2.9616604064254206e-05, + "loss": 0.3641, + "step": 26080 + }, + { + "epoch": 9.848999622499056, + "grad_norm": 0.6913762092590332, + "learning_rate": 2.9578865201453732e-05, + "loss": 0.3614, + "step": 26090 + }, + { + "epoch": 9.85277463193658, + "grad_norm": 0.7300879955291748, + "learning_rate": 2.9541140296306335e-05, + "loss": 0.3701, + "step": 26100 + }, + { + "epoch": 9.856549641374103, + "grad_norm": 0.6649528741836548, + "learning_rate": 2.9503429374596627e-05, + "loss": 0.3746, + "step": 26110 + }, + { + "epoch": 9.860324650811627, + "grad_norm": 0.693498432636261, + "learning_rate": 2.946573246209967e-05, + "loss": 0.3665, + "step": 26120 + }, + { + "epoch": 9.86409966024915, + "grad_norm": 0.8667472004890442, + "learning_rate": 2.942804958458094e-05, + "loss": 0.3788, + "step": 26130 + }, + { + "epoch": 9.867874669686675, + "grad_norm": 0.6629686951637268, + "learning_rate": 2.9390380767796343e-05, + "loss": 0.3673, + "step": 26140 + }, + { + "epoch": 9.871649679124198, + "grad_norm": 0.6008899211883545, + "learning_rate": 2.9352726037492174e-05, + "loss": 0.3666, + "step": 26150 + }, + { + "epoch": 9.875424688561722, + "grad_norm": 0.7541015148162842, + "learning_rate": 2.9315085419405052e-05, + "loss": 0.364, + "step": 26160 + }, + { + "epoch": 9.879199697999246, + "grad_norm": 0.6057034730911255, + "learning_rate": 2.927745893926199e-05, + "loss": 0.3711, + "step": 26170 + }, + { + "epoch": 9.88297470743677, + "grad_norm": 0.6768243908882141, + "learning_rate": 2.9239846622780358e-05, + "loss": 0.3707, + "step": 26180 + }, + { + "epoch": 9.886749716874291, + "grad_norm": 0.7239327430725098, + "learning_rate": 2.9202248495667788e-05, + "loss": 0.3791, + "step": 26190 + }, + { + "epoch": 9.890524726311815, + "grad_norm": 0.5899133086204529, + "learning_rate": 2.916466458362227e-05, + "loss": 0.3661, + "step": 26200 + }, + { + "epoch": 9.894299735749339, + "grad_norm": 0.5926426649093628, + "learning_rate": 2.9127094912332033e-05, + "loss": 0.3753, + "step": 26210 + }, + { + "epoch": 9.898074745186863, + "grad_norm": 2.781858205795288, + "learning_rate": 2.9089539507475606e-05, + "loss": 0.3749, + "step": 26220 + }, + { + "epoch": 9.901849754624386, + "grad_norm": 0.6369478106498718, + "learning_rate": 2.9051998394721748e-05, + "loss": 0.3652, + "step": 26230 + }, + { + "epoch": 9.90562476406191, + "grad_norm": 0.6876001954078674, + "learning_rate": 2.901447159972948e-05, + "loss": 0.3613, + "step": 26240 + }, + { + "epoch": 9.909399773499434, + "grad_norm": 0.6942611932754517, + "learning_rate": 2.8976959148148e-05, + "loss": 0.372, + "step": 26250 + }, + { + "epoch": 9.913174782936958, + "grad_norm": 0.7924140095710754, + "learning_rate": 2.8939461065616674e-05, + "loss": 0.3695, + "step": 26260 + }, + { + "epoch": 9.916949792374481, + "grad_norm": 0.6020333766937256, + "learning_rate": 2.8901977377765127e-05, + "loss": 0.3688, + "step": 26270 + }, + { + "epoch": 9.920724801812005, + "grad_norm": 0.6307262778282166, + "learning_rate": 2.8864508110213094e-05, + "loss": 0.3649, + "step": 26280 + }, + { + "epoch": 9.924499811249529, + "grad_norm": 0.6264989972114563, + "learning_rate": 2.8827053288570503e-05, + "loss": 0.3745, + "step": 26290 + }, + { + "epoch": 9.92827482068705, + "grad_norm": 0.7633573412895203, + "learning_rate": 2.8789612938437315e-05, + "loss": 0.3692, + "step": 26300 + }, + { + "epoch": 9.932049830124575, + "grad_norm": 0.9851648211479187, + "learning_rate": 2.8752187085403683e-05, + "loss": 0.3746, + "step": 26310 + }, + { + "epoch": 9.935824839562098, + "grad_norm": 0.6989134550094604, + "learning_rate": 2.8714775755049818e-05, + "loss": 0.3723, + "step": 26320 + }, + { + "epoch": 9.939599848999622, + "grad_norm": 0.6544426679611206, + "learning_rate": 2.867737897294604e-05, + "loss": 0.3744, + "step": 26330 + }, + { + "epoch": 9.943374858437146, + "grad_norm": 0.6202349662780762, + "learning_rate": 2.8639996764652653e-05, + "loss": 0.3753, + "step": 26340 + }, + { + "epoch": 9.94714986787467, + "grad_norm": 0.6523613929748535, + "learning_rate": 2.8602629155720084e-05, + "loss": 0.3651, + "step": 26350 + }, + { + "epoch": 9.950924877312193, + "grad_norm": 0.6727064251899719, + "learning_rate": 2.8565276171688703e-05, + "loss": 0.3732, + "step": 26360 + }, + { + "epoch": 9.954699886749717, + "grad_norm": 0.7078951001167297, + "learning_rate": 2.8527937838088943e-05, + "loss": 0.3648, + "step": 26370 + }, + { + "epoch": 9.95847489618724, + "grad_norm": 0.7242457270622253, + "learning_rate": 2.84906141804412e-05, + "loss": 0.3718, + "step": 26380 + }, + { + "epoch": 9.962249905624764, + "grad_norm": 0.6599700450897217, + "learning_rate": 2.8453305224255867e-05, + "loss": 0.364, + "step": 26390 + }, + { + "epoch": 9.966024915062288, + "grad_norm": 0.6147913336753845, + "learning_rate": 2.8416010995033216e-05, + "loss": 0.3598, + "step": 26400 + }, + { + "epoch": 9.969799924499812, + "grad_norm": 0.6449320316314697, + "learning_rate": 2.8378731518263524e-05, + "loss": 0.3718, + "step": 26410 + }, + { + "epoch": 9.973574933937336, + "grad_norm": 0.6980818510055542, + "learning_rate": 2.834146681942696e-05, + "loss": 0.3695, + "step": 26420 + }, + { + "epoch": 9.977349943374858, + "grad_norm": 0.6812340617179871, + "learning_rate": 2.8304216923993622e-05, + "loss": 0.3734, + "step": 26430 + }, + { + "epoch": 9.981124952812381, + "grad_norm": 0.644469678401947, + "learning_rate": 2.8266981857423413e-05, + "loss": 0.3704, + "step": 26440 + }, + { + "epoch": 9.984899962249905, + "grad_norm": 0.6593320369720459, + "learning_rate": 2.8229761645166197e-05, + "loss": 0.3713, + "step": 26450 + }, + { + "epoch": 9.988674971687429, + "grad_norm": 0.5889549255371094, + "learning_rate": 2.81925563126616e-05, + "loss": 0.3609, + "step": 26460 + }, + { + "epoch": 9.992449981124953, + "grad_norm": 0.6138685345649719, + "learning_rate": 2.8155365885339124e-05, + "loss": 0.3712, + "step": 26470 + }, + { + "epoch": 9.996224990562476, + "grad_norm": 0.6867060661315918, + "learning_rate": 2.8118190388618093e-05, + "loss": 0.3694, + "step": 26480 + }, + { + "epoch": 10.0, + "grad_norm": 0.6916232705116272, + "learning_rate": 2.8081029847907614e-05, + "loss": 0.375, + "step": 26490 + }, + { + "epoch": 10.003775009437524, + "grad_norm": 0.6438423991203308, + "learning_rate": 2.8043884288606525e-05, + "loss": 0.3668, + "step": 26500 + }, + { + "epoch": 10.007550018875047, + "grad_norm": 0.5959725975990295, + "learning_rate": 2.8006753736103496e-05, + "loss": 0.3637, + "step": 26510 + }, + { + "epoch": 10.011325028312571, + "grad_norm": 0.5706154108047485, + "learning_rate": 2.7969638215776918e-05, + "loss": 0.3712, + "step": 26520 + }, + { + "epoch": 10.015100037750095, + "grad_norm": 0.6247308254241943, + "learning_rate": 2.793253775299487e-05, + "loss": 0.3686, + "step": 26530 + }, + { + "epoch": 10.018875047187619, + "grad_norm": 0.6258418560028076, + "learning_rate": 2.7895452373115184e-05, + "loss": 0.3733, + "step": 26540 + }, + { + "epoch": 10.022650056625142, + "grad_norm": 0.6243855357170105, + "learning_rate": 2.785838210148539e-05, + "loss": 0.3663, + "step": 26550 + }, + { + "epoch": 10.026425066062664, + "grad_norm": 0.6902307868003845, + "learning_rate": 2.782132696344263e-05, + "loss": 0.3783, + "step": 26560 + }, + { + "epoch": 10.030200075500188, + "grad_norm": 0.650276243686676, + "learning_rate": 2.7784286984313745e-05, + "loss": 0.368, + "step": 26570 + }, + { + "epoch": 10.033975084937712, + "grad_norm": 0.6885844469070435, + "learning_rate": 2.7747262189415236e-05, + "loss": 0.3603, + "step": 26580 + }, + { + "epoch": 10.037750094375236, + "grad_norm": 0.6648432612419128, + "learning_rate": 2.7710252604053205e-05, + "loss": 0.3738, + "step": 26590 + }, + { + "epoch": 10.04152510381276, + "grad_norm": 0.6929357051849365, + "learning_rate": 2.767325825352332e-05, + "loss": 0.3692, + "step": 26600 + }, + { + "epoch": 10.045300113250283, + "grad_norm": 0.6674989461898804, + "learning_rate": 2.7636279163110913e-05, + "loss": 0.3668, + "step": 26610 + }, + { + "epoch": 10.049075122687807, + "grad_norm": 0.6698505282402039, + "learning_rate": 2.7599315358090795e-05, + "loss": 0.3727, + "step": 26620 + }, + { + "epoch": 10.05285013212533, + "grad_norm": 0.6277247071266174, + "learning_rate": 2.7562366863727407e-05, + "loss": 0.3664, + "step": 26630 + }, + { + "epoch": 10.056625141562854, + "grad_norm": 0.6725890636444092, + "learning_rate": 2.7525433705274695e-05, + "loss": 0.3661, + "step": 26640 + }, + { + "epoch": 10.060400151000378, + "grad_norm": 0.6743810772895813, + "learning_rate": 2.748851590797614e-05, + "loss": 0.3725, + "step": 26650 + }, + { + "epoch": 10.064175160437902, + "grad_norm": 0.7060372829437256, + "learning_rate": 2.7451613497064675e-05, + "loss": 0.3743, + "step": 26660 + }, + { + "epoch": 10.067950169875425, + "grad_norm": 0.6378623843193054, + "learning_rate": 2.7414726497762765e-05, + "loss": 0.3688, + "step": 26670 + }, + { + "epoch": 10.071725179312947, + "grad_norm": 0.6196885704994202, + "learning_rate": 2.737785493528232e-05, + "loss": 0.3732, + "step": 26680 + }, + { + "epoch": 10.075500188750471, + "grad_norm": 0.770141065120697, + "learning_rate": 2.7340998834824745e-05, + "loss": 0.3697, + "step": 26690 + }, + { + "epoch": 10.079275198187995, + "grad_norm": 0.6975536942481995, + "learning_rate": 2.7304158221580777e-05, + "loss": 0.3688, + "step": 26700 + }, + { + "epoch": 10.083050207625519, + "grad_norm": 0.6799348592758179, + "learning_rate": 2.7267333120730675e-05, + "loss": 0.3716, + "step": 26710 + }, + { + "epoch": 10.086825217063042, + "grad_norm": 0.6629582643508911, + "learning_rate": 2.7230523557444017e-05, + "loss": 0.3653, + "step": 26720 + }, + { + "epoch": 10.090600226500566, + "grad_norm": 0.612046480178833, + "learning_rate": 2.7193729556879798e-05, + "loss": 0.3646, + "step": 26730 + }, + { + "epoch": 10.09437523593809, + "grad_norm": 0.8744679093360901, + "learning_rate": 2.715695114418637e-05, + "loss": 0.3704, + "step": 26740 + }, + { + "epoch": 10.098150245375614, + "grad_norm": 0.6572837829589844, + "learning_rate": 2.7120188344501475e-05, + "loss": 0.368, + "step": 26750 + }, + { + "epoch": 10.101925254813137, + "grad_norm": 0.680645763874054, + "learning_rate": 2.7083441182952067e-05, + "loss": 0.3746, + "step": 26760 + }, + { + "epoch": 10.105700264250661, + "grad_norm": 0.6419723629951477, + "learning_rate": 2.7046709684654527e-05, + "loss": 0.366, + "step": 26770 + }, + { + "epoch": 10.109475273688185, + "grad_norm": 0.6527169346809387, + "learning_rate": 2.700999387471448e-05, + "loss": 0.3715, + "step": 26780 + }, + { + "epoch": 10.113250283125709, + "grad_norm": 0.6019080281257629, + "learning_rate": 2.6973293778226854e-05, + "loss": 0.3676, + "step": 26790 + }, + { + "epoch": 10.117025292563232, + "grad_norm": 0.6170861721038818, + "learning_rate": 2.6936609420275804e-05, + "loss": 0.381, + "step": 26800 + }, + { + "epoch": 10.120800302000754, + "grad_norm": 0.8691985607147217, + "learning_rate": 2.689994082593472e-05, + "loss": 0.3656, + "step": 26810 + }, + { + "epoch": 10.124575311438278, + "grad_norm": 0.6585285067558289, + "learning_rate": 2.6863288020266264e-05, + "loss": 0.368, + "step": 26820 + }, + { + "epoch": 10.128350320875802, + "grad_norm": 0.6726837158203125, + "learning_rate": 2.682665102832228e-05, + "loss": 0.3652, + "step": 26830 + }, + { + "epoch": 10.132125330313325, + "grad_norm": 0.7205042839050293, + "learning_rate": 2.67900298751438e-05, + "loss": 0.3722, + "step": 26840 + }, + { + "epoch": 10.13590033975085, + "grad_norm": 0.6744375228881836, + "learning_rate": 2.6753424585761067e-05, + "loss": 0.3751, + "step": 26850 + }, + { + "epoch": 10.139675349188373, + "grad_norm": 3.0684523582458496, + "learning_rate": 2.671683518519341e-05, + "loss": 0.3647, + "step": 26860 + }, + { + "epoch": 10.143450358625897, + "grad_norm": 0.585762083530426, + "learning_rate": 2.668026169844936e-05, + "loss": 0.3621, + "step": 26870 + }, + { + "epoch": 10.14722536806342, + "grad_norm": 0.6915805339813232, + "learning_rate": 2.6643704150526538e-05, + "loss": 0.3648, + "step": 26880 + }, + { + "epoch": 10.151000377500944, + "grad_norm": 0.7378256916999817, + "learning_rate": 2.6607162566411716e-05, + "loss": 0.3619, + "step": 26890 + }, + { + "epoch": 10.154775386938468, + "grad_norm": 0.750008761882782, + "learning_rate": 2.6570636971080697e-05, + "loss": 0.3755, + "step": 26900 + }, + { + "epoch": 10.158550396375992, + "grad_norm": 0.6996826529502869, + "learning_rate": 2.6534127389498364e-05, + "loss": 0.3635, + "step": 26910 + }, + { + "epoch": 10.162325405813515, + "grad_norm": 0.6714365482330322, + "learning_rate": 2.6497633846618696e-05, + "loss": 0.3672, + "step": 26920 + }, + { + "epoch": 10.166100415251037, + "grad_norm": 3.8461427688598633, + "learning_rate": 2.6461156367384677e-05, + "loss": 0.3705, + "step": 26930 + }, + { + "epoch": 10.169875424688561, + "grad_norm": 0.6795220375061035, + "learning_rate": 2.6424694976728316e-05, + "loss": 0.3727, + "step": 26940 + }, + { + "epoch": 10.173650434126085, + "grad_norm": 0.7694344520568848, + "learning_rate": 2.6388249699570667e-05, + "loss": 0.3705, + "step": 26950 + }, + { + "epoch": 10.177425443563608, + "grad_norm": 0.6592543721199036, + "learning_rate": 2.6351820560821672e-05, + "loss": 0.3717, + "step": 26960 + }, + { + "epoch": 10.181200453001132, + "grad_norm": 0.7184135317802429, + "learning_rate": 2.631540758538034e-05, + "loss": 0.3713, + "step": 26970 + }, + { + "epoch": 10.184975462438656, + "grad_norm": 0.6492671966552734, + "learning_rate": 2.6279010798134597e-05, + "loss": 0.3637, + "step": 26980 + }, + { + "epoch": 10.18875047187618, + "grad_norm": 0.6501529812812805, + "learning_rate": 2.6242630223961305e-05, + "loss": 0.3774, + "step": 26990 + }, + { + "epoch": 10.192525481313703, + "grad_norm": 0.8947603106498718, + "learning_rate": 2.6206265887726244e-05, + "loss": 0.369, + "step": 27000 + }, + { + "epoch": 10.196300490751227, + "grad_norm": 0.7419344186782837, + "learning_rate": 2.6169917814284066e-05, + "loss": 0.3672, + "step": 27010 + }, + { + "epoch": 10.20007550018875, + "grad_norm": 0.703224241733551, + "learning_rate": 2.6133586028478364e-05, + "loss": 0.3683, + "step": 27020 + }, + { + "epoch": 10.203850509626275, + "grad_norm": 0.6347489356994629, + "learning_rate": 2.609727055514155e-05, + "loss": 0.367, + "step": 27030 + }, + { + "epoch": 10.207625519063798, + "grad_norm": 0.6552822589874268, + "learning_rate": 2.606097141909494e-05, + "loss": 0.3647, + "step": 27040 + }, + { + "epoch": 10.211400528501322, + "grad_norm": 0.6825968623161316, + "learning_rate": 2.6024688645148644e-05, + "loss": 0.3713, + "step": 27050 + }, + { + "epoch": 10.215175537938844, + "grad_norm": 0.7292717695236206, + "learning_rate": 2.5988422258101564e-05, + "loss": 0.3673, + "step": 27060 + }, + { + "epoch": 10.218950547376368, + "grad_norm": 1.0935133695602417, + "learning_rate": 2.5952172282741453e-05, + "loss": 0.3688, + "step": 27070 + }, + { + "epoch": 10.222725556813892, + "grad_norm": 0.6263819932937622, + "learning_rate": 2.5915938743844853e-05, + "loss": 0.3764, + "step": 27080 + }, + { + "epoch": 10.226500566251415, + "grad_norm": 0.6874544024467468, + "learning_rate": 2.5879721666177003e-05, + "loss": 0.3675, + "step": 27090 + }, + { + "epoch": 10.230275575688939, + "grad_norm": 0.640953004360199, + "learning_rate": 2.5843521074491972e-05, + "loss": 0.3691, + "step": 27100 + }, + { + "epoch": 10.234050585126463, + "grad_norm": 0.6374151110649109, + "learning_rate": 2.5807336993532487e-05, + "loss": 0.3623, + "step": 27110 + }, + { + "epoch": 10.237825594563986, + "grad_norm": 0.6693562269210815, + "learning_rate": 2.577116944803004e-05, + "loss": 0.3702, + "step": 27120 + }, + { + "epoch": 10.24160060400151, + "grad_norm": 0.6881791949272156, + "learning_rate": 2.5735018462704818e-05, + "loss": 0.3695, + "step": 27130 + }, + { + "epoch": 10.245375613439034, + "grad_norm": 0.6473117470741272, + "learning_rate": 2.5698884062265665e-05, + "loss": 0.3652, + "step": 27140 + }, + { + "epoch": 10.249150622876558, + "grad_norm": 0.7239118218421936, + "learning_rate": 2.5662766271410134e-05, + "loss": 0.3715, + "step": 27150 + }, + { + "epoch": 10.252925632314081, + "grad_norm": 0.6931194067001343, + "learning_rate": 2.5626665114824343e-05, + "loss": 0.3679, + "step": 27160 + }, + { + "epoch": 10.256700641751605, + "grad_norm": 0.7126404643058777, + "learning_rate": 2.5590580617183148e-05, + "loss": 0.3729, + "step": 27170 + }, + { + "epoch": 10.260475651189129, + "grad_norm": 0.612730085849762, + "learning_rate": 2.5554512803149912e-05, + "loss": 0.3626, + "step": 27180 + }, + { + "epoch": 10.26425066062665, + "grad_norm": 0.631647527217865, + "learning_rate": 2.5518461697376662e-05, + "loss": 0.3656, + "step": 27190 + }, + { + "epoch": 10.268025670064175, + "grad_norm": 0.692253053188324, + "learning_rate": 2.548242732450402e-05, + "loss": 0.3658, + "step": 27200 + }, + { + "epoch": 10.271800679501698, + "grad_norm": 0.8286940455436707, + "learning_rate": 2.5446409709161095e-05, + "loss": 0.3708, + "step": 27210 + }, + { + "epoch": 10.275575688939222, + "grad_norm": 1.055858850479126, + "learning_rate": 2.541040887596561e-05, + "loss": 0.362, + "step": 27220 + }, + { + "epoch": 10.279350698376746, + "grad_norm": 0.6639923453330994, + "learning_rate": 2.537442484952378e-05, + "loss": 0.3684, + "step": 27230 + }, + { + "epoch": 10.28312570781427, + "grad_norm": 0.6410587430000305, + "learning_rate": 2.533845765443037e-05, + "loss": 0.3736, + "step": 27240 + }, + { + "epoch": 10.286900717251793, + "grad_norm": 0.6842541694641113, + "learning_rate": 2.530250731526863e-05, + "loss": 0.3624, + "step": 27250 + }, + { + "epoch": 10.290675726689317, + "grad_norm": 0.6937667727470398, + "learning_rate": 2.5266573856610253e-05, + "loss": 0.3588, + "step": 27260 + }, + { + "epoch": 10.29445073612684, + "grad_norm": 0.6150068044662476, + "learning_rate": 2.5230657303015403e-05, + "loss": 0.3604, + "step": 27270 + }, + { + "epoch": 10.298225745564364, + "grad_norm": 0.6743102669715881, + "learning_rate": 2.5194757679032728e-05, + "loss": 0.3679, + "step": 27280 + }, + { + "epoch": 10.302000755001888, + "grad_norm": 0.6566803455352783, + "learning_rate": 2.5158875009199278e-05, + "loss": 0.3657, + "step": 27290 + }, + { + "epoch": 10.305775764439412, + "grad_norm": 0.6627777218818665, + "learning_rate": 2.5123009318040537e-05, + "loss": 0.3677, + "step": 27300 + }, + { + "epoch": 10.309550773876934, + "grad_norm": 0.7576314806938171, + "learning_rate": 2.508716063007034e-05, + "loss": 0.3617, + "step": 27310 + }, + { + "epoch": 10.313325783314458, + "grad_norm": 1.1635149717330933, + "learning_rate": 2.5051328969790934e-05, + "loss": 0.367, + "step": 27320 + }, + { + "epoch": 10.317100792751981, + "grad_norm": 0.6241893768310547, + "learning_rate": 2.501551436169292e-05, + "loss": 0.3721, + "step": 27330 + }, + { + "epoch": 10.320875802189505, + "grad_norm": 0.5901798605918884, + "learning_rate": 2.4979716830255255e-05, + "loss": 0.3629, + "step": 27340 + }, + { + "epoch": 10.324650811627029, + "grad_norm": 0.6239721179008484, + "learning_rate": 2.4943936399945233e-05, + "loss": 0.3647, + "step": 27350 + }, + { + "epoch": 10.328425821064553, + "grad_norm": 0.6076030135154724, + "learning_rate": 2.4908173095218412e-05, + "loss": 0.3602, + "step": 27360 + }, + { + "epoch": 10.332200830502076, + "grad_norm": 0.6556723117828369, + "learning_rate": 2.4872426940518663e-05, + "loss": 0.369, + "step": 27370 + }, + { + "epoch": 10.3359758399396, + "grad_norm": 0.6872203350067139, + "learning_rate": 2.4836697960278156e-05, + "loss": 0.3669, + "step": 27380 + }, + { + "epoch": 10.339750849377124, + "grad_norm": 0.6859211921691895, + "learning_rate": 2.480098617891732e-05, + "loss": 0.3678, + "step": 27390 + }, + { + "epoch": 10.343525858814647, + "grad_norm": 0.6686061024665833, + "learning_rate": 2.4765291620844837e-05, + "loss": 0.3558, + "step": 27400 + }, + { + "epoch": 10.347300868252171, + "grad_norm": 0.6350464820861816, + "learning_rate": 2.472961431045756e-05, + "loss": 0.3808, + "step": 27410 + }, + { + "epoch": 10.351075877689695, + "grad_norm": 0.6436896324157715, + "learning_rate": 2.4693954272140622e-05, + "loss": 0.3644, + "step": 27420 + }, + { + "epoch": 10.354850887127217, + "grad_norm": 0.7389644980430603, + "learning_rate": 2.4658311530267315e-05, + "loss": 0.368, + "step": 27430 + }, + { + "epoch": 10.35862589656474, + "grad_norm": 0.7080246210098267, + "learning_rate": 2.4622686109199124e-05, + "loss": 0.368, + "step": 27440 + }, + { + "epoch": 10.362400906002264, + "grad_norm": 0.6405688524246216, + "learning_rate": 2.4587078033285695e-05, + "loss": 0.3731, + "step": 27450 + }, + { + "epoch": 10.366175915439788, + "grad_norm": 0.6456767320632935, + "learning_rate": 2.45514873268648e-05, + "loss": 0.3615, + "step": 27460 + }, + { + "epoch": 10.369950924877312, + "grad_norm": 0.6688347458839417, + "learning_rate": 2.4515914014262336e-05, + "loss": 0.3618, + "step": 27470 + }, + { + "epoch": 10.373725934314836, + "grad_norm": 0.7001455426216125, + "learning_rate": 2.4480358119792345e-05, + "loss": 0.3669, + "step": 27480 + }, + { + "epoch": 10.37750094375236, + "grad_norm": 0.6570714712142944, + "learning_rate": 2.4444819667756942e-05, + "loss": 0.3574, + "step": 27490 + }, + { + "epoch": 10.381275953189883, + "grad_norm": 0.6306593418121338, + "learning_rate": 2.4409298682446346e-05, + "loss": 0.3646, + "step": 27500 + }, + { + "epoch": 10.385050962627407, + "grad_norm": 0.6927255988121033, + "learning_rate": 2.437379518813877e-05, + "loss": 0.3671, + "step": 27510 + }, + { + "epoch": 10.38882597206493, + "grad_norm": 0.7374937534332275, + "learning_rate": 2.4338309209100547e-05, + "loss": 0.3626, + "step": 27520 + }, + { + "epoch": 10.392600981502454, + "grad_norm": 1.0554702281951904, + "learning_rate": 2.4302840769586004e-05, + "loss": 0.3635, + "step": 27530 + }, + { + "epoch": 10.396375990939978, + "grad_norm": 0.6217535734176636, + "learning_rate": 2.42673898938375e-05, + "loss": 0.3728, + "step": 27540 + }, + { + "epoch": 10.400151000377502, + "grad_norm": 0.654321551322937, + "learning_rate": 2.4231956606085343e-05, + "loss": 0.3616, + "step": 27550 + }, + { + "epoch": 10.403926009815024, + "grad_norm": 0.7654052972793579, + "learning_rate": 2.419654093054789e-05, + "loss": 0.3723, + "step": 27560 + }, + { + "epoch": 10.407701019252547, + "grad_norm": 0.6208662390708923, + "learning_rate": 2.4161142891431375e-05, + "loss": 0.3639, + "step": 27570 + }, + { + "epoch": 10.411476028690071, + "grad_norm": 0.7695934772491455, + "learning_rate": 2.412576251293005e-05, + "loss": 0.3657, + "step": 27580 + }, + { + "epoch": 10.415251038127595, + "grad_norm": 0.6474165916442871, + "learning_rate": 2.4090399819226068e-05, + "loss": 0.3664, + "step": 27590 + }, + { + "epoch": 10.419026047565119, + "grad_norm": 0.716998815536499, + "learning_rate": 2.4055054834489514e-05, + "loss": 0.3695, + "step": 27600 + }, + { + "epoch": 10.422801057002642, + "grad_norm": 0.6260504722595215, + "learning_rate": 2.401972758287832e-05, + "loss": 0.3659, + "step": 27610 + }, + { + "epoch": 10.426576066440166, + "grad_norm": 0.668073296546936, + "learning_rate": 2.398441808853834e-05, + "loss": 0.3555, + "step": 27620 + }, + { + "epoch": 10.43035107587769, + "grad_norm": 0.6870536208152771, + "learning_rate": 2.3949126375603288e-05, + "loss": 0.3675, + "step": 27630 + }, + { + "epoch": 10.434126085315214, + "grad_norm": 0.6690881848335266, + "learning_rate": 2.3913852468194724e-05, + "loss": 0.3678, + "step": 27640 + }, + { + "epoch": 10.437901094752737, + "grad_norm": 0.6636070013046265, + "learning_rate": 2.387859639042201e-05, + "loss": 0.3693, + "step": 27650 + }, + { + "epoch": 10.441676104190261, + "grad_norm": 0.7105987668037415, + "learning_rate": 2.3843358166382368e-05, + "loss": 0.3618, + "step": 27660 + }, + { + "epoch": 10.445451113627785, + "grad_norm": 0.6671977639198303, + "learning_rate": 2.3808137820160757e-05, + "loss": 0.3639, + "step": 27670 + }, + { + "epoch": 10.449226123065309, + "grad_norm": 0.6227991580963135, + "learning_rate": 2.3772935375829975e-05, + "loss": 0.3587, + "step": 27680 + }, + { + "epoch": 10.45300113250283, + "grad_norm": 0.5635626316070557, + "learning_rate": 2.3737750857450553e-05, + "loss": 0.3723, + "step": 27690 + }, + { + "epoch": 10.456776141940354, + "grad_norm": 0.6364959478378296, + "learning_rate": 2.3702584289070805e-05, + "loss": 0.3624, + "step": 27700 + }, + { + "epoch": 10.460551151377878, + "grad_norm": 0.6209636330604553, + "learning_rate": 2.36674356947267e-05, + "loss": 0.3678, + "step": 27710 + }, + { + "epoch": 10.464326160815402, + "grad_norm": 0.6616617441177368, + "learning_rate": 2.3632305098442004e-05, + "loss": 0.3583, + "step": 27720 + }, + { + "epoch": 10.468101170252925, + "grad_norm": 0.7123914361000061, + "learning_rate": 2.3597192524228156e-05, + "loss": 0.3636, + "step": 27730 + }, + { + "epoch": 10.47187617969045, + "grad_norm": 0.6567232012748718, + "learning_rate": 2.356209799608424e-05, + "loss": 0.3654, + "step": 27740 + }, + { + "epoch": 10.475651189127973, + "grad_norm": 0.6329348087310791, + "learning_rate": 2.352702153799704e-05, + "loss": 0.3665, + "step": 27750 + }, + { + "epoch": 10.479426198565497, + "grad_norm": 0.6293632984161377, + "learning_rate": 2.3491963173941018e-05, + "loss": 0.3626, + "step": 27760 + }, + { + "epoch": 10.48320120800302, + "grad_norm": 0.652357816696167, + "learning_rate": 2.3456922927878196e-05, + "loss": 0.3759, + "step": 27770 + }, + { + "epoch": 10.486976217440544, + "grad_norm": 0.6584979295730591, + "learning_rate": 2.3421900823758257e-05, + "loss": 0.3652, + "step": 27780 + }, + { + "epoch": 10.490751226878068, + "grad_norm": 0.6862519383430481, + "learning_rate": 2.3386896885518496e-05, + "loss": 0.3648, + "step": 27790 + }, + { + "epoch": 10.494526236315592, + "grad_norm": 0.6221929788589478, + "learning_rate": 2.335191113708378e-05, + "loss": 0.3721, + "step": 27800 + }, + { + "epoch": 10.498301245753115, + "grad_norm": 0.6476244330406189, + "learning_rate": 2.331694360236651e-05, + "loss": 0.3746, + "step": 27810 + }, + { + "epoch": 10.502076255190637, + "grad_norm": 0.626964807510376, + "learning_rate": 2.3281994305266702e-05, + "loss": 0.3616, + "step": 27820 + }, + { + "epoch": 10.505851264628161, + "grad_norm": 0.7498915791511536, + "learning_rate": 2.3247063269671826e-05, + "loss": 0.3689, + "step": 27830 + }, + { + "epoch": 10.509626274065685, + "grad_norm": 0.6770737767219543, + "learning_rate": 2.321215051945695e-05, + "loss": 0.3672, + "step": 27840 + }, + { + "epoch": 10.513401283503208, + "grad_norm": 0.7251414656639099, + "learning_rate": 2.3177256078484588e-05, + "loss": 0.3664, + "step": 27850 + }, + { + "epoch": 10.517176292940732, + "grad_norm": 0.7046555876731873, + "learning_rate": 2.3142379970604798e-05, + "loss": 0.361, + "step": 27860 + }, + { + "epoch": 10.520951302378256, + "grad_norm": 0.6266787648200989, + "learning_rate": 2.3107522219655025e-05, + "loss": 0.3649, + "step": 27870 + }, + { + "epoch": 10.52472631181578, + "grad_norm": 0.6274697780609131, + "learning_rate": 2.3072682849460236e-05, + "loss": 0.3601, + "step": 27880 + }, + { + "epoch": 10.528501321253303, + "grad_norm": 0.6082983613014221, + "learning_rate": 2.303786188383281e-05, + "loss": 0.3688, + "step": 27890 + }, + { + "epoch": 10.532276330690827, + "grad_norm": 0.6668601632118225, + "learning_rate": 2.300305934657257e-05, + "loss": 0.361, + "step": 27900 + }, + { + "epoch": 10.53605134012835, + "grad_norm": 0.6825070977210999, + "learning_rate": 2.2968275261466677e-05, + "loss": 0.3637, + "step": 27910 + }, + { + "epoch": 10.539826349565875, + "grad_norm": 0.6331124901771545, + "learning_rate": 2.293350965228977e-05, + "loss": 0.3639, + "step": 27920 + }, + { + "epoch": 10.543601359003397, + "grad_norm": 0.6340346932411194, + "learning_rate": 2.2898762542803776e-05, + "loss": 0.3625, + "step": 27930 + }, + { + "epoch": 10.54737636844092, + "grad_norm": 0.7296791672706604, + "learning_rate": 2.286403395675803e-05, + "loss": 0.3667, + "step": 27940 + }, + { + "epoch": 10.551151377878444, + "grad_norm": 0.6600122451782227, + "learning_rate": 2.28293239178892e-05, + "loss": 0.3671, + "step": 27950 + }, + { + "epoch": 10.554926387315968, + "grad_norm": 0.7184299230575562, + "learning_rate": 2.2794632449921287e-05, + "loss": 0.3702, + "step": 27960 + }, + { + "epoch": 10.558701396753492, + "grad_norm": 0.6096726655960083, + "learning_rate": 2.275995957656555e-05, + "loss": 0.3645, + "step": 27970 + }, + { + "epoch": 10.562476406191015, + "grad_norm": 0.6670506000518799, + "learning_rate": 2.272530532152058e-05, + "loss": 0.3646, + "step": 27980 + }, + { + "epoch": 10.566251415628539, + "grad_norm": 0.6431169509887695, + "learning_rate": 2.2690669708472233e-05, + "loss": 0.361, + "step": 27990 + }, + { + "epoch": 10.570026425066063, + "grad_norm": 0.6406874060630798, + "learning_rate": 2.2656052761093655e-05, + "loss": 0.3677, + "step": 28000 + }, + { + "epoch": 10.573801434503586, + "grad_norm": 0.6442010402679443, + "learning_rate": 2.262145450304517e-05, + "loss": 0.3661, + "step": 28010 + }, + { + "epoch": 10.57757644394111, + "grad_norm": 0.6408098340034485, + "learning_rate": 2.2586874957974352e-05, + "loss": 0.3727, + "step": 28020 + }, + { + "epoch": 10.581351453378634, + "grad_norm": 0.648128092288971, + "learning_rate": 2.2552314149516012e-05, + "loss": 0.3656, + "step": 28030 + }, + { + "epoch": 10.585126462816158, + "grad_norm": 0.7829309105873108, + "learning_rate": 2.2517772101292133e-05, + "loss": 0.3665, + "step": 28040 + }, + { + "epoch": 10.588901472253681, + "grad_norm": 0.768104612827301, + "learning_rate": 2.248324883691188e-05, + "loss": 0.3687, + "step": 28050 + }, + { + "epoch": 10.592676481691203, + "grad_norm": 0.631535530090332, + "learning_rate": 2.24487443799716e-05, + "loss": 0.3683, + "step": 28060 + }, + { + "epoch": 10.596451491128727, + "grad_norm": 0.7682189345359802, + "learning_rate": 2.241425875405472e-05, + "loss": 0.3705, + "step": 28070 + }, + { + "epoch": 10.60022650056625, + "grad_norm": 0.6943038702011108, + "learning_rate": 2.2379791982731868e-05, + "loss": 0.4653, + "step": 28080 + }, + { + "epoch": 10.604001510003775, + "grad_norm": 0.634013295173645, + "learning_rate": 2.2345344089560756e-05, + "loss": 0.3621, + "step": 28090 + }, + { + "epoch": 10.607776519441298, + "grad_norm": 0.7249836325645447, + "learning_rate": 2.2310915098086206e-05, + "loss": 0.3655, + "step": 28100 + }, + { + "epoch": 10.611551528878822, + "grad_norm": 0.8200324773788452, + "learning_rate": 2.227650503184009e-05, + "loss": 0.373, + "step": 28110 + }, + { + "epoch": 10.615326538316346, + "grad_norm": 0.8548815250396729, + "learning_rate": 2.2242113914341357e-05, + "loss": 0.3581, + "step": 28120 + }, + { + "epoch": 10.61910154775387, + "grad_norm": 0.6571340560913086, + "learning_rate": 2.220774176909602e-05, + "loss": 0.3629, + "step": 28130 + }, + { + "epoch": 10.622876557191393, + "grad_norm": 0.6676787734031677, + "learning_rate": 2.2173388619597114e-05, + "loss": 0.3682, + "step": 28140 + }, + { + "epoch": 10.626651566628917, + "grad_norm": 0.7401626706123352, + "learning_rate": 2.21390544893247e-05, + "loss": 0.3667, + "step": 28150 + }, + { + "epoch": 10.63042657606644, + "grad_norm": 0.6297852396965027, + "learning_rate": 2.210473940174585e-05, + "loss": 0.3739, + "step": 28160 + }, + { + "epoch": 10.634201585503964, + "grad_norm": 0.6574814319610596, + "learning_rate": 2.207044338031456e-05, + "loss": 0.3613, + "step": 28170 + }, + { + "epoch": 10.637976594941488, + "grad_norm": 0.5910388827323914, + "learning_rate": 2.203616644847186e-05, + "loss": 0.3778, + "step": 28180 + }, + { + "epoch": 10.64175160437901, + "grad_norm": 0.6458864808082581, + "learning_rate": 2.200190862964571e-05, + "loss": 0.3678, + "step": 28190 + }, + { + "epoch": 10.645526613816534, + "grad_norm": 0.7278522849082947, + "learning_rate": 2.1967669947251024e-05, + "loss": 0.364, + "step": 28200 + }, + { + "epoch": 10.649301623254058, + "grad_norm": 0.6762776970863342, + "learning_rate": 2.1933450424689583e-05, + "loss": 0.3594, + "step": 28210 + }, + { + "epoch": 10.653076632691581, + "grad_norm": 1.226162075996399, + "learning_rate": 2.1899250085350142e-05, + "loss": 0.3733, + "step": 28220 + }, + { + "epoch": 10.656851642129105, + "grad_norm": 0.5935925841331482, + "learning_rate": 2.1865068952608277e-05, + "loss": 0.3628, + "step": 28230 + }, + { + "epoch": 10.660626651566629, + "grad_norm": 0.7492917776107788, + "learning_rate": 2.1830907049826487e-05, + "loss": 0.3611, + "step": 28240 + }, + { + "epoch": 10.664401661004153, + "grad_norm": 0.7051345705986023, + "learning_rate": 2.179676440035411e-05, + "loss": 0.3602, + "step": 28250 + }, + { + "epoch": 10.668176670441676, + "grad_norm": 0.6424268484115601, + "learning_rate": 2.1762641027527337e-05, + "loss": 0.366, + "step": 28260 + }, + { + "epoch": 10.6719516798792, + "grad_norm": 0.8368542790412903, + "learning_rate": 2.1728536954669143e-05, + "loss": 0.3646, + "step": 28270 + }, + { + "epoch": 10.675726689316724, + "grad_norm": 0.7216016054153442, + "learning_rate": 2.169445220508936e-05, + "loss": 0.36, + "step": 28280 + }, + { + "epoch": 10.679501698754247, + "grad_norm": 0.6066749691963196, + "learning_rate": 2.166038680208461e-05, + "loss": 0.3587, + "step": 28290 + }, + { + "epoch": 10.683276708191771, + "grad_norm": 0.5810279250144958, + "learning_rate": 2.162634076893823e-05, + "loss": 0.3697, + "step": 28300 + }, + { + "epoch": 10.687051717629295, + "grad_norm": 0.6657140254974365, + "learning_rate": 2.1592314128920388e-05, + "loss": 0.3678, + "step": 28310 + }, + { + "epoch": 10.690826727066817, + "grad_norm": 0.636150598526001, + "learning_rate": 2.155830690528799e-05, + "loss": 0.3654, + "step": 28320 + }, + { + "epoch": 10.69460173650434, + "grad_norm": 0.6530934572219849, + "learning_rate": 2.1524319121284613e-05, + "loss": 0.3573, + "step": 28330 + }, + { + "epoch": 10.698376745941864, + "grad_norm": 0.6528118848800659, + "learning_rate": 2.1490350800140607e-05, + "loss": 0.3712, + "step": 28340 + }, + { + "epoch": 10.702151755379388, + "grad_norm": 0.6614232659339905, + "learning_rate": 2.1456401965073002e-05, + "loss": 0.3647, + "step": 28350 + }, + { + "epoch": 10.705926764816912, + "grad_norm": 0.6931934356689453, + "learning_rate": 2.1422472639285524e-05, + "loss": 0.3641, + "step": 28360 + }, + { + "epoch": 10.709701774254436, + "grad_norm": 0.6107961535453796, + "learning_rate": 2.13885628459685e-05, + "loss": 0.3576, + "step": 28370 + }, + { + "epoch": 10.71347678369196, + "grad_norm": 0.6835388541221619, + "learning_rate": 2.135467260829901e-05, + "loss": 0.364, + "step": 28380 + }, + { + "epoch": 10.717251793129483, + "grad_norm": 0.655625581741333, + "learning_rate": 2.1320801949440654e-05, + "loss": 0.3629, + "step": 28390 + }, + { + "epoch": 10.721026802567007, + "grad_norm": 0.7267212867736816, + "learning_rate": 2.1286950892543744e-05, + "loss": 0.3641, + "step": 28400 + }, + { + "epoch": 10.72480181200453, + "grad_norm": 0.6693907380104065, + "learning_rate": 2.125311946074515e-05, + "loss": 0.3657, + "step": 28410 + }, + { + "epoch": 10.728576821442054, + "grad_norm": 0.6558740735054016, + "learning_rate": 2.1219307677168355e-05, + "loss": 0.3646, + "step": 28420 + }, + { + "epoch": 10.732351830879578, + "grad_norm": 0.7080862522125244, + "learning_rate": 2.118551556492336e-05, + "loss": 0.361, + "step": 28430 + }, + { + "epoch": 10.7361268403171, + "grad_norm": 0.6530542969703674, + "learning_rate": 2.1151743147106774e-05, + "loss": 0.3648, + "step": 28440 + }, + { + "epoch": 10.739901849754624, + "grad_norm": 0.705801784992218, + "learning_rate": 2.111799044680172e-05, + "loss": 0.3622, + "step": 28450 + }, + { + "epoch": 10.743676859192147, + "grad_norm": 0.6428266763687134, + "learning_rate": 2.1084257487077873e-05, + "loss": 0.3669, + "step": 28460 + }, + { + "epoch": 10.747451868629671, + "grad_norm": 0.7444311380386353, + "learning_rate": 2.1050544290991357e-05, + "loss": 0.3596, + "step": 28470 + }, + { + "epoch": 10.751226878067195, + "grad_norm": 0.6238864660263062, + "learning_rate": 2.101685088158486e-05, + "loss": 0.3694, + "step": 28480 + }, + { + "epoch": 10.755001887504719, + "grad_norm": 0.681848406791687, + "learning_rate": 2.0983177281887472e-05, + "loss": 0.369, + "step": 28490 + }, + { + "epoch": 10.758776896942242, + "grad_norm": 0.6995017528533936, + "learning_rate": 2.0949523514914798e-05, + "loss": 0.3644, + "step": 28500 + }, + { + "epoch": 10.762551906379766, + "grad_norm": 0.6024167537689209, + "learning_rate": 2.0915889603668876e-05, + "loss": 0.357, + "step": 28510 + }, + { + "epoch": 10.76632691581729, + "grad_norm": 0.6610404849052429, + "learning_rate": 2.0882275571138175e-05, + "loss": 0.3634, + "step": 28520 + }, + { + "epoch": 10.770101925254814, + "grad_norm": 0.629758894443512, + "learning_rate": 2.0848681440297545e-05, + "loss": 0.3644, + "step": 28530 + }, + { + "epoch": 10.773876934692337, + "grad_norm": 0.6765365600585938, + "learning_rate": 2.081510723410827e-05, + "loss": 0.3606, + "step": 28540 + }, + { + "epoch": 10.777651944129861, + "grad_norm": 0.7556784152984619, + "learning_rate": 2.0781552975518003e-05, + "loss": 0.3672, + "step": 28550 + }, + { + "epoch": 10.781426953567383, + "grad_norm": 0.7298506498336792, + "learning_rate": 2.074801868746078e-05, + "loss": 0.368, + "step": 28560 + }, + { + "epoch": 10.785201963004907, + "grad_norm": 0.7526068091392517, + "learning_rate": 2.0714504392856955e-05, + "loss": 0.368, + "step": 28570 + }, + { + "epoch": 10.78897697244243, + "grad_norm": 0.6001387238502502, + "learning_rate": 2.0681010114613215e-05, + "loss": 0.3641, + "step": 28580 + }, + { + "epoch": 10.792751981879954, + "grad_norm": 0.662828803062439, + "learning_rate": 2.0647535875622597e-05, + "loss": 0.362, + "step": 28590 + }, + { + "epoch": 10.796526991317478, + "grad_norm": 2.932863712310791, + "learning_rate": 2.0614081698764432e-05, + "loss": 0.3642, + "step": 28600 + }, + { + "epoch": 10.800302000755002, + "grad_norm": 0.7236543297767639, + "learning_rate": 2.0580647606904334e-05, + "loss": 0.3654, + "step": 28610 + }, + { + "epoch": 10.804077010192525, + "grad_norm": 0.6587764620780945, + "learning_rate": 2.0547233622894208e-05, + "loss": 0.3629, + "step": 28620 + }, + { + "epoch": 10.80785201963005, + "grad_norm": 0.6333833932876587, + "learning_rate": 2.0513839769572157e-05, + "loss": 0.3685, + "step": 28630 + }, + { + "epoch": 10.811627029067573, + "grad_norm": 0.596711277961731, + "learning_rate": 2.0480466069762584e-05, + "loss": 0.3737, + "step": 28640 + }, + { + "epoch": 10.815402038505097, + "grad_norm": 0.7744527459144592, + "learning_rate": 2.0447112546276104e-05, + "loss": 0.3628, + "step": 28650 + }, + { + "epoch": 10.81917704794262, + "grad_norm": 0.6719459295272827, + "learning_rate": 2.0413779221909547e-05, + "loss": 0.3714, + "step": 28660 + }, + { + "epoch": 10.822952057380144, + "grad_norm": 0.6692177653312683, + "learning_rate": 2.0380466119445912e-05, + "loss": 0.3652, + "step": 28670 + }, + { + "epoch": 10.826727066817668, + "grad_norm": 0.6114712953567505, + "learning_rate": 2.0347173261654373e-05, + "loss": 0.367, + "step": 28680 + }, + { + "epoch": 10.83050207625519, + "grad_norm": 0.6642469763755798, + "learning_rate": 2.03139006712903e-05, + "loss": 0.3674, + "step": 28690 + }, + { + "epoch": 10.834277085692714, + "grad_norm": 0.6511169075965881, + "learning_rate": 2.028064837109519e-05, + "loss": 0.3648, + "step": 28700 + }, + { + "epoch": 10.838052095130237, + "grad_norm": 0.7235956788063049, + "learning_rate": 2.0247416383796685e-05, + "loss": 0.3712, + "step": 28710 + }, + { + "epoch": 10.841827104567761, + "grad_norm": 0.6243189573287964, + "learning_rate": 2.0214204732108548e-05, + "loss": 0.3573, + "step": 28720 + }, + { + "epoch": 10.845602114005285, + "grad_norm": 0.6611545085906982, + "learning_rate": 2.0181013438730596e-05, + "loss": 0.3603, + "step": 28730 + }, + { + "epoch": 10.849377123442808, + "grad_norm": 0.6944743990898132, + "learning_rate": 2.0147842526348783e-05, + "loss": 0.3589, + "step": 28740 + }, + { + "epoch": 10.853152132880332, + "grad_norm": 0.6612570881843567, + "learning_rate": 2.011469201763511e-05, + "loss": 0.3635, + "step": 28750 + }, + { + "epoch": 10.856927142317856, + "grad_norm": 0.6813220977783203, + "learning_rate": 2.0081561935247665e-05, + "loss": 0.3619, + "step": 28760 + }, + { + "epoch": 10.86070215175538, + "grad_norm": 0.638427734375, + "learning_rate": 2.0048452301830523e-05, + "loss": 0.3605, + "step": 28770 + }, + { + "epoch": 10.864477161192903, + "grad_norm": 0.607269823551178, + "learning_rate": 2.0015363140013788e-05, + "loss": 0.3669, + "step": 28780 + }, + { + "epoch": 10.868252170630427, + "grad_norm": 0.6158018708229065, + "learning_rate": 1.9982294472413606e-05, + "loss": 0.3642, + "step": 28790 + }, + { + "epoch": 10.87202718006795, + "grad_norm": 0.650993287563324, + "learning_rate": 1.9949246321632103e-05, + "loss": 0.3563, + "step": 28800 + }, + { + "epoch": 10.875802189505475, + "grad_norm": 0.6607373952865601, + "learning_rate": 1.9916218710257377e-05, + "loss": 0.3608, + "step": 28810 + }, + { + "epoch": 10.879577198942997, + "grad_norm": 0.7220985889434814, + "learning_rate": 1.988321166086351e-05, + "loss": 0.3646, + "step": 28820 + }, + { + "epoch": 10.88335220838052, + "grad_norm": 0.7125725150108337, + "learning_rate": 1.9850225196010468e-05, + "loss": 0.3605, + "step": 28830 + }, + { + "epoch": 10.887127217818044, + "grad_norm": 0.6737269759178162, + "learning_rate": 1.981725933824421e-05, + "loss": 0.363, + "step": 28840 + }, + { + "epoch": 10.890902227255568, + "grad_norm": 0.6922343373298645, + "learning_rate": 1.978431411009661e-05, + "loss": 0.3687, + "step": 28850 + }, + { + "epoch": 10.894677236693092, + "grad_norm": 4.779869079589844, + "learning_rate": 1.9751389534085375e-05, + "loss": 0.3591, + "step": 28860 + }, + { + "epoch": 10.898452246130615, + "grad_norm": 0.6346884369850159, + "learning_rate": 1.9718485632714184e-05, + "loss": 0.364, + "step": 28870 + }, + { + "epoch": 10.902227255568139, + "grad_norm": 0.6880037188529968, + "learning_rate": 1.968560242847251e-05, + "loss": 0.3693, + "step": 28880 + }, + { + "epoch": 10.906002265005663, + "grad_norm": 0.7235897183418274, + "learning_rate": 1.965273994383573e-05, + "loss": 0.3605, + "step": 28890 + }, + { + "epoch": 10.909777274443186, + "grad_norm": 0.6026926636695862, + "learning_rate": 1.961989820126504e-05, + "loss": 0.3659, + "step": 28900 + }, + { + "epoch": 10.91355228388071, + "grad_norm": 0.6524337530136108, + "learning_rate": 1.958707722320746e-05, + "loss": 0.3559, + "step": 28910 + }, + { + "epoch": 10.917327293318234, + "grad_norm": 0.6593090295791626, + "learning_rate": 1.955427703209584e-05, + "loss": 0.366, + "step": 28920 + }, + { + "epoch": 10.921102302755758, + "grad_norm": 0.6755902767181396, + "learning_rate": 1.9521497650348764e-05, + "loss": 0.3599, + "step": 28930 + }, + { + "epoch": 10.924877312193281, + "grad_norm": 0.7313739061355591, + "learning_rate": 1.948873910037067e-05, + "loss": 0.363, + "step": 28940 + }, + { + "epoch": 10.928652321630803, + "grad_norm": 0.686374306678772, + "learning_rate": 1.9456001404551678e-05, + "loss": 0.377, + "step": 28950 + }, + { + "epoch": 10.932427331068327, + "grad_norm": 0.789500892162323, + "learning_rate": 1.942328458526771e-05, + "loss": 0.3672, + "step": 28960 + }, + { + "epoch": 10.93620234050585, + "grad_norm": 0.6305020451545715, + "learning_rate": 1.9390588664880427e-05, + "loss": 0.3595, + "step": 28970 + }, + { + "epoch": 10.939977349943375, + "grad_norm": 0.6587271690368652, + "learning_rate": 1.9357913665737145e-05, + "loss": 0.3596, + "step": 28980 + }, + { + "epoch": 10.943752359380898, + "grad_norm": 0.6776351928710938, + "learning_rate": 1.932525961017093e-05, + "loss": 0.3646, + "step": 28990 + }, + { + "epoch": 10.947527368818422, + "grad_norm": 0.6554774641990662, + "learning_rate": 1.9292626520500533e-05, + "loss": 0.3653, + "step": 29000 + }, + { + "epoch": 10.951302378255946, + "grad_norm": 0.702714204788208, + "learning_rate": 1.9260014419030354e-05, + "loss": 0.3693, + "step": 29010 + }, + { + "epoch": 10.95507738769347, + "grad_norm": 0.692285418510437, + "learning_rate": 1.9227423328050475e-05, + "loss": 0.3628, + "step": 29020 + }, + { + "epoch": 10.958852397130993, + "grad_norm": 0.6136350631713867, + "learning_rate": 1.9194853269836582e-05, + "loss": 0.3628, + "step": 29030 + }, + { + "epoch": 10.962627406568517, + "grad_norm": 0.8171948790550232, + "learning_rate": 1.916230426664999e-05, + "loss": 0.3799, + "step": 29040 + }, + { + "epoch": 10.96640241600604, + "grad_norm": 0.6025060415267944, + "learning_rate": 1.912977634073765e-05, + "loss": 0.3664, + "step": 29050 + }, + { + "epoch": 10.970177425443563, + "grad_norm": 0.678777277469635, + "learning_rate": 1.9097269514332083e-05, + "loss": 0.3663, + "step": 29060 + }, + { + "epoch": 10.973952434881086, + "grad_norm": 0.6317446827888489, + "learning_rate": 1.9064783809651433e-05, + "loss": 0.3644, + "step": 29070 + }, + { + "epoch": 10.97772744431861, + "grad_norm": 0.6572142243385315, + "learning_rate": 1.9032319248899333e-05, + "loss": 0.3561, + "step": 29080 + }, + { + "epoch": 10.981502453756134, + "grad_norm": 0.5952383279800415, + "learning_rate": 1.8999875854265015e-05, + "loss": 0.3578, + "step": 29090 + }, + { + "epoch": 10.985277463193658, + "grad_norm": 0.6633411645889282, + "learning_rate": 1.8967453647923232e-05, + "loss": 0.362, + "step": 29100 + }, + { + "epoch": 10.989052472631181, + "grad_norm": 0.6560747027397156, + "learning_rate": 1.893505265203427e-05, + "loss": 0.3602, + "step": 29110 + }, + { + "epoch": 10.992827482068705, + "grad_norm": 0.6242331862449646, + "learning_rate": 1.8902672888743907e-05, + "loss": 0.3589, + "step": 29120 + }, + { + "epoch": 10.996602491506229, + "grad_norm": 0.6771829724311829, + "learning_rate": 1.8870314380183396e-05, + "loss": 0.3618, + "step": 29130 + }, + { + "epoch": 11.000377500943753, + "grad_norm": 1.5967614650726318, + "learning_rate": 1.8837977148469448e-05, + "loss": 0.3566, + "step": 29140 + }, + { + "epoch": 11.004152510381276, + "grad_norm": 0.6474411487579346, + "learning_rate": 1.880566121570429e-05, + "loss": 0.3596, + "step": 29150 + }, + { + "epoch": 11.0079275198188, + "grad_norm": 0.6497718691825867, + "learning_rate": 1.877336660397554e-05, + "loss": 0.3564, + "step": 29160 + }, + { + "epoch": 11.011702529256324, + "grad_norm": 0.6360942125320435, + "learning_rate": 1.874109333535628e-05, + "loss": 0.3681, + "step": 29170 + }, + { + "epoch": 11.015477538693847, + "grad_norm": 0.7708438634872437, + "learning_rate": 1.870884143190496e-05, + "loss": 0.3651, + "step": 29180 + }, + { + "epoch": 11.01925254813137, + "grad_norm": 0.8585805892944336, + "learning_rate": 1.867661091566546e-05, + "loss": 0.3656, + "step": 29190 + }, + { + "epoch": 11.023027557568893, + "grad_norm": 2.161421537399292, + "learning_rate": 1.864440180866704e-05, + "loss": 0.3588, + "step": 29200 + }, + { + "epoch": 11.026802567006417, + "grad_norm": 0.6585597991943359, + "learning_rate": 1.8612214132924317e-05, + "loss": 0.3567, + "step": 29210 + }, + { + "epoch": 11.03057757644394, + "grad_norm": 0.6775062084197998, + "learning_rate": 1.858004791043728e-05, + "loss": 0.3581, + "step": 29220 + }, + { + "epoch": 11.034352585881464, + "grad_norm": 0.8132985234260559, + "learning_rate": 1.854790316319123e-05, + "loss": 0.3758, + "step": 29230 + }, + { + "epoch": 11.038127595318988, + "grad_norm": 0.6738925576210022, + "learning_rate": 1.8515779913156766e-05, + "loss": 0.3583, + "step": 29240 + }, + { + "epoch": 11.041902604756512, + "grad_norm": 0.6279427409172058, + "learning_rate": 1.848367818228986e-05, + "loss": 0.3697, + "step": 29250 + }, + { + "epoch": 11.045677614194036, + "grad_norm": 0.6961294412612915, + "learning_rate": 1.8451597992531733e-05, + "loss": 0.3619, + "step": 29260 + }, + { + "epoch": 11.04945262363156, + "grad_norm": 0.8088012337684631, + "learning_rate": 1.8419539365808914e-05, + "loss": 0.3755, + "step": 29270 + }, + { + "epoch": 11.053227633069083, + "grad_norm": 0.6103382706642151, + "learning_rate": 1.838750232403313e-05, + "loss": 0.3592, + "step": 29280 + }, + { + "epoch": 11.057002642506607, + "grad_norm": 0.6636168956756592, + "learning_rate": 1.835548688910142e-05, + "loss": 0.3726, + "step": 29290 + }, + { + "epoch": 11.06077765194413, + "grad_norm": 0.6306051015853882, + "learning_rate": 1.8323493082896037e-05, + "loss": 0.3665, + "step": 29300 + }, + { + "epoch": 11.064552661381654, + "grad_norm": 0.6422659754753113, + "learning_rate": 1.8291520927284454e-05, + "loss": 0.3643, + "step": 29310 + }, + { + "epoch": 11.068327670819176, + "grad_norm": 0.6881163716316223, + "learning_rate": 1.8259570444119305e-05, + "loss": 0.363, + "step": 29320 + }, + { + "epoch": 11.0721026802567, + "grad_norm": 0.5860868096351624, + "learning_rate": 1.8227641655238488e-05, + "loss": 0.3639, + "step": 29330 + }, + { + "epoch": 11.075877689694224, + "grad_norm": 0.6115444302558899, + "learning_rate": 1.819573458246498e-05, + "loss": 0.3687, + "step": 29340 + }, + { + "epoch": 11.079652699131747, + "grad_norm": 0.6907765865325928, + "learning_rate": 1.816384924760699e-05, + "loss": 0.367, + "step": 29350 + }, + { + "epoch": 11.083427708569271, + "grad_norm": 0.7525956034660339, + "learning_rate": 1.813198567245784e-05, + "loss": 0.3615, + "step": 29360 + }, + { + "epoch": 11.087202718006795, + "grad_norm": 0.625262975692749, + "learning_rate": 1.8100143878796006e-05, + "loss": 0.3565, + "step": 29370 + }, + { + "epoch": 11.090977727444319, + "grad_norm": 0.6794587969779968, + "learning_rate": 1.8068323888385015e-05, + "loss": 0.363, + "step": 29380 + }, + { + "epoch": 11.094752736881842, + "grad_norm": 0.6202002167701721, + "learning_rate": 1.803652572297355e-05, + "loss": 0.3639, + "step": 29390 + }, + { + "epoch": 11.098527746319366, + "grad_norm": 0.679071307182312, + "learning_rate": 1.8004749404295353e-05, + "loss": 0.3658, + "step": 29400 + }, + { + "epoch": 11.10230275575689, + "grad_norm": 0.6545715928077698, + "learning_rate": 1.797299495406926e-05, + "loss": 0.3692, + "step": 29410 + }, + { + "epoch": 11.106077765194414, + "grad_norm": 0.6335902810096741, + "learning_rate": 1.7941262393999103e-05, + "loss": 0.3591, + "step": 29420 + }, + { + "epoch": 11.109852774631937, + "grad_norm": 1.0289889574050903, + "learning_rate": 1.7909551745773816e-05, + "loss": 0.3531, + "step": 29430 + }, + { + "epoch": 11.113627784069461, + "grad_norm": 0.6615308523178101, + "learning_rate": 1.7877863031067304e-05, + "loss": 0.3552, + "step": 29440 + }, + { + "epoch": 11.117402793506983, + "grad_norm": 0.6820453405380249, + "learning_rate": 1.7846196271538516e-05, + "loss": 0.3688, + "step": 29450 + }, + { + "epoch": 11.121177802944507, + "grad_norm": 0.6518820524215698, + "learning_rate": 1.7814551488831384e-05, + "loss": 0.3598, + "step": 29460 + }, + { + "epoch": 11.12495281238203, + "grad_norm": 0.725908637046814, + "learning_rate": 1.7782928704574835e-05, + "loss": 0.3621, + "step": 29470 + }, + { + "epoch": 11.128727821819554, + "grad_norm": 2.0189263820648193, + "learning_rate": 1.775132794038271e-05, + "loss": 0.3616, + "step": 29480 + }, + { + "epoch": 11.132502831257078, + "grad_norm": 0.6810202598571777, + "learning_rate": 1.7719749217853855e-05, + "loss": 0.3604, + "step": 29490 + }, + { + "epoch": 11.136277840694602, + "grad_norm": 0.6146163940429688, + "learning_rate": 1.7688192558572038e-05, + "loss": 0.3593, + "step": 29500 + }, + { + "epoch": 11.140052850132125, + "grad_norm": 0.7058233022689819, + "learning_rate": 1.7656657984105906e-05, + "loss": 0.3563, + "step": 29510 + }, + { + "epoch": 11.14382785956965, + "grad_norm": 0.6673465967178345, + "learning_rate": 1.7625145516009068e-05, + "loss": 0.3613, + "step": 29520 + }, + { + "epoch": 11.147602869007173, + "grad_norm": 0.6134123802185059, + "learning_rate": 1.7593655175820005e-05, + "loss": 0.3697, + "step": 29530 + }, + { + "epoch": 11.151377878444697, + "grad_norm": 0.7190696597099304, + "learning_rate": 1.7562186985062046e-05, + "loss": 0.3666, + "step": 29540 + }, + { + "epoch": 11.15515288788222, + "grad_norm": 0.7087527513504028, + "learning_rate": 1.7530740965243403e-05, + "loss": 0.3696, + "step": 29550 + }, + { + "epoch": 11.158927897319744, + "grad_norm": 1.1012113094329834, + "learning_rate": 1.7499317137857153e-05, + "loss": 0.3582, + "step": 29560 + }, + { + "epoch": 11.162702906757266, + "grad_norm": 0.6906222701072693, + "learning_rate": 1.7467915524381184e-05, + "loss": 0.3591, + "step": 29570 + }, + { + "epoch": 11.16647791619479, + "grad_norm": 0.6418552398681641, + "learning_rate": 1.7436536146278182e-05, + "loss": 0.3702, + "step": 29580 + }, + { + "epoch": 11.170252925632314, + "grad_norm": 1.7984684705734253, + "learning_rate": 1.7405179024995688e-05, + "loss": 0.3667, + "step": 29590 + }, + { + "epoch": 11.174027935069837, + "grad_norm": 1.6302399635314941, + "learning_rate": 1.737384418196596e-05, + "loss": 0.3828, + "step": 29600 + }, + { + "epoch": 11.177802944507361, + "grad_norm": 0.7185459136962891, + "learning_rate": 1.734253163860609e-05, + "loss": 0.3602, + "step": 29610 + }, + { + "epoch": 11.181577953944885, + "grad_norm": 0.6619534492492676, + "learning_rate": 1.7311241416317896e-05, + "loss": 0.3618, + "step": 29620 + }, + { + "epoch": 11.185352963382408, + "grad_norm": 0.6776392459869385, + "learning_rate": 1.7279973536487982e-05, + "loss": 0.3626, + "step": 29630 + }, + { + "epoch": 11.189127972819932, + "grad_norm": 0.8006789088249207, + "learning_rate": 1.724872802048761e-05, + "loss": 0.3607, + "step": 29640 + }, + { + "epoch": 11.192902982257456, + "grad_norm": 0.6797590255737305, + "learning_rate": 1.7217504889672803e-05, + "loss": 0.3616, + "step": 29650 + }, + { + "epoch": 11.19667799169498, + "grad_norm": 0.6782128810882568, + "learning_rate": 1.7186304165384287e-05, + "loss": 0.3651, + "step": 29660 + }, + { + "epoch": 11.200453001132503, + "grad_norm": 0.7670999765396118, + "learning_rate": 1.7155125868947475e-05, + "loss": 0.3615, + "step": 29670 + }, + { + "epoch": 11.204228010570027, + "grad_norm": 0.6768984198570251, + "learning_rate": 1.7123970021672404e-05, + "loss": 0.3597, + "step": 29680 + }, + { + "epoch": 11.20800302000755, + "grad_norm": 0.6669506430625916, + "learning_rate": 1.709283664485384e-05, + "loss": 0.3612, + "step": 29690 + }, + { + "epoch": 11.211778029445073, + "grad_norm": 0.6505656838417053, + "learning_rate": 1.7061725759771113e-05, + "loss": 0.3583, + "step": 29700 + }, + { + "epoch": 11.215553038882597, + "grad_norm": 0.6229871511459351, + "learning_rate": 1.7030637387688248e-05, + "loss": 0.364, + "step": 29710 + }, + { + "epoch": 11.21932804832012, + "grad_norm": 2.578798770904541, + "learning_rate": 1.6999571549853836e-05, + "loss": 0.361, + "step": 29720 + }, + { + "epoch": 11.223103057757644, + "grad_norm": 0.6974425911903381, + "learning_rate": 1.696852826750112e-05, + "loss": 0.3677, + "step": 29730 + }, + { + "epoch": 11.226878067195168, + "grad_norm": 0.6888813376426697, + "learning_rate": 1.6937507561847844e-05, + "loss": 0.3687, + "step": 29740 + }, + { + "epoch": 11.230653076632692, + "grad_norm": 0.6524753570556641, + "learning_rate": 1.6906509454096385e-05, + "loss": 0.3633, + "step": 29750 + }, + { + "epoch": 11.234428086070215, + "grad_norm": 0.7103590965270996, + "learning_rate": 1.687553396543367e-05, + "loss": 0.3671, + "step": 29760 + }, + { + "epoch": 11.238203095507739, + "grad_norm": 0.6661110520362854, + "learning_rate": 1.6844581117031154e-05, + "loss": 0.3597, + "step": 29770 + }, + { + "epoch": 11.241978104945263, + "grad_norm": 1.3560032844543457, + "learning_rate": 1.681365093004481e-05, + "loss": 0.3731, + "step": 29780 + }, + { + "epoch": 11.245753114382786, + "grad_norm": 0.6432333588600159, + "learning_rate": 1.678274342561511e-05, + "loss": 0.3623, + "step": 29790 + }, + { + "epoch": 11.24952812382031, + "grad_norm": 0.6939373016357422, + "learning_rate": 1.675185862486706e-05, + "loss": 0.367, + "step": 29800 + }, + { + "epoch": 11.253303133257834, + "grad_norm": 1.7712675333023071, + "learning_rate": 1.6720996548910127e-05, + "loss": 0.3675, + "step": 29810 + }, + { + "epoch": 11.257078142695356, + "grad_norm": 0.7062354683876038, + "learning_rate": 1.6690157218838247e-05, + "loss": 0.3667, + "step": 29820 + }, + { + "epoch": 11.26085315213288, + "grad_norm": 0.6343713402748108, + "learning_rate": 1.665934065572984e-05, + "loss": 0.3529, + "step": 29830 + }, + { + "epoch": 11.264628161570403, + "grad_norm": 0.6509836316108704, + "learning_rate": 1.6628546880647688e-05, + "loss": 0.3628, + "step": 29840 + }, + { + "epoch": 11.268403171007927, + "grad_norm": 0.6391473412513733, + "learning_rate": 1.6597775914639076e-05, + "loss": 0.3665, + "step": 29850 + }, + { + "epoch": 11.27217818044545, + "grad_norm": 0.7817753553390503, + "learning_rate": 1.6567027778735654e-05, + "loss": 0.3619, + "step": 29860 + }, + { + "epoch": 11.275953189882975, + "grad_norm": 0.5986528992652893, + "learning_rate": 1.653630249395351e-05, + "loss": 0.3592, + "step": 29870 + }, + { + "epoch": 11.279728199320498, + "grad_norm": 0.6571201682090759, + "learning_rate": 1.6505600081293072e-05, + "loss": 0.3613, + "step": 29880 + }, + { + "epoch": 11.283503208758022, + "grad_norm": 0.7287798523902893, + "learning_rate": 1.647492056173912e-05, + "loss": 0.3705, + "step": 29890 + }, + { + "epoch": 11.287278218195546, + "grad_norm": 0.6623769998550415, + "learning_rate": 1.6444263956260848e-05, + "loss": 0.3559, + "step": 29900 + }, + { + "epoch": 11.29105322763307, + "grad_norm": 0.6253150105476379, + "learning_rate": 1.641363028581175e-05, + "loss": 0.363, + "step": 29910 + }, + { + "epoch": 11.294828237070593, + "grad_norm": 0.6631579399108887, + "learning_rate": 1.638301957132965e-05, + "loss": 0.3649, + "step": 29920 + }, + { + "epoch": 11.298603246508117, + "grad_norm": 0.7632430791854858, + "learning_rate": 1.6352431833736703e-05, + "loss": 0.3576, + "step": 29930 + }, + { + "epoch": 11.30237825594564, + "grad_norm": 0.8125953674316406, + "learning_rate": 1.6321867093939298e-05, + "loss": 0.3624, + "step": 29940 + }, + { + "epoch": 11.306153265383163, + "grad_norm": 0.6205570101737976, + "learning_rate": 1.629132537282817e-05, + "loss": 0.3575, + "step": 29950 + }, + { + "epoch": 11.309928274820686, + "grad_norm": 0.6771644949913025, + "learning_rate": 1.62608066912783e-05, + "loss": 0.3572, + "step": 29960 + }, + { + "epoch": 11.31370328425821, + "grad_norm": 0.6554526090621948, + "learning_rate": 1.623031107014893e-05, + "loss": 0.3662, + "step": 29970 + }, + { + "epoch": 11.317478293695734, + "grad_norm": 0.6397532820701599, + "learning_rate": 1.619983853028351e-05, + "loss": 0.3668, + "step": 29980 + }, + { + "epoch": 11.321253303133258, + "grad_norm": 0.6637176871299744, + "learning_rate": 1.6169389092509724e-05, + "loss": 0.3611, + "step": 29990 + }, + { + "epoch": 11.325028312570781, + "grad_norm": 0.6467320322990417, + "learning_rate": 1.6138962777639494e-05, + "loss": 0.3625, + "step": 30000 + }, + { + "epoch": 11.328803322008305, + "grad_norm": 0.6978408694267273, + "learning_rate": 1.610855960646891e-05, + "loss": 0.3677, + "step": 30010 + }, + { + "epoch": 11.332578331445829, + "grad_norm": 0.6771886348724365, + "learning_rate": 1.607817959977826e-05, + "loss": 0.3654, + "step": 30020 + }, + { + "epoch": 11.336353340883353, + "grad_norm": 0.691331684589386, + "learning_rate": 1.6047822778332005e-05, + "loss": 0.363, + "step": 30030 + }, + { + "epoch": 11.340128350320876, + "grad_norm": 0.6371560096740723, + "learning_rate": 1.6017489162878713e-05, + "loss": 0.3671, + "step": 30040 + }, + { + "epoch": 11.3439033597584, + "grad_norm": 0.6647712588310242, + "learning_rate": 1.5987178774151147e-05, + "loss": 0.3557, + "step": 30050 + }, + { + "epoch": 11.347678369195924, + "grad_norm": 0.6369796991348267, + "learning_rate": 1.5956891632866184e-05, + "loss": 0.3607, + "step": 30060 + }, + { + "epoch": 11.351453378633446, + "grad_norm": 0.6717673540115356, + "learning_rate": 1.592662775972476e-05, + "loss": 0.3559, + "step": 30070 + }, + { + "epoch": 11.35522838807097, + "grad_norm": 0.6065929532051086, + "learning_rate": 1.5896387175411986e-05, + "loss": 0.3572, + "step": 30080 + }, + { + "epoch": 11.359003397508493, + "grad_norm": 0.7193185091018677, + "learning_rate": 1.586616990059699e-05, + "loss": 0.3617, + "step": 30090 + }, + { + "epoch": 11.362778406946017, + "grad_norm": 0.7065017223358154, + "learning_rate": 1.5835975955933002e-05, + "loss": 0.3536, + "step": 30100 + }, + { + "epoch": 11.36655341638354, + "grad_norm": 0.7551389336585999, + "learning_rate": 1.5805805362057295e-05, + "loss": 0.3586, + "step": 30110 + }, + { + "epoch": 11.370328425821064, + "grad_norm": 0.6497236490249634, + "learning_rate": 1.57756581395912e-05, + "loss": 0.3634, + "step": 30120 + }, + { + "epoch": 11.374103435258588, + "grad_norm": 0.6430761814117432, + "learning_rate": 1.574553430914006e-05, + "loss": 0.3627, + "step": 30130 + }, + { + "epoch": 11.377878444696112, + "grad_norm": 0.6875909566879272, + "learning_rate": 1.5715433891293206e-05, + "loss": 0.3609, + "step": 30140 + }, + { + "epoch": 11.381653454133636, + "grad_norm": 0.61098313331604, + "learning_rate": 1.5685356906624016e-05, + "loss": 0.3585, + "step": 30150 + }, + { + "epoch": 11.38542846357116, + "grad_norm": 0.6427933573722839, + "learning_rate": 1.56553033756898e-05, + "loss": 0.3601, + "step": 30160 + }, + { + "epoch": 11.389203473008683, + "grad_norm": 0.662592351436615, + "learning_rate": 1.5625273319031873e-05, + "loss": 0.364, + "step": 30170 + }, + { + "epoch": 11.392978482446207, + "grad_norm": 0.7252653241157532, + "learning_rate": 1.5595266757175518e-05, + "loss": 0.3614, + "step": 30180 + }, + { + "epoch": 11.39675349188373, + "grad_norm": 0.6587193608283997, + "learning_rate": 1.55652837106299e-05, + "loss": 0.3618, + "step": 30190 + }, + { + "epoch": 11.400528501321253, + "grad_norm": 0.7090880274772644, + "learning_rate": 1.5535324199888164e-05, + "loss": 0.3572, + "step": 30200 + }, + { + "epoch": 11.404303510758776, + "grad_norm": 0.6730210781097412, + "learning_rate": 1.5505388245427355e-05, + "loss": 0.3621, + "step": 30210 + }, + { + "epoch": 11.4080785201963, + "grad_norm": 0.6814711689949036, + "learning_rate": 1.5475475867708418e-05, + "loss": 0.3647, + "step": 30220 + }, + { + "epoch": 11.411853529633824, + "grad_norm": 0.6800294518470764, + "learning_rate": 1.5445587087176195e-05, + "loss": 0.3635, + "step": 30230 + }, + { + "epoch": 11.415628539071347, + "grad_norm": 0.6638886332511902, + "learning_rate": 1.5415721924259346e-05, + "loss": 0.3634, + "step": 30240 + }, + { + "epoch": 11.419403548508871, + "grad_norm": 0.6355138421058655, + "learning_rate": 1.538588039937047e-05, + "loss": 0.36, + "step": 30250 + }, + { + "epoch": 11.423178557946395, + "grad_norm": 0.6450951099395752, + "learning_rate": 1.535606253290594e-05, + "loss": 0.3644, + "step": 30260 + }, + { + "epoch": 11.426953567383919, + "grad_norm": 0.640677273273468, + "learning_rate": 1.5326268345245993e-05, + "loss": 0.3594, + "step": 30270 + }, + { + "epoch": 11.430728576821442, + "grad_norm": 0.7016587257385254, + "learning_rate": 1.5296497856754693e-05, + "loss": 0.3645, + "step": 30280 + }, + { + "epoch": 11.434503586258966, + "grad_norm": 0.6003674864768982, + "learning_rate": 1.5266751087779858e-05, + "loss": 0.3645, + "step": 30290 + }, + { + "epoch": 11.43827859569649, + "grad_norm": 0.6319372057914734, + "learning_rate": 1.5237028058653142e-05, + "loss": 0.3662, + "step": 30300 + }, + { + "epoch": 11.442053605134014, + "grad_norm": 0.6442000865936279, + "learning_rate": 1.5207328789689946e-05, + "loss": 0.3606, + "step": 30310 + }, + { + "epoch": 11.445828614571536, + "grad_norm": 0.6055095791816711, + "learning_rate": 1.5177653301189454e-05, + "loss": 0.3652, + "step": 30320 + }, + { + "epoch": 11.44960362400906, + "grad_norm": 0.7226874232292175, + "learning_rate": 1.5148001613434587e-05, + "loss": 0.3553, + "step": 30330 + }, + { + "epoch": 11.453378633446583, + "grad_norm": 0.7227387428283691, + "learning_rate": 1.5118373746691983e-05, + "loss": 0.3628, + "step": 30340 + }, + { + "epoch": 11.457153642884107, + "grad_norm": 0.6881961226463318, + "learning_rate": 1.5088769721211992e-05, + "loss": 0.359, + "step": 30350 + }, + { + "epoch": 11.46092865232163, + "grad_norm": 0.7442403435707092, + "learning_rate": 1.505918955722871e-05, + "loss": 0.3632, + "step": 30360 + }, + { + "epoch": 11.464703661759154, + "grad_norm": 0.6394659876823425, + "learning_rate": 1.5029633274959886e-05, + "loss": 0.3559, + "step": 30370 + }, + { + "epoch": 11.468478671196678, + "grad_norm": 0.6404333114624023, + "learning_rate": 1.5000100894606989e-05, + "loss": 0.3605, + "step": 30380 + }, + { + "epoch": 11.472253680634202, + "grad_norm": 0.7091559171676636, + "learning_rate": 1.4970592436355079e-05, + "loss": 0.3546, + "step": 30390 + }, + { + "epoch": 11.476028690071725, + "grad_norm": 0.6807871460914612, + "learning_rate": 1.4941107920372938e-05, + "loss": 0.3703, + "step": 30400 + }, + { + "epoch": 11.47980369950925, + "grad_norm": 0.6545217633247375, + "learning_rate": 1.4911647366812947e-05, + "loss": 0.3574, + "step": 30410 + }, + { + "epoch": 11.483578708946773, + "grad_norm": 0.6775568723678589, + "learning_rate": 1.4882210795811114e-05, + "loss": 0.3653, + "step": 30420 + }, + { + "epoch": 11.487353718384297, + "grad_norm": 0.7269086837768555, + "learning_rate": 1.485279822748708e-05, + "loss": 0.3608, + "step": 30430 + }, + { + "epoch": 11.49112872782182, + "grad_norm": 0.7882838249206543, + "learning_rate": 1.4823409681944034e-05, + "loss": 0.3708, + "step": 30440 + }, + { + "epoch": 11.494903737259342, + "grad_norm": 0.6561617255210876, + "learning_rate": 1.4794045179268763e-05, + "loss": 0.3586, + "step": 30450 + }, + { + "epoch": 11.498678746696866, + "grad_norm": 0.6408365368843079, + "learning_rate": 1.4764704739531643e-05, + "loss": 0.3661, + "step": 30460 + }, + { + "epoch": 11.50245375613439, + "grad_norm": 0.6767753958702087, + "learning_rate": 1.473538838278658e-05, + "loss": 0.3576, + "step": 30470 + }, + { + "epoch": 11.506228765571914, + "grad_norm": 0.6685946583747864, + "learning_rate": 1.4706096129071045e-05, + "loss": 0.3657, + "step": 30480 + }, + { + "epoch": 11.510003775009437, + "grad_norm": 0.7614819407463074, + "learning_rate": 1.4676827998405996e-05, + "loss": 0.3657, + "step": 30490 + }, + { + "epoch": 11.513778784446961, + "grad_norm": 0.6719235181808472, + "learning_rate": 1.4647584010795928e-05, + "loss": 0.3587, + "step": 30500 + }, + { + "epoch": 11.517553793884485, + "grad_norm": 0.6133652925491333, + "learning_rate": 1.4618364186228834e-05, + "loss": 0.3656, + "step": 30510 + }, + { + "epoch": 11.521328803322008, + "grad_norm": 0.6366606950759888, + "learning_rate": 1.4589168544676207e-05, + "loss": 0.3592, + "step": 30520 + }, + { + "epoch": 11.525103812759532, + "grad_norm": 0.6164776086807251, + "learning_rate": 1.455999710609296e-05, + "loss": 0.3579, + "step": 30530 + }, + { + "epoch": 11.528878822197056, + "grad_norm": 0.5788912177085876, + "learning_rate": 1.453084989041753e-05, + "loss": 0.3612, + "step": 30540 + }, + { + "epoch": 11.53265383163458, + "grad_norm": 0.6258981227874756, + "learning_rate": 1.4501726917571728e-05, + "loss": 0.3599, + "step": 30550 + }, + { + "epoch": 11.536428841072103, + "grad_norm": 0.8489433526992798, + "learning_rate": 1.4472628207460848e-05, + "loss": 0.3561, + "step": 30560 + }, + { + "epoch": 11.540203850509627, + "grad_norm": 0.7995439767837524, + "learning_rate": 1.4443553779973584e-05, + "loss": 0.3608, + "step": 30570 + }, + { + "epoch": 11.54397885994715, + "grad_norm": 0.642866313457489, + "learning_rate": 1.4414503654982037e-05, + "loss": 0.3593, + "step": 30580 + }, + { + "epoch": 11.547753869384673, + "grad_norm": 0.688869059085846, + "learning_rate": 1.4385477852341706e-05, + "loss": 0.3612, + "step": 30590 + }, + { + "epoch": 11.551528878822197, + "grad_norm": 0.6555159091949463, + "learning_rate": 1.4356476391891421e-05, + "loss": 0.3577, + "step": 30600 + }, + { + "epoch": 11.55530388825972, + "grad_norm": 0.8551262021064758, + "learning_rate": 1.4327499293453423e-05, + "loss": 0.371, + "step": 30610 + }, + { + "epoch": 11.559078897697244, + "grad_norm": 0.6704701781272888, + "learning_rate": 1.4298546576833304e-05, + "loss": 0.3564, + "step": 30620 + }, + { + "epoch": 11.562853907134768, + "grad_norm": 0.7134976387023926, + "learning_rate": 1.426961826181994e-05, + "loss": 0.3619, + "step": 30630 + }, + { + "epoch": 11.566628916572292, + "grad_norm": 0.6610779166221619, + "learning_rate": 1.4240714368185592e-05, + "loss": 0.3642, + "step": 30640 + }, + { + "epoch": 11.570403926009815, + "grad_norm": 0.6555269956588745, + "learning_rate": 1.4211834915685767e-05, + "loss": 0.3627, + "step": 30650 + }, + { + "epoch": 11.574178935447339, + "grad_norm": 0.677355170249939, + "learning_rate": 1.4182979924059313e-05, + "loss": 0.3527, + "step": 30660 + }, + { + "epoch": 11.577953944884863, + "grad_norm": 0.6868191361427307, + "learning_rate": 1.415414941302835e-05, + "loss": 0.3604, + "step": 30670 + }, + { + "epoch": 11.581728954322386, + "grad_norm": 0.7704634666442871, + "learning_rate": 1.4125343402298257e-05, + "loss": 0.3605, + "step": 30680 + }, + { + "epoch": 11.58550396375991, + "grad_norm": 0.6135473251342773, + "learning_rate": 1.4096561911557688e-05, + "loss": 0.36, + "step": 30690 + }, + { + "epoch": 11.589278973197432, + "grad_norm": 1.2090810537338257, + "learning_rate": 1.4067804960478481e-05, + "loss": 0.3648, + "step": 30700 + }, + { + "epoch": 11.593053982634956, + "grad_norm": 0.6949913501739502, + "learning_rate": 1.4039072568715784e-05, + "loss": 0.3637, + "step": 30710 + }, + { + "epoch": 11.59682899207248, + "grad_norm": 0.6498703360557556, + "learning_rate": 1.401036475590788e-05, + "loss": 0.3651, + "step": 30720 + }, + { + "epoch": 11.600604001510003, + "grad_norm": 0.7282187342643738, + "learning_rate": 1.3981681541676301e-05, + "loss": 0.3629, + "step": 30730 + }, + { + "epoch": 11.604379010947527, + "grad_norm": 0.6524527668952942, + "learning_rate": 1.3953022945625772e-05, + "loss": 0.3581, + "step": 30740 + }, + { + "epoch": 11.60815402038505, + "grad_norm": 0.6055677533149719, + "learning_rate": 1.3924388987344138e-05, + "loss": 0.3575, + "step": 30750 + }, + { + "epoch": 11.611929029822575, + "grad_norm": 0.717749834060669, + "learning_rate": 1.3895779686402465e-05, + "loss": 0.3602, + "step": 30760 + }, + { + "epoch": 11.615704039260098, + "grad_norm": 0.7012677788734436, + "learning_rate": 1.3867195062354931e-05, + "loss": 0.3574, + "step": 30770 + }, + { + "epoch": 11.619479048697622, + "grad_norm": 0.7111396193504333, + "learning_rate": 1.3838635134738858e-05, + "loss": 0.3537, + "step": 30780 + }, + { + "epoch": 11.623254058135146, + "grad_norm": 0.7217731475830078, + "learning_rate": 1.3810099923074704e-05, + "loss": 0.3599, + "step": 30790 + }, + { + "epoch": 11.62702906757267, + "grad_norm": 0.609724760055542, + "learning_rate": 1.3781589446866006e-05, + "loss": 0.3743, + "step": 30800 + }, + { + "epoch": 11.630804077010193, + "grad_norm": 0.6935325860977173, + "learning_rate": 1.3753103725599387e-05, + "loss": 0.3577, + "step": 30810 + }, + { + "epoch": 11.634579086447715, + "grad_norm": 0.7039898037910461, + "learning_rate": 1.3724642778744579e-05, + "loss": 0.3615, + "step": 30820 + }, + { + "epoch": 11.638354095885239, + "grad_norm": 0.6811739802360535, + "learning_rate": 1.3696206625754376e-05, + "loss": 0.3589, + "step": 30830 + }, + { + "epoch": 11.642129105322763, + "grad_norm": 0.6018708348274231, + "learning_rate": 1.3667795286064639e-05, + "loss": 0.3626, + "step": 30840 + }, + { + "epoch": 11.645904114760286, + "grad_norm": 0.6382670998573303, + "learning_rate": 1.3639408779094215e-05, + "loss": 0.3649, + "step": 30850 + }, + { + "epoch": 11.64967912419781, + "grad_norm": 0.6437296867370605, + "learning_rate": 1.361104712424503e-05, + "loss": 0.364, + "step": 30860 + }, + { + "epoch": 11.653454133635334, + "grad_norm": 0.6563718318939209, + "learning_rate": 1.358271034090201e-05, + "loss": 0.3597, + "step": 30870 + }, + { + "epoch": 11.657229143072858, + "grad_norm": 0.6579415798187256, + "learning_rate": 1.3554398448433076e-05, + "loss": 0.3694, + "step": 30880 + }, + { + "epoch": 11.661004152510381, + "grad_norm": 0.7728137373924255, + "learning_rate": 1.3526111466189156e-05, + "loss": 0.3608, + "step": 30890 + }, + { + "epoch": 11.664779161947905, + "grad_norm": 0.649925708770752, + "learning_rate": 1.3497849413504121e-05, + "loss": 0.3571, + "step": 30900 + }, + { + "epoch": 11.668554171385429, + "grad_norm": 0.6554239392280579, + "learning_rate": 1.3469612309694802e-05, + "loss": 0.3653, + "step": 30910 + }, + { + "epoch": 11.672329180822953, + "grad_norm": 0.6548380851745605, + "learning_rate": 1.3441400174061008e-05, + "loss": 0.367, + "step": 30920 + }, + { + "epoch": 11.676104190260476, + "grad_norm": 0.7009982466697693, + "learning_rate": 1.3413213025885469e-05, + "loss": 0.362, + "step": 30930 + }, + { + "epoch": 11.679879199698, + "grad_norm": 0.6359342336654663, + "learning_rate": 1.3385050884433853e-05, + "loss": 0.3625, + "step": 30940 + }, + { + "epoch": 11.683654209135522, + "grad_norm": 0.6045933365821838, + "learning_rate": 1.3356913768954676e-05, + "loss": 0.3534, + "step": 30950 + }, + { + "epoch": 11.687429218573046, + "grad_norm": 0.6957305669784546, + "learning_rate": 1.3328801698679415e-05, + "loss": 0.3594, + "step": 30960 + }, + { + "epoch": 11.69120422801057, + "grad_norm": 1.19200599193573, + "learning_rate": 1.3300714692822413e-05, + "loss": 0.3681, + "step": 30970 + }, + { + "epoch": 11.694979237448093, + "grad_norm": 0.6715503931045532, + "learning_rate": 1.3272652770580857e-05, + "loss": 0.3561, + "step": 30980 + }, + { + "epoch": 11.698754246885617, + "grad_norm": 0.679587721824646, + "learning_rate": 1.3244615951134831e-05, + "loss": 0.3577, + "step": 30990 + }, + { + "epoch": 11.70252925632314, + "grad_norm": 0.624066948890686, + "learning_rate": 1.3216604253647219e-05, + "loss": 0.3624, + "step": 31000 + }, + { + "epoch": 11.706304265760664, + "grad_norm": 0.7792338728904724, + "learning_rate": 1.3188617697263738e-05, + "loss": 0.3621, + "step": 31010 + }, + { + "epoch": 11.710079275198188, + "grad_norm": 0.6704654693603516, + "learning_rate": 1.3160656301112956e-05, + "loss": 0.3506, + "step": 31020 + }, + { + "epoch": 11.713854284635712, + "grad_norm": 0.6014306545257568, + "learning_rate": 1.3132720084306211e-05, + "loss": 0.3666, + "step": 31030 + }, + { + "epoch": 11.717629294073236, + "grad_norm": 0.6405402421951294, + "learning_rate": 1.3104809065937668e-05, + "loss": 0.3574, + "step": 31040 + }, + { + "epoch": 11.72140430351076, + "grad_norm": 0.6713264584541321, + "learning_rate": 1.3076923265084212e-05, + "loss": 0.3619, + "step": 31050 + }, + { + "epoch": 11.725179312948283, + "grad_norm": 0.6099828481674194, + "learning_rate": 1.3049062700805543e-05, + "loss": 0.3555, + "step": 31060 + }, + { + "epoch": 11.728954322385807, + "grad_norm": 0.6921255588531494, + "learning_rate": 1.3021227392144092e-05, + "loss": 0.3596, + "step": 31070 + }, + { + "epoch": 11.732729331823329, + "grad_norm": 0.6106016635894775, + "learning_rate": 1.2993417358125037e-05, + "loss": 0.3541, + "step": 31080 + }, + { + "epoch": 11.736504341260853, + "grad_norm": 0.5807941555976868, + "learning_rate": 1.2965632617756257e-05, + "loss": 0.3613, + "step": 31090 + }, + { + "epoch": 11.740279350698376, + "grad_norm": 0.5835058093070984, + "learning_rate": 1.2937873190028387e-05, + "loss": 0.3577, + "step": 31100 + }, + { + "epoch": 11.7440543601359, + "grad_norm": 0.5961214303970337, + "learning_rate": 1.2910139093914697e-05, + "loss": 0.3567, + "step": 31110 + }, + { + "epoch": 11.747829369573424, + "grad_norm": 0.696262538433075, + "learning_rate": 1.2882430348371205e-05, + "loss": 0.3555, + "step": 31120 + }, + { + "epoch": 11.751604379010947, + "grad_norm": 0.6330527663230896, + "learning_rate": 1.2854746972336567e-05, + "loss": 0.3644, + "step": 31130 + }, + { + "epoch": 11.755379388448471, + "grad_norm": 0.6530078649520874, + "learning_rate": 1.2827088984732132e-05, + "loss": 0.3728, + "step": 31140 + }, + { + "epoch": 11.759154397885995, + "grad_norm": 0.596555769443512, + "learning_rate": 1.279945640446184e-05, + "loss": 0.3646, + "step": 31150 + }, + { + "epoch": 11.762929407323519, + "grad_norm": 0.6525673270225525, + "learning_rate": 1.2771849250412315e-05, + "loss": 0.3603, + "step": 31160 + }, + { + "epoch": 11.766704416761042, + "grad_norm": 0.716068685054779, + "learning_rate": 1.2744267541452792e-05, + "loss": 0.3579, + "step": 31170 + }, + { + "epoch": 11.770479426198566, + "grad_norm": 0.5986026525497437, + "learning_rate": 1.2716711296435112e-05, + "loss": 0.3612, + "step": 31180 + }, + { + "epoch": 11.77425443563609, + "grad_norm": 0.6966813206672668, + "learning_rate": 1.268918053419369e-05, + "loss": 0.3645, + "step": 31190 + }, + { + "epoch": 11.778029445073614, + "grad_norm": 0.6633387207984924, + "learning_rate": 1.2661675273545564e-05, + "loss": 0.3607, + "step": 31200 + }, + { + "epoch": 11.781804454511136, + "grad_norm": 0.7241559028625488, + "learning_rate": 1.26341955332903e-05, + "loss": 0.3615, + "step": 31210 + }, + { + "epoch": 11.78557946394866, + "grad_norm": 0.6759303212165833, + "learning_rate": 1.2606741332210048e-05, + "loss": 0.3611, + "step": 31220 + }, + { + "epoch": 11.789354473386183, + "grad_norm": 0.642245888710022, + "learning_rate": 1.2579312689069495e-05, + "loss": 0.3591, + "step": 31230 + }, + { + "epoch": 11.793129482823707, + "grad_norm": 0.6876919865608215, + "learning_rate": 1.2551909622615881e-05, + "loss": 0.3537, + "step": 31240 + }, + { + "epoch": 11.79690449226123, + "grad_norm": 0.6188132762908936, + "learning_rate": 1.252453215157891e-05, + "loss": 0.3683, + "step": 31250 + }, + { + "epoch": 11.800679501698754, + "grad_norm": 0.6617509722709656, + "learning_rate": 1.2497180294670841e-05, + "loss": 0.3575, + "step": 31260 + }, + { + "epoch": 11.804454511136278, + "grad_norm": 0.6628468036651611, + "learning_rate": 1.2469854070586423e-05, + "loss": 0.3568, + "step": 31270 + }, + { + "epoch": 11.808229520573802, + "grad_norm": 0.6406071186065674, + "learning_rate": 1.2442553498002845e-05, + "loss": 0.3705, + "step": 31280 + }, + { + "epoch": 11.812004530011325, + "grad_norm": 0.704318106174469, + "learning_rate": 1.241527859557981e-05, + "loss": 0.3602, + "step": 31290 + }, + { + "epoch": 11.81577953944885, + "grad_norm": 0.6852730512619019, + "learning_rate": 1.238802938195946e-05, + "loss": 0.3549, + "step": 31300 + }, + { + "epoch": 11.819554548886373, + "grad_norm": 0.7002149820327759, + "learning_rate": 1.2360805875766357e-05, + "loss": 0.3582, + "step": 31310 + }, + { + "epoch": 11.823329558323895, + "grad_norm": 0.6709305047988892, + "learning_rate": 1.2333608095607524e-05, + "loss": 0.3598, + "step": 31320 + }, + { + "epoch": 11.827104567761419, + "grad_norm": 0.6867847442626953, + "learning_rate": 1.230643606007238e-05, + "loss": 0.3582, + "step": 31330 + }, + { + "epoch": 11.830879577198942, + "grad_norm": 0.7019827961921692, + "learning_rate": 1.2279289787732778e-05, + "loss": 0.3513, + "step": 31340 + }, + { + "epoch": 11.834654586636466, + "grad_norm": 0.5999962687492371, + "learning_rate": 1.2252169297142906e-05, + "loss": 0.3575, + "step": 31350 + }, + { + "epoch": 11.83842959607399, + "grad_norm": 0.6360159516334534, + "learning_rate": 1.2225074606839392e-05, + "loss": 0.3611, + "step": 31360 + }, + { + "epoch": 11.842204605511514, + "grad_norm": 0.5966977477073669, + "learning_rate": 1.2198005735341173e-05, + "loss": 0.364, + "step": 31370 + }, + { + "epoch": 11.845979614949037, + "grad_norm": 0.6092012524604797, + "learning_rate": 1.2170962701149584e-05, + "loss": 0.3562, + "step": 31380 + }, + { + "epoch": 11.849754624386561, + "grad_norm": 0.7129442095756531, + "learning_rate": 1.2143945522748285e-05, + "loss": 0.3621, + "step": 31390 + }, + { + "epoch": 11.853529633824085, + "grad_norm": 0.6315842270851135, + "learning_rate": 1.2116954218603266e-05, + "loss": 0.3636, + "step": 31400 + }, + { + "epoch": 11.857304643261608, + "grad_norm": 0.7170038223266602, + "learning_rate": 1.2089988807162817e-05, + "loss": 0.3601, + "step": 31410 + }, + { + "epoch": 11.861079652699132, + "grad_norm": 0.5954352021217346, + "learning_rate": 1.2063049306857543e-05, + "loss": 0.3585, + "step": 31420 + }, + { + "epoch": 11.864854662136656, + "grad_norm": 0.6874489784240723, + "learning_rate": 1.2036135736100346e-05, + "loss": 0.3589, + "step": 31430 + }, + { + "epoch": 11.86862967157418, + "grad_norm": 0.6595900058746338, + "learning_rate": 1.200924811328641e-05, + "loss": 0.3526, + "step": 31440 + }, + { + "epoch": 11.872404681011702, + "grad_norm": 0.6593577861785889, + "learning_rate": 1.1982386456793143e-05, + "loss": 0.3656, + "step": 31450 + }, + { + "epoch": 11.876179690449225, + "grad_norm": 0.6922234296798706, + "learning_rate": 1.1955550784980257e-05, + "loss": 0.3593, + "step": 31460 + }, + { + "epoch": 11.87995469988675, + "grad_norm": 0.6410038471221924, + "learning_rate": 1.1928741116189663e-05, + "loss": 0.3544, + "step": 31470 + }, + { + "epoch": 11.883729709324273, + "grad_norm": 0.6991416811943054, + "learning_rate": 1.1901957468745529e-05, + "loss": 0.3648, + "step": 31480 + }, + { + "epoch": 11.887504718761797, + "grad_norm": 0.6780281662940979, + "learning_rate": 1.1875199860954217e-05, + "loss": 0.3662, + "step": 31490 + }, + { + "epoch": 11.89127972819932, + "grad_norm": 0.8736029863357544, + "learning_rate": 1.1848468311104316e-05, + "loss": 0.3568, + "step": 31500 + }, + { + "epoch": 11.895054737636844, + "grad_norm": 0.5960100889205933, + "learning_rate": 1.1821762837466566e-05, + "loss": 0.3573, + "step": 31510 + }, + { + "epoch": 11.898829747074368, + "grad_norm": 0.6374369263648987, + "learning_rate": 1.179508345829392e-05, + "loss": 0.3708, + "step": 31520 + }, + { + "epoch": 11.902604756511892, + "grad_norm": 0.7211053967475891, + "learning_rate": 1.1768430191821478e-05, + "loss": 0.3584, + "step": 31530 + }, + { + "epoch": 11.906379765949415, + "grad_norm": 0.8091753721237183, + "learning_rate": 1.174180305626651e-05, + "loss": 0.3602, + "step": 31540 + }, + { + "epoch": 11.910154775386939, + "grad_norm": 0.6501530408859253, + "learning_rate": 1.1715202069828401e-05, + "loss": 0.3593, + "step": 31550 + }, + { + "epoch": 11.913929784824463, + "grad_norm": 0.7378227710723877, + "learning_rate": 1.1688627250688661e-05, + "loss": 0.3638, + "step": 31560 + }, + { + "epoch": 11.917704794261986, + "grad_norm": 0.6729764342308044, + "learning_rate": 1.1662078617010935e-05, + "loss": 0.3598, + "step": 31570 + }, + { + "epoch": 11.921479803699508, + "grad_norm": 0.6561091542243958, + "learning_rate": 1.1635556186940972e-05, + "loss": 0.3649, + "step": 31580 + }, + { + "epoch": 11.925254813137032, + "grad_norm": 0.6765795946121216, + "learning_rate": 1.1609059978606602e-05, + "loss": 0.3709, + "step": 31590 + }, + { + "epoch": 11.929029822574556, + "grad_norm": 0.641923189163208, + "learning_rate": 1.1582590010117745e-05, + "loss": 0.3595, + "step": 31600 + }, + { + "epoch": 11.93280483201208, + "grad_norm": 0.6592920422554016, + "learning_rate": 1.1556146299566345e-05, + "loss": 0.3543, + "step": 31610 + }, + { + "epoch": 11.936579841449603, + "grad_norm": 0.6286612153053284, + "learning_rate": 1.1529728865026452e-05, + "loss": 0.3685, + "step": 31620 + }, + { + "epoch": 11.940354850887127, + "grad_norm": 1.297844648361206, + "learning_rate": 1.150333772455412e-05, + "loss": 0.3626, + "step": 31630 + }, + { + "epoch": 11.94412986032465, + "grad_norm": 0.6182207465171814, + "learning_rate": 1.1476972896187466e-05, + "loss": 0.355, + "step": 31640 + }, + { + "epoch": 11.947904869762175, + "grad_norm": 0.7070791125297546, + "learning_rate": 1.1450634397946592e-05, + "loss": 0.359, + "step": 31650 + }, + { + "epoch": 11.951679879199698, + "grad_norm": 0.6540588140487671, + "learning_rate": 1.142432224783359e-05, + "loss": 0.358, + "step": 31660 + }, + { + "epoch": 11.955454888637222, + "grad_norm": 0.6664989590644836, + "learning_rate": 1.1398036463832584e-05, + "loss": 0.3555, + "step": 31670 + }, + { + "epoch": 11.959229898074746, + "grad_norm": 0.6335260272026062, + "learning_rate": 1.1371777063909661e-05, + "loss": 0.3611, + "step": 31680 + }, + { + "epoch": 11.96300490751227, + "grad_norm": 0.614297091960907, + "learning_rate": 1.1345544066012874e-05, + "loss": 0.3628, + "step": 31690 + }, + { + "epoch": 11.966779916949793, + "grad_norm": 0.746405839920044, + "learning_rate": 1.131933748807224e-05, + "loss": 0.3556, + "step": 31700 + }, + { + "epoch": 11.970554926387315, + "grad_norm": 0.6299577951431274, + "learning_rate": 1.1293157347999683e-05, + "loss": 0.3573, + "step": 31710 + }, + { + "epoch": 11.974329935824839, + "grad_norm": 0.643743634223938, + "learning_rate": 1.1267003663689097e-05, + "loss": 0.3597, + "step": 31720 + }, + { + "epoch": 11.978104945262363, + "grad_norm": 0.7774217128753662, + "learning_rate": 1.1240876453016275e-05, + "loss": 0.3816, + "step": 31730 + }, + { + "epoch": 11.981879954699886, + "grad_norm": 0.6441808342933655, + "learning_rate": 1.1214775733838933e-05, + "loss": 0.3589, + "step": 31740 + }, + { + "epoch": 11.98565496413741, + "grad_norm": 0.6050808429718018, + "learning_rate": 1.118870152399666e-05, + "loss": 0.3601, + "step": 31750 + }, + { + "epoch": 11.989429973574934, + "grad_norm": 0.6663985252380371, + "learning_rate": 1.1162653841310905e-05, + "loss": 0.3602, + "step": 31760 + }, + { + "epoch": 11.993204983012458, + "grad_norm": 0.6749792695045471, + "learning_rate": 1.1136632703585043e-05, + "loss": 0.3576, + "step": 31770 + }, + { + "epoch": 11.996979992449981, + "grad_norm": 0.731533408164978, + "learning_rate": 1.111063812860426e-05, + "loss": 0.3565, + "step": 31780 + }, + { + "epoch": 12.000755001887505, + "grad_norm": 0.6132811903953552, + "learning_rate": 1.1084670134135616e-05, + "loss": 0.3547, + "step": 31790 + }, + { + "epoch": 12.004530011325029, + "grad_norm": 0.6406649947166443, + "learning_rate": 1.1058728737927998e-05, + "loss": 0.3593, + "step": 31800 + }, + { + "epoch": 12.008305020762553, + "grad_norm": 0.6679116487503052, + "learning_rate": 1.1032813957712086e-05, + "loss": 0.3594, + "step": 31810 + }, + { + "epoch": 12.012080030200076, + "grad_norm": 0.6831429600715637, + "learning_rate": 1.1006925811200385e-05, + "loss": 0.355, + "step": 31820 + }, + { + "epoch": 12.015855039637598, + "grad_norm": 0.6503385305404663, + "learning_rate": 1.0981064316087231e-05, + "loss": 0.3616, + "step": 31830 + }, + { + "epoch": 12.019630049075122, + "grad_norm": 0.6955201625823975, + "learning_rate": 1.0955229490048675e-05, + "loss": 0.3606, + "step": 31840 + }, + { + "epoch": 12.023405058512646, + "grad_norm": 0.6384215950965881, + "learning_rate": 1.0929421350742603e-05, + "loss": 0.3632, + "step": 31850 + }, + { + "epoch": 12.02718006795017, + "grad_norm": 0.6311360001564026, + "learning_rate": 1.090363991580861e-05, + "loss": 0.3628, + "step": 31860 + }, + { + "epoch": 12.030955077387693, + "grad_norm": 0.8894734382629395, + "learning_rate": 1.0877885202868083e-05, + "loss": 0.3631, + "step": 31870 + }, + { + "epoch": 12.034730086825217, + "grad_norm": 0.6360461115837097, + "learning_rate": 1.0852157229524118e-05, + "loss": 0.3779, + "step": 31880 + }, + { + "epoch": 12.03850509626274, + "grad_norm": 0.7593185901641846, + "learning_rate": 1.0826456013361542e-05, + "loss": 0.3516, + "step": 31890 + }, + { + "epoch": 12.042280105700264, + "grad_norm": 0.681585431098938, + "learning_rate": 1.080078157194691e-05, + "loss": 0.3682, + "step": 31900 + }, + { + "epoch": 12.046055115137788, + "grad_norm": 0.7083263397216797, + "learning_rate": 1.077513392282844e-05, + "loss": 0.3588, + "step": 31910 + }, + { + "epoch": 12.049830124575312, + "grad_norm": 0.7412534952163696, + "learning_rate": 1.0749513083536073e-05, + "loss": 0.3593, + "step": 31920 + }, + { + "epoch": 12.053605134012836, + "grad_norm": 0.6712055802345276, + "learning_rate": 1.072391907158139e-05, + "loss": 0.3601, + "step": 31930 + }, + { + "epoch": 12.05738014345036, + "grad_norm": 0.6988897323608398, + "learning_rate": 1.0698351904457676e-05, + "loss": 0.3553, + "step": 31940 + }, + { + "epoch": 12.061155152887883, + "grad_norm": 0.7310452461242676, + "learning_rate": 1.0672811599639849e-05, + "loss": 0.3612, + "step": 31950 + }, + { + "epoch": 12.064930162325405, + "grad_norm": 0.6422081589698792, + "learning_rate": 1.0647298174584453e-05, + "loss": 0.3648, + "step": 31960 + }, + { + "epoch": 12.068705171762929, + "grad_norm": 0.6467820405960083, + "learning_rate": 1.0621811646729673e-05, + "loss": 0.3573, + "step": 31970 + }, + { + "epoch": 12.072480181200453, + "grad_norm": 0.5941371917724609, + "learning_rate": 1.0596352033495328e-05, + "loss": 0.3646, + "step": 31980 + }, + { + "epoch": 12.076255190637976, + "grad_norm": 0.6265065670013428, + "learning_rate": 1.057091935228281e-05, + "loss": 0.3674, + "step": 31990 + }, + { + "epoch": 12.0800302000755, + "grad_norm": 0.7404178380966187, + "learning_rate": 1.0545513620475134e-05, + "loss": 0.3639, + "step": 32000 + }, + { + "epoch": 12.083805209513024, + "grad_norm": 0.6679591536521912, + "learning_rate": 1.0520134855436863e-05, + "loss": 0.3705, + "step": 32010 + }, + { + "epoch": 12.087580218950547, + "grad_norm": 0.6543688774108887, + "learning_rate": 1.049478307451413e-05, + "loss": 0.3586, + "step": 32020 + }, + { + "epoch": 12.091355228388071, + "grad_norm": 0.6841951012611389, + "learning_rate": 1.0469458295034656e-05, + "loss": 0.3585, + "step": 32030 + }, + { + "epoch": 12.095130237825595, + "grad_norm": 0.7727581858634949, + "learning_rate": 1.044416053430769e-05, + "loss": 0.3591, + "step": 32040 + }, + { + "epoch": 12.098905247263119, + "grad_norm": 0.6680077910423279, + "learning_rate": 1.0418889809624021e-05, + "loss": 0.3614, + "step": 32050 + }, + { + "epoch": 12.102680256700642, + "grad_norm": 0.5810276865959167, + "learning_rate": 1.0393646138255925e-05, + "loss": 0.3641, + "step": 32060 + }, + { + "epoch": 12.106455266138166, + "grad_norm": 0.6576191782951355, + "learning_rate": 1.036842953745723e-05, + "loss": 0.3583, + "step": 32070 + }, + { + "epoch": 12.110230275575688, + "grad_norm": 0.6342170834541321, + "learning_rate": 1.034324002446324e-05, + "loss": 0.355, + "step": 32080 + }, + { + "epoch": 12.114005285013212, + "grad_norm": 0.6298282146453857, + "learning_rate": 1.0318077616490751e-05, + "loss": 0.3631, + "step": 32090 + }, + { + "epoch": 12.117780294450736, + "grad_norm": 0.6232679486274719, + "learning_rate": 1.0292942330738054e-05, + "loss": 0.3566, + "step": 32100 + }, + { + "epoch": 12.12155530388826, + "grad_norm": 0.663240373134613, + "learning_rate": 1.026783418438485e-05, + "loss": 0.3568, + "step": 32110 + }, + { + "epoch": 12.125330313325783, + "grad_norm": 0.6801791191101074, + "learning_rate": 1.0242753194592319e-05, + "loss": 0.3593, + "step": 32120 + }, + { + "epoch": 12.129105322763307, + "grad_norm": 0.6788033246994019, + "learning_rate": 1.0217699378503092e-05, + "loss": 0.3492, + "step": 32130 + }, + { + "epoch": 12.13288033220083, + "grad_norm": 0.6879256963729858, + "learning_rate": 1.0192672753241218e-05, + "loss": 0.3575, + "step": 32140 + }, + { + "epoch": 12.136655341638354, + "grad_norm": 0.623914897441864, + "learning_rate": 1.0167673335912164e-05, + "loss": 0.3612, + "step": 32150 + }, + { + "epoch": 12.140430351075878, + "grad_norm": 0.6156293749809265, + "learning_rate": 1.0142701143602779e-05, + "loss": 0.3612, + "step": 32160 + }, + { + "epoch": 12.144205360513402, + "grad_norm": 0.6880207657814026, + "learning_rate": 1.0117756193381328e-05, + "loss": 0.3544, + "step": 32170 + }, + { + "epoch": 12.147980369950925, + "grad_norm": 0.6038033962249756, + "learning_rate": 1.0092838502297453e-05, + "loss": 0.3633, + "step": 32180 + }, + { + "epoch": 12.15175537938845, + "grad_norm": 0.6807112693786621, + "learning_rate": 1.0067948087382162e-05, + "loss": 0.3554, + "step": 32190 + }, + { + "epoch": 12.155530388825973, + "grad_norm": 0.6536206007003784, + "learning_rate": 1.0043084965647826e-05, + "loss": 0.3642, + "step": 32200 + }, + { + "epoch": 12.159305398263495, + "grad_norm": 0.685413122177124, + "learning_rate": 1.001824915408815e-05, + "loss": 0.4406, + "step": 32210 + }, + { + "epoch": 12.163080407701019, + "grad_norm": 0.6319077014923096, + "learning_rate": 9.99344066967815e-06, + "loss": 0.3569, + "step": 32220 + }, + { + "epoch": 12.166855417138542, + "grad_norm": 0.6181957125663757, + "learning_rate": 9.968659529374219e-06, + "loss": 0.3557, + "step": 32230 + }, + { + "epoch": 12.170630426576066, + "grad_norm": 0.6865522861480713, + "learning_rate": 9.943905750114025e-06, + "loss": 0.358, + "step": 32240 + }, + { + "epoch": 12.17440543601359, + "grad_norm": 0.7125778794288635, + "learning_rate": 9.919179348816559e-06, + "loss": 0.3643, + "step": 32250 + }, + { + "epoch": 12.178180445451114, + "grad_norm": 0.654378354549408, + "learning_rate": 9.894480342382056e-06, + "loss": 0.3583, + "step": 32260 + }, + { + "epoch": 12.181955454888637, + "grad_norm": 0.6028674244880676, + "learning_rate": 9.869808747692061e-06, + "loss": 0.3586, + "step": 32270 + }, + { + "epoch": 12.185730464326161, + "grad_norm": 0.6447907090187073, + "learning_rate": 9.84516458160939e-06, + "loss": 0.3629, + "step": 32280 + }, + { + "epoch": 12.189505473763685, + "grad_norm": 0.687196671962738, + "learning_rate": 9.820547860978097e-06, + "loss": 0.3575, + "step": 32290 + }, + { + "epoch": 12.193280483201208, + "grad_norm": 0.6808565258979797, + "learning_rate": 9.795958602623461e-06, + "loss": 0.3607, + "step": 32300 + }, + { + "epoch": 12.197055492638732, + "grad_norm": 0.6441514492034912, + "learning_rate": 9.77139682335203e-06, + "loss": 0.3544, + "step": 32310 + }, + { + "epoch": 12.200830502076256, + "grad_norm": 0.5570544600486755, + "learning_rate": 9.74686253995153e-06, + "loss": 0.3567, + "step": 32320 + }, + { + "epoch": 12.204605511513778, + "grad_norm": 0.639428436756134, + "learning_rate": 9.722355769190911e-06, + "loss": 0.3592, + "step": 32330 + }, + { + "epoch": 12.208380520951302, + "grad_norm": 0.636038064956665, + "learning_rate": 9.697876527820333e-06, + "loss": 0.3634, + "step": 32340 + }, + { + "epoch": 12.212155530388825, + "grad_norm": 0.6335992813110352, + "learning_rate": 9.673424832571132e-06, + "loss": 0.3599, + "step": 32350 + }, + { + "epoch": 12.21593053982635, + "grad_norm": 0.6807220578193665, + "learning_rate": 9.649000700155791e-06, + "loss": 0.3591, + "step": 32360 + }, + { + "epoch": 12.219705549263873, + "grad_norm": 0.7252523303031921, + "learning_rate": 9.624604147267984e-06, + "loss": 0.3608, + "step": 32370 + }, + { + "epoch": 12.223480558701397, + "grad_norm": 0.6216152906417847, + "learning_rate": 9.600235190582524e-06, + "loss": 0.358, + "step": 32380 + }, + { + "epoch": 12.22725556813892, + "grad_norm": 0.6576191186904907, + "learning_rate": 9.57589384675538e-06, + "loss": 0.3652, + "step": 32390 + }, + { + "epoch": 12.231030577576444, + "grad_norm": 0.626547634601593, + "learning_rate": 9.551580132423598e-06, + "loss": 0.3683, + "step": 32400 + }, + { + "epoch": 12.234805587013968, + "grad_norm": 0.6060603857040405, + "learning_rate": 9.527294064205405e-06, + "loss": 0.363, + "step": 32410 + }, + { + "epoch": 12.238580596451492, + "grad_norm": 0.6533463001251221, + "learning_rate": 9.503035658700072e-06, + "loss": 0.3593, + "step": 32420 + }, + { + "epoch": 12.242355605889015, + "grad_norm": 0.6849529147148132, + "learning_rate": 9.478804932488e-06, + "loss": 0.3578, + "step": 32430 + }, + { + "epoch": 12.246130615326539, + "grad_norm": 0.6339350938796997, + "learning_rate": 9.454601902130667e-06, + "loss": 0.3487, + "step": 32440 + }, + { + "epoch": 12.249905624764063, + "grad_norm": 0.7915511727333069, + "learning_rate": 9.430426584170626e-06, + "loss": 0.3567, + "step": 32450 + }, + { + "epoch": 12.253680634201585, + "grad_norm": 0.6535844206809998, + "learning_rate": 9.40627899513145e-06, + "loss": 0.3514, + "step": 32460 + }, + { + "epoch": 12.257455643639108, + "grad_norm": 0.6630949378013611, + "learning_rate": 9.38215915151781e-06, + "loss": 0.3596, + "step": 32470 + }, + { + "epoch": 12.261230653076632, + "grad_norm": 0.9704565405845642, + "learning_rate": 9.358067069815396e-06, + "loss": 0.3575, + "step": 32480 + }, + { + "epoch": 12.265005662514156, + "grad_norm": 0.6713346838951111, + "learning_rate": 9.334002766490901e-06, + "loss": 0.3604, + "step": 32490 + }, + { + "epoch": 12.26878067195168, + "grad_norm": 0.6596808433532715, + "learning_rate": 9.309966257992054e-06, + "loss": 0.3557, + "step": 32500 + }, + { + "epoch": 12.272555681389203, + "grad_norm": 0.6223334670066833, + "learning_rate": 9.2859575607476e-06, + "loss": 0.3603, + "step": 32510 + }, + { + "epoch": 12.276330690826727, + "grad_norm": 0.638768196105957, + "learning_rate": 9.26197669116723e-06, + "loss": 0.3637, + "step": 32520 + }, + { + "epoch": 12.28010570026425, + "grad_norm": 0.6910896897315979, + "learning_rate": 9.238023665641655e-06, + "loss": 0.352, + "step": 32530 + }, + { + "epoch": 12.283880709701775, + "grad_norm": 0.7031621932983398, + "learning_rate": 9.21409850054254e-06, + "loss": 0.3612, + "step": 32540 + }, + { + "epoch": 12.287655719139298, + "grad_norm": 0.6571682095527649, + "learning_rate": 9.190201212222526e-06, + "loss": 0.3605, + "step": 32550 + }, + { + "epoch": 12.291430728576822, + "grad_norm": 0.670776903629303, + "learning_rate": 9.166331817015155e-06, + "loss": 0.3516, + "step": 32560 + }, + { + "epoch": 12.295205738014346, + "grad_norm": 0.5901675820350647, + "learning_rate": 9.142490331234965e-06, + "loss": 0.358, + "step": 32570 + }, + { + "epoch": 12.298980747451868, + "grad_norm": 0.7036232948303223, + "learning_rate": 9.11867677117736e-06, + "loss": 0.3604, + "step": 32580 + }, + { + "epoch": 12.302755756889391, + "grad_norm": 0.6272095441818237, + "learning_rate": 9.09489115311869e-06, + "loss": 0.3574, + "step": 32590 + }, + { + "epoch": 12.306530766326915, + "grad_norm": 0.7180802226066589, + "learning_rate": 9.071133493316202e-06, + "loss": 0.3626, + "step": 32600 + }, + { + "epoch": 12.310305775764439, + "grad_norm": 0.6323314309120178, + "learning_rate": 9.047403808008053e-06, + "loss": 0.3604, + "step": 32610 + }, + { + "epoch": 12.314080785201963, + "grad_norm": 0.6044765114784241, + "learning_rate": 9.023702113413224e-06, + "loss": 0.3679, + "step": 32620 + }, + { + "epoch": 12.317855794639486, + "grad_norm": 0.7297821044921875, + "learning_rate": 9.00002842573162e-06, + "loss": 0.3595, + "step": 32630 + }, + { + "epoch": 12.32163080407701, + "grad_norm": 0.8278955221176147, + "learning_rate": 8.976382761143981e-06, + "loss": 0.3556, + "step": 32640 + }, + { + "epoch": 12.325405813514534, + "grad_norm": 0.6725118160247803, + "learning_rate": 8.952765135811909e-06, + "loss": 0.3635, + "step": 32650 + }, + { + "epoch": 12.329180822952058, + "grad_norm": 0.6396539807319641, + "learning_rate": 8.929175565877795e-06, + "loss": 0.3559, + "step": 32660 + }, + { + "epoch": 12.332955832389581, + "grad_norm": 1.543081521987915, + "learning_rate": 8.905614067464924e-06, + "loss": 0.3599, + "step": 32670 + }, + { + "epoch": 12.336730841827105, + "grad_norm": 0.6082138419151306, + "learning_rate": 8.882080656677333e-06, + "loss": 0.3559, + "step": 32680 + }, + { + "epoch": 12.340505851264629, + "grad_norm": 0.6150057911872864, + "learning_rate": 8.85857534959989e-06, + "loss": 0.3602, + "step": 32690 + }, + { + "epoch": 12.344280860702153, + "grad_norm": 0.8048917651176453, + "learning_rate": 8.835098162298256e-06, + "loss": 0.3605, + "step": 32700 + }, + { + "epoch": 12.348055870139675, + "grad_norm": 0.735430896282196, + "learning_rate": 8.811649110818875e-06, + "loss": 0.353, + "step": 32710 + }, + { + "epoch": 12.351830879577198, + "grad_norm": 0.6606729030609131, + "learning_rate": 8.78822821118893e-06, + "loss": 0.3521, + "step": 32720 + }, + { + "epoch": 12.355605889014722, + "grad_norm": 0.6922546029090881, + "learning_rate": 8.764835479416405e-06, + "loss": 0.359, + "step": 32730 + }, + { + "epoch": 12.359380898452246, + "grad_norm": 0.6572328805923462, + "learning_rate": 8.741470931490003e-06, + "loss": 0.3557, + "step": 32740 + }, + { + "epoch": 12.36315590788977, + "grad_norm": 0.6792584657669067, + "learning_rate": 8.718134583379189e-06, + "loss": 0.354, + "step": 32750 + }, + { + "epoch": 12.366930917327293, + "grad_norm": 0.6346094012260437, + "learning_rate": 8.69482645103412e-06, + "loss": 0.3549, + "step": 32760 + }, + { + "epoch": 12.370705926764817, + "grad_norm": 0.6417624950408936, + "learning_rate": 8.671546550385679e-06, + "loss": 0.3598, + "step": 32770 + }, + { + "epoch": 12.37448093620234, + "grad_norm": 0.7251724004745483, + "learning_rate": 8.648294897345466e-06, + "loss": 0.352, + "step": 32780 + }, + { + "epoch": 12.378255945639864, + "grad_norm": 0.628129243850708, + "learning_rate": 8.62507150780576e-06, + "loss": 0.3574, + "step": 32790 + }, + { + "epoch": 12.382030955077388, + "grad_norm": 0.6672044992446899, + "learning_rate": 8.601876397639542e-06, + "loss": 0.3662, + "step": 32800 + }, + { + "epoch": 12.385805964514912, + "grad_norm": 0.614966630935669, + "learning_rate": 8.578709582700456e-06, + "loss": 0.3499, + "step": 32810 + }, + { + "epoch": 12.389580973952436, + "grad_norm": 0.629789412021637, + "learning_rate": 8.555571078822783e-06, + "loss": 0.358, + "step": 32820 + }, + { + "epoch": 12.39335598338996, + "grad_norm": 0.6185657978057861, + "learning_rate": 8.532460901821477e-06, + "loss": 0.3626, + "step": 32830 + }, + { + "epoch": 12.397130992827481, + "grad_norm": 0.6272408366203308, + "learning_rate": 8.509379067492128e-06, + "loss": 0.3543, + "step": 32840 + }, + { + "epoch": 12.400906002265005, + "grad_norm": 0.7130616903305054, + "learning_rate": 8.486325591610967e-06, + "loss": 0.3542, + "step": 32850 + }, + { + "epoch": 12.404681011702529, + "grad_norm": 0.6526590585708618, + "learning_rate": 8.463300489934811e-06, + "loss": 0.3627, + "step": 32860 + }, + { + "epoch": 12.408456021140053, + "grad_norm": 0.6703090071678162, + "learning_rate": 8.440303778201093e-06, + "loss": 0.3571, + "step": 32870 + }, + { + "epoch": 12.412231030577576, + "grad_norm": 0.6281063556671143, + "learning_rate": 8.417335472127858e-06, + "loss": 0.3667, + "step": 32880 + }, + { + "epoch": 12.4160060400151, + "grad_norm": 0.6609539985656738, + "learning_rate": 8.394395587413722e-06, + "loss": 0.3607, + "step": 32890 + }, + { + "epoch": 12.419781049452624, + "grad_norm": 0.6523950099945068, + "learning_rate": 8.371484139737889e-06, + "loss": 0.3728, + "step": 32900 + }, + { + "epoch": 12.423556058890147, + "grad_norm": 0.6022361516952515, + "learning_rate": 8.348601144760127e-06, + "loss": 0.3586, + "step": 32910 + }, + { + "epoch": 12.427331068327671, + "grad_norm": 0.6702349185943604, + "learning_rate": 8.325746618120716e-06, + "loss": 0.3576, + "step": 32920 + }, + { + "epoch": 12.431106077765195, + "grad_norm": 0.6322970986366272, + "learning_rate": 8.302920575440531e-06, + "loss": 0.356, + "step": 32930 + }, + { + "epoch": 12.434881087202719, + "grad_norm": 0.6621061563491821, + "learning_rate": 8.280123032320953e-06, + "loss": 0.3598, + "step": 32940 + }, + { + "epoch": 12.438656096640242, + "grad_norm": 2.3764612674713135, + "learning_rate": 8.257354004343892e-06, + "loss": 0.3598, + "step": 32950 + }, + { + "epoch": 12.442431106077764, + "grad_norm": 0.6495422720909119, + "learning_rate": 8.234613507071754e-06, + "loss": 0.3675, + "step": 32960 + }, + { + "epoch": 12.446206115515288, + "grad_norm": 0.5902871489524841, + "learning_rate": 8.211901556047475e-06, + "loss": 0.3548, + "step": 32970 + }, + { + "epoch": 12.449981124952812, + "grad_norm": 0.597322404384613, + "learning_rate": 8.189218166794427e-06, + "loss": 0.35, + "step": 32980 + }, + { + "epoch": 12.453756134390336, + "grad_norm": 0.656929075717926, + "learning_rate": 8.166563354816509e-06, + "loss": 0.3605, + "step": 32990 + }, + { + "epoch": 12.45753114382786, + "grad_norm": 0.6605633497238159, + "learning_rate": 8.143937135598074e-06, + "loss": 0.3512, + "step": 33000 + }, + { + "epoch": 12.461306153265383, + "grad_norm": 0.6512786149978638, + "learning_rate": 8.121339524603943e-06, + "loss": 0.3524, + "step": 33010 + }, + { + "epoch": 12.465081162702907, + "grad_norm": 0.8367393016815186, + "learning_rate": 8.098770537279338e-06, + "loss": 0.3565, + "step": 33020 + }, + { + "epoch": 12.46885617214043, + "grad_norm": 0.602468729019165, + "learning_rate": 8.076230189049966e-06, + "loss": 0.3513, + "step": 33030 + }, + { + "epoch": 12.472631181577954, + "grad_norm": 0.7088872790336609, + "learning_rate": 8.05371849532196e-06, + "loss": 0.3584, + "step": 33040 + }, + { + "epoch": 12.476406191015478, + "grad_norm": 0.7384507060050964, + "learning_rate": 8.031235471481813e-06, + "loss": 0.3559, + "step": 33050 + }, + { + "epoch": 12.480181200453002, + "grad_norm": 0.714253842830658, + "learning_rate": 8.008781132896486e-06, + "loss": 0.3569, + "step": 33060 + }, + { + "epoch": 12.483956209890525, + "grad_norm": 0.6598066091537476, + "learning_rate": 7.986355494913312e-06, + "loss": 0.3623, + "step": 33070 + }, + { + "epoch": 12.487731219328047, + "grad_norm": 0.7399073839187622, + "learning_rate": 7.963958572859975e-06, + "loss": 0.3538, + "step": 33080 + }, + { + "epoch": 12.491506228765571, + "grad_norm": 0.6438301205635071, + "learning_rate": 7.941590382044583e-06, + "loss": 0.3642, + "step": 33090 + }, + { + "epoch": 12.495281238203095, + "grad_norm": 0.6220930814743042, + "learning_rate": 7.919250937755574e-06, + "loss": 0.3514, + "step": 33100 + }, + { + "epoch": 12.499056247640619, + "grad_norm": 0.5921124815940857, + "learning_rate": 7.896940255261764e-06, + "loss": 0.3619, + "step": 33110 + }, + { + "epoch": 12.502831257078142, + "grad_norm": 0.6638910174369812, + "learning_rate": 7.874658349812269e-06, + "loss": 0.3648, + "step": 33120 + }, + { + "epoch": 12.506606266515666, + "grad_norm": 0.6646683812141418, + "learning_rate": 7.85240523663659e-06, + "loss": 0.3603, + "step": 33130 + }, + { + "epoch": 12.51038127595319, + "grad_norm": 0.6426159739494324, + "learning_rate": 7.830180930944487e-06, + "loss": 0.3619, + "step": 33140 + }, + { + "epoch": 12.514156285390714, + "grad_norm": 0.6459843516349792, + "learning_rate": 7.80798544792608e-06, + "loss": 0.3581, + "step": 33150 + }, + { + "epoch": 12.517931294828237, + "grad_norm": 0.7059284448623657, + "learning_rate": 7.785818802751776e-06, + "loss": 0.3593, + "step": 33160 + }, + { + "epoch": 12.521706304265761, + "grad_norm": 0.5886619687080383, + "learning_rate": 7.763681010572277e-06, + "loss": 0.3711, + "step": 33170 + }, + { + "epoch": 12.525481313703285, + "grad_norm": 0.6966323852539062, + "learning_rate": 7.741572086518528e-06, + "loss": 0.3547, + "step": 33180 + }, + { + "epoch": 12.529256323140809, + "grad_norm": 0.6798071265220642, + "learning_rate": 7.719492045701782e-06, + "loss": 0.358, + "step": 33190 + }, + { + "epoch": 12.533031332578332, + "grad_norm": 0.7177457213401794, + "learning_rate": 7.697440903213549e-06, + "loss": 0.3652, + "step": 33200 + }, + { + "epoch": 12.536806342015854, + "grad_norm": 0.7038207054138184, + "learning_rate": 7.675418674125573e-06, + "loss": 0.3558, + "step": 33210 + }, + { + "epoch": 12.540581351453378, + "grad_norm": 0.7092708945274353, + "learning_rate": 7.65342537348983e-06, + "loss": 0.3569, + "step": 33220 + }, + { + "epoch": 12.544356360890902, + "grad_norm": 0.6563994884490967, + "learning_rate": 7.631461016338549e-06, + "loss": 0.3582, + "step": 33230 + }, + { + "epoch": 12.548131370328425, + "grad_norm": 0.6491038799285889, + "learning_rate": 7.609525617684138e-06, + "loss": 0.3556, + "step": 33240 + }, + { + "epoch": 12.55190637976595, + "grad_norm": 0.6468068957328796, + "learning_rate": 7.587619192519252e-06, + "loss": 0.3535, + "step": 33250 + }, + { + "epoch": 12.555681389203473, + "grad_norm": 0.6689091920852661, + "learning_rate": 7.565741755816725e-06, + "loss": 0.363, + "step": 33260 + }, + { + "epoch": 12.559456398640997, + "grad_norm": 0.6048831939697266, + "learning_rate": 7.543893322529588e-06, + "loss": 0.358, + "step": 33270 + }, + { + "epoch": 12.56323140807852, + "grad_norm": 0.6832661032676697, + "learning_rate": 7.522073907591015e-06, + "loss": 0.3602, + "step": 33280 + }, + { + "epoch": 12.567006417516044, + "grad_norm": 0.6341314911842346, + "learning_rate": 7.500283525914381e-06, + "loss": 0.358, + "step": 33290 + }, + { + "epoch": 12.570781426953568, + "grad_norm": 0.6825064420700073, + "learning_rate": 7.478522192393211e-06, + "loss": 0.3489, + "step": 33300 + }, + { + "epoch": 12.574556436391092, + "grad_norm": 0.5636059045791626, + "learning_rate": 7.456789921901181e-06, + "loss": 0.3566, + "step": 33310 + }, + { + "epoch": 12.578331445828615, + "grad_norm": 0.7039142847061157, + "learning_rate": 7.435086729292079e-06, + "loss": 0.357, + "step": 33320 + }, + { + "epoch": 12.582106455266139, + "grad_norm": 0.6117770075798035, + "learning_rate": 7.413412629399824e-06, + "loss": 0.3609, + "step": 33330 + }, + { + "epoch": 12.585881464703661, + "grad_norm": 0.6217195391654968, + "learning_rate": 7.391767637038466e-06, + "loss": 0.3557, + "step": 33340 + }, + { + "epoch": 12.589656474141185, + "grad_norm": 0.7603038549423218, + "learning_rate": 7.370151767002159e-06, + "loss": 0.3557, + "step": 33350 + }, + { + "epoch": 12.593431483578708, + "grad_norm": 0.660520613193512, + "learning_rate": 7.348565034065147e-06, + "loss": 0.3648, + "step": 33360 + }, + { + "epoch": 12.597206493016232, + "grad_norm": 0.6169815063476562, + "learning_rate": 7.327007452981765e-06, + "loss": 0.3573, + "step": 33370 + }, + { + "epoch": 12.600981502453756, + "grad_norm": 0.59542316198349, + "learning_rate": 7.305479038486396e-06, + "loss": 0.3576, + "step": 33380 + }, + { + "epoch": 12.60475651189128, + "grad_norm": 0.6083225011825562, + "learning_rate": 7.283979805293517e-06, + "loss": 0.353, + "step": 33390 + }, + { + "epoch": 12.608531521328803, + "grad_norm": 0.6213321089744568, + "learning_rate": 7.262509768097653e-06, + "loss": 0.3552, + "step": 33400 + }, + { + "epoch": 12.612306530766327, + "grad_norm": 0.6928298473358154, + "learning_rate": 7.241068941573381e-06, + "loss": 0.3615, + "step": 33410 + }, + { + "epoch": 12.61608154020385, + "grad_norm": 0.6654199957847595, + "learning_rate": 7.219657340375296e-06, + "loss": 0.3566, + "step": 33420 + }, + { + "epoch": 12.619856549641375, + "grad_norm": 0.6036897301673889, + "learning_rate": 7.198274979138014e-06, + "loss": 0.3679, + "step": 33430 + }, + { + "epoch": 12.623631559078898, + "grad_norm": 0.685053825378418, + "learning_rate": 7.176921872476189e-06, + "loss": 0.3523, + "step": 33440 + }, + { + "epoch": 12.627406568516422, + "grad_norm": 0.6758593320846558, + "learning_rate": 7.1555980349844674e-06, + "loss": 0.3555, + "step": 33450 + }, + { + "epoch": 12.631181577953946, + "grad_norm": 1.1203943490982056, + "learning_rate": 7.134303481237492e-06, + "loss": 0.3602, + "step": 33460 + }, + { + "epoch": 12.634956587391468, + "grad_norm": 0.6802670955657959, + "learning_rate": 7.113038225789892e-06, + "loss": 0.3648, + "step": 33470 + }, + { + "epoch": 12.638731596828991, + "grad_norm": 0.6041070222854614, + "learning_rate": 7.091802283176258e-06, + "loss": 0.3545, + "step": 33480 + }, + { + "epoch": 12.642506606266515, + "grad_norm": 0.6177239418029785, + "learning_rate": 7.070595667911162e-06, + "loss": 0.3596, + "step": 33490 + }, + { + "epoch": 12.646281615704039, + "grad_norm": 0.6666644215583801, + "learning_rate": 7.049418394489127e-06, + "loss": 0.3587, + "step": 33500 + }, + { + "epoch": 12.650056625141563, + "grad_norm": 0.7733400464057922, + "learning_rate": 7.028270477384636e-06, + "loss": 0.3585, + "step": 33510 + }, + { + "epoch": 12.653831634579086, + "grad_norm": 0.8025512099266052, + "learning_rate": 7.0071519310520695e-06, + "loss": 0.3509, + "step": 33520 + }, + { + "epoch": 12.65760664401661, + "grad_norm": 0.6963528990745544, + "learning_rate": 6.98606276992575e-06, + "loss": 0.3549, + "step": 33530 + }, + { + "epoch": 12.661381653454134, + "grad_norm": 0.705504298210144, + "learning_rate": 6.965003008419929e-06, + "loss": 0.3559, + "step": 33540 + }, + { + "epoch": 12.665156662891658, + "grad_norm": 0.7005383968353271, + "learning_rate": 6.943972660928755e-06, + "loss": 0.3532, + "step": 33550 + }, + { + "epoch": 12.668931672329181, + "grad_norm": 0.7059851884841919, + "learning_rate": 6.922971741826273e-06, + "loss": 0.3553, + "step": 33560 + }, + { + "epoch": 12.672706681766705, + "grad_norm": 0.6162865161895752, + "learning_rate": 6.902000265466424e-06, + "loss": 0.3571, + "step": 33570 + }, + { + "epoch": 12.676481691204227, + "grad_norm": 0.6487360596656799, + "learning_rate": 6.881058246182998e-06, + "loss": 0.36, + "step": 33580 + }, + { + "epoch": 12.68025670064175, + "grad_norm": 0.63123619556427, + "learning_rate": 6.860145698289666e-06, + "loss": 0.3617, + "step": 33590 + }, + { + "epoch": 12.684031710079275, + "grad_norm": 0.6987262964248657, + "learning_rate": 6.839262636079985e-06, + "loss": 0.3586, + "step": 33600 + }, + { + "epoch": 12.687806719516798, + "grad_norm": 0.6346328258514404, + "learning_rate": 6.818409073827292e-06, + "loss": 0.3589, + "step": 33610 + }, + { + "epoch": 12.691581728954322, + "grad_norm": 0.6731489896774292, + "learning_rate": 6.79758502578483e-06, + "loss": 0.361, + "step": 33620 + }, + { + "epoch": 12.695356738391846, + "grad_norm": 0.6909154653549194, + "learning_rate": 6.776790506185621e-06, + "loss": 0.3558, + "step": 33630 + }, + { + "epoch": 12.69913174782937, + "grad_norm": 0.6457541584968567, + "learning_rate": 6.756025529242533e-06, + "loss": 0.3564, + "step": 33640 + }, + { + "epoch": 12.702906757266893, + "grad_norm": 0.6108506321907043, + "learning_rate": 6.735290109148223e-06, + "loss": 0.3597, + "step": 33650 + }, + { + "epoch": 12.706681766704417, + "grad_norm": 0.699229896068573, + "learning_rate": 6.714584260075163e-06, + "loss": 0.3595, + "step": 33660 + }, + { + "epoch": 12.71045677614194, + "grad_norm": 0.6555954813957214, + "learning_rate": 6.693907996175619e-06, + "loss": 0.3443, + "step": 33670 + }, + { + "epoch": 12.714231785579464, + "grad_norm": 0.6641535758972168, + "learning_rate": 6.673261331581593e-06, + "loss": 0.3564, + "step": 33680 + }, + { + "epoch": 12.718006795016988, + "grad_norm": 0.6798548698425293, + "learning_rate": 6.652644280404912e-06, + "loss": 0.3535, + "step": 33690 + }, + { + "epoch": 12.721781804454512, + "grad_norm": 0.6146652698516846, + "learning_rate": 6.6320568567371175e-06, + "loss": 0.3541, + "step": 33700 + }, + { + "epoch": 12.725556813892034, + "grad_norm": 0.6196102499961853, + "learning_rate": 6.611499074649518e-06, + "loss": 0.3546, + "step": 33710 + }, + { + "epoch": 12.729331823329558, + "grad_norm": 0.7105995416641235, + "learning_rate": 6.590970948193187e-06, + "loss": 0.3663, + "step": 33720 + }, + { + "epoch": 12.733106832767081, + "grad_norm": 0.6891978979110718, + "learning_rate": 6.5704724913988825e-06, + "loss": 0.3672, + "step": 33730 + }, + { + "epoch": 12.736881842204605, + "grad_norm": 0.6553974151611328, + "learning_rate": 6.550003718277109e-06, + "loss": 0.356, + "step": 33740 + }, + { + "epoch": 12.740656851642129, + "grad_norm": 0.699963390827179, + "learning_rate": 6.529564642818087e-06, + "loss": 0.3572, + "step": 33750 + }, + { + "epoch": 12.744431861079653, + "grad_norm": 0.6198428869247437, + "learning_rate": 6.50915527899173e-06, + "loss": 0.3634, + "step": 33760 + }, + { + "epoch": 12.748206870517176, + "grad_norm": 0.6034395098686218, + "learning_rate": 6.4887756407476575e-06, + "loss": 0.3631, + "step": 33770 + }, + { + "epoch": 12.7519818799547, + "grad_norm": 0.6118060946464539, + "learning_rate": 6.468425742015149e-06, + "loss": 0.3553, + "step": 33780 + }, + { + "epoch": 12.755756889392224, + "grad_norm": 0.6215893626213074, + "learning_rate": 6.448105596703152e-06, + "loss": 0.3587, + "step": 33790 + }, + { + "epoch": 12.759531898829747, + "grad_norm": 0.6398429274559021, + "learning_rate": 6.427815218700317e-06, + "loss": 0.3582, + "step": 33800 + }, + { + "epoch": 12.763306908267271, + "grad_norm": 0.6768361330032349, + "learning_rate": 6.407554621874911e-06, + "loss": 0.3592, + "step": 33810 + }, + { + "epoch": 12.767081917704795, + "grad_norm": 0.6989614963531494, + "learning_rate": 6.3873238200748864e-06, + "loss": 0.3553, + "step": 33820 + }, + { + "epoch": 12.770856927142319, + "grad_norm": 0.7120910882949829, + "learning_rate": 6.367122827127775e-06, + "loss": 0.3587, + "step": 33830 + }, + { + "epoch": 12.77463193657984, + "grad_norm": 0.7099307179450989, + "learning_rate": 6.346951656840777e-06, + "loss": 0.3575, + "step": 33840 + }, + { + "epoch": 12.778406946017364, + "grad_norm": 0.6789335012435913, + "learning_rate": 6.326810323000698e-06, + "loss": 0.3565, + "step": 33850 + }, + { + "epoch": 12.782181955454888, + "grad_norm": 0.6464335322380066, + "learning_rate": 6.306698839373948e-06, + "loss": 0.3583, + "step": 33860 + }, + { + "epoch": 12.785956964892412, + "grad_norm": 0.7024797201156616, + "learning_rate": 6.286617219706547e-06, + "loss": 0.3623, + "step": 33870 + }, + { + "epoch": 12.789731974329936, + "grad_norm": 0.6883578896522522, + "learning_rate": 6.26656547772409e-06, + "loss": 0.3593, + "step": 33880 + }, + { + "epoch": 12.79350698376746, + "grad_norm": 0.656173050403595, + "learning_rate": 6.246543627131734e-06, + "loss": 0.3568, + "step": 33890 + }, + { + "epoch": 12.797281993204983, + "grad_norm": 0.593108594417572, + "learning_rate": 6.22655168161424e-06, + "loss": 0.3583, + "step": 33900 + }, + { + "epoch": 12.801057002642507, + "grad_norm": 0.7652211785316467, + "learning_rate": 6.20658965483591e-06, + "loss": 0.3587, + "step": 33910 + }, + { + "epoch": 12.80483201208003, + "grad_norm": 0.6598030924797058, + "learning_rate": 6.186657560440612e-06, + "loss": 0.3575, + "step": 33920 + }, + { + "epoch": 12.808607021517554, + "grad_norm": 0.6479567289352417, + "learning_rate": 6.166755412051734e-06, + "loss": 0.3528, + "step": 33930 + }, + { + "epoch": 12.812382030955078, + "grad_norm": 0.7282853722572327, + "learning_rate": 6.1468832232721985e-06, + "loss": 0.3615, + "step": 33940 + }, + { + "epoch": 12.816157040392602, + "grad_norm": 0.670918345451355, + "learning_rate": 6.127041007684476e-06, + "loss": 0.364, + "step": 33950 + }, + { + "epoch": 12.819932049830125, + "grad_norm": 0.6073424220085144, + "learning_rate": 6.107228778850527e-06, + "loss": 0.3621, + "step": 33960 + }, + { + "epoch": 12.823707059267647, + "grad_norm": 0.6667119264602661, + "learning_rate": 6.087446550311837e-06, + "loss": 0.3633, + "step": 33970 + }, + { + "epoch": 12.827482068705171, + "grad_norm": 0.6431853175163269, + "learning_rate": 6.067694335589353e-06, + "loss": 0.3569, + "step": 33980 + }, + { + "epoch": 12.831257078142695, + "grad_norm": 0.74583899974823, + "learning_rate": 6.047972148183528e-06, + "loss": 0.3595, + "step": 33990 + }, + { + "epoch": 12.835032087580219, + "grad_norm": 0.7069852948188782, + "learning_rate": 6.0282800015742936e-06, + "loss": 0.3536, + "step": 34000 + }, + { + "epoch": 12.838807097017742, + "grad_norm": 0.6726012825965881, + "learning_rate": 6.008617909221054e-06, + "loss": 0.3586, + "step": 34010 + }, + { + "epoch": 12.842582106455266, + "grad_norm": 0.704440712928772, + "learning_rate": 5.988985884562664e-06, + "loss": 0.357, + "step": 34020 + }, + { + "epoch": 12.84635711589279, + "grad_norm": 0.6290958523750305, + "learning_rate": 5.969383941017414e-06, + "loss": 0.3569, + "step": 34030 + }, + { + "epoch": 12.850132125330314, + "grad_norm": 0.6034786105155945, + "learning_rate": 5.949812091983059e-06, + "loss": 0.3541, + "step": 34040 + }, + { + "epoch": 12.853907134767837, + "grad_norm": 0.6244705319404602, + "learning_rate": 5.9302703508367695e-06, + "loss": 0.3569, + "step": 34050 + }, + { + "epoch": 12.857682144205361, + "grad_norm": 0.6468330025672913, + "learning_rate": 5.910758730935157e-06, + "loss": 0.3611, + "step": 34060 + }, + { + "epoch": 12.861457153642885, + "grad_norm": 0.6652206182479858, + "learning_rate": 5.8912772456142016e-06, + "loss": 0.3511, + "step": 34070 + }, + { + "epoch": 12.865232163080409, + "grad_norm": 0.6235213875770569, + "learning_rate": 5.871825908189338e-06, + "loss": 0.3538, + "step": 34080 + }, + { + "epoch": 12.86900717251793, + "grad_norm": 0.6621882319450378, + "learning_rate": 5.852404731955352e-06, + "loss": 0.3609, + "step": 34090 + }, + { + "epoch": 12.872782181955454, + "grad_norm": 0.6129846572875977, + "learning_rate": 5.833013730186443e-06, + "loss": 0.3619, + "step": 34100 + }, + { + "epoch": 12.876557191392978, + "grad_norm": 0.6356860399246216, + "learning_rate": 5.813652916136181e-06, + "loss": 0.3535, + "step": 34110 + }, + { + "epoch": 12.880332200830502, + "grad_norm": 0.6019640564918518, + "learning_rate": 5.794322303037508e-06, + "loss": 0.3631, + "step": 34120 + }, + { + "epoch": 12.884107210268025, + "grad_norm": 0.7462950348854065, + "learning_rate": 5.775021904102701e-06, + "loss": 0.3548, + "step": 34130 + }, + { + "epoch": 12.88788221970555, + "grad_norm": 0.8151988983154297, + "learning_rate": 5.755751732523401e-06, + "loss": 0.3544, + "step": 34140 + }, + { + "epoch": 12.891657229143073, + "grad_norm": 0.6986622214317322, + "learning_rate": 5.736511801470595e-06, + "loss": 0.3494, + "step": 34150 + }, + { + "epoch": 12.895432238580597, + "grad_norm": 0.6462591290473938, + "learning_rate": 5.717302124094603e-06, + "loss": 0.368, + "step": 34160 + }, + { + "epoch": 12.89920724801812, + "grad_norm": 0.6205435991287231, + "learning_rate": 5.6981227135250396e-06, + "loss": 0.3582, + "step": 34170 + }, + { + "epoch": 12.902982257455644, + "grad_norm": 4.8062543869018555, + "learning_rate": 5.678973582870867e-06, + "loss": 0.3628, + "step": 34180 + }, + { + "epoch": 12.906757266893168, + "grad_norm": 0.6735921502113342, + "learning_rate": 5.659854745220317e-06, + "loss": 0.3619, + "step": 34190 + }, + { + "epoch": 12.910532276330692, + "grad_norm": 0.5726591944694519, + "learning_rate": 5.640766213640947e-06, + "loss": 0.3549, + "step": 34200 + }, + { + "epoch": 12.914307285768214, + "grad_norm": 0.6301986575126648, + "learning_rate": 5.6217080011795934e-06, + "loss": 0.3588, + "step": 34210 + }, + { + "epoch": 12.918082295205737, + "grad_norm": 0.6712170243263245, + "learning_rate": 5.602680120862364e-06, + "loss": 0.3573, + "step": 34220 + }, + { + "epoch": 12.921857304643261, + "grad_norm": 0.7288546562194824, + "learning_rate": 5.583682585694616e-06, + "loss": 0.3583, + "step": 34230 + }, + { + "epoch": 12.925632314080785, + "grad_norm": 0.6534035801887512, + "learning_rate": 5.564715408661009e-06, + "loss": 0.3538, + "step": 34240 + }, + { + "epoch": 12.929407323518308, + "grad_norm": 0.6488263607025146, + "learning_rate": 5.5457786027254265e-06, + "loss": 0.3587, + "step": 34250 + }, + { + "epoch": 12.933182332955832, + "grad_norm": 0.6462862491607666, + "learning_rate": 5.526872180830977e-06, + "loss": 0.3626, + "step": 34260 + }, + { + "epoch": 12.936957342393356, + "grad_norm": 0.6145511269569397, + "learning_rate": 5.507996155900041e-06, + "loss": 0.3568, + "step": 34270 + }, + { + "epoch": 12.94073235183088, + "grad_norm": 0.6510082483291626, + "learning_rate": 5.489150540834203e-06, + "loss": 0.3537, + "step": 34280 + }, + { + "epoch": 12.944507361268403, + "grad_norm": 0.7098113298416138, + "learning_rate": 5.47033534851425e-06, + "loss": 0.3498, + "step": 34290 + }, + { + "epoch": 12.948282370705927, + "grad_norm": 0.6072222590446472, + "learning_rate": 5.451550591800198e-06, + "loss": 0.3544, + "step": 34300 + }, + { + "epoch": 12.95205738014345, + "grad_norm": 0.6960251331329346, + "learning_rate": 5.43279628353125e-06, + "loss": 0.3611, + "step": 34310 + }, + { + "epoch": 12.955832389580975, + "grad_norm": 0.6805595755577087, + "learning_rate": 5.4140724365258075e-06, + "loss": 0.3595, + "step": 34320 + }, + { + "epoch": 12.959607399018498, + "grad_norm": 0.651698648929596, + "learning_rate": 5.395379063581424e-06, + "loss": 0.3566, + "step": 34330 + }, + { + "epoch": 12.96338240845602, + "grad_norm": 0.6734188199043274, + "learning_rate": 5.376716177474866e-06, + "loss": 0.3475, + "step": 34340 + }, + { + "epoch": 12.967157417893544, + "grad_norm": 0.6215294599533081, + "learning_rate": 5.358083790962015e-06, + "loss": 0.3579, + "step": 34350 + }, + { + "epoch": 12.970932427331068, + "grad_norm": 0.7030113339424133, + "learning_rate": 5.33948191677795e-06, + "loss": 0.356, + "step": 34360 + }, + { + "epoch": 12.974707436768592, + "grad_norm": 0.6129488945007324, + "learning_rate": 5.320910567636866e-06, + "loss": 0.3583, + "step": 34370 + }, + { + "epoch": 12.978482446206115, + "grad_norm": 2.6713273525238037, + "learning_rate": 5.302369756232123e-06, + "loss": 0.3658, + "step": 34380 + }, + { + "epoch": 12.982257455643639, + "grad_norm": 0.6619277596473694, + "learning_rate": 5.283859495236165e-06, + "loss": 0.3596, + "step": 34390 + }, + { + "epoch": 12.986032465081163, + "grad_norm": 0.6460644006729126, + "learning_rate": 5.265379797300602e-06, + "loss": 0.352, + "step": 34400 + }, + { + "epoch": 12.989807474518686, + "grad_norm": 0.6166976690292358, + "learning_rate": 5.246930675056117e-06, + "loss": 0.364, + "step": 34410 + }, + { + "epoch": 12.99358248395621, + "grad_norm": 0.5918395519256592, + "learning_rate": 5.228512141112535e-06, + "loss": 0.3565, + "step": 34420 + }, + { + "epoch": 12.997357493393734, + "grad_norm": 3.736072063446045, + "learning_rate": 5.21012420805872e-06, + "loss": 0.4422, + "step": 34430 + }, + { + "epoch": 13.001132502831258, + "grad_norm": 0.6360698342323303, + "learning_rate": 5.191766888462679e-06, + "loss": 0.3608, + "step": 34440 + }, + { + "epoch": 13.004907512268781, + "grad_norm": 1.15024733543396, + "learning_rate": 5.173440194871437e-06, + "loss": 0.3568, + "step": 34450 + }, + { + "epoch": 13.008682521706305, + "grad_norm": 0.6446855068206787, + "learning_rate": 5.155144139811141e-06, + "loss": 0.3572, + "step": 34460 + }, + { + "epoch": 13.012457531143827, + "grad_norm": 0.6619588732719421, + "learning_rate": 5.13687873578696e-06, + "loss": 0.3568, + "step": 34470 + }, + { + "epoch": 13.01623254058135, + "grad_norm": 0.7428315281867981, + "learning_rate": 5.11864399528314e-06, + "loss": 0.3547, + "step": 34480 + }, + { + "epoch": 13.020007550018875, + "grad_norm": 0.6570870876312256, + "learning_rate": 5.1004399307629346e-06, + "loss": 0.3606, + "step": 34490 + }, + { + "epoch": 13.023782559456398, + "grad_norm": 0.660350501537323, + "learning_rate": 5.082266554668652e-06, + "loss": 0.3647, + "step": 34500 + }, + { + "epoch": 13.027557568893922, + "grad_norm": 0.8094748258590698, + "learning_rate": 5.06412387942164e-06, + "loss": 0.3574, + "step": 34510 + }, + { + "epoch": 13.031332578331446, + "grad_norm": 0.669248640537262, + "learning_rate": 5.046011917422239e-06, + "loss": 0.3524, + "step": 34520 + }, + { + "epoch": 13.03510758776897, + "grad_norm": 0.6466529369354248, + "learning_rate": 5.0279306810498026e-06, + "loss": 0.3646, + "step": 34530 + }, + { + "epoch": 13.038882597206493, + "grad_norm": 0.6580777168273926, + "learning_rate": 5.009880182662679e-06, + "loss": 0.3482, + "step": 34540 + }, + { + "epoch": 13.042657606644017, + "grad_norm": 1.0861129760742188, + "learning_rate": 4.991860434598217e-06, + "loss": 0.3527, + "step": 34550 + }, + { + "epoch": 13.04643261608154, + "grad_norm": 0.6849576830863953, + "learning_rate": 4.973871449172746e-06, + "loss": 0.3616, + "step": 34560 + }, + { + "epoch": 13.050207625519064, + "grad_norm": 0.6503666043281555, + "learning_rate": 4.9559132386815695e-06, + "loss": 0.3614, + "step": 34570 + }, + { + "epoch": 13.053982634956588, + "grad_norm": 0.6452770233154297, + "learning_rate": 4.93798581539896e-06, + "loss": 0.3598, + "step": 34580 + }, + { + "epoch": 13.05775764439411, + "grad_norm": 0.6467193961143494, + "learning_rate": 4.920089191578131e-06, + "loss": 0.36, + "step": 34590 + }, + { + "epoch": 13.061532653831634, + "grad_norm": 0.6155425906181335, + "learning_rate": 4.90222337945126e-06, + "loss": 0.3565, + "step": 34600 + }, + { + "epoch": 13.065307663269158, + "grad_norm": 0.687741219997406, + "learning_rate": 4.8843883912294605e-06, + "loss": 0.3541, + "step": 34610 + }, + { + "epoch": 13.069082672706681, + "grad_norm": 0.6692128777503967, + "learning_rate": 4.866584239102795e-06, + "loss": 0.3546, + "step": 34620 + }, + { + "epoch": 13.072857682144205, + "grad_norm": 1.0037407875061035, + "learning_rate": 4.8488109352402164e-06, + "loss": 0.3644, + "step": 34630 + }, + { + "epoch": 13.076632691581729, + "grad_norm": 0.6341879963874817, + "learning_rate": 4.831068491789603e-06, + "loss": 0.3578, + "step": 34640 + }, + { + "epoch": 13.080407701019253, + "grad_norm": 0.6491614580154419, + "learning_rate": 4.81335692087776e-06, + "loss": 0.3579, + "step": 34650 + }, + { + "epoch": 13.084182710456776, + "grad_norm": 0.6379111409187317, + "learning_rate": 4.795676234610374e-06, + "loss": 0.3643, + "step": 34660 + }, + { + "epoch": 13.0879577198943, + "grad_norm": 0.678970992565155, + "learning_rate": 4.778026445072037e-06, + "loss": 0.3638, + "step": 34670 + }, + { + "epoch": 13.091732729331824, + "grad_norm": 0.6357844471931458, + "learning_rate": 4.760407564326208e-06, + "loss": 0.3521, + "step": 34680 + }, + { + "epoch": 13.095507738769347, + "grad_norm": 0.5853003859519958, + "learning_rate": 4.742819604415222e-06, + "loss": 0.3604, + "step": 34690 + }, + { + "epoch": 13.099282748206871, + "grad_norm": 0.7209833264350891, + "learning_rate": 4.725262577360284e-06, + "loss": 0.3599, + "step": 34700 + }, + { + "epoch": 13.103057757644395, + "grad_norm": 0.6747405529022217, + "learning_rate": 4.707736495161458e-06, + "loss": 0.3538, + "step": 34710 + }, + { + "epoch": 13.106832767081917, + "grad_norm": 0.6171832084655762, + "learning_rate": 4.690241369797671e-06, + "loss": 0.3574, + "step": 34720 + }, + { + "epoch": 13.11060777651944, + "grad_norm": 0.9365268349647522, + "learning_rate": 4.672777213226659e-06, + "loss": 0.3562, + "step": 34730 + }, + { + "epoch": 13.114382785956964, + "grad_norm": 0.6216146945953369, + "learning_rate": 4.655344037385001e-06, + "loss": 0.3577, + "step": 34740 + }, + { + "epoch": 13.118157795394488, + "grad_norm": 0.6404061317443848, + "learning_rate": 4.6379418541881215e-06, + "loss": 0.3617, + "step": 34750 + }, + { + "epoch": 13.121932804832012, + "grad_norm": 0.638793408870697, + "learning_rate": 4.620570675530245e-06, + "loss": 0.3593, + "step": 34760 + }, + { + "epoch": 13.125707814269536, + "grad_norm": 0.6440224647521973, + "learning_rate": 4.603230513284407e-06, + "loss": 0.3637, + "step": 34770 + }, + { + "epoch": 13.12948282370706, + "grad_norm": 0.7139766216278076, + "learning_rate": 4.58592137930246e-06, + "loss": 0.3537, + "step": 34780 + }, + { + "epoch": 13.133257833144583, + "grad_norm": 0.752666711807251, + "learning_rate": 4.568643285415009e-06, + "loss": 0.3647, + "step": 34790 + }, + { + "epoch": 13.137032842582107, + "grad_norm": 0.7453562617301941, + "learning_rate": 4.551396243431483e-06, + "loss": 0.3593, + "step": 34800 + }, + { + "epoch": 13.14080785201963, + "grad_norm": 0.6269282698631287, + "learning_rate": 4.534180265140075e-06, + "loss": 0.3643, + "step": 34810 + }, + { + "epoch": 13.144582861457154, + "grad_norm": 0.7206990122795105, + "learning_rate": 4.5169953623077356e-06, + "loss": 0.3554, + "step": 34820 + }, + { + "epoch": 13.148357870894678, + "grad_norm": 0.6602156758308411, + "learning_rate": 4.499841546680189e-06, + "loss": 0.3563, + "step": 34830 + }, + { + "epoch": 13.1521328803322, + "grad_norm": 0.6244385838508606, + "learning_rate": 4.4827188299819025e-06, + "loss": 0.348, + "step": 34840 + }, + { + "epoch": 13.155907889769724, + "grad_norm": 0.6173793077468872, + "learning_rate": 4.465627223916086e-06, + "loss": 0.359, + "step": 34850 + }, + { + "epoch": 13.159682899207247, + "grad_norm": 0.6792297959327698, + "learning_rate": 4.448566740164706e-06, + "loss": 0.3597, + "step": 34860 + }, + { + "epoch": 13.163457908644771, + "grad_norm": 0.6277993321418762, + "learning_rate": 4.431537390388429e-06, + "loss": 0.3617, + "step": 34870 + }, + { + "epoch": 13.167232918082295, + "grad_norm": 0.6655922532081604, + "learning_rate": 4.414539186226674e-06, + "loss": 0.353, + "step": 34880 + }, + { + "epoch": 13.171007927519819, + "grad_norm": 0.670093834400177, + "learning_rate": 4.397572139297529e-06, + "loss": 0.3579, + "step": 34890 + }, + { + "epoch": 13.174782936957342, + "grad_norm": 0.6460064649581909, + "learning_rate": 4.380636261197829e-06, + "loss": 0.3587, + "step": 34900 + }, + { + "epoch": 13.178557946394866, + "grad_norm": 0.5938677191734314, + "learning_rate": 4.36373156350307e-06, + "loss": 0.3583, + "step": 34910 + }, + { + "epoch": 13.18233295583239, + "grad_norm": 0.6573899388313293, + "learning_rate": 4.3468580577674665e-06, + "loss": 0.3568, + "step": 34920 + }, + { + "epoch": 13.186107965269914, + "grad_norm": 0.6724887490272522, + "learning_rate": 4.330015755523908e-06, + "loss": 0.3636, + "step": 34930 + }, + { + "epoch": 13.189882974707437, + "grad_norm": 0.8144491910934448, + "learning_rate": 4.313204668283926e-06, + "loss": 0.3654, + "step": 34940 + }, + { + "epoch": 13.193657984144961, + "grad_norm": 0.5686742663383484, + "learning_rate": 4.296424807537752e-06, + "loss": 0.3628, + "step": 34950 + }, + { + "epoch": 13.197432993582485, + "grad_norm": 0.9146296977996826, + "learning_rate": 4.279676184754267e-06, + "loss": 0.3596, + "step": 34960 + }, + { + "epoch": 13.201208003020007, + "grad_norm": 0.639041006565094, + "learning_rate": 4.262958811380985e-06, + "loss": 0.3575, + "step": 34970 + }, + { + "epoch": 13.20498301245753, + "grad_norm": 0.7431603074073792, + "learning_rate": 4.246272698844095e-06, + "loss": 0.3555, + "step": 34980 + }, + { + "epoch": 13.208758021895054, + "grad_norm": 0.6070488691329956, + "learning_rate": 4.229617858548374e-06, + "loss": 0.3574, + "step": 34990 + }, + { + "epoch": 13.212533031332578, + "grad_norm": 1.7812660932540894, + "learning_rate": 4.212994301877271e-06, + "loss": 0.3628, + "step": 35000 + }, + { + "epoch": 13.216308040770102, + "grad_norm": 0.763363242149353, + "learning_rate": 4.196402040192804e-06, + "loss": 0.3554, + "step": 35010 + }, + { + "epoch": 13.220083050207625, + "grad_norm": 1.0130776166915894, + "learning_rate": 4.179841084835639e-06, + "loss": 0.3557, + "step": 35020 + }, + { + "epoch": 13.22385805964515, + "grad_norm": 0.64314204454422, + "learning_rate": 4.163311447125046e-06, + "loss": 0.3546, + "step": 35030 + }, + { + "epoch": 13.227633069082673, + "grad_norm": 0.6824278831481934, + "learning_rate": 4.146813138358851e-06, + "loss": 0.3567, + "step": 35040 + }, + { + "epoch": 13.231408078520197, + "grad_norm": 0.6351913213729858, + "learning_rate": 4.130346169813509e-06, + "loss": 0.3514, + "step": 35050 + }, + { + "epoch": 13.23518308795772, + "grad_norm": 0.6978806853294373, + "learning_rate": 4.113910552744038e-06, + "loss": 0.3568, + "step": 35060 + }, + { + "epoch": 13.238958097395244, + "grad_norm": 0.6495206356048584, + "learning_rate": 4.097506298384024e-06, + "loss": 0.3525, + "step": 35070 + }, + { + "epoch": 13.242733106832768, + "grad_norm": 0.6304375529289246, + "learning_rate": 4.081133417945626e-06, + "loss": 0.3564, + "step": 35080 + }, + { + "epoch": 13.246508116270292, + "grad_norm": 0.6751596927642822, + "learning_rate": 4.0647919226195484e-06, + "loss": 0.3574, + "step": 35090 + }, + { + "epoch": 13.250283125707814, + "grad_norm": 0.6097880005836487, + "learning_rate": 4.04848182357504e-06, + "loss": 0.3571, + "step": 35100 + }, + { + "epoch": 13.254058135145337, + "grad_norm": 0.6437689661979675, + "learning_rate": 4.032203131959911e-06, + "loss": 0.3664, + "step": 35110 + }, + { + "epoch": 13.257833144582861, + "grad_norm": 0.6554880142211914, + "learning_rate": 4.015955858900483e-06, + "loss": 0.3576, + "step": 35120 + }, + { + "epoch": 13.261608154020385, + "grad_norm": 0.6221573352813721, + "learning_rate": 3.999740015501641e-06, + "loss": 0.3552, + "step": 35130 + }, + { + "epoch": 13.265383163457908, + "grad_norm": 0.8590520024299622, + "learning_rate": 3.9835556128467236e-06, + "loss": 0.3618, + "step": 35140 + }, + { + "epoch": 13.269158172895432, + "grad_norm": 0.6513279676437378, + "learning_rate": 3.967402661997644e-06, + "loss": 0.3588, + "step": 35150 + }, + { + "epoch": 13.272933182332956, + "grad_norm": 0.7151180505752563, + "learning_rate": 3.951281173994781e-06, + "loss": 0.3559, + "step": 35160 + }, + { + "epoch": 13.27670819177048, + "grad_norm": 0.6879009008407593, + "learning_rate": 3.935191159857016e-06, + "loss": 0.3538, + "step": 35170 + }, + { + "epoch": 13.280483201208003, + "grad_norm": 0.6725070476531982, + "learning_rate": 3.919132630581746e-06, + "loss": 0.364, + "step": 35180 + }, + { + "epoch": 13.284258210645527, + "grad_norm": 0.6446594595909119, + "learning_rate": 3.903105597144802e-06, + "loss": 0.3523, + "step": 35190 + }, + { + "epoch": 13.28803322008305, + "grad_norm": 0.6886351704597473, + "learning_rate": 3.887110070500505e-06, + "loss": 0.3607, + "step": 35200 + }, + { + "epoch": 13.291808229520575, + "grad_norm": 0.6238877177238464, + "learning_rate": 3.871146061581654e-06, + "loss": 0.3507, + "step": 35210 + }, + { + "epoch": 13.295583238958097, + "grad_norm": 0.640650749206543, + "learning_rate": 3.855213581299505e-06, + "loss": 0.3612, + "step": 35220 + }, + { + "epoch": 13.29935824839562, + "grad_norm": 0.665745198726654, + "learning_rate": 3.839312640543763e-06, + "loss": 0.359, + "step": 35230 + }, + { + "epoch": 13.303133257833144, + "grad_norm": 0.6206352114677429, + "learning_rate": 3.823443250182551e-06, + "loss": 0.3547, + "step": 35240 + }, + { + "epoch": 13.306908267270668, + "grad_norm": 1.118415117263794, + "learning_rate": 3.80760542106246e-06, + "loss": 0.3534, + "step": 35250 + }, + { + "epoch": 13.310683276708192, + "grad_norm": 0.6065353751182556, + "learning_rate": 3.7917991640084972e-06, + "loss": 0.3532, + "step": 35260 + }, + { + "epoch": 13.314458286145715, + "grad_norm": 0.6253830194473267, + "learning_rate": 3.7760244898240925e-06, + "loss": 0.4729, + "step": 35270 + }, + { + "epoch": 13.318233295583239, + "grad_norm": 0.5790948271751404, + "learning_rate": 3.760281409291078e-06, + "loss": 0.3582, + "step": 35280 + }, + { + "epoch": 13.322008305020763, + "grad_norm": 0.6971233487129211, + "learning_rate": 3.7445699331697113e-06, + "loss": 0.3591, + "step": 35290 + }, + { + "epoch": 13.325783314458286, + "grad_norm": 0.7019782066345215, + "learning_rate": 3.7288900721986262e-06, + "loss": 0.3583, + "step": 35300 + }, + { + "epoch": 13.32955832389581, + "grad_norm": 0.682455837726593, + "learning_rate": 3.7132418370948663e-06, + "loss": 0.3568, + "step": 35310 + }, + { + "epoch": 13.333333333333334, + "grad_norm": 0.6666685938835144, + "learning_rate": 3.6976252385538535e-06, + "loss": 0.358, + "step": 35320 + }, + { + "epoch": 13.337108342770858, + "grad_norm": 0.634299099445343, + "learning_rate": 3.6820402872493832e-06, + "loss": 0.3542, + "step": 35330 + }, + { + "epoch": 13.34088335220838, + "grad_norm": 0.6098415851593018, + "learning_rate": 3.6664869938336343e-06, + "loss": 0.3598, + "step": 35340 + }, + { + "epoch": 13.344658361645903, + "grad_norm": 0.6019866466522217, + "learning_rate": 3.650965368937115e-06, + "loss": 0.3575, + "step": 35350 + }, + { + "epoch": 13.348433371083427, + "grad_norm": 0.6554697155952454, + "learning_rate": 3.6354754231687214e-06, + "loss": 0.3604, + "step": 35360 + }, + { + "epoch": 13.35220838052095, + "grad_norm": 1.0949931144714355, + "learning_rate": 3.620017167115697e-06, + "loss": 0.3586, + "step": 35370 + }, + { + "epoch": 13.355983389958475, + "grad_norm": 0.663237988948822, + "learning_rate": 3.604590611343589e-06, + "loss": 0.3582, + "step": 35380 + }, + { + "epoch": 13.359758399395998, + "grad_norm": 0.6626077890396118, + "learning_rate": 3.589195766396325e-06, + "loss": 0.3581, + "step": 35390 + }, + { + "epoch": 13.363533408833522, + "grad_norm": 0.5932807326316833, + "learning_rate": 3.573832642796121e-06, + "loss": 0.3545, + "step": 35400 + }, + { + "epoch": 13.367308418271046, + "grad_norm": 0.6529201865196228, + "learning_rate": 3.5585012510435336e-06, + "loss": 0.3567, + "step": 35410 + }, + { + "epoch": 13.37108342770857, + "grad_norm": 0.6512988805770874, + "learning_rate": 3.5432016016174184e-06, + "loss": 0.3592, + "step": 35420 + }, + { + "epoch": 13.374858437146093, + "grad_norm": 0.6883192658424377, + "learning_rate": 3.5279337049749506e-06, + "loss": 0.3495, + "step": 35430 + }, + { + "epoch": 13.378633446583617, + "grad_norm": 0.625953733921051, + "learning_rate": 3.5126975715515985e-06, + "loss": 0.3524, + "step": 35440 + }, + { + "epoch": 13.38240845602114, + "grad_norm": 0.7120693325996399, + "learning_rate": 3.4974932117610994e-06, + "loss": 0.3546, + "step": 35450 + }, + { + "epoch": 13.386183465458664, + "grad_norm": 0.6227659583091736, + "learning_rate": 3.4823206359955065e-06, + "loss": 0.3585, + "step": 35460 + }, + { + "epoch": 13.389958474896186, + "grad_norm": 0.6581076979637146, + "learning_rate": 3.4671798546251133e-06, + "loss": 0.3584, + "step": 35470 + }, + { + "epoch": 13.39373348433371, + "grad_norm": 0.6351528167724609, + "learning_rate": 3.4520708779985088e-06, + "loss": 0.3486, + "step": 35480 + }, + { + "epoch": 13.397508493771234, + "grad_norm": 0.6142309308052063, + "learning_rate": 3.4369937164425493e-06, + "loss": 0.3578, + "step": 35490 + }, + { + "epoch": 13.401283503208758, + "grad_norm": 0.6407093405723572, + "learning_rate": 3.421948380262313e-06, + "loss": 0.3559, + "step": 35500 + }, + { + "epoch": 13.405058512646281, + "grad_norm": 0.6450949311256409, + "learning_rate": 3.406934879741158e-06, + "loss": 0.3575, + "step": 35510 + }, + { + "epoch": 13.408833522083805, + "grad_norm": 0.7444202899932861, + "learning_rate": 3.391953225140665e-06, + "loss": 0.3688, + "step": 35520 + }, + { + "epoch": 13.412608531521329, + "grad_norm": 0.706688642501831, + "learning_rate": 3.3770034267006533e-06, + "loss": 0.3587, + "step": 35530 + }, + { + "epoch": 13.416383540958853, + "grad_norm": 0.6597870588302612, + "learning_rate": 3.362085494639178e-06, + "loss": 0.3541, + "step": 35540 + }, + { + "epoch": 13.420158550396376, + "grad_norm": 0.5902731418609619, + "learning_rate": 3.3471994391524986e-06, + "loss": 0.3547, + "step": 35550 + }, + { + "epoch": 13.4239335598339, + "grad_norm": 0.6152647733688354, + "learning_rate": 3.3323452704150825e-06, + "loss": 0.3527, + "step": 35560 + }, + { + "epoch": 13.427708569271424, + "grad_norm": 0.6410410404205322, + "learning_rate": 3.3175229985796253e-06, + "loss": 0.3597, + "step": 35570 + }, + { + "epoch": 13.431483578708947, + "grad_norm": 0.6362301111221313, + "learning_rate": 3.3027326337770115e-06, + "loss": 0.3563, + "step": 35580 + }, + { + "epoch": 13.435258588146471, + "grad_norm": 0.6383989453315735, + "learning_rate": 3.2879741861163216e-06, + "loss": 0.3579, + "step": 35590 + }, + { + "epoch": 13.439033597583993, + "grad_norm": 0.6493092179298401, + "learning_rate": 3.2732476656847976e-06, + "loss": 0.3608, + "step": 35600 + }, + { + "epoch": 13.442808607021517, + "grad_norm": 0.6834067106246948, + "learning_rate": 3.2585530825478883e-06, + "loss": 0.3529, + "step": 35610 + }, + { + "epoch": 13.44658361645904, + "grad_norm": 0.7077740430831909, + "learning_rate": 3.243890446749204e-06, + "loss": 0.3575, + "step": 35620 + }, + { + "epoch": 13.450358625896564, + "grad_norm": 0.6362556219100952, + "learning_rate": 3.2292597683105107e-06, + "loss": 0.3594, + "step": 35630 + }, + { + "epoch": 13.454133635334088, + "grad_norm": 0.6774179339408875, + "learning_rate": 3.214661057231755e-06, + "loss": 0.3668, + "step": 35640 + }, + { + "epoch": 13.457908644771612, + "grad_norm": 0.7050675749778748, + "learning_rate": 3.2000943234910097e-06, + "loss": 0.3517, + "step": 35650 + }, + { + "epoch": 13.461683654209136, + "grad_norm": 0.8577159643173218, + "learning_rate": 3.1855595770444835e-06, + "loss": 0.3551, + "step": 35660 + }, + { + "epoch": 13.46545866364666, + "grad_norm": 0.6533629298210144, + "learning_rate": 3.1710568278265574e-06, + "loss": 0.347, + "step": 35670 + }, + { + "epoch": 13.469233673084183, + "grad_norm": 0.63999342918396, + "learning_rate": 3.156586085749719e-06, + "loss": 0.3484, + "step": 35680 + }, + { + "epoch": 13.473008682521707, + "grad_norm": 0.6630726456642151, + "learning_rate": 3.1421473607045893e-06, + "loss": 0.3556, + "step": 35690 + }, + { + "epoch": 13.47678369195923, + "grad_norm": 1.4107182025909424, + "learning_rate": 3.127740662559886e-06, + "loss": 0.355, + "step": 35700 + }, + { + "epoch": 13.480558701396754, + "grad_norm": 0.7354855537414551, + "learning_rate": 3.1133660011624655e-06, + "loss": 0.3602, + "step": 35710 + }, + { + "epoch": 13.484333710834276, + "grad_norm": 0.6842527389526367, + "learning_rate": 3.099023386337263e-06, + "loss": 0.3602, + "step": 35720 + }, + { + "epoch": 13.4881087202718, + "grad_norm": 0.6937289237976074, + "learning_rate": 3.0847128278873317e-06, + "loss": 0.3585, + "step": 35730 + }, + { + "epoch": 13.491883729709324, + "grad_norm": 0.7109084725379944, + "learning_rate": 3.0704343355938027e-06, + "loss": 0.3568, + "step": 35740 + }, + { + "epoch": 13.495658739146847, + "grad_norm": 0.7426275610923767, + "learning_rate": 3.056187919215886e-06, + "loss": 0.3545, + "step": 35750 + }, + { + "epoch": 13.499433748584371, + "grad_norm": 0.7047889828681946, + "learning_rate": 3.041973588490865e-06, + "loss": 0.3583, + "step": 35760 + }, + { + "epoch": 13.503208758021895, + "grad_norm": 1.050673007965088, + "learning_rate": 3.0277913531341128e-06, + "loss": 0.3619, + "step": 35770 + }, + { + "epoch": 13.506983767459419, + "grad_norm": 0.6489849090576172, + "learning_rate": 3.0136412228390476e-06, + "loss": 0.3544, + "step": 35780 + }, + { + "epoch": 13.510758776896942, + "grad_norm": 0.635733425617218, + "learning_rate": 2.999523207277155e-06, + "loss": 0.3575, + "step": 35790 + }, + { + "epoch": 13.514533786334466, + "grad_norm": 0.6572074890136719, + "learning_rate": 2.9854373160979552e-06, + "loss": 0.3506, + "step": 35800 + }, + { + "epoch": 13.51830879577199, + "grad_norm": 0.6802811622619629, + "learning_rate": 2.9713835589290362e-06, + "loss": 0.3496, + "step": 35810 + }, + { + "epoch": 13.522083805209514, + "grad_norm": 0.6372085213661194, + "learning_rate": 2.9573619453759913e-06, + "loss": 0.3557, + "step": 35820 + }, + { + "epoch": 13.525858814647037, + "grad_norm": 0.6938003897666931, + "learning_rate": 2.943372485022489e-06, + "loss": 0.3472, + "step": 35830 + }, + { + "epoch": 13.52963382408456, + "grad_norm": 0.6606130599975586, + "learning_rate": 2.9294151874301623e-06, + "loss": 0.3722, + "step": 35840 + }, + { + "epoch": 13.533408833522083, + "grad_norm": 0.6342771649360657, + "learning_rate": 2.9154900621387093e-06, + "loss": 0.355, + "step": 35850 + }, + { + "epoch": 13.537183842959607, + "grad_norm": 0.691672682762146, + "learning_rate": 2.9015971186658164e-06, + "loss": 0.3586, + "step": 35860 + }, + { + "epoch": 13.54095885239713, + "grad_norm": 0.5848689675331116, + "learning_rate": 2.887736366507182e-06, + "loss": 0.3591, + "step": 35870 + }, + { + "epoch": 13.544733861834654, + "grad_norm": 0.5893767476081848, + "learning_rate": 2.873907815136501e-06, + "loss": 0.3476, + "step": 35880 + }, + { + "epoch": 13.548508871272178, + "grad_norm": 0.6813375949859619, + "learning_rate": 2.8601114740054556e-06, + "loss": 0.3556, + "step": 35890 + }, + { + "epoch": 13.552283880709702, + "grad_norm": 0.629166841506958, + "learning_rate": 2.8463473525437155e-06, + "loss": 0.357, + "step": 35900 + }, + { + "epoch": 13.556058890147225, + "grad_norm": 0.6545192003250122, + "learning_rate": 2.8326154601589216e-06, + "loss": 0.3553, + "step": 35910 + }, + { + "epoch": 13.55983389958475, + "grad_norm": 0.639487087726593, + "learning_rate": 2.818915806236705e-06, + "loss": 0.3558, + "step": 35920 + }, + { + "epoch": 13.563608909022273, + "grad_norm": 0.6719843149185181, + "learning_rate": 2.8052484001406478e-06, + "loss": 0.3527, + "step": 35930 + }, + { + "epoch": 13.567383918459797, + "grad_norm": 0.598436713218689, + "learning_rate": 2.791613251212277e-06, + "loss": 0.3582, + "step": 35940 + }, + { + "epoch": 13.57115892789732, + "grad_norm": 0.6548103094100952, + "learning_rate": 2.7780103687711166e-06, + "loss": 0.3549, + "step": 35950 + }, + { + "epoch": 13.574933937334844, + "grad_norm": 0.6896668076515198, + "learning_rate": 2.764439762114579e-06, + "loss": 0.3508, + "step": 35960 + }, + { + "epoch": 13.578708946772366, + "grad_norm": 0.6703106164932251, + "learning_rate": 2.750901440518061e-06, + "loss": 0.3556, + "step": 35970 + }, + { + "epoch": 13.58248395620989, + "grad_norm": 0.664481520652771, + "learning_rate": 2.7373954132348776e-06, + "loss": 0.356, + "step": 35980 + }, + { + "epoch": 13.586258965647414, + "grad_norm": 0.6433864831924438, + "learning_rate": 2.7239216894962784e-06, + "loss": 0.3526, + "step": 35990 + }, + { + "epoch": 13.590033975084937, + "grad_norm": 0.6484426259994507, + "learning_rate": 2.710480278511407e-06, + "loss": 0.3545, + "step": 36000 + }, + { + "epoch": 13.593808984522461, + "grad_norm": 5.11317777633667, + "learning_rate": 2.6970711894673483e-06, + "loss": 0.3684, + "step": 36010 + }, + { + "epoch": 13.597583993959985, + "grad_norm": 0.7673658132553101, + "learning_rate": 2.6836944315291047e-06, + "loss": 0.3563, + "step": 36020 + }, + { + "epoch": 13.601359003397508, + "grad_norm": 0.6486364603042603, + "learning_rate": 2.67035001383954e-06, + "loss": 0.3549, + "step": 36030 + }, + { + "epoch": 13.605134012835032, + "grad_norm": 0.5933120250701904, + "learning_rate": 2.6570379455194527e-06, + "loss": 0.3573, + "step": 36040 + }, + { + "epoch": 13.608909022272556, + "grad_norm": 0.7512244582176208, + "learning_rate": 2.6437582356675206e-06, + "loss": 0.3507, + "step": 36050 + }, + { + "epoch": 13.61268403171008, + "grad_norm": 0.7129798531532288, + "learning_rate": 2.6305108933602775e-06, + "loss": 0.362, + "step": 36060 + }, + { + "epoch": 13.616459041147603, + "grad_norm": 0.5950183868408203, + "learning_rate": 2.6172959276521803e-06, + "loss": 0.3546, + "step": 36070 + }, + { + "epoch": 13.620234050585127, + "grad_norm": 0.6639828085899353, + "learning_rate": 2.6041133475755153e-06, + "loss": 0.3559, + "step": 36080 + }, + { + "epoch": 13.62400906002265, + "grad_norm": 0.654403567314148, + "learning_rate": 2.590963162140475e-06, + "loss": 0.3544, + "step": 36090 + }, + { + "epoch": 13.627784069460173, + "grad_norm": 0.6647920608520508, + "learning_rate": 2.5778453803350587e-06, + "loss": 0.3572, + "step": 36100 + }, + { + "epoch": 13.631559078897697, + "grad_norm": 0.6501339077949524, + "learning_rate": 2.5647600111251723e-06, + "loss": 0.3574, + "step": 36110 + }, + { + "epoch": 13.63533408833522, + "grad_norm": 0.6624252200126648, + "learning_rate": 2.551707063454528e-06, + "loss": 0.3604, + "step": 36120 + }, + { + "epoch": 13.639109097772744, + "grad_norm": 1.6550086736679077, + "learning_rate": 2.5386865462446897e-06, + "loss": 0.3647, + "step": 36130 + }, + { + "epoch": 13.642884107210268, + "grad_norm": 0.6511169672012329, + "learning_rate": 2.52569846839506e-06, + "loss": 0.3546, + "step": 36140 + }, + { + "epoch": 13.646659116647792, + "grad_norm": 0.6163169741630554, + "learning_rate": 2.512742838782883e-06, + "loss": 0.3556, + "step": 36150 + }, + { + "epoch": 13.650434126085315, + "grad_norm": 0.6587780117988586, + "learning_rate": 2.4998196662631924e-06, + "loss": 0.3611, + "step": 36160 + }, + { + "epoch": 13.654209135522839, + "grad_norm": 0.6356183290481567, + "learning_rate": 2.4869289596688604e-06, + "loss": 0.35, + "step": 36170 + }, + { + "epoch": 13.657984144960363, + "grad_norm": 0.6465098857879639, + "learning_rate": 2.4740707278105567e-06, + "loss": 0.3539, + "step": 36180 + }, + { + "epoch": 13.661759154397886, + "grad_norm": 0.6922428011894226, + "learning_rate": 2.461244979476779e-06, + "loss": 0.3567, + "step": 36190 + }, + { + "epoch": 13.66553416383541, + "grad_norm": 0.6384284496307373, + "learning_rate": 2.4484517234337866e-06, + "loss": 0.3517, + "step": 36200 + }, + { + "epoch": 13.669309173272934, + "grad_norm": 0.6639804244041443, + "learning_rate": 2.4356909684256634e-06, + "loss": 0.3593, + "step": 36210 + }, + { + "epoch": 13.673084182710458, + "grad_norm": 0.6959599852561951, + "learning_rate": 2.4229627231742545e-06, + "loss": 0.3552, + "step": 36220 + }, + { + "epoch": 13.67685919214798, + "grad_norm": 0.6804720163345337, + "learning_rate": 2.4102669963791956e-06, + "loss": 0.3578, + "step": 36230 + }, + { + "epoch": 13.680634201585503, + "grad_norm": 0.6466209888458252, + "learning_rate": 2.3976037967179065e-06, + "loss": 0.3557, + "step": 36240 + }, + { + "epoch": 13.684409211023027, + "grad_norm": 0.6484111547470093, + "learning_rate": 2.3849731328455582e-06, + "loss": 0.3582, + "step": 36250 + }, + { + "epoch": 13.68818422046055, + "grad_norm": 0.637500524520874, + "learning_rate": 2.3723750133950894e-06, + "loss": 0.3571, + "step": 36260 + }, + { + "epoch": 13.691959229898075, + "grad_norm": 3.0127627849578857, + "learning_rate": 2.3598094469771957e-06, + "loss": 0.3572, + "step": 36270 + }, + { + "epoch": 13.695734239335598, + "grad_norm": 0.6502493619918823, + "learning_rate": 2.347276442180324e-06, + "loss": 0.3556, + "step": 36280 + }, + { + "epoch": 13.699509248773122, + "grad_norm": 0.6446804404258728, + "learning_rate": 2.3347760075706715e-06, + "loss": 0.3613, + "step": 36290 + }, + { + "epoch": 13.703284258210646, + "grad_norm": 0.6583272218704224, + "learning_rate": 2.3223081516921595e-06, + "loss": 0.3598, + "step": 36300 + }, + { + "epoch": 13.70705926764817, + "grad_norm": 0.5799362063407898, + "learning_rate": 2.3098728830664494e-06, + "loss": 0.3619, + "step": 36310 + }, + { + "epoch": 13.710834277085693, + "grad_norm": 0.7087305188179016, + "learning_rate": 2.2974702101929314e-06, + "loss": 0.3569, + "step": 36320 + }, + { + "epoch": 13.714609286523217, + "grad_norm": 0.6843621730804443, + "learning_rate": 2.2851001415487185e-06, + "loss": 0.3589, + "step": 36330 + }, + { + "epoch": 13.71838429596074, + "grad_norm": 0.71535325050354, + "learning_rate": 2.272762685588631e-06, + "loss": 0.3547, + "step": 36340 + }, + { + "epoch": 13.722159305398263, + "grad_norm": 0.6820167303085327, + "learning_rate": 2.260457850745218e-06, + "loss": 0.3498, + "step": 36350 + }, + { + "epoch": 13.725934314835786, + "grad_norm": 0.6012173295021057, + "learning_rate": 2.2481856454286963e-06, + "loss": 0.3575, + "step": 36360 + }, + { + "epoch": 13.72970932427331, + "grad_norm": 0.687364399433136, + "learning_rate": 2.235946078027018e-06, + "loss": 0.3519, + "step": 36370 + }, + { + "epoch": 13.733484333710834, + "grad_norm": 0.6643616557121277, + "learning_rate": 2.223739156905802e-06, + "loss": 0.3559, + "step": 36380 + }, + { + "epoch": 13.737259343148358, + "grad_norm": 0.6354239583015442, + "learning_rate": 2.2115648904083796e-06, + "loss": 0.3561, + "step": 36390 + }, + { + "epoch": 13.741034352585881, + "grad_norm": 0.6376845836639404, + "learning_rate": 2.199423286855734e-06, + "loss": 0.3633, + "step": 36400 + }, + { + "epoch": 13.744809362023405, + "grad_norm": 0.6230825185775757, + "learning_rate": 2.1873143545465323e-06, + "loss": 0.3594, + "step": 36410 + }, + { + "epoch": 13.748584371460929, + "grad_norm": 0.644271969795227, + "learning_rate": 2.1752381017571267e-06, + "loss": 0.3556, + "step": 36420 + }, + { + "epoch": 13.752359380898453, + "grad_norm": 0.6420625448226929, + "learning_rate": 2.1631945367415142e-06, + "loss": 0.3586, + "step": 36430 + }, + { + "epoch": 13.756134390335976, + "grad_norm": 0.8240686655044556, + "learning_rate": 2.15118366773136e-06, + "loss": 0.3611, + "step": 36440 + }, + { + "epoch": 13.7599093997735, + "grad_norm": 0.6971853971481323, + "learning_rate": 2.1392055029359926e-06, + "loss": 0.3599, + "step": 36450 + }, + { + "epoch": 13.763684409211024, + "grad_norm": 0.6388399004936218, + "learning_rate": 2.127260050542357e-06, + "loss": 0.3603, + "step": 36460 + }, + { + "epoch": 13.767459418648546, + "grad_norm": 0.6157034039497375, + "learning_rate": 2.1153473187150608e-06, + "loss": 0.3599, + "step": 36470 + }, + { + "epoch": 13.77123442808607, + "grad_norm": 0.6063554286956787, + "learning_rate": 2.1034673155963467e-06, + "loss": 0.3592, + "step": 36480 + }, + { + "epoch": 13.775009437523593, + "grad_norm": 0.6081372499465942, + "learning_rate": 2.0916200493060976e-06, + "loss": 0.3581, + "step": 36490 + }, + { + "epoch": 13.778784446961117, + "grad_norm": 0.6699353456497192, + "learning_rate": 2.0798055279417916e-06, + "loss": 0.3602, + "step": 36500 + }, + { + "epoch": 13.78255945639864, + "grad_norm": 0.6695933938026428, + "learning_rate": 2.0680237595785356e-06, + "loss": 0.3552, + "step": 36510 + }, + { + "epoch": 13.786334465836164, + "grad_norm": 0.5835472941398621, + "learning_rate": 2.0562747522690774e-06, + "loss": 0.3573, + "step": 36520 + }, + { + "epoch": 13.790109475273688, + "grad_norm": 0.7515608072280884, + "learning_rate": 2.0445585140437373e-06, + "loss": 0.3574, + "step": 36530 + }, + { + "epoch": 13.793884484711212, + "grad_norm": 0.6449537873268127, + "learning_rate": 2.0328750529104602e-06, + "loss": 0.3573, + "step": 36540 + }, + { + "epoch": 13.797659494148736, + "grad_norm": 0.6437288522720337, + "learning_rate": 2.021224376854791e-06, + "loss": 0.3579, + "step": 36550 + }, + { + "epoch": 13.80143450358626, + "grad_norm": 0.6661025285720825, + "learning_rate": 2.0096064938398386e-06, + "loss": 0.3663, + "step": 36560 + }, + { + "epoch": 13.805209513023783, + "grad_norm": 0.6766248941421509, + "learning_rate": 1.998021411806328e-06, + "loss": 0.3595, + "step": 36570 + }, + { + "epoch": 13.808984522461307, + "grad_norm": 0.6111851930618286, + "learning_rate": 1.9864691386725586e-06, + "loss": 0.3578, + "step": 36580 + }, + { + "epoch": 13.81275953189883, + "grad_norm": 0.7111583948135376, + "learning_rate": 1.9749496823343814e-06, + "loss": 0.3598, + "step": 36590 + }, + { + "epoch": 13.816534541336353, + "grad_norm": 0.7276173830032349, + "learning_rate": 1.963463050665254e-06, + "loss": 0.349, + "step": 36600 + }, + { + "epoch": 13.820309550773876, + "grad_norm": 0.6221917271614075, + "learning_rate": 1.9520092515161682e-06, + "loss": 0.3696, + "step": 36610 + }, + { + "epoch": 13.8240845602114, + "grad_norm": 0.7171021699905396, + "learning_rate": 1.940588292715695e-06, + "loss": 0.355, + "step": 36620 + }, + { + "epoch": 13.827859569648924, + "grad_norm": 0.7668521404266357, + "learning_rate": 1.929200182069951e-06, + "loss": 0.3491, + "step": 36630 + }, + { + "epoch": 13.831634579086447, + "grad_norm": 0.9479874968528748, + "learning_rate": 1.9178449273625997e-06, + "loss": 0.3594, + "step": 36640 + }, + { + "epoch": 13.835409588523971, + "grad_norm": 0.6809915900230408, + "learning_rate": 1.90652253635486e-06, + "loss": 0.3518, + "step": 36650 + }, + { + "epoch": 13.839184597961495, + "grad_norm": 0.6257880330085754, + "learning_rate": 1.8952330167854693e-06, + "loss": 0.358, + "step": 36660 + }, + { + "epoch": 13.842959607399019, + "grad_norm": 0.6460468769073486, + "learning_rate": 1.8839763763707107e-06, + "loss": 0.3555, + "step": 36670 + }, + { + "epoch": 13.846734616836542, + "grad_norm": 0.7003082036972046, + "learning_rate": 1.8727526228043912e-06, + "loss": 0.3521, + "step": 36680 + }, + { + "epoch": 13.850509626274066, + "grad_norm": 0.6699946522712708, + "learning_rate": 1.8615617637578408e-06, + "loss": 0.3566, + "step": 36690 + }, + { + "epoch": 13.85428463571159, + "grad_norm": 0.6396785378456116, + "learning_rate": 1.850403806879908e-06, + "loss": 0.3517, + "step": 36700 + }, + { + "epoch": 13.858059645149114, + "grad_norm": 0.6421851515769958, + "learning_rate": 1.8392787597969475e-06, + "loss": 0.3532, + "step": 36710 + }, + { + "epoch": 13.861834654586637, + "grad_norm": 0.6195537447929382, + "learning_rate": 1.828186630112827e-06, + "loss": 0.3673, + "step": 36720 + }, + { + "epoch": 13.86560966402416, + "grad_norm": 0.5870434045791626, + "learning_rate": 1.8171274254089098e-06, + "loss": 0.3594, + "step": 36730 + }, + { + "epoch": 13.869384673461683, + "grad_norm": 0.7309495806694031, + "learning_rate": 1.806101153244061e-06, + "loss": 0.36, + "step": 36740 + }, + { + "epoch": 13.873159682899207, + "grad_norm": 0.6230713129043579, + "learning_rate": 1.7951078211546401e-06, + "loss": 0.3546, + "step": 36750 + }, + { + "epoch": 13.87693469233673, + "grad_norm": 0.6043184399604797, + "learning_rate": 1.7841474366544764e-06, + "loss": 0.3515, + "step": 36760 + }, + { + "epoch": 13.880709701774254, + "grad_norm": 0.6820389032363892, + "learning_rate": 1.7732200072348825e-06, + "loss": 0.3657, + "step": 36770 + }, + { + "epoch": 13.884484711211778, + "grad_norm": 0.6838846206665039, + "learning_rate": 1.7623255403646677e-06, + "loss": 0.3519, + "step": 36780 + }, + { + "epoch": 13.888259720649302, + "grad_norm": 0.6790589094161987, + "learning_rate": 1.7514640434900865e-06, + "loss": 0.3544, + "step": 36790 + }, + { + "epoch": 13.892034730086825, + "grad_norm": 0.6140289902687073, + "learning_rate": 1.7406355240348838e-06, + "loss": 0.353, + "step": 36800 + }, + { + "epoch": 13.89580973952435, + "grad_norm": 0.6230329871177673, + "learning_rate": 1.7298399894002392e-06, + "loss": 0.3547, + "step": 36810 + }, + { + "epoch": 13.899584748961873, + "grad_norm": 0.674197793006897, + "learning_rate": 1.7190774469647952e-06, + "loss": 0.3633, + "step": 36820 + }, + { + "epoch": 13.903359758399397, + "grad_norm": 0.8990030884742737, + "learning_rate": 1.7083479040846617e-06, + "loss": 0.3618, + "step": 36830 + }, + { + "epoch": 13.90713476783692, + "grad_norm": 0.6309157609939575, + "learning_rate": 1.697651368093367e-06, + "loss": 0.3521, + "step": 36840 + }, + { + "epoch": 13.910909777274444, + "grad_norm": 0.6408490538597107, + "learning_rate": 1.6869878463019017e-06, + "loss": 0.357, + "step": 36850 + }, + { + "epoch": 13.914684786711966, + "grad_norm": 0.6487820148468018, + "learning_rate": 1.6763573459986858e-06, + "loss": 0.3681, + "step": 36860 + }, + { + "epoch": 13.91845979614949, + "grad_norm": 0.663537859916687, + "learning_rate": 1.6657598744495516e-06, + "loss": 0.3592, + "step": 36870 + }, + { + "epoch": 13.922234805587014, + "grad_norm": 0.629589319229126, + "learning_rate": 1.6551954388977774e-06, + "loss": 0.3481, + "step": 36880 + }, + { + "epoch": 13.926009815024537, + "grad_norm": 0.6090066432952881, + "learning_rate": 1.644664046564054e-06, + "loss": 0.3581, + "step": 36890 + }, + { + "epoch": 13.929784824462061, + "grad_norm": 0.765868067741394, + "learning_rate": 1.634165704646501e-06, + "loss": 0.3518, + "step": 36900 + }, + { + "epoch": 13.933559833899585, + "grad_norm": 0.6522102355957031, + "learning_rate": 1.6237004203206175e-06, + "loss": 0.3544, + "step": 36910 + }, + { + "epoch": 13.937334843337108, + "grad_norm": 0.6589536666870117, + "learning_rate": 1.6132682007393373e-06, + "loss": 0.3544, + "step": 36920 + }, + { + "epoch": 13.941109852774632, + "grad_norm": 0.6735085248947144, + "learning_rate": 1.6028690530329847e-06, + "loss": 0.3427, + "step": 36930 + }, + { + "epoch": 13.944884862212156, + "grad_norm": 0.6307739615440369, + "learning_rate": 1.5925029843092742e-06, + "loss": 0.3557, + "step": 36940 + }, + { + "epoch": 13.94865987164968, + "grad_norm": 0.6922662854194641, + "learning_rate": 1.5821700016533214e-06, + "loss": 0.3629, + "step": 36950 + }, + { + "epoch": 13.952434881087203, + "grad_norm": 0.6589555740356445, + "learning_rate": 1.5718701121276268e-06, + "loss": 0.3487, + "step": 36960 + }, + { + "epoch": 13.956209890524725, + "grad_norm": 0.7357824444770813, + "learning_rate": 1.5616033227720483e-06, + "loss": 0.3495, + "step": 36970 + }, + { + "epoch": 13.95998489996225, + "grad_norm": 0.6229332089424133, + "learning_rate": 1.5513696406038558e-06, + "loss": 0.3577, + "step": 36980 + }, + { + "epoch": 13.963759909399773, + "grad_norm": 0.5904605388641357, + "learning_rate": 1.5411690726176653e-06, + "loss": 0.3531, + "step": 36990 + }, + { + "epoch": 13.967534918837297, + "grad_norm": 0.7504417300224304, + "learning_rate": 1.5310016257854832e-06, + "loss": 0.3604, + "step": 37000 + }, + { + "epoch": 13.97130992827482, + "grad_norm": 0.6638484001159668, + "learning_rate": 1.5208673070566504e-06, + "loss": 0.3546, + "step": 37010 + }, + { + "epoch": 13.975084937712344, + "grad_norm": 0.6932189464569092, + "learning_rate": 1.5107661233578818e-06, + "loss": 0.353, + "step": 37020 + }, + { + "epoch": 13.978859947149868, + "grad_norm": 0.871865451335907, + "learning_rate": 1.5006980815932437e-06, + "loss": 0.3684, + "step": 37030 + }, + { + "epoch": 13.982634956587392, + "grad_norm": 0.6513112783432007, + "learning_rate": 1.490663188644148e-06, + "loss": 0.3602, + "step": 37040 + }, + { + "epoch": 13.986409966024915, + "grad_norm": 0.6713985204696655, + "learning_rate": 1.4806614513693418e-06, + "loss": 0.3541, + "step": 37050 + }, + { + "epoch": 13.990184975462439, + "grad_norm": 0.681988537311554, + "learning_rate": 1.4706928766049289e-06, + "loss": 0.3581, + "step": 37060 + }, + { + "epoch": 13.993959984899963, + "grad_norm": 0.948577344417572, + "learning_rate": 1.4607574711643257e-06, + "loss": 0.3532, + "step": 37070 + }, + { + "epoch": 13.997734994337486, + "grad_norm": 0.7015256881713867, + "learning_rate": 1.4508552418382949e-06, + "loss": 0.355, + "step": 37080 + }, + { + "epoch": 14.00151000377501, + "grad_norm": 0.6600831151008606, + "learning_rate": 1.4409861953949112e-06, + "loss": 0.3522, + "step": 37090 + }, + { + "epoch": 14.005285013212532, + "grad_norm": 0.6127675771713257, + "learning_rate": 1.43115033857959e-06, + "loss": 0.3566, + "step": 37100 + }, + { + "epoch": 14.009060022650056, + "grad_norm": 0.6839843988418579, + "learning_rate": 1.4213476781150258e-06, + "loss": 0.3503, + "step": 37110 + }, + { + "epoch": 14.01283503208758, + "grad_norm": 0.6490364074707031, + "learning_rate": 1.411578220701254e-06, + "loss": 0.3628, + "step": 37120 + }, + { + "epoch": 14.016610041525103, + "grad_norm": 0.6181484460830688, + "learning_rate": 1.4018419730156108e-06, + "loss": 0.3498, + "step": 37130 + }, + { + "epoch": 14.020385050962627, + "grad_norm": 0.638804018497467, + "learning_rate": 1.3921389417127283e-06, + "loss": 0.3595, + "step": 37140 + }, + { + "epoch": 14.02416006040015, + "grad_norm": 0.6494945287704468, + "learning_rate": 1.3824691334245298e-06, + "loss": 0.36, + "step": 37150 + }, + { + "epoch": 14.027935069837675, + "grad_norm": 0.6143277287483215, + "learning_rate": 1.3728325547602505e-06, + "loss": 0.3568, + "step": 37160 + }, + { + "epoch": 14.031710079275198, + "grad_norm": 0.6732072234153748, + "learning_rate": 1.3632292123063828e-06, + "loss": 0.3524, + "step": 37170 + }, + { + "epoch": 14.035485088712722, + "grad_norm": 0.6264946460723877, + "learning_rate": 1.3536591126267317e-06, + "loss": 0.3535, + "step": 37180 + }, + { + "epoch": 14.039260098150246, + "grad_norm": 0.6948634386062622, + "learning_rate": 1.3441222622623761e-06, + "loss": 0.3541, + "step": 37190 + }, + { + "epoch": 14.04303510758777, + "grad_norm": 0.6478399038314819, + "learning_rate": 1.3346186677316575e-06, + "loss": 0.3537, + "step": 37200 + }, + { + "epoch": 14.046810117025293, + "grad_norm": 0.727102518081665, + "learning_rate": 1.3251483355301852e-06, + "loss": 0.3499, + "step": 37210 + }, + { + "epoch": 14.050585126462817, + "grad_norm": 0.5733349919319153, + "learning_rate": 1.315711272130854e-06, + "loss": 0.3598, + "step": 37220 + }, + { + "epoch": 14.054360135900339, + "grad_norm": 0.6830695867538452, + "learning_rate": 1.306307483983804e-06, + "loss": 0.3582, + "step": 37230 + }, + { + "epoch": 14.058135145337863, + "grad_norm": 0.6603965163230896, + "learning_rate": 1.2969369775164275e-06, + "loss": 0.3528, + "step": 37240 + }, + { + "epoch": 14.061910154775386, + "grad_norm": 0.672484278678894, + "learning_rate": 1.28759975913339e-06, + "loss": 0.353, + "step": 37250 + }, + { + "epoch": 14.06568516421291, + "grad_norm": 0.6874468922615051, + "learning_rate": 1.2782958352165863e-06, + "loss": 0.3559, + "step": 37260 + }, + { + "epoch": 14.069460173650434, + "grad_norm": 0.6835435032844543, + "learning_rate": 1.2690252121251633e-06, + "loss": 0.359, + "step": 37270 + }, + { + "epoch": 14.073235183087958, + "grad_norm": 0.6037148237228394, + "learning_rate": 1.2597878961954968e-06, + "loss": 0.3595, + "step": 37280 + }, + { + "epoch": 14.077010192525481, + "grad_norm": 0.7436287999153137, + "learning_rate": 1.250583893741214e-06, + "loss": 0.3606, + "step": 37290 + }, + { + "epoch": 14.080785201963005, + "grad_norm": 0.6740867495536804, + "learning_rate": 1.2414132110531607e-06, + "loss": 0.3607, + "step": 37300 + }, + { + "epoch": 14.084560211400529, + "grad_norm": 0.7421841025352478, + "learning_rate": 1.2322758543994062e-06, + "loss": 0.3554, + "step": 37310 + }, + { + "epoch": 14.088335220838053, + "grad_norm": 0.7035681009292603, + "learning_rate": 1.2231718300252603e-06, + "loss": 0.3539, + "step": 37320 + }, + { + "epoch": 14.092110230275576, + "grad_norm": 0.6054702401161194, + "learning_rate": 1.214101144153229e-06, + "loss": 0.3691, + "step": 37330 + }, + { + "epoch": 14.0958852397131, + "grad_norm": 0.6276698112487793, + "learning_rate": 1.205063802983042e-06, + "loss": 0.3569, + "step": 37340 + }, + { + "epoch": 14.099660249150624, + "grad_norm": 0.652378261089325, + "learning_rate": 1.1960598126916301e-06, + "loss": 0.3559, + "step": 37350 + }, + { + "epoch": 14.103435258588146, + "grad_norm": 0.6906107664108276, + "learning_rate": 1.187089179433154e-06, + "loss": 0.3614, + "step": 37360 + }, + { + "epoch": 14.10721026802567, + "grad_norm": 0.6413514018058777, + "learning_rate": 1.178151909338937e-06, + "loss": 0.3563, + "step": 37370 + }, + { + "epoch": 14.110985277463193, + "grad_norm": 0.6062707901000977, + "learning_rate": 1.1692480085175317e-06, + "loss": 0.3571, + "step": 37380 + }, + { + "epoch": 14.114760286900717, + "grad_norm": 0.6939738392829895, + "learning_rate": 1.1603774830546587e-06, + "loss": 0.3524, + "step": 37390 + }, + { + "epoch": 14.11853529633824, + "grad_norm": 0.6625202894210815, + "learning_rate": 1.1515403390132573e-06, + "loss": 0.3564, + "step": 37400 + }, + { + "epoch": 14.122310305775764, + "grad_norm": 0.7371194362640381, + "learning_rate": 1.1427365824334124e-06, + "loss": 0.3535, + "step": 37410 + }, + { + "epoch": 14.126085315213288, + "grad_norm": 0.5587255954742432, + "learning_rate": 1.133966219332422e-06, + "loss": 0.3549, + "step": 37420 + }, + { + "epoch": 14.129860324650812, + "grad_norm": 0.6779747605323792, + "learning_rate": 1.1252292557047405e-06, + "loss": 0.3618, + "step": 37430 + }, + { + "epoch": 14.133635334088336, + "grad_norm": 0.6948127150535583, + "learning_rate": 1.1165256975219974e-06, + "loss": 0.3502, + "step": 37440 + }, + { + "epoch": 14.13741034352586, + "grad_norm": 0.796514630317688, + "learning_rate": 1.1078555507330002e-06, + "loss": 0.3551, + "step": 37450 + }, + { + "epoch": 14.141185352963383, + "grad_norm": 0.6783875226974487, + "learning_rate": 1.0992188212637089e-06, + "loss": 0.355, + "step": 37460 + }, + { + "epoch": 14.144960362400907, + "grad_norm": 0.6731642484664917, + "learning_rate": 1.0906155150172458e-06, + "loss": 0.351, + "step": 37470 + }, + { + "epoch": 14.148735371838429, + "grad_norm": 0.670313835144043, + "learning_rate": 1.0820456378738852e-06, + "loss": 0.3541, + "step": 37480 + }, + { + "epoch": 14.152510381275953, + "grad_norm": 0.768552303314209, + "learning_rate": 1.0735091956910692e-06, + "loss": 0.3524, + "step": 37490 + }, + { + "epoch": 14.156285390713476, + "grad_norm": 0.6623304486274719, + "learning_rate": 1.0650061943033695e-06, + "loss": 0.3575, + "step": 37500 + }, + { + "epoch": 14.160060400151, + "grad_norm": 0.632932186126709, + "learning_rate": 1.0565366395225041e-06, + "loss": 0.3482, + "step": 37510 + }, + { + "epoch": 14.163835409588524, + "grad_norm": 0.633401095867157, + "learning_rate": 1.0481005371373366e-06, + "loss": 0.3536, + "step": 37520 + }, + { + "epoch": 14.167610419026047, + "grad_norm": 0.6039905548095703, + "learning_rate": 1.0396978929138602e-06, + "loss": 0.3505, + "step": 37530 + }, + { + "epoch": 14.171385428463571, + "grad_norm": 0.6414741277694702, + "learning_rate": 1.0313287125952032e-06, + "loss": 0.3634, + "step": 37540 + }, + { + "epoch": 14.175160437901095, + "grad_norm": 0.6331318020820618, + "learning_rate": 1.0229930019016288e-06, + "loss": 0.3648, + "step": 37550 + }, + { + "epoch": 14.178935447338619, + "grad_norm": 0.6243364214897156, + "learning_rate": 1.0146907665305183e-06, + "loss": 0.3647, + "step": 37560 + }, + { + "epoch": 14.182710456776142, + "grad_norm": 0.6351826786994934, + "learning_rate": 1.0064220121563551e-06, + "loss": 0.358, + "step": 37570 + }, + { + "epoch": 14.186485466213666, + "grad_norm": 0.6927258372306824, + "learning_rate": 9.98186744430768e-07, + "loss": 0.3517, + "step": 37580 + }, + { + "epoch": 14.19026047565119, + "grad_norm": 0.6401153206825256, + "learning_rate": 9.899849689824826e-07, + "loss": 0.3611, + "step": 37590 + }, + { + "epoch": 14.194035485088712, + "grad_norm": 0.7551748752593994, + "learning_rate": 9.818166914173365e-07, + "loss": 0.3524, + "step": 37600 + }, + { + "epoch": 14.197810494526236, + "grad_norm": 0.6238972544670105, + "learning_rate": 9.736819173182642e-07, + "loss": 0.3578, + "step": 37610 + }, + { + "epoch": 14.20158550396376, + "grad_norm": 0.795691728591919, + "learning_rate": 9.655806522453126e-07, + "loss": 0.3544, + "step": 37620 + }, + { + "epoch": 14.205360513401283, + "grad_norm": 0.558307945728302, + "learning_rate": 9.575129017356187e-07, + "loss": 0.3547, + "step": 37630 + }, + { + "epoch": 14.209135522838807, + "grad_norm": 0.6478992104530334, + "learning_rate": 9.494786713034109e-07, + "loss": 0.3494, + "step": 37640 + }, + { + "epoch": 14.21291053227633, + "grad_norm": 0.646564245223999, + "learning_rate": 9.414779664400075e-07, + "loss": 0.3572, + "step": 37650 + }, + { + "epoch": 14.216685541713854, + "grad_norm": 0.6614131927490234, + "learning_rate": 9.335107926138342e-07, + "loss": 0.3513, + "step": 37660 + }, + { + "epoch": 14.220460551151378, + "grad_norm": 0.7705824375152588, + "learning_rate": 9.255771552703519e-07, + "loss": 0.3504, + "step": 37670 + }, + { + "epoch": 14.224235560588902, + "grad_norm": 0.6299923658370972, + "learning_rate": 9.176770598321449e-07, + "loss": 0.3535, + "step": 37680 + }, + { + "epoch": 14.228010570026425, + "grad_norm": 0.6774992346763611, + "learning_rate": 9.098105116988498e-07, + "loss": 0.3546, + "step": 37690 + }, + { + "epoch": 14.23178557946395, + "grad_norm": 0.6203901171684265, + "learning_rate": 9.019775162471767e-07, + "loss": 0.3477, + "step": 37700 + }, + { + "epoch": 14.235560588901473, + "grad_norm": 0.590056300163269, + "learning_rate": 8.94178078830904e-07, + "loss": 0.3571, + "step": 37710 + }, + { + "epoch": 14.239335598338997, + "grad_norm": 0.6835026741027832, + "learning_rate": 8.864122047808731e-07, + "loss": 0.362, + "step": 37720 + }, + { + "epoch": 14.243110607776519, + "grad_norm": 0.7431061863899231, + "learning_rate": 8.786798994049882e-07, + "loss": 0.3595, + "step": 37730 + }, + { + "epoch": 14.246885617214042, + "grad_norm": 0.6774893999099731, + "learning_rate": 8.709811679882052e-07, + "loss": 0.3546, + "step": 37740 + }, + { + "epoch": 14.250660626651566, + "grad_norm": 0.6696170568466187, + "learning_rate": 8.633160157925313e-07, + "loss": 0.3597, + "step": 37750 + }, + { + "epoch": 14.25443563608909, + "grad_norm": 0.592789351940155, + "learning_rate": 8.556844480570369e-07, + "loss": 0.3576, + "step": 37760 + }, + { + "epoch": 14.258210645526614, + "grad_norm": 0.6565343141555786, + "learning_rate": 8.480864699978164e-07, + "loss": 0.3634, + "step": 37770 + }, + { + "epoch": 14.261985654964137, + "grad_norm": 0.5975145697593689, + "learning_rate": 8.405220868080154e-07, + "loss": 0.3502, + "step": 37780 + }, + { + "epoch": 14.265760664401661, + "grad_norm": 0.9825413227081299, + "learning_rate": 8.329913036578263e-07, + "loss": 0.3514, + "step": 37790 + }, + { + "epoch": 14.269535673839185, + "grad_norm": 0.6201307773590088, + "learning_rate": 8.254941256944593e-07, + "loss": 0.358, + "step": 37800 + }, + { + "epoch": 14.273310683276708, + "grad_norm": 0.6925731897354126, + "learning_rate": 8.18030558042171e-07, + "loss": 0.3512, + "step": 37810 + }, + { + "epoch": 14.277085692714232, + "grad_norm": 0.6344226002693176, + "learning_rate": 8.106006058022419e-07, + "loss": 0.3585, + "step": 37820 + }, + { + "epoch": 14.280860702151756, + "grad_norm": 0.6603698134422302, + "learning_rate": 8.03204274052971e-07, + "loss": 0.3566, + "step": 37830 + }, + { + "epoch": 14.28463571158928, + "grad_norm": 0.7116165161132812, + "learning_rate": 7.958415678496756e-07, + "loss": 0.359, + "step": 37840 + }, + { + "epoch": 14.288410721026803, + "grad_norm": 0.6361221075057983, + "learning_rate": 7.885124922247022e-07, + "loss": 0.3539, + "step": 37850 + }, + { + "epoch": 14.292185730464325, + "grad_norm": 0.6573535203933716, + "learning_rate": 7.812170521874051e-07, + "loss": 0.3601, + "step": 37860 + }, + { + "epoch": 14.29596073990185, + "grad_norm": 0.6462699770927429, + "learning_rate": 7.739552527241456e-07, + "loss": 0.3573, + "step": 37870 + }, + { + "epoch": 14.299735749339373, + "grad_norm": 0.587371826171875, + "learning_rate": 7.667270987983033e-07, + "loss": 0.3561, + "step": 37880 + }, + { + "epoch": 14.303510758776897, + "grad_norm": 0.6839852929115295, + "learning_rate": 7.595325953502375e-07, + "loss": 0.3578, + "step": 37890 + }, + { + "epoch": 14.30728576821442, + "grad_norm": 0.5843148231506348, + "learning_rate": 7.523717472973368e-07, + "loss": 0.3582, + "step": 37900 + }, + { + "epoch": 14.311060777651944, + "grad_norm": 0.6995810866355896, + "learning_rate": 7.452445595339696e-07, + "loss": 0.3567, + "step": 37910 + }, + { + "epoch": 14.314835787089468, + "grad_norm": 0.6964356899261475, + "learning_rate": 7.381510369315059e-07, + "loss": 0.3596, + "step": 37920 + }, + { + "epoch": 14.318610796526992, + "grad_norm": 1.5723952054977417, + "learning_rate": 7.310911843382895e-07, + "loss": 0.3567, + "step": 37930 + }, + { + "epoch": 14.322385805964515, + "grad_norm": 0.7851320505142212, + "learning_rate": 7.240650065796716e-07, + "loss": 0.3499, + "step": 37940 + }, + { + "epoch": 14.326160815402039, + "grad_norm": 0.6869962811470032, + "learning_rate": 7.170725084579777e-07, + "loss": 0.3588, + "step": 37950 + }, + { + "epoch": 14.329935824839563, + "grad_norm": 0.6377101540565491, + "learning_rate": 7.101136947525233e-07, + "loss": 0.3529, + "step": 37960 + }, + { + "epoch": 14.333710834277086, + "grad_norm": 0.713228166103363, + "learning_rate": 7.031885702195707e-07, + "loss": 0.357, + "step": 37970 + }, + { + "epoch": 14.337485843714608, + "grad_norm": 0.6053051352500916, + "learning_rate": 6.962971395924e-07, + "loss": 0.3571, + "step": 37980 + }, + { + "epoch": 14.341260853152132, + "grad_norm": 0.6226997971534729, + "learning_rate": 6.894394075812217e-07, + "loss": 0.3558, + "step": 37990 + }, + { + "epoch": 14.345035862589656, + "grad_norm": 0.6585350036621094, + "learning_rate": 6.826153788732359e-07, + "loss": 0.3541, + "step": 38000 + }, + { + "epoch": 14.34881087202718, + "grad_norm": 0.7358768582344055, + "learning_rate": 6.758250581326009e-07, + "loss": 0.3588, + "step": 38010 + }, + { + "epoch": 14.352585881464703, + "grad_norm": 0.7400320172309875, + "learning_rate": 6.690684500004485e-07, + "loss": 0.3553, + "step": 38020 + }, + { + "epoch": 14.356360890902227, + "grad_norm": 0.6965455412864685, + "learning_rate": 6.623455590948458e-07, + "loss": 0.3536, + "step": 38030 + }, + { + "epoch": 14.36013590033975, + "grad_norm": 0.6424039006233215, + "learning_rate": 6.556563900108226e-07, + "loss": 0.3559, + "step": 38040 + }, + { + "epoch": 14.363910909777275, + "grad_norm": 0.664570152759552, + "learning_rate": 6.490009473203662e-07, + "loss": 0.3596, + "step": 38050 + }, + { + "epoch": 14.367685919214798, + "grad_norm": 0.6233497262001038, + "learning_rate": 6.423792355724156e-07, + "loss": 0.3518, + "step": 38060 + }, + { + "epoch": 14.371460928652322, + "grad_norm": 0.6261218190193176, + "learning_rate": 6.35791259292845e-07, + "loss": 0.3594, + "step": 38070 + }, + { + "epoch": 14.375235938089846, + "grad_norm": 0.7287961840629578, + "learning_rate": 6.292370229844691e-07, + "loss": 0.3546, + "step": 38080 + }, + { + "epoch": 14.37901094752737, + "grad_norm": 0.5852963328361511, + "learning_rate": 6.227165311270488e-07, + "loss": 0.3556, + "step": 38090 + }, + { + "epoch": 14.382785956964893, + "grad_norm": 0.6326406002044678, + "learning_rate": 6.16229788177275e-07, + "loss": 0.3512, + "step": 38100 + }, + { + "epoch": 14.386560966402415, + "grad_norm": 0.6186873912811279, + "learning_rate": 6.097767985687841e-07, + "loss": 0.3541, + "step": 38110 + }, + { + "epoch": 14.390335975839939, + "grad_norm": 1.111286997795105, + "learning_rate": 6.033575667121317e-07, + "loss": 0.3576, + "step": 38120 + }, + { + "epoch": 14.394110985277463, + "grad_norm": 0.6550930738449097, + "learning_rate": 5.969720969947967e-07, + "loss": 0.355, + "step": 38130 + }, + { + "epoch": 14.397885994714986, + "grad_norm": 0.6559245586395264, + "learning_rate": 5.906203937811883e-07, + "loss": 0.3628, + "step": 38140 + }, + { + "epoch": 14.40166100415251, + "grad_norm": 0.6878376007080078, + "learning_rate": 5.843024614126447e-07, + "loss": 0.3554, + "step": 38150 + }, + { + "epoch": 14.405436013590034, + "grad_norm": 0.644935667514801, + "learning_rate": 5.780183042074061e-07, + "loss": 0.3574, + "step": 38160 + }, + { + "epoch": 14.409211023027558, + "grad_norm": 0.6619318723678589, + "learning_rate": 5.717679264606312e-07, + "loss": 0.3569, + "step": 38170 + }, + { + "epoch": 14.412986032465081, + "grad_norm": 0.6411536931991577, + "learning_rate": 5.655513324444028e-07, + "loss": 0.3592, + "step": 38180 + }, + { + "epoch": 14.416761041902605, + "grad_norm": 0.6617749929428101, + "learning_rate": 5.593685264076942e-07, + "loss": 0.3463, + "step": 38190 + }, + { + "epoch": 14.420536051340129, + "grad_norm": 0.6915404796600342, + "learning_rate": 5.532195125764029e-07, + "loss": 0.3586, + "step": 38200 + }, + { + "epoch": 14.424311060777653, + "grad_norm": 0.6411628723144531, + "learning_rate": 5.471042951533167e-07, + "loss": 0.3425, + "step": 38210 + }, + { + "epoch": 14.428086070215176, + "grad_norm": 0.645287275314331, + "learning_rate": 5.410228783181371e-07, + "loss": 0.3582, + "step": 38220 + }, + { + "epoch": 14.431861079652698, + "grad_norm": 0.6621134281158447, + "learning_rate": 5.349752662274444e-07, + "loss": 0.3567, + "step": 38230 + }, + { + "epoch": 14.435636089090222, + "grad_norm": 0.6822801232337952, + "learning_rate": 5.289614630147321e-07, + "loss": 0.3463, + "step": 38240 + }, + { + "epoch": 14.439411098527746, + "grad_norm": 0.6010865569114685, + "learning_rate": 5.229814727903737e-07, + "loss": 0.3603, + "step": 38250 + }, + { + "epoch": 14.44318610796527, + "grad_norm": 0.7438214421272278, + "learning_rate": 5.170352996416439e-07, + "loss": 0.3529, + "step": 38260 + }, + { + "epoch": 14.446961117402793, + "grad_norm": 0.6098829507827759, + "learning_rate": 5.111229476326918e-07, + "loss": 0.3516, + "step": 38270 + }, + { + "epoch": 14.450736126840317, + "grad_norm": 0.6266598105430603, + "learning_rate": 5.052444208045459e-07, + "loss": 0.3569, + "step": 38280 + }, + { + "epoch": 14.45451113627784, + "grad_norm": 0.6430060267448425, + "learning_rate": 4.993997231751312e-07, + "loss": 0.3495, + "step": 38290 + }, + { + "epoch": 14.458286145715364, + "grad_norm": 0.6154651045799255, + "learning_rate": 4.935888587392468e-07, + "loss": 0.3546, + "step": 38300 + }, + { + "epoch": 14.462061155152888, + "grad_norm": 0.6602094173431396, + "learning_rate": 4.878118314685598e-07, + "loss": 0.3539, + "step": 38310 + }, + { + "epoch": 14.465836164590412, + "grad_norm": 0.7772398591041565, + "learning_rate": 4.820686453116119e-07, + "loss": 0.3564, + "step": 38320 + }, + { + "epoch": 14.469611174027936, + "grad_norm": 0.629912257194519, + "learning_rate": 4.763593041938186e-07, + "loss": 0.355, + "step": 38330 + }, + { + "epoch": 14.47338618346546, + "grad_norm": 0.6255464553833008, + "learning_rate": 4.706838120174584e-07, + "loss": 0.3516, + "step": 38340 + }, + { + "epoch": 14.477161192902983, + "grad_norm": 0.6643255949020386, + "learning_rate": 4.650421726616838e-07, + "loss": 0.3522, + "step": 38350 + }, + { + "epoch": 14.480936202340505, + "grad_norm": 0.6950973868370056, + "learning_rate": 4.594343899824993e-07, + "loss": 0.3541, + "step": 38360 + }, + { + "epoch": 14.484711211778029, + "grad_norm": 0.7115216851234436, + "learning_rate": 4.538604678127667e-07, + "loss": 0.3539, + "step": 38370 + }, + { + "epoch": 14.488486221215553, + "grad_norm": 0.6523749232292175, + "learning_rate": 4.483204099622107e-07, + "loss": 0.3555, + "step": 38380 + }, + { + "epoch": 14.492261230653076, + "grad_norm": 0.6457560658454895, + "learning_rate": 4.4281422021740793e-07, + "loss": 0.3583, + "step": 38390 + }, + { + "epoch": 14.4960362400906, + "grad_norm": 0.6862673163414001, + "learning_rate": 4.373419023417924e-07, + "loss": 0.3614, + "step": 38400 + }, + { + "epoch": 14.499811249528124, + "grad_norm": 0.623674750328064, + "learning_rate": 4.319034600756389e-07, + "loss": 0.3476, + "step": 38410 + }, + { + "epoch": 14.503586258965647, + "grad_norm": 0.6614101529121399, + "learning_rate": 4.264988971360739e-07, + "loss": 0.3646, + "step": 38420 + }, + { + "epoch": 14.507361268403171, + "grad_norm": 0.650454044342041, + "learning_rate": 4.211282172170594e-07, + "loss": 0.3581, + "step": 38430 + }, + { + "epoch": 14.511136277840695, + "grad_norm": 2.998237133026123, + "learning_rate": 4.1579142398940894e-07, + "loss": 0.3612, + "step": 38440 + }, + { + "epoch": 14.514911287278219, + "grad_norm": 0.6962543725967407, + "learning_rate": 4.1048852110076585e-07, + "loss": 0.351, + "step": 38450 + }, + { + "epoch": 14.518686296715742, + "grad_norm": 0.6083894371986389, + "learning_rate": 4.052195121756197e-07, + "loss": 0.3617, + "step": 38460 + }, + { + "epoch": 14.522461306153266, + "grad_norm": 0.6192116737365723, + "learning_rate": 3.9998440081528974e-07, + "loss": 0.3561, + "step": 38470 + }, + { + "epoch": 14.52623631559079, + "grad_norm": 0.755118727684021, + "learning_rate": 3.9478319059791936e-07, + "loss": 0.3502, + "step": 38480 + }, + { + "epoch": 14.530011325028312, + "grad_norm": 0.606930673122406, + "learning_rate": 3.8961588507848143e-07, + "loss": 0.3569, + "step": 38490 + }, + { + "epoch": 14.533786334465836, + "grad_norm": 0.6492167115211487, + "learning_rate": 3.844824877887954e-07, + "loss": 0.3622, + "step": 38500 + }, + { + "epoch": 14.53756134390336, + "grad_norm": 0.771045982837677, + "learning_rate": 3.7938300223748224e-07, + "loss": 0.3567, + "step": 38510 + }, + { + "epoch": 14.541336353340883, + "grad_norm": 0.6617088317871094, + "learning_rate": 3.743174319099929e-07, + "loss": 0.3589, + "step": 38520 + }, + { + "epoch": 14.545111362778407, + "grad_norm": 0.7514125108718872, + "learning_rate": 3.6928578026859673e-07, + "loss": 0.3504, + "step": 38530 + }, + { + "epoch": 14.54888637221593, + "grad_norm": 0.797964870929718, + "learning_rate": 3.642880507523816e-07, + "loss": 0.3609, + "step": 38540 + }, + { + "epoch": 14.552661381653454, + "grad_norm": 0.6243863105773926, + "learning_rate": 3.5932424677724285e-07, + "loss": 0.3564, + "step": 38550 + }, + { + "epoch": 14.556436391090978, + "grad_norm": 0.6841285824775696, + "learning_rate": 3.5439437173589996e-07, + "loss": 0.3466, + "step": 38560 + }, + { + "epoch": 14.560211400528502, + "grad_norm": 0.6735568642616272, + "learning_rate": 3.4949842899787423e-07, + "loss": 0.3542, + "step": 38570 + }, + { + "epoch": 14.563986409966025, + "grad_norm": 0.6594800353050232, + "learning_rate": 3.446364219094944e-07, + "loss": 0.3551, + "step": 38580 + }, + { + "epoch": 14.56776141940355, + "grad_norm": 0.6403582692146301, + "learning_rate": 3.3980835379389675e-07, + "loss": 0.3547, + "step": 38590 + }, + { + "epoch": 14.571536428841073, + "grad_norm": 0.7487730979919434, + "learning_rate": 3.350142279510249e-07, + "loss": 0.3494, + "step": 38600 + }, + { + "epoch": 14.575311438278595, + "grad_norm": 0.6844361424446106, + "learning_rate": 3.3025404765761324e-07, + "loss": 0.3558, + "step": 38610 + }, + { + "epoch": 14.579086447716119, + "grad_norm": 0.6371558904647827, + "learning_rate": 3.255278161672037e-07, + "loss": 0.354, + "step": 38620 + }, + { + "epoch": 14.582861457153642, + "grad_norm": 0.6130198836326599, + "learning_rate": 3.20835536710129e-07, + "loss": 0.3504, + "step": 38630 + }, + { + "epoch": 14.586636466591166, + "grad_norm": 0.6150995492935181, + "learning_rate": 3.1617721249351804e-07, + "loss": 0.3579, + "step": 38640 + }, + { + "epoch": 14.59041147602869, + "grad_norm": 0.6839898824691772, + "learning_rate": 3.115528467012963e-07, + "loss": 0.3568, + "step": 38650 + }, + { + "epoch": 14.594186485466214, + "grad_norm": 0.6125872731208801, + "learning_rate": 3.0696244249416305e-07, + "loss": 0.3575, + "step": 38660 + }, + { + "epoch": 14.597961494903737, + "grad_norm": 0.6067168712615967, + "learning_rate": 3.024060030096254e-07, + "loss": 0.3578, + "step": 38670 + }, + { + "epoch": 14.601736504341261, + "grad_norm": 0.6161952614784241, + "learning_rate": 2.9788353136196435e-07, + "loss": 0.3523, + "step": 38680 + }, + { + "epoch": 14.605511513778785, + "grad_norm": 0.6272019147872925, + "learning_rate": 2.9339503064224615e-07, + "loss": 0.3504, + "step": 38690 + }, + { + "epoch": 14.609286523216308, + "grad_norm": 1.410046935081482, + "learning_rate": 2.889405039183224e-07, + "loss": 0.3514, + "step": 38700 + }, + { + "epoch": 14.613061532653832, + "grad_norm": 0.6313207745552063, + "learning_rate": 2.8451995423481315e-07, + "loss": 0.3526, + "step": 38710 + }, + { + "epoch": 14.616836542091356, + "grad_norm": 0.5658713579177856, + "learning_rate": 2.801333846131293e-07, + "loss": 0.3529, + "step": 38720 + }, + { + "epoch": 14.620611551528878, + "grad_norm": 0.6897971034049988, + "learning_rate": 2.757807980514504e-07, + "loss": 0.3529, + "step": 38730 + }, + { + "epoch": 14.624386560966402, + "grad_norm": 0.6845794320106506, + "learning_rate": 2.714621975247189e-07, + "loss": 0.3556, + "step": 38740 + }, + { + "epoch": 14.628161570403925, + "grad_norm": 0.6772444844245911, + "learning_rate": 2.671775859846626e-07, + "loss": 0.3499, + "step": 38750 + }, + { + "epoch": 14.63193657984145, + "grad_norm": 0.5830336213111877, + "learning_rate": 2.629269663597722e-07, + "loss": 0.3561, + "step": 38760 + }, + { + "epoch": 14.635711589278973, + "grad_norm": 1.1887177228927612, + "learning_rate": 2.587103415553127e-07, + "loss": 0.3635, + "step": 38770 + }, + { + "epoch": 14.639486598716497, + "grad_norm": 0.6681121587753296, + "learning_rate": 2.545277144533009e-07, + "loss": 0.3584, + "step": 38780 + }, + { + "epoch": 14.64326160815402, + "grad_norm": 0.6925423741340637, + "learning_rate": 2.5037908791252227e-07, + "loss": 0.352, + "step": 38790 + }, + { + "epoch": 14.647036617591544, + "grad_norm": 0.6650035977363586, + "learning_rate": 2.462644647685308e-07, + "loss": 0.357, + "step": 38800 + }, + { + "epoch": 14.650811627029068, + "grad_norm": 0.7153099775314331, + "learning_rate": 2.421838478336269e-07, + "loss": 0.3473, + "step": 38810 + }, + { + "epoch": 14.654586636466592, + "grad_norm": 0.738753616809845, + "learning_rate": 2.3813723989687398e-07, + "loss": 0.358, + "step": 38820 + }, + { + "epoch": 14.658361645904115, + "grad_norm": 0.6535517573356628, + "learning_rate": 2.3412464372409293e-07, + "loss": 0.36, + "step": 38830 + }, + { + "epoch": 14.662136655341639, + "grad_norm": 0.6161585450172424, + "learning_rate": 2.301460620578455e-07, + "loss": 0.3573, + "step": 38840 + }, + { + "epoch": 14.665911664779163, + "grad_norm": 0.6572152972221375, + "learning_rate": 2.262014976174731e-07, + "loss": 0.349, + "step": 38850 + }, + { + "epoch": 14.669686674216685, + "grad_norm": 0.6994696855545044, + "learning_rate": 2.2229095309903025e-07, + "loss": 0.364, + "step": 38860 + }, + { + "epoch": 14.673461683654208, + "grad_norm": 0.6057741641998291, + "learning_rate": 2.184144311753511e-07, + "loss": 0.3513, + "step": 38870 + }, + { + "epoch": 14.677236693091732, + "grad_norm": 0.6180564165115356, + "learning_rate": 2.14571934495994e-07, + "loss": 0.3535, + "step": 38880 + }, + { + "epoch": 14.681011702529256, + "grad_norm": 0.7175959944725037, + "learning_rate": 2.1076346568726923e-07, + "loss": 0.3528, + "step": 38890 + }, + { + "epoch": 14.68478671196678, + "grad_norm": 0.7563894391059875, + "learning_rate": 2.0698902735223346e-07, + "loss": 0.3621, + "step": 38900 + }, + { + "epoch": 14.688561721404303, + "grad_norm": 0.6452277302742004, + "learning_rate": 2.0324862207067863e-07, + "loss": 0.3579, + "step": 38910 + }, + { + "epoch": 14.692336730841827, + "grad_norm": 0.613015353679657, + "learning_rate": 1.9954225239913195e-07, + "loss": 0.353, + "step": 38920 + }, + { + "epoch": 14.69611174027935, + "grad_norm": 0.5917133688926697, + "learning_rate": 1.9586992087087252e-07, + "loss": 0.3599, + "step": 38930 + }, + { + "epoch": 14.699886749716875, + "grad_norm": 0.61355060338974, + "learning_rate": 1.9223162999588705e-07, + "loss": 0.3512, + "step": 38940 + }, + { + "epoch": 14.703661759154398, + "grad_norm": 0.6601895689964294, + "learning_rate": 1.886273822609308e-07, + "loss": 0.3498, + "step": 38950 + }, + { + "epoch": 14.707436768591922, + "grad_norm": 0.6723988056182861, + "learning_rate": 1.8505718012946094e-07, + "loss": 0.3566, + "step": 38960 + }, + { + "epoch": 14.711211778029446, + "grad_norm": 0.5990282893180847, + "learning_rate": 1.8152102604168664e-07, + "loss": 0.3544, + "step": 38970 + }, + { + "epoch": 14.71498678746697, + "grad_norm": 0.7967637181282043, + "learning_rate": 1.7801892241452457e-07, + "loss": 0.3581, + "step": 38980 + }, + { + "epoch": 14.718761796904491, + "grad_norm": 0.7710434198379517, + "learning_rate": 1.745508716416433e-07, + "loss": 0.3552, + "step": 38990 + }, + { + "epoch": 14.722536806342015, + "grad_norm": 0.7385621666908264, + "learning_rate": 1.7111687609341342e-07, + "loss": 0.3571, + "step": 39000 + }, + { + "epoch": 14.726311815779539, + "grad_norm": 0.6296396851539612, + "learning_rate": 1.6771693811693523e-07, + "loss": 0.358, + "step": 39010 + }, + { + "epoch": 14.730086825217063, + "grad_norm": 0.6770869493484497, + "learning_rate": 1.643510600360443e-07, + "loss": 0.3515, + "step": 39020 + }, + { + "epoch": 14.733861834654586, + "grad_norm": 0.6164085865020752, + "learning_rate": 1.610192441512781e-07, + "loss": 0.3659, + "step": 39030 + }, + { + "epoch": 14.73763684409211, + "grad_norm": 1.3167994022369385, + "learning_rate": 1.5772149273990956e-07, + "loss": 0.3633, + "step": 39040 + }, + { + "epoch": 14.741411853529634, + "grad_norm": 0.6471686959266663, + "learning_rate": 1.5445780805591337e-07, + "loss": 0.3429, + "step": 39050 + }, + { + "epoch": 14.745186862967158, + "grad_norm": 0.6361897587776184, + "learning_rate": 1.5122819232998852e-07, + "loss": 0.3591, + "step": 39060 + }, + { + "epoch": 14.748961872404681, + "grad_norm": 0.6572710275650024, + "learning_rate": 1.4803264776955262e-07, + "loss": 0.3654, + "step": 39070 + }, + { + "epoch": 14.752736881842205, + "grad_norm": 0.6715047359466553, + "learning_rate": 1.448711765587196e-07, + "loss": 0.3587, + "step": 39080 + }, + { + "epoch": 14.756511891279729, + "grad_norm": 0.6003199219703674, + "learning_rate": 1.4174378085833885e-07, + "loss": 0.3599, + "step": 39090 + }, + { + "epoch": 14.760286900717253, + "grad_norm": 0.637787938117981, + "learning_rate": 1.3865046280594486e-07, + "loss": 0.3602, + "step": 39100 + }, + { + "epoch": 14.764061910154776, + "grad_norm": 0.6328327655792236, + "learning_rate": 1.3559122451579643e-07, + "loss": 0.354, + "step": 39110 + }, + { + "epoch": 14.767836919592298, + "grad_norm": 0.6023069024085999, + "learning_rate": 1.3256606807884874e-07, + "loss": 0.3549, + "step": 39120 + }, + { + "epoch": 14.771611929029822, + "grad_norm": 0.631928563117981, + "learning_rate": 1.2957499556277563e-07, + "loss": 0.3608, + "step": 39130 + }, + { + "epoch": 14.775386938467346, + "grad_norm": 0.6594300866127014, + "learning_rate": 1.266180090119473e-07, + "loss": 0.3556, + "step": 39140 + }, + { + "epoch": 14.77916194790487, + "grad_norm": 0.6091174483299255, + "learning_rate": 1.2369511044742488e-07, + "loss": 0.3608, + "step": 39150 + }, + { + "epoch": 14.782936957342393, + "grad_norm": 0.7767238020896912, + "learning_rate": 1.2080630186698805e-07, + "loss": 0.3599, + "step": 39160 + }, + { + "epoch": 14.786711966779917, + "grad_norm": 0.6567763090133667, + "learning_rate": 1.1795158524511851e-07, + "loss": 0.3513, + "step": 39170 + }, + { + "epoch": 14.79048697621744, + "grad_norm": 0.6258882880210876, + "learning_rate": 1.1513096253297773e-07, + "loss": 0.3553, + "step": 39180 + }, + { + "epoch": 14.794261985654964, + "grad_norm": 0.6142001152038574, + "learning_rate": 1.1234443565843466e-07, + "loss": 0.3534, + "step": 39190 + }, + { + "epoch": 14.798036995092488, + "grad_norm": 0.6969514489173889, + "learning_rate": 1.0959200652606028e-07, + "loss": 0.3575, + "step": 39200 + }, + { + "epoch": 14.801812004530012, + "grad_norm": 0.729709804058075, + "learning_rate": 1.0687367701711082e-07, + "loss": 0.347, + "step": 39210 + }, + { + "epoch": 14.805587013967536, + "grad_norm": 0.6622199416160583, + "learning_rate": 1.0418944898953898e-07, + "loss": 0.3567, + "step": 39220 + }, + { + "epoch": 14.809362023405058, + "grad_norm": 0.7699572443962097, + "learning_rate": 1.0153932427798829e-07, + "loss": 0.3591, + "step": 39230 + }, + { + "epoch": 14.813137032842581, + "grad_norm": 0.6490731239318848, + "learning_rate": 9.892330469379874e-08, + "loss": 0.359, + "step": 39240 + }, + { + "epoch": 14.816912042280105, + "grad_norm": 0.6460186839103699, + "learning_rate": 9.634139202498449e-08, + "loss": 0.3545, + "step": 39250 + }, + { + "epoch": 14.820687051717629, + "grad_norm": 0.789674699306488, + "learning_rate": 9.379358803627281e-08, + "loss": 0.3515, + "step": 39260 + }, + { + "epoch": 14.824462061155153, + "grad_norm": 0.6353222131729126, + "learning_rate": 9.12798944690485e-08, + "loss": 0.354, + "step": 39270 + }, + { + "epoch": 14.828237070592676, + "grad_norm": 0.6650903820991516, + "learning_rate": 8.880031304141501e-08, + "loss": 0.349, + "step": 39280 + }, + { + "epoch": 14.8320120800302, + "grad_norm": 0.7060394287109375, + "learning_rate": 8.635484544812777e-08, + "loss": 0.3572, + "step": 39290 + }, + { + "epoch": 14.835787089467724, + "grad_norm": 2.0902068614959717, + "learning_rate": 8.394349336064422e-08, + "loss": 0.3551, + "step": 39300 + }, + { + "epoch": 14.839562098905247, + "grad_norm": 0.7247877717018127, + "learning_rate": 8.156625842710153e-08, + "loss": 0.3518, + "step": 39310 + }, + { + "epoch": 14.843337108342771, + "grad_norm": 0.6578109264373779, + "learning_rate": 7.92231422723222e-08, + "loss": 0.3499, + "step": 39320 + }, + { + "epoch": 14.847112117780295, + "grad_norm": 0.630346953868866, + "learning_rate": 7.691414649780293e-08, + "loss": 0.351, + "step": 39330 + }, + { + "epoch": 14.850887127217819, + "grad_norm": 0.6793875098228455, + "learning_rate": 7.463927268171467e-08, + "loss": 0.3486, + "step": 39340 + }, + { + "epoch": 14.854662136655342, + "grad_norm": 0.6113439798355103, + "learning_rate": 7.239852237891365e-08, + "loss": 0.3605, + "step": 39350 + }, + { + "epoch": 14.858437146092864, + "grad_norm": 0.7049407958984375, + "learning_rate": 7.019189712093588e-08, + "loss": 0.3501, + "step": 39360 + }, + { + "epoch": 14.862212155530388, + "grad_norm": 0.7017509937286377, + "learning_rate": 6.801939841598603e-08, + "loss": 0.3599, + "step": 39370 + }, + { + "epoch": 14.865987164967912, + "grad_norm": 0.6390900015830994, + "learning_rate": 6.588102774894856e-08, + "loss": 0.3557, + "step": 39380 + }, + { + "epoch": 14.869762174405436, + "grad_norm": 0.6362788081169128, + "learning_rate": 6.377678658137653e-08, + "loss": 0.3548, + "step": 39390 + }, + { + "epoch": 14.87353718384296, + "grad_norm": 0.6280719041824341, + "learning_rate": 6.17066763514973e-08, + "loss": 0.3493, + "step": 39400 + }, + { + "epoch": 14.877312193280483, + "grad_norm": 0.6526157855987549, + "learning_rate": 5.967069847421792e-08, + "loss": 0.3532, + "step": 39410 + }, + { + "epoch": 14.881087202718007, + "grad_norm": 0.7428958415985107, + "learning_rate": 5.7668854341108583e-08, + "loss": 0.3563, + "step": 39420 + }, + { + "epoch": 14.88486221215553, + "grad_norm": 0.6039178371429443, + "learning_rate": 5.570114532040815e-08, + "loss": 0.3567, + "step": 39430 + }, + { + "epoch": 14.888637221593054, + "grad_norm": 1.0244649648666382, + "learning_rate": 5.3767572757029706e-08, + "loss": 0.3521, + "step": 39440 + }, + { + "epoch": 14.892412231030578, + "grad_norm": 0.6426689028739929, + "learning_rate": 5.186813797254386e-08, + "loss": 0.3525, + "step": 39450 + }, + { + "epoch": 14.896187240468102, + "grad_norm": 0.7617814540863037, + "learning_rate": 5.000284226521213e-08, + "loss": 0.3529, + "step": 39460 + }, + { + "epoch": 14.899962249905625, + "grad_norm": 0.7546370625495911, + "learning_rate": 4.8171686909931346e-08, + "loss": 0.3519, + "step": 39470 + }, + { + "epoch": 14.90373725934315, + "grad_norm": 0.6495998501777649, + "learning_rate": 4.637467315828925e-08, + "loss": 0.358, + "step": 39480 + }, + { + "epoch": 14.907512268780671, + "grad_norm": 0.6548603177070618, + "learning_rate": 4.461180223852002e-08, + "loss": 0.3483, + "step": 39490 + }, + { + "epoch": 14.911287278218195, + "grad_norm": 0.6339406371116638, + "learning_rate": 4.288307535553204e-08, + "loss": 0.3604, + "step": 39500 + }, + { + "epoch": 14.915062287655719, + "grad_norm": 0.6532941460609436, + "learning_rate": 4.1188493690896836e-08, + "loss": 0.3598, + "step": 39510 + }, + { + "epoch": 14.918837297093242, + "grad_norm": 0.6393380761146545, + "learning_rate": 3.952805840284346e-08, + "loss": 0.3521, + "step": 39520 + }, + { + "epoch": 14.922612306530766, + "grad_norm": 0.6625932455062866, + "learning_rate": 3.790177062625855e-08, + "loss": 0.3567, + "step": 39530 + }, + { + "epoch": 14.92638731596829, + "grad_norm": 0.6500564217567444, + "learning_rate": 3.6309631472702946e-08, + "loss": 0.3549, + "step": 39540 + }, + { + "epoch": 14.930162325405814, + "grad_norm": 0.6394986510276794, + "learning_rate": 3.475164203038395e-08, + "loss": 0.3602, + "step": 39550 + }, + { + "epoch": 14.933937334843337, + "grad_norm": 0.6065540313720703, + "learning_rate": 3.3227803364171974e-08, + "loss": 0.3435, + "step": 39560 + }, + { + "epoch": 14.937712344280861, + "grad_norm": 0.6980146765708923, + "learning_rate": 3.173811651560055e-08, + "loss": 0.3484, + "step": 39570 + }, + { + "epoch": 14.941487353718385, + "grad_norm": 0.6110873222351074, + "learning_rate": 3.0282582502855204e-08, + "loss": 0.3528, + "step": 39580 + }, + { + "epoch": 14.945262363155909, + "grad_norm": 0.74271559715271, + "learning_rate": 2.8861202320773496e-08, + "loss": 0.3547, + "step": 39590 + }, + { + "epoch": 14.949037372593432, + "grad_norm": 0.5941160321235657, + "learning_rate": 2.7473976940867175e-08, + "loss": 0.3606, + "step": 39600 + }, + { + "epoch": 14.952812382030956, + "grad_norm": 0.6520451903343201, + "learning_rate": 2.6120907311277808e-08, + "loss": 0.351, + "step": 39610 + }, + { + "epoch": 14.956587391468478, + "grad_norm": 0.6158818006515503, + "learning_rate": 2.480199435682673e-08, + "loss": 0.3502, + "step": 39620 + }, + { + "epoch": 14.960362400906002, + "grad_norm": 0.6889585852622986, + "learning_rate": 2.3517238978976174e-08, + "loss": 0.3592, + "step": 39630 + }, + { + "epoch": 14.964137410343525, + "grad_norm": 0.5970800518989563, + "learning_rate": 2.2266642055834842e-08, + "loss": 0.359, + "step": 39640 + }, + { + "epoch": 14.96791241978105, + "grad_norm": 0.65416419506073, + "learning_rate": 2.105020444219119e-08, + "loss": 0.3527, + "step": 39650 + }, + { + "epoch": 14.971687429218573, + "grad_norm": 0.664504885673523, + "learning_rate": 1.9867926969452388e-08, + "loss": 0.364, + "step": 39660 + }, + { + "epoch": 14.975462438656097, + "grad_norm": 0.6672757863998413, + "learning_rate": 1.8719810445705366e-08, + "loss": 0.3531, + "step": 39670 + }, + { + "epoch": 14.97923744809362, + "grad_norm": 0.6902869343757629, + "learning_rate": 1.760585565567241e-08, + "loss": 0.3564, + "step": 39680 + }, + { + "epoch": 14.983012457531144, + "grad_norm": 0.7175024151802063, + "learning_rate": 1.6526063360727816e-08, + "loss": 0.3559, + "step": 39690 + }, + { + "epoch": 14.986787466968668, + "grad_norm": 0.6465026140213013, + "learning_rate": 1.5480434298908997e-08, + "loss": 0.3539, + "step": 39700 + }, + { + "epoch": 14.990562476406192, + "grad_norm": 0.6712453961372375, + "learning_rate": 1.4468969184883164e-08, + "loss": 0.3535, + "step": 39710 + }, + { + "epoch": 14.994337485843715, + "grad_norm": 0.620090126991272, + "learning_rate": 1.3491668709980643e-08, + "loss": 0.3544, + "step": 39720 + }, + { + "epoch": 14.998112495281239, + "grad_norm": 0.6814067363739014, + "learning_rate": 1.254853354217822e-08, + "loss": 0.3551, + "step": 39730 + }, + { + "epoch": 15.001887504718761, + "grad_norm": 0.8120419979095459, + "learning_rate": 1.1639564326099139e-08, + "loss": 0.3553, + "step": 39740 + }, + { + "epoch": 15.005662514156285, + "grad_norm": 0.6603171825408936, + "learning_rate": 1.0764761683018653e-08, + "loss": 0.3537, + "step": 39750 + }, + { + "epoch": 15.009437523593808, + "grad_norm": 0.6315841674804688, + "learning_rate": 9.924126210852925e-09, + "loss": 0.3503, + "step": 39760 + }, + { + "epoch": 15.013212533031332, + "grad_norm": 0.6496405601501465, + "learning_rate": 9.117658484170122e-09, + "loss": 0.3577, + "step": 39770 + }, + { + "epoch": 15.016987542468856, + "grad_norm": 0.7022120356559753, + "learning_rate": 8.345359054179324e-09, + "loss": 0.3678, + "step": 39780 + }, + { + "epoch": 15.02076255190638, + "grad_norm": 0.7122491598129272, + "learning_rate": 7.60722844874162e-09, + "loss": 0.3561, + "step": 39790 + }, + { + "epoch": 15.024537561343903, + "grad_norm": 0.625100314617157, + "learning_rate": 6.903267172364558e-09, + "loss": 0.347, + "step": 39800 + }, + { + "epoch": 15.028312570781427, + "grad_norm": 0.6396900415420532, + "learning_rate": 6.23347570619659e-09, + "loss": 0.3605, + "step": 39810 + }, + { + "epoch": 15.03208758021895, + "grad_norm": 0.5994987487792969, + "learning_rate": 5.59785450803818e-09, + "loss": 0.3573, + "step": 39820 + }, + { + "epoch": 15.035862589656475, + "grad_norm": 0.7558580636978149, + "learning_rate": 4.996404012325151e-09, + "loss": 0.3536, + "step": 39830 + }, + { + "epoch": 15.039637599093998, + "grad_norm": 0.6943445801734924, + "learning_rate": 4.429124630145331e-09, + "loss": 0.3561, + "step": 39840 + }, + { + "epoch": 15.043412608531522, + "grad_norm": 0.9574018716812134, + "learning_rate": 3.89601674922746e-09, + "loss": 0.362, + "step": 39850 + }, + { + "epoch": 15.047187617969046, + "grad_norm": 0.6449063420295715, + "learning_rate": 3.3970807339467336e-09, + "loss": 0.3537, + "step": 39860 + }, + { + "epoch": 15.050962627406568, + "grad_norm": 0.5976826548576355, + "learning_rate": 2.9323169253248073e-09, + "loss": 0.3513, + "step": 39870 + }, + { + "epoch": 15.054737636844091, + "grad_norm": 0.6099610328674316, + "learning_rate": 2.501725641018693e-09, + "loss": 0.3545, + "step": 39880 + }, + { + "epoch": 15.058512646281615, + "grad_norm": 0.7549223899841309, + "learning_rate": 2.105307175331861e-09, + "loss": 0.3567, + "step": 39890 + }, + { + "epoch": 15.062287655719139, + "grad_norm": 0.6407955288887024, + "learning_rate": 1.7430617992197918e-09, + "loss": 0.3558, + "step": 39900 + }, + { + "epoch": 15.066062665156663, + "grad_norm": 0.6533637046813965, + "learning_rate": 1.4149897602677708e-09, + "loss": 0.3484, + "step": 39910 + }, + { + "epoch": 15.069837674594186, + "grad_norm": 0.604796826839447, + "learning_rate": 1.1210912827130937e-09, + "loss": 0.3582, + "step": 39920 + }, + { + "epoch": 15.07361268403171, + "grad_norm": 0.5980611443519592, + "learning_rate": 8.613665674284121e-10, + "loss": 0.3548, + "step": 39930 + }, + { + "epoch": 15.077387693469234, + "grad_norm": 0.6383752226829529, + "learning_rate": 6.358157919383878e-10, + "loss": 0.3544, + "step": 39940 + }, + { + "epoch": 15.081162702906758, + "grad_norm": 0.7199327349662781, + "learning_rate": 4.4443911040303877e-10, + "loss": 0.3526, + "step": 39950 + }, + { + "epoch": 15.084937712344281, + "grad_norm": 0.7080827951431274, + "learning_rate": 2.8723665362329066e-10, + "loss": 0.3549, + "step": 39960 + }, + { + "epoch": 15.088712721781805, + "grad_norm": 0.6483982801437378, + "learning_rate": 1.6420852905207895e-10, + "loss": 0.3583, + "step": 39970 + }, + { + "epoch": 15.092487731219329, + "grad_norm": 0.6438269019126892, + "learning_rate": 7.535482076659328e-11, + "loss": 0.3451, + "step": 39980 + }, + { + "epoch": 15.09626274065685, + "grad_norm": 0.6255228519439697, + "learning_rate": 2.0675589512686443e-11, + "loss": 0.3557, + "step": 39990 + }, + { + "epoch": 15.100037750094375, + "grad_norm": 0.8365612626075745, + "learning_rate": 1.7087265491433358e-13, + "loss": 0.353, + "step": 40000 + } + ], + "logging_steps": 10, + "max_steps": 40000, + "num_input_tokens_seen": 0, + "num_train_epochs": 16, + "save_steps": 4000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.0739727642197072e+20, + "train_batch_size": 32, + "trial_name": null, + "trial_params": null +}