{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 31.746031746031747, "eval_steps": 500, "global_step": 30000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010582010582010581, "grad_norm": 8.936141014099121, "learning_rate": 6.000000000000001e-07, "loss": 0.8416, "step": 10 }, { "epoch": 0.021164021164021163, "grad_norm": 9.36695671081543, "learning_rate": 1.2666666666666667e-06, "loss": 0.8339, "step": 20 }, { "epoch": 0.031746031746031744, "grad_norm": 5.117445468902588, "learning_rate": 1.9333333333333336e-06, "loss": 0.6132, "step": 30 }, { "epoch": 0.042328042328042326, "grad_norm": 1.8590303659439087, "learning_rate": 2.6e-06, "loss": 0.3853, "step": 40 }, { "epoch": 0.05291005291005291, "grad_norm": 2.014482021331787, "learning_rate": 3.2666666666666666e-06, "loss": 0.3082, "step": 50 }, { "epoch": 0.06349206349206349, "grad_norm": 1.1662486791610718, "learning_rate": 3.9333333333333335e-06, "loss": 0.2301, "step": 60 }, { "epoch": 0.07407407407407407, "grad_norm": 1.1701092720031738, "learning_rate": 4.6e-06, "loss": 0.196, "step": 70 }, { "epoch": 0.08465608465608465, "grad_norm": 0.814723551273346, "learning_rate": 5.266666666666667e-06, "loss": 0.1607, "step": 80 }, { "epoch": 0.09523809523809523, "grad_norm": 0.9235273599624634, "learning_rate": 5.933333333333334e-06, "loss": 0.1452, "step": 90 }, { "epoch": 0.10582010582010581, "grad_norm": 0.8884081840515137, "learning_rate": 6.6e-06, "loss": 0.1274, "step": 100 }, { "epoch": 0.1164021164021164, "grad_norm": 0.5123031139373779, "learning_rate": 7.266666666666668e-06, "loss": 0.1246, "step": 110 }, { "epoch": 0.12698412698412698, "grad_norm": 0.5364177823066711, "learning_rate": 7.933333333333334e-06, "loss": 0.1129, "step": 120 }, { "epoch": 0.13756613756613756, "grad_norm": 0.579659104347229, "learning_rate": 8.599999999999999e-06, "loss": 0.1031, "step": 130 }, { "epoch": 0.14814814814814814, "grad_norm": 0.8478009104728699, "learning_rate": 9.266666666666667e-06, "loss": 0.0944, "step": 140 }, { "epoch": 0.15873015873015872, "grad_norm": 0.591192364692688, "learning_rate": 9.933333333333334e-06, "loss": 0.0949, "step": 150 }, { "epoch": 0.1693121693121693, "grad_norm": 0.7813432216644287, "learning_rate": 1.06e-05, "loss": 0.0919, "step": 160 }, { "epoch": 0.17989417989417988, "grad_norm": 0.9157405495643616, "learning_rate": 1.1266666666666667e-05, "loss": 0.0868, "step": 170 }, { "epoch": 0.19047619047619047, "grad_norm": 1.449837327003479, "learning_rate": 1.1933333333333333e-05, "loss": 0.0854, "step": 180 }, { "epoch": 0.20105820105820105, "grad_norm": 0.7559556365013123, "learning_rate": 1.2600000000000001e-05, "loss": 0.0818, "step": 190 }, { "epoch": 0.21164021164021163, "grad_norm": 0.5587271451950073, "learning_rate": 1.3266666666666666e-05, "loss": 0.0825, "step": 200 }, { "epoch": 0.2222222222222222, "grad_norm": 0.7865070700645447, "learning_rate": 1.3933333333333334e-05, "loss": 0.0809, "step": 210 }, { "epoch": 0.2328042328042328, "grad_norm": 0.3796408176422119, "learning_rate": 1.4599999999999999e-05, "loss": 0.075, "step": 220 }, { "epoch": 0.24338624338624337, "grad_norm": 1.1076023578643799, "learning_rate": 1.5266666666666667e-05, "loss": 0.0757, "step": 230 }, { "epoch": 0.25396825396825395, "grad_norm": 0.6831932067871094, "learning_rate": 1.5933333333333332e-05, "loss": 0.0767, "step": 240 }, { "epoch": 0.26455026455026454, "grad_norm": 0.425535649061203, "learning_rate": 1.66e-05, "loss": 0.0699, "step": 250 }, { "epoch": 0.2751322751322751, "grad_norm": 0.6111235022544861, "learning_rate": 1.726666666666667e-05, "loss": 0.071, "step": 260 }, { "epoch": 0.2857142857142857, "grad_norm": 0.4874987006187439, "learning_rate": 1.7933333333333337e-05, "loss": 0.072, "step": 270 }, { "epoch": 0.2962962962962963, "grad_norm": 0.39111024141311646, "learning_rate": 1.86e-05, "loss": 0.0639, "step": 280 }, { "epoch": 0.30687830687830686, "grad_norm": 0.3845563232898712, "learning_rate": 1.926666666666667e-05, "loss": 0.0661, "step": 290 }, { "epoch": 0.31746031746031744, "grad_norm": 0.6963057518005371, "learning_rate": 1.9933333333333334e-05, "loss": 0.0728, "step": 300 }, { "epoch": 0.328042328042328, "grad_norm": 0.5722204446792603, "learning_rate": 2.06e-05, "loss": 0.0643, "step": 310 }, { "epoch": 0.3386243386243386, "grad_norm": 0.6162781715393066, "learning_rate": 2.1266666666666667e-05, "loss": 0.0635, "step": 320 }, { "epoch": 0.3492063492063492, "grad_norm": 0.789279043674469, "learning_rate": 2.1933333333333332e-05, "loss": 0.0708, "step": 330 }, { "epoch": 0.35978835978835977, "grad_norm": 0.8498165607452393, "learning_rate": 2.26e-05, "loss": 0.0665, "step": 340 }, { "epoch": 0.37037037037037035, "grad_norm": 0.4812783896923065, "learning_rate": 2.326666666666667e-05, "loss": 0.0632, "step": 350 }, { "epoch": 0.38095238095238093, "grad_norm": 0.6319581866264343, "learning_rate": 2.3933333333333337e-05, "loss": 0.0697, "step": 360 }, { "epoch": 0.3915343915343915, "grad_norm": 0.5971472263336182, "learning_rate": 2.46e-05, "loss": 0.0546, "step": 370 }, { "epoch": 0.4021164021164021, "grad_norm": 0.7531677484512329, "learning_rate": 2.5266666666666666e-05, "loss": 0.0632, "step": 380 }, { "epoch": 0.4126984126984127, "grad_norm": 0.5624972581863403, "learning_rate": 2.5933333333333338e-05, "loss": 0.0623, "step": 390 }, { "epoch": 0.42328042328042326, "grad_norm": 0.4891939163208008, "learning_rate": 2.6600000000000003e-05, "loss": 0.0603, "step": 400 }, { "epoch": 0.43386243386243384, "grad_norm": 0.39211660623550415, "learning_rate": 2.7266666666666668e-05, "loss": 0.0594, "step": 410 }, { "epoch": 0.4444444444444444, "grad_norm": 0.5586352348327637, "learning_rate": 2.7933333333333332e-05, "loss": 0.0559, "step": 420 }, { "epoch": 0.455026455026455, "grad_norm": 0.5189799070358276, "learning_rate": 2.86e-05, "loss": 0.0574, "step": 430 }, { "epoch": 0.4656084656084656, "grad_norm": 0.48110270500183105, "learning_rate": 2.926666666666667e-05, "loss": 0.0585, "step": 440 }, { "epoch": 0.47619047619047616, "grad_norm": 0.4026082456111908, "learning_rate": 2.9933333333333337e-05, "loss": 0.0549, "step": 450 }, { "epoch": 0.48677248677248675, "grad_norm": 0.7211326360702515, "learning_rate": 3.06e-05, "loss": 0.0515, "step": 460 }, { "epoch": 0.4973544973544973, "grad_norm": 0.6826295852661133, "learning_rate": 3.126666666666666e-05, "loss": 0.0571, "step": 470 }, { "epoch": 0.5079365079365079, "grad_norm": 0.35797232389450073, "learning_rate": 3.1933333333333335e-05, "loss": 0.0579, "step": 480 }, { "epoch": 0.5185185185185185, "grad_norm": 0.657314121723175, "learning_rate": 3.26e-05, "loss": 0.0531, "step": 490 }, { "epoch": 0.5291005291005291, "grad_norm": 0.6749728918075562, "learning_rate": 3.326666666666667e-05, "loss": 0.0545, "step": 500 }, { "epoch": 0.5396825396825397, "grad_norm": 0.6911066770553589, "learning_rate": 3.3933333333333336e-05, "loss": 0.0525, "step": 510 }, { "epoch": 0.5502645502645502, "grad_norm": 0.4255806803703308, "learning_rate": 3.46e-05, "loss": 0.0542, "step": 520 }, { "epoch": 0.5608465608465608, "grad_norm": 0.5804958343505859, "learning_rate": 3.526666666666667e-05, "loss": 0.0531, "step": 530 }, { "epoch": 0.5714285714285714, "grad_norm": 0.7385868430137634, "learning_rate": 3.593333333333334e-05, "loss": 0.0557, "step": 540 }, { "epoch": 0.582010582010582, "grad_norm": 0.4954354465007782, "learning_rate": 3.66e-05, "loss": 0.0457, "step": 550 }, { "epoch": 0.5925925925925926, "grad_norm": 0.4444650709629059, "learning_rate": 3.726666666666667e-05, "loss": 0.0499, "step": 560 }, { "epoch": 0.6031746031746031, "grad_norm": 0.5095195770263672, "learning_rate": 3.793333333333334e-05, "loss": 0.052, "step": 570 }, { "epoch": 0.6137566137566137, "grad_norm": 0.41383302211761475, "learning_rate": 3.86e-05, "loss": 0.0486, "step": 580 }, { "epoch": 0.6243386243386243, "grad_norm": 0.4939655661582947, "learning_rate": 3.926666666666667e-05, "loss": 0.0492, "step": 590 }, { "epoch": 0.6349206349206349, "grad_norm": 0.5334852933883667, "learning_rate": 3.993333333333333e-05, "loss": 0.0507, "step": 600 }, { "epoch": 0.6455026455026455, "grad_norm": 0.5378982424736023, "learning_rate": 4.0600000000000004e-05, "loss": 0.049, "step": 610 }, { "epoch": 0.656084656084656, "grad_norm": 0.5856459140777588, "learning_rate": 4.126666666666667e-05, "loss": 0.0472, "step": 620 }, { "epoch": 0.6666666666666666, "grad_norm": 0.5753624439239502, "learning_rate": 4.1933333333333334e-05, "loss": 0.0499, "step": 630 }, { "epoch": 0.6772486772486772, "grad_norm": 0.48166850209236145, "learning_rate": 4.26e-05, "loss": 0.0511, "step": 640 }, { "epoch": 0.6878306878306878, "grad_norm": 0.6529409885406494, "learning_rate": 4.3266666666666664e-05, "loss": 0.0574, "step": 650 }, { "epoch": 0.6984126984126984, "grad_norm": 0.5158888697624207, "learning_rate": 4.3933333333333335e-05, "loss": 0.0534, "step": 660 }, { "epoch": 0.708994708994709, "grad_norm": 0.3487018346786499, "learning_rate": 4.46e-05, "loss": 0.0453, "step": 670 }, { "epoch": 0.7195767195767195, "grad_norm": 0.5029096603393555, "learning_rate": 4.526666666666667e-05, "loss": 0.0459, "step": 680 }, { "epoch": 0.7301587301587301, "grad_norm": 0.443412184715271, "learning_rate": 4.5933333333333336e-05, "loss": 0.0422, "step": 690 }, { "epoch": 0.7407407407407407, "grad_norm": 0.4283904433250427, "learning_rate": 4.660000000000001e-05, "loss": 0.0427, "step": 700 }, { "epoch": 0.7513227513227513, "grad_norm": 0.606509804725647, "learning_rate": 4.726666666666667e-05, "loss": 0.0463, "step": 710 }, { "epoch": 0.7619047619047619, "grad_norm": 0.4075102210044861, "learning_rate": 4.793333333333334e-05, "loss": 0.0445, "step": 720 }, { "epoch": 0.7724867724867724, "grad_norm": 0.44399550557136536, "learning_rate": 4.86e-05, "loss": 0.0451, "step": 730 }, { "epoch": 0.783068783068783, "grad_norm": 0.5372384786605835, "learning_rate": 4.926666666666667e-05, "loss": 0.0477, "step": 740 }, { "epoch": 0.7936507936507936, "grad_norm": 0.47048407793045044, "learning_rate": 4.993333333333334e-05, "loss": 0.0437, "step": 750 }, { "epoch": 0.8042328042328042, "grad_norm": 0.4803128242492676, "learning_rate": 5.0600000000000003e-05, "loss": 0.0456, "step": 760 }, { "epoch": 0.8148148148148148, "grad_norm": 0.4746180772781372, "learning_rate": 5.1266666666666675e-05, "loss": 0.0524, "step": 770 }, { "epoch": 0.8253968253968254, "grad_norm": 0.5961106419563293, "learning_rate": 5.193333333333333e-05, "loss": 0.0529, "step": 780 }, { "epoch": 0.8359788359788359, "grad_norm": 0.43920084834098816, "learning_rate": 5.2600000000000005e-05, "loss": 0.0477, "step": 790 }, { "epoch": 0.8465608465608465, "grad_norm": 0.4560076594352722, "learning_rate": 5.326666666666666e-05, "loss": 0.0446, "step": 800 }, { "epoch": 0.8571428571428571, "grad_norm": 0.4857955873012543, "learning_rate": 5.3933333333333334e-05, "loss": 0.0413, "step": 810 }, { "epoch": 0.8677248677248677, "grad_norm": 0.7737778425216675, "learning_rate": 5.4600000000000006e-05, "loss": 0.0486, "step": 820 }, { "epoch": 0.8783068783068783, "grad_norm": 0.42554351687431335, "learning_rate": 5.5266666666666664e-05, "loss": 0.0457, "step": 830 }, { "epoch": 0.8888888888888888, "grad_norm": 0.3324168920516968, "learning_rate": 5.5933333333333335e-05, "loss": 0.0443, "step": 840 }, { "epoch": 0.8994708994708994, "grad_norm": 0.4079880118370056, "learning_rate": 5.66e-05, "loss": 0.0475, "step": 850 }, { "epoch": 0.91005291005291, "grad_norm": 0.4060775935649872, "learning_rate": 5.726666666666667e-05, "loss": 0.0445, "step": 860 }, { "epoch": 0.9206349206349206, "grad_norm": 0.30285122990608215, "learning_rate": 5.7933333333333337e-05, "loss": 0.0415, "step": 870 }, { "epoch": 0.9312169312169312, "grad_norm": 0.5555624961853027, "learning_rate": 5.86e-05, "loss": 0.049, "step": 880 }, { "epoch": 0.9417989417989417, "grad_norm": 0.5379679203033447, "learning_rate": 5.926666666666667e-05, "loss": 0.0482, "step": 890 }, { "epoch": 0.9523809523809523, "grad_norm": 0.32665514945983887, "learning_rate": 5.9933333333333345e-05, "loss": 0.0455, "step": 900 }, { "epoch": 0.9629629629629629, "grad_norm": 0.5463411211967468, "learning_rate": 6.06e-05, "loss": 0.0416, "step": 910 }, { "epoch": 0.9735449735449735, "grad_norm": 0.3586346507072449, "learning_rate": 6.126666666666667e-05, "loss": 0.0401, "step": 920 }, { "epoch": 0.9841269841269841, "grad_norm": 0.4292141795158386, "learning_rate": 6.193333333333333e-05, "loss": 0.0428, "step": 930 }, { "epoch": 0.9947089947089947, "grad_norm": 0.5075591206550598, "learning_rate": 6.26e-05, "loss": 0.0431, "step": 940 }, { "epoch": 1.0052910052910053, "grad_norm": 0.468470960855484, "learning_rate": 6.326666666666667e-05, "loss": 0.044, "step": 950 }, { "epoch": 1.0158730158730158, "grad_norm": 0.5031815767288208, "learning_rate": 6.393333333333333e-05, "loss": 0.045, "step": 960 }, { "epoch": 1.0264550264550265, "grad_norm": 0.5218284726142883, "learning_rate": 6.460000000000001e-05, "loss": 0.0419, "step": 970 }, { "epoch": 1.037037037037037, "grad_norm": 0.24055950343608856, "learning_rate": 6.526666666666666e-05, "loss": 0.0409, "step": 980 }, { "epoch": 1.0476190476190477, "grad_norm": 0.3444627523422241, "learning_rate": 6.593333333333334e-05, "loss": 0.0461, "step": 990 }, { "epoch": 1.0582010582010581, "grad_norm": 0.3532026708126068, "learning_rate": 6.66e-05, "loss": 0.0369, "step": 1000 }, { "epoch": 1.0687830687830688, "grad_norm": 0.5975197553634644, "learning_rate": 6.726666666666667e-05, "loss": 0.0384, "step": 1010 }, { "epoch": 1.0793650793650793, "grad_norm": 0.5111613869667053, "learning_rate": 6.793333333333334e-05, "loss": 0.0426, "step": 1020 }, { "epoch": 1.08994708994709, "grad_norm": 0.24810604751110077, "learning_rate": 6.860000000000001e-05, "loss": 0.0407, "step": 1030 }, { "epoch": 1.1005291005291005, "grad_norm": 0.4800681471824646, "learning_rate": 6.926666666666667e-05, "loss": 0.0371, "step": 1040 }, { "epoch": 1.1111111111111112, "grad_norm": 0.2537323832511902, "learning_rate": 6.993333333333334e-05, "loss": 0.0383, "step": 1050 }, { "epoch": 1.1216931216931216, "grad_norm": 0.41719233989715576, "learning_rate": 7.06e-05, "loss": 0.0386, "step": 1060 }, { "epoch": 1.1322751322751323, "grad_norm": 0.2658056616783142, "learning_rate": 7.126666666666667e-05, "loss": 0.038, "step": 1070 }, { "epoch": 1.1428571428571428, "grad_norm": 0.38473737239837646, "learning_rate": 7.193333333333334e-05, "loss": 0.0426, "step": 1080 }, { "epoch": 1.1534391534391535, "grad_norm": 0.46496281027793884, "learning_rate": 7.26e-05, "loss": 0.041, "step": 1090 }, { "epoch": 1.164021164021164, "grad_norm": 0.4313805103302002, "learning_rate": 7.326666666666667e-05, "loss": 0.0426, "step": 1100 }, { "epoch": 1.1746031746031746, "grad_norm": 0.5299776196479797, "learning_rate": 7.393333333333333e-05, "loss": 0.04, "step": 1110 }, { "epoch": 1.1851851851851851, "grad_norm": 0.30032458901405334, "learning_rate": 7.46e-05, "loss": 0.0353, "step": 1120 }, { "epoch": 1.1957671957671958, "grad_norm": 0.38715532422065735, "learning_rate": 7.526666666666668e-05, "loss": 0.0373, "step": 1130 }, { "epoch": 1.2063492063492063, "grad_norm": 0.3915134370326996, "learning_rate": 7.593333333333334e-05, "loss": 0.042, "step": 1140 }, { "epoch": 1.216931216931217, "grad_norm": 0.35134851932525635, "learning_rate": 7.66e-05, "loss": 0.0386, "step": 1150 }, { "epoch": 1.2275132275132274, "grad_norm": 0.36247336864471436, "learning_rate": 7.726666666666667e-05, "loss": 0.0386, "step": 1160 }, { "epoch": 1.2380952380952381, "grad_norm": 0.40519481897354126, "learning_rate": 7.793333333333333e-05, "loss": 0.0408, "step": 1170 }, { "epoch": 1.2486772486772486, "grad_norm": 0.29457229375839233, "learning_rate": 7.860000000000001e-05, "loss": 0.0431, "step": 1180 }, { "epoch": 1.2592592592592593, "grad_norm": 0.5451787710189819, "learning_rate": 7.926666666666666e-05, "loss": 0.0433, "step": 1190 }, { "epoch": 1.2698412698412698, "grad_norm": 0.284775048494339, "learning_rate": 7.993333333333334e-05, "loss": 0.0444, "step": 1200 }, { "epoch": 1.2804232804232805, "grad_norm": 0.32610365748405457, "learning_rate": 8.060000000000001e-05, "loss": 0.0382, "step": 1210 }, { "epoch": 1.291005291005291, "grad_norm": 0.4211512506008148, "learning_rate": 8.126666666666667e-05, "loss": 0.041, "step": 1220 }, { "epoch": 1.3015873015873016, "grad_norm": 0.30921512842178345, "learning_rate": 8.193333333333334e-05, "loss": 0.0382, "step": 1230 }, { "epoch": 1.312169312169312, "grad_norm": 0.24003291130065918, "learning_rate": 8.26e-05, "loss": 0.0341, "step": 1240 }, { "epoch": 1.3227513227513228, "grad_norm": 0.5141268968582153, "learning_rate": 8.326666666666667e-05, "loss": 0.0351, "step": 1250 }, { "epoch": 1.3333333333333333, "grad_norm": 0.6243072152137756, "learning_rate": 8.393333333333335e-05, "loss": 0.0378, "step": 1260 }, { "epoch": 1.343915343915344, "grad_norm": 0.29815471172332764, "learning_rate": 8.46e-05, "loss": 0.0348, "step": 1270 }, { "epoch": 1.3544973544973544, "grad_norm": 0.46788886189460754, "learning_rate": 8.526666666666667e-05, "loss": 0.04, "step": 1280 }, { "epoch": 1.3650793650793651, "grad_norm": 0.27657946944236755, "learning_rate": 8.593333333333333e-05, "loss": 0.0392, "step": 1290 }, { "epoch": 1.3756613756613756, "grad_norm": 0.2719906270503998, "learning_rate": 8.66e-05, "loss": 0.0386, "step": 1300 }, { "epoch": 1.3862433862433863, "grad_norm": 0.4385339319705963, "learning_rate": 8.726666666666667e-05, "loss": 0.0446, "step": 1310 }, { "epoch": 1.3968253968253967, "grad_norm": 0.3614080250263214, "learning_rate": 8.793333333333333e-05, "loss": 0.0392, "step": 1320 }, { "epoch": 1.4074074074074074, "grad_norm": 0.2880930006504059, "learning_rate": 8.86e-05, "loss": 0.0462, "step": 1330 }, { "epoch": 1.417989417989418, "grad_norm": 0.2485153079032898, "learning_rate": 8.926666666666668e-05, "loss": 0.0437, "step": 1340 }, { "epoch": 1.4285714285714286, "grad_norm": 0.25833091139793396, "learning_rate": 8.993333333333334e-05, "loss": 0.0402, "step": 1350 }, { "epoch": 1.439153439153439, "grad_norm": 0.352584570646286, "learning_rate": 9.06e-05, "loss": 0.0398, "step": 1360 }, { "epoch": 1.4497354497354498, "grad_norm": 0.44882410764694214, "learning_rate": 9.126666666666667e-05, "loss": 0.037, "step": 1370 }, { "epoch": 1.4603174603174602, "grad_norm": 0.3905755281448364, "learning_rate": 9.193333333333334e-05, "loss": 0.0368, "step": 1380 }, { "epoch": 1.470899470899471, "grad_norm": 0.2982556223869324, "learning_rate": 9.260000000000001e-05, "loss": 0.0373, "step": 1390 }, { "epoch": 1.4814814814814814, "grad_norm": 0.3537434935569763, "learning_rate": 9.326666666666667e-05, "loss": 0.038, "step": 1400 }, { "epoch": 1.492063492063492, "grad_norm": 0.2808147072792053, "learning_rate": 9.393333333333334e-05, "loss": 0.0392, "step": 1410 }, { "epoch": 1.5026455026455028, "grad_norm": 0.2846652865409851, "learning_rate": 9.46e-05, "loss": 0.0361, "step": 1420 }, { "epoch": 1.5132275132275133, "grad_norm": 0.3075992166996002, "learning_rate": 9.526666666666667e-05, "loss": 0.0348, "step": 1430 }, { "epoch": 1.5238095238095237, "grad_norm": 0.3038991391658783, "learning_rate": 9.593333333333334e-05, "loss": 0.0338, "step": 1440 }, { "epoch": 1.5343915343915344, "grad_norm": 0.25168052315711975, "learning_rate": 9.66e-05, "loss": 0.0367, "step": 1450 }, { "epoch": 1.544973544973545, "grad_norm": 0.4394334852695465, "learning_rate": 9.726666666666667e-05, "loss": 0.0365, "step": 1460 }, { "epoch": 1.5555555555555556, "grad_norm": 0.40322718024253845, "learning_rate": 9.793333333333333e-05, "loss": 0.0348, "step": 1470 }, { "epoch": 1.566137566137566, "grad_norm": 0.2442326545715332, "learning_rate": 9.86e-05, "loss": 0.0361, "step": 1480 }, { "epoch": 1.5767195767195767, "grad_norm": 0.36882537603378296, "learning_rate": 9.926666666666668e-05, "loss": 0.0408, "step": 1490 }, { "epoch": 1.5873015873015874, "grad_norm": 0.3560725152492523, "learning_rate": 9.993333333333334e-05, "loss": 0.0381, "step": 1500 }, { "epoch": 1.597883597883598, "grad_norm": 0.39973852038383484, "learning_rate": 9.999997539434007e-05, "loss": 0.0377, "step": 1510 }, { "epoch": 1.6084656084656084, "grad_norm": 0.4991508424282074, "learning_rate": 9.999989033776898e-05, "loss": 0.0367, "step": 1520 }, { "epoch": 1.619047619047619, "grad_norm": 0.5416008234024048, "learning_rate": 9.999974452661641e-05, "loss": 0.044, "step": 1530 }, { "epoch": 1.6296296296296298, "grad_norm": 0.3502684235572815, "learning_rate": 9.999953796105959e-05, "loss": 0.0421, "step": 1540 }, { "epoch": 1.6402116402116402, "grad_norm": 0.3063826262950897, "learning_rate": 9.999927064134949e-05, "loss": 0.0339, "step": 1550 }, { "epoch": 1.6507936507936507, "grad_norm": 0.3482449948787689, "learning_rate": 9.999894256781095e-05, "loss": 0.0413, "step": 1560 }, { "epoch": 1.6613756613756614, "grad_norm": 0.36818769574165344, "learning_rate": 9.99985537408426e-05, "loss": 0.0357, "step": 1570 }, { "epoch": 1.671957671957672, "grad_norm": 0.27622190117836, "learning_rate": 9.999810416091688e-05, "loss": 0.0339, "step": 1580 }, { "epoch": 1.6825396825396826, "grad_norm": 0.2747170925140381, "learning_rate": 9.99975938285801e-05, "loss": 0.0347, "step": 1590 }, { "epoch": 1.693121693121693, "grad_norm": 0.2372732013463974, "learning_rate": 9.999702274445236e-05, "loss": 0.041, "step": 1600 }, { "epoch": 1.7037037037037037, "grad_norm": 0.28219789266586304, "learning_rate": 9.999639090922756e-05, "loss": 0.0332, "step": 1610 }, { "epoch": 1.7142857142857144, "grad_norm": 0.41299858689308167, "learning_rate": 9.999569832367346e-05, "loss": 0.0354, "step": 1620 }, { "epoch": 1.7248677248677249, "grad_norm": 0.4950662851333618, "learning_rate": 9.999494498863162e-05, "loss": 0.0443, "step": 1630 }, { "epoch": 1.7354497354497354, "grad_norm": 0.453332781791687, "learning_rate": 9.99941309050174e-05, "loss": 0.038, "step": 1640 }, { "epoch": 1.746031746031746, "grad_norm": 0.309845507144928, "learning_rate": 9.999325607381999e-05, "loss": 0.0378, "step": 1650 }, { "epoch": 1.7566137566137567, "grad_norm": 0.36152681708335876, "learning_rate": 9.999232049610238e-05, "loss": 0.0392, "step": 1660 }, { "epoch": 1.7671957671957672, "grad_norm": 0.35339006781578064, "learning_rate": 9.999132417300142e-05, "loss": 0.0391, "step": 1670 }, { "epoch": 1.7777777777777777, "grad_norm": 0.24615387618541718, "learning_rate": 9.99902671057277e-05, "loss": 0.0433, "step": 1680 }, { "epoch": 1.7883597883597884, "grad_norm": 0.27722659707069397, "learning_rate": 9.998914929556569e-05, "loss": 0.0399, "step": 1690 }, { "epoch": 1.798941798941799, "grad_norm": 0.34978312253952026, "learning_rate": 9.998797074387361e-05, "loss": 0.0371, "step": 1700 }, { "epoch": 1.8095238095238095, "grad_norm": 0.29569652676582336, "learning_rate": 9.99867314520835e-05, "loss": 0.0315, "step": 1710 }, { "epoch": 1.82010582010582, "grad_norm": 0.28561344742774963, "learning_rate": 9.998543142170126e-05, "loss": 0.0388, "step": 1720 }, { "epoch": 1.8306878306878307, "grad_norm": 0.28362226486206055, "learning_rate": 9.99840706543065e-05, "loss": 0.0442, "step": 1730 }, { "epoch": 1.8412698412698414, "grad_norm": 0.35063615441322327, "learning_rate": 9.998264915155274e-05, "loss": 0.0434, "step": 1740 }, { "epoch": 1.8518518518518519, "grad_norm": 0.2760818898677826, "learning_rate": 9.998116691516718e-05, "loss": 0.0437, "step": 1750 }, { "epoch": 1.8624338624338623, "grad_norm": 0.26761332154273987, "learning_rate": 9.997962394695091e-05, "loss": 0.0344, "step": 1760 }, { "epoch": 1.873015873015873, "grad_norm": 0.2880273461341858, "learning_rate": 9.997802024877875e-05, "loss": 0.0359, "step": 1770 }, { "epoch": 1.8835978835978837, "grad_norm": 0.3273584544658661, "learning_rate": 9.99763558225994e-05, "loss": 0.0381, "step": 1780 }, { "epoch": 1.8941798941798942, "grad_norm": 0.21250487864017487, "learning_rate": 9.997463067043526e-05, "loss": 0.0379, "step": 1790 }, { "epoch": 1.9047619047619047, "grad_norm": 0.3019721806049347, "learning_rate": 9.997284479438253e-05, "loss": 0.0371, "step": 1800 }, { "epoch": 1.9153439153439153, "grad_norm": 0.2825166881084442, "learning_rate": 9.997099819661127e-05, "loss": 0.035, "step": 1810 }, { "epoch": 1.925925925925926, "grad_norm": 0.21520768105983734, "learning_rate": 9.996909087936524e-05, "loss": 0.0333, "step": 1820 }, { "epoch": 1.9365079365079365, "grad_norm": 0.20333130657672882, "learning_rate": 9.996712284496201e-05, "loss": 0.0366, "step": 1830 }, { "epoch": 1.947089947089947, "grad_norm": 0.30042168498039246, "learning_rate": 9.996509409579293e-05, "loss": 0.0362, "step": 1840 }, { "epoch": 1.9576719576719577, "grad_norm": 0.2078566998243332, "learning_rate": 9.996300463432312e-05, "loss": 0.0355, "step": 1850 }, { "epoch": 1.9682539682539684, "grad_norm": 0.20628400146961212, "learning_rate": 9.996085446309148e-05, "loss": 0.0346, "step": 1860 }, { "epoch": 1.9788359788359788, "grad_norm": 0.27391737699508667, "learning_rate": 9.995864358471066e-05, "loss": 0.0304, "step": 1870 }, { "epoch": 1.9894179894179893, "grad_norm": 0.3650337755680084, "learning_rate": 9.99563720018671e-05, "loss": 0.0337, "step": 1880 }, { "epoch": 2.0, "grad_norm": 0.47546714544296265, "learning_rate": 9.995403971732098e-05, "loss": 0.0367, "step": 1890 }, { "epoch": 2.0105820105820107, "grad_norm": 0.2382199615240097, "learning_rate": 9.995164673390625e-05, "loss": 0.0349, "step": 1900 }, { "epoch": 2.0211640211640214, "grad_norm": 0.3941175043582916, "learning_rate": 9.994919305453059e-05, "loss": 0.0352, "step": 1910 }, { "epoch": 2.0317460317460316, "grad_norm": 0.39152991771698, "learning_rate": 9.994667868217548e-05, "loss": 0.0313, "step": 1920 }, { "epoch": 2.0423280423280423, "grad_norm": 0.557101845741272, "learning_rate": 9.99441036198961e-05, "loss": 0.0362, "step": 1930 }, { "epoch": 2.052910052910053, "grad_norm": 0.3060609698295593, "learning_rate": 9.99414678708214e-05, "loss": 0.0297, "step": 1940 }, { "epoch": 2.0634920634920633, "grad_norm": 0.3889998495578766, "learning_rate": 9.993877143815407e-05, "loss": 0.0371, "step": 1950 }, { "epoch": 2.074074074074074, "grad_norm": 0.3242912292480469, "learning_rate": 9.993601432517053e-05, "loss": 0.0349, "step": 1960 }, { "epoch": 2.0846560846560847, "grad_norm": 0.20132385194301605, "learning_rate": 9.993319653522091e-05, "loss": 0.0326, "step": 1970 }, { "epoch": 2.0952380952380953, "grad_norm": 0.2600923776626587, "learning_rate": 9.993031807172911e-05, "loss": 0.0363, "step": 1980 }, { "epoch": 2.105820105820106, "grad_norm": 0.24345675110816956, "learning_rate": 9.992737893819273e-05, "loss": 0.0341, "step": 1990 }, { "epoch": 2.1164021164021163, "grad_norm": 0.16455724835395813, "learning_rate": 9.992437913818312e-05, "loss": 0.0336, "step": 2000 }, { "epoch": 2.126984126984127, "grad_norm": 0.3164030611515045, "learning_rate": 9.992131867534526e-05, "loss": 0.03, "step": 2010 }, { "epoch": 2.1375661375661377, "grad_norm": 0.2891181707382202, "learning_rate": 9.991819755339796e-05, "loss": 0.0359, "step": 2020 }, { "epoch": 2.148148148148148, "grad_norm": 0.29365161061286926, "learning_rate": 9.991501577613365e-05, "loss": 0.0324, "step": 2030 }, { "epoch": 2.1587301587301586, "grad_norm": 0.2762247622013092, "learning_rate": 9.99117733474185e-05, "loss": 0.0306, "step": 2040 }, { "epoch": 2.1693121693121693, "grad_norm": 0.249510258436203, "learning_rate": 9.990847027119234e-05, "loss": 0.0318, "step": 2050 }, { "epoch": 2.17989417989418, "grad_norm": 0.3514985144138336, "learning_rate": 9.990510655146877e-05, "loss": 0.0295, "step": 2060 }, { "epoch": 2.1904761904761907, "grad_norm": 0.2896115183830261, "learning_rate": 9.990168219233496e-05, "loss": 0.0315, "step": 2070 }, { "epoch": 2.201058201058201, "grad_norm": 0.19914424419403076, "learning_rate": 9.989819719795188e-05, "loss": 0.0287, "step": 2080 }, { "epoch": 2.2116402116402116, "grad_norm": 0.23942850530147552, "learning_rate": 9.989465157255412e-05, "loss": 0.0322, "step": 2090 }, { "epoch": 2.2222222222222223, "grad_norm": 0.2262566238641739, "learning_rate": 9.989104532044994e-05, "loss": 0.0279, "step": 2100 }, { "epoch": 2.2328042328042326, "grad_norm": 0.2357911318540573, "learning_rate": 9.988737844602128e-05, "loss": 0.0274, "step": 2110 }, { "epoch": 2.2433862433862433, "grad_norm": 0.207014799118042, "learning_rate": 9.988365095372372e-05, "loss": 0.0277, "step": 2120 }, { "epoch": 2.253968253968254, "grad_norm": 0.3012750446796417, "learning_rate": 9.987986284808654e-05, "loss": 0.0332, "step": 2130 }, { "epoch": 2.2645502645502646, "grad_norm": 0.37401503324508667, "learning_rate": 9.987601413371264e-05, "loss": 0.0316, "step": 2140 }, { "epoch": 2.2751322751322753, "grad_norm": 0.19214403629302979, "learning_rate": 9.987210481527855e-05, "loss": 0.0311, "step": 2150 }, { "epoch": 2.2857142857142856, "grad_norm": 0.24868172407150269, "learning_rate": 9.98681348975345e-05, "loss": 0.0304, "step": 2160 }, { "epoch": 2.2962962962962963, "grad_norm": 0.26443183422088623, "learning_rate": 9.986410438530427e-05, "loss": 0.0303, "step": 2170 }, { "epoch": 2.306878306878307, "grad_norm": 0.32223764061927795, "learning_rate": 9.986001328348534e-05, "loss": 0.0295, "step": 2180 }, { "epoch": 2.317460317460317, "grad_norm": 0.2191818207502365, "learning_rate": 9.985586159704878e-05, "loss": 0.029, "step": 2190 }, { "epoch": 2.328042328042328, "grad_norm": 0.2986072301864624, "learning_rate": 9.985164933103929e-05, "loss": 0.0323, "step": 2200 }, { "epoch": 2.3386243386243386, "grad_norm": 0.2174193263053894, "learning_rate": 9.984737649057513e-05, "loss": 0.0317, "step": 2210 }, { "epoch": 2.3492063492063493, "grad_norm": 0.3140878677368164, "learning_rate": 9.984304308084827e-05, "loss": 0.0407, "step": 2220 }, { "epoch": 2.35978835978836, "grad_norm": 0.2489406019449234, "learning_rate": 9.983864910712416e-05, "loss": 0.0333, "step": 2230 }, { "epoch": 2.3703703703703702, "grad_norm": 0.30106040835380554, "learning_rate": 9.98341945747419e-05, "loss": 0.0303, "step": 2240 }, { "epoch": 2.380952380952381, "grad_norm": 0.2998392581939697, "learning_rate": 9.98296794891142e-05, "loss": 0.0294, "step": 2250 }, { "epoch": 2.3915343915343916, "grad_norm": 0.25180819630622864, "learning_rate": 9.982510385572725e-05, "loss": 0.0306, "step": 2260 }, { "epoch": 2.402116402116402, "grad_norm": 0.23264649510383606, "learning_rate": 9.982046768014094e-05, "loss": 0.028, "step": 2270 }, { "epoch": 2.4126984126984126, "grad_norm": 0.2753427028656006, "learning_rate": 9.981577096798863e-05, "loss": 0.0259, "step": 2280 }, { "epoch": 2.4232804232804233, "grad_norm": 0.2087375819683075, "learning_rate": 9.981101372497727e-05, "loss": 0.032, "step": 2290 }, { "epoch": 2.433862433862434, "grad_norm": 0.27186354994773865, "learning_rate": 9.980619595688737e-05, "loss": 0.0293, "step": 2300 }, { "epoch": 2.4444444444444446, "grad_norm": 0.19205987453460693, "learning_rate": 9.980131766957295e-05, "loss": 0.0312, "step": 2310 }, { "epoch": 2.455026455026455, "grad_norm": 0.24335825443267822, "learning_rate": 9.979637886896163e-05, "loss": 0.029, "step": 2320 }, { "epoch": 2.4656084656084656, "grad_norm": 0.19624657928943634, "learning_rate": 9.979137956105447e-05, "loss": 0.0337, "step": 2330 }, { "epoch": 2.4761904761904763, "grad_norm": 0.21104492247104645, "learning_rate": 9.978631975192613e-05, "loss": 0.0253, "step": 2340 }, { "epoch": 2.4867724867724865, "grad_norm": 0.291870653629303, "learning_rate": 9.978119944772475e-05, "loss": 0.0353, "step": 2350 }, { "epoch": 2.497354497354497, "grad_norm": 0.20102587342262268, "learning_rate": 9.977601865467197e-05, "loss": 0.0288, "step": 2360 }, { "epoch": 2.507936507936508, "grad_norm": 0.1770666092634201, "learning_rate": 9.977077737906297e-05, "loss": 0.0282, "step": 2370 }, { "epoch": 2.5185185185185186, "grad_norm": 0.27207255363464355, "learning_rate": 9.976547562726636e-05, "loss": 0.0275, "step": 2380 }, { "epoch": 2.5291005291005293, "grad_norm": 0.20059968531131744, "learning_rate": 9.976011340572429e-05, "loss": 0.0331, "step": 2390 }, { "epoch": 2.5396825396825395, "grad_norm": 0.23637302219867706, "learning_rate": 9.975469072095237e-05, "loss": 0.0287, "step": 2400 }, { "epoch": 2.5502645502645502, "grad_norm": 0.30920884013175964, "learning_rate": 9.974920757953965e-05, "loss": 0.0304, "step": 2410 }, { "epoch": 2.560846560846561, "grad_norm": 0.2816784679889679, "learning_rate": 9.97436639881487e-05, "loss": 0.038, "step": 2420 }, { "epoch": 2.571428571428571, "grad_norm": 0.2768605649471283, "learning_rate": 9.973805995351545e-05, "loss": 0.037, "step": 2430 }, { "epoch": 2.582010582010582, "grad_norm": 0.33137211203575134, "learning_rate": 9.973239548244939e-05, "loss": 0.0322, "step": 2440 }, { "epoch": 2.5925925925925926, "grad_norm": 0.39114734530448914, "learning_rate": 9.972667058183333e-05, "loss": 0.0295, "step": 2450 }, { "epoch": 2.6031746031746033, "grad_norm": 0.313888281583786, "learning_rate": 9.972088525862362e-05, "loss": 0.0284, "step": 2460 }, { "epoch": 2.613756613756614, "grad_norm": 0.26245546340942383, "learning_rate": 9.971503951984995e-05, "loss": 0.0333, "step": 2470 }, { "epoch": 2.624338624338624, "grad_norm": 0.2396690845489502, "learning_rate": 9.970913337261543e-05, "loss": 0.0327, "step": 2480 }, { "epoch": 2.634920634920635, "grad_norm": 0.24638260900974274, "learning_rate": 9.97031668240966e-05, "loss": 0.0317, "step": 2490 }, { "epoch": 2.6455026455026456, "grad_norm": 0.23108890652656555, "learning_rate": 9.969713988154339e-05, "loss": 0.0304, "step": 2500 }, { "epoch": 2.656084656084656, "grad_norm": 0.28835028409957886, "learning_rate": 9.969105255227906e-05, "loss": 0.0307, "step": 2510 }, { "epoch": 2.6666666666666665, "grad_norm": 0.29013532400131226, "learning_rate": 9.968490484370035e-05, "loss": 0.0298, "step": 2520 }, { "epoch": 2.677248677248677, "grad_norm": 0.2208428531885147, "learning_rate": 9.967869676327726e-05, "loss": 0.0283, "step": 2530 }, { "epoch": 2.687830687830688, "grad_norm": 0.19803676009178162, "learning_rate": 9.96724283185532e-05, "loss": 0.0247, "step": 2540 }, { "epoch": 2.6984126984126986, "grad_norm": 0.199357807636261, "learning_rate": 9.966609951714494e-05, "loss": 0.034, "step": 2550 }, { "epoch": 2.708994708994709, "grad_norm": 0.33765938878059387, "learning_rate": 9.965971036674255e-05, "loss": 0.0274, "step": 2560 }, { "epoch": 2.7195767195767195, "grad_norm": 0.2613915503025055, "learning_rate": 9.965326087510947e-05, "loss": 0.0276, "step": 2570 }, { "epoch": 2.7301587301587302, "grad_norm": 0.19864608347415924, "learning_rate": 9.964675105008244e-05, "loss": 0.028, "step": 2580 }, { "epoch": 2.7407407407407405, "grad_norm": 0.33229854702949524, "learning_rate": 9.964018089957147e-05, "loss": 0.0268, "step": 2590 }, { "epoch": 2.751322751322751, "grad_norm": 0.22380158305168152, "learning_rate": 9.963355043155996e-05, "loss": 0.0312, "step": 2600 }, { "epoch": 2.761904761904762, "grad_norm": 0.3922805190086365, "learning_rate": 9.962685965410455e-05, "loss": 0.0274, "step": 2610 }, { "epoch": 2.7724867724867726, "grad_norm": 0.2782486379146576, "learning_rate": 9.962010857533514e-05, "loss": 0.0289, "step": 2620 }, { "epoch": 2.7830687830687832, "grad_norm": 0.3146815896034241, "learning_rate": 9.961329720345493e-05, "loss": 0.031, "step": 2630 }, { "epoch": 2.7936507936507935, "grad_norm": 0.16208294034004211, "learning_rate": 9.96064255467404e-05, "loss": 0.0293, "step": 2640 }, { "epoch": 2.804232804232804, "grad_norm": 0.33112767338752747, "learning_rate": 9.959949361354126e-05, "loss": 0.0298, "step": 2650 }, { "epoch": 2.814814814814815, "grad_norm": 0.2928871810436249, "learning_rate": 9.959250141228045e-05, "loss": 0.0281, "step": 2660 }, { "epoch": 2.825396825396825, "grad_norm": 0.23492847383022308, "learning_rate": 9.958544895145414e-05, "loss": 0.0264, "step": 2670 }, { "epoch": 2.835978835978836, "grad_norm": 0.16989514231681824, "learning_rate": 9.957833623963177e-05, "loss": 0.0304, "step": 2680 }, { "epoch": 2.8465608465608465, "grad_norm": 0.22784683108329773, "learning_rate": 9.957116328545593e-05, "loss": 0.0306, "step": 2690 }, { "epoch": 2.857142857142857, "grad_norm": 0.2510013282299042, "learning_rate": 9.956393009764244e-05, "loss": 0.0254, "step": 2700 }, { "epoch": 2.867724867724868, "grad_norm": 0.14823012053966522, "learning_rate": 9.955663668498032e-05, "loss": 0.0229, "step": 2710 }, { "epoch": 2.878306878306878, "grad_norm": 0.176004096865654, "learning_rate": 9.954928305633173e-05, "loss": 0.0254, "step": 2720 }, { "epoch": 2.888888888888889, "grad_norm": 0.1823616921901703, "learning_rate": 9.954186922063204e-05, "loss": 0.0319, "step": 2730 }, { "epoch": 2.8994708994708995, "grad_norm": 0.17964163422584534, "learning_rate": 9.953439518688974e-05, "loss": 0.0275, "step": 2740 }, { "epoch": 2.91005291005291, "grad_norm": 0.19006715714931488, "learning_rate": 9.952686096418652e-05, "loss": 0.028, "step": 2750 }, { "epoch": 2.9206349206349205, "grad_norm": 0.3385632038116455, "learning_rate": 9.951926656167714e-05, "loss": 0.0265, "step": 2760 }, { "epoch": 2.931216931216931, "grad_norm": 0.2010369598865509, "learning_rate": 9.951161198858953e-05, "loss": 0.0264, "step": 2770 }, { "epoch": 2.941798941798942, "grad_norm": 0.21000969409942627, "learning_rate": 9.950389725422471e-05, "loss": 0.0243, "step": 2780 }, { "epoch": 2.9523809523809526, "grad_norm": 0.15133130550384521, "learning_rate": 9.949612236795682e-05, "loss": 0.0227, "step": 2790 }, { "epoch": 2.962962962962963, "grad_norm": 0.24288538098335266, "learning_rate": 9.948828733923305e-05, "loss": 0.0271, "step": 2800 }, { "epoch": 2.9735449735449735, "grad_norm": 0.20132894814014435, "learning_rate": 9.948039217757374e-05, "loss": 0.0266, "step": 2810 }, { "epoch": 2.984126984126984, "grad_norm": 0.21461914479732513, "learning_rate": 9.947243689257225e-05, "loss": 0.0287, "step": 2820 }, { "epoch": 2.9947089947089944, "grad_norm": 0.31207385659217834, "learning_rate": 9.946442149389497e-05, "loss": 0.0275, "step": 2830 }, { "epoch": 3.005291005291005, "grad_norm": 0.18261997401714325, "learning_rate": 9.945634599128139e-05, "loss": 0.028, "step": 2840 }, { "epoch": 3.015873015873016, "grad_norm": 0.20634686946868896, "learning_rate": 9.944821039454402e-05, "loss": 0.0261, "step": 2850 }, { "epoch": 3.0264550264550265, "grad_norm": 0.239877849817276, "learning_rate": 9.944001471356835e-05, "loss": 0.0286, "step": 2860 }, { "epoch": 3.037037037037037, "grad_norm": 0.15490078926086426, "learning_rate": 9.94317589583129e-05, "loss": 0.0273, "step": 2870 }, { "epoch": 3.0476190476190474, "grad_norm": 0.25720202922821045, "learning_rate": 9.942344313880922e-05, "loss": 0.0291, "step": 2880 }, { "epoch": 3.058201058201058, "grad_norm": 0.303619384765625, "learning_rate": 9.941506726516179e-05, "loss": 0.0273, "step": 2890 }, { "epoch": 3.068783068783069, "grad_norm": 0.23992906510829926, "learning_rate": 9.94066313475481e-05, "loss": 0.0289, "step": 2900 }, { "epoch": 3.0793650793650795, "grad_norm": 0.29250290989875793, "learning_rate": 9.939813539621857e-05, "loss": 0.0261, "step": 2910 }, { "epoch": 3.0899470899470898, "grad_norm": 0.13921523094177246, "learning_rate": 9.93895794214966e-05, "loss": 0.0272, "step": 2920 }, { "epoch": 3.1005291005291005, "grad_norm": 0.31076255440711975, "learning_rate": 9.938096343377852e-05, "loss": 0.03, "step": 2930 }, { "epoch": 3.111111111111111, "grad_norm": 0.19056645035743713, "learning_rate": 9.937228744353353e-05, "loss": 0.0269, "step": 2940 }, { "epoch": 3.121693121693122, "grad_norm": 0.18811097741127014, "learning_rate": 9.936355146130379e-05, "loss": 0.0277, "step": 2950 }, { "epoch": 3.132275132275132, "grad_norm": 0.18918462097644806, "learning_rate": 9.935475549770435e-05, "loss": 0.024, "step": 2960 }, { "epoch": 3.142857142857143, "grad_norm": 0.16715602576732635, "learning_rate": 9.934589956342315e-05, "loss": 0.0236, "step": 2970 }, { "epoch": 3.1534391534391535, "grad_norm": 0.3086070716381073, "learning_rate": 9.933698366922093e-05, "loss": 0.0293, "step": 2980 }, { "epoch": 3.164021164021164, "grad_norm": 0.2112513780593872, "learning_rate": 9.93280078259314e-05, "loss": 0.0283, "step": 2990 }, { "epoch": 3.1746031746031744, "grad_norm": 0.2307254821062088, "learning_rate": 9.931897204446104e-05, "loss": 0.0248, "step": 3000 }, { "epoch": 3.185185185185185, "grad_norm": 0.29472124576568604, "learning_rate": 9.930987633578915e-05, "loss": 0.029, "step": 3010 }, { "epoch": 3.195767195767196, "grad_norm": 0.25854524970054626, "learning_rate": 9.93007207109679e-05, "loss": 0.0281, "step": 3020 }, { "epoch": 3.2063492063492065, "grad_norm": 0.21072831749916077, "learning_rate": 9.929150518112224e-05, "loss": 0.0245, "step": 3030 }, { "epoch": 3.2169312169312168, "grad_norm": 0.19457313418388367, "learning_rate": 9.928222975744991e-05, "loss": 0.028, "step": 3040 }, { "epoch": 3.2275132275132274, "grad_norm": 0.1913045346736908, "learning_rate": 9.92728944512214e-05, "loss": 0.0228, "step": 3050 }, { "epoch": 3.238095238095238, "grad_norm": 0.24282030761241913, "learning_rate": 9.926349927378001e-05, "loss": 0.0233, "step": 3060 }, { "epoch": 3.248677248677249, "grad_norm": 0.17764173448085785, "learning_rate": 9.925404423654174e-05, "loss": 0.0245, "step": 3070 }, { "epoch": 3.259259259259259, "grad_norm": 0.26288723945617676, "learning_rate": 9.924452935099537e-05, "loss": 0.0253, "step": 3080 }, { "epoch": 3.2698412698412698, "grad_norm": 0.24634087085723877, "learning_rate": 9.92349546287024e-05, "loss": 0.0252, "step": 3090 }, { "epoch": 3.2804232804232805, "grad_norm": 0.33013099431991577, "learning_rate": 9.9225320081297e-05, "loss": 0.0276, "step": 3100 }, { "epoch": 3.291005291005291, "grad_norm": 0.20951156318187714, "learning_rate": 9.921562572048606e-05, "loss": 0.0254, "step": 3110 }, { "epoch": 3.3015873015873014, "grad_norm": 0.171208918094635, "learning_rate": 9.920587155804913e-05, "loss": 0.0301, "step": 3120 }, { "epoch": 3.312169312169312, "grad_norm": 0.19983966648578644, "learning_rate": 9.919605760583845e-05, "loss": 0.029, "step": 3130 }, { "epoch": 3.322751322751323, "grad_norm": 0.1933031529188156, "learning_rate": 9.91861838757789e-05, "loss": 0.0265, "step": 3140 }, { "epoch": 3.3333333333333335, "grad_norm": 0.23514573276042938, "learning_rate": 9.917625037986798e-05, "loss": 0.0258, "step": 3150 }, { "epoch": 3.3439153439153437, "grad_norm": 0.23969581723213196, "learning_rate": 9.916625713017583e-05, "loss": 0.0252, "step": 3160 }, { "epoch": 3.3544973544973544, "grad_norm": 0.19836458563804626, "learning_rate": 9.915620413884519e-05, "loss": 0.0231, "step": 3170 }, { "epoch": 3.365079365079365, "grad_norm": 0.21828173100948334, "learning_rate": 9.914609141809139e-05, "loss": 0.0255, "step": 3180 }, { "epoch": 3.375661375661376, "grad_norm": 0.2646982967853546, "learning_rate": 9.913591898020235e-05, "loss": 0.0273, "step": 3190 }, { "epoch": 3.386243386243386, "grad_norm": 0.23596763610839844, "learning_rate": 9.912568683753853e-05, "loss": 0.0242, "step": 3200 }, { "epoch": 3.3968253968253967, "grad_norm": 0.21948781609535217, "learning_rate": 9.911539500253295e-05, "loss": 0.0294, "step": 3210 }, { "epoch": 3.4074074074074074, "grad_norm": 0.281631201505661, "learning_rate": 9.910504348769118e-05, "loss": 0.0251, "step": 3220 }, { "epoch": 3.417989417989418, "grad_norm": 0.22955290973186493, "learning_rate": 9.909463230559127e-05, "loss": 0.0241, "step": 3230 }, { "epoch": 3.4285714285714284, "grad_norm": 0.24076218903064728, "learning_rate": 9.908416146888376e-05, "loss": 0.0274, "step": 3240 }, { "epoch": 3.439153439153439, "grad_norm": 0.217754527926445, "learning_rate": 9.907363099029175e-05, "loss": 0.0265, "step": 3250 }, { "epoch": 3.4497354497354498, "grad_norm": 0.15226824581623077, "learning_rate": 9.906304088261073e-05, "loss": 0.0276, "step": 3260 }, { "epoch": 3.4603174603174605, "grad_norm": 0.2203245311975479, "learning_rate": 9.905239115870872e-05, "loss": 0.0287, "step": 3270 }, { "epoch": 3.4708994708994707, "grad_norm": 0.21335138380527496, "learning_rate": 9.90416818315261e-05, "loss": 0.0272, "step": 3280 }, { "epoch": 3.4814814814814814, "grad_norm": 0.22707191109657288, "learning_rate": 9.903091291407573e-05, "loss": 0.0278, "step": 3290 }, { "epoch": 3.492063492063492, "grad_norm": 0.27302494645118713, "learning_rate": 9.902008441944286e-05, "loss": 0.0252, "step": 3300 }, { "epoch": 3.502645502645503, "grad_norm": 0.25003916025161743, "learning_rate": 9.900919636078512e-05, "loss": 0.0275, "step": 3310 }, { "epoch": 3.5132275132275135, "grad_norm": 0.255531907081604, "learning_rate": 9.899824875133255e-05, "loss": 0.0259, "step": 3320 }, { "epoch": 3.5238095238095237, "grad_norm": 0.14549410343170166, "learning_rate": 9.898724160438749e-05, "loss": 0.0258, "step": 3330 }, { "epoch": 3.5343915343915344, "grad_norm": 0.23905743658542633, "learning_rate": 9.89761749333247e-05, "loss": 0.0272, "step": 3340 }, { "epoch": 3.544973544973545, "grad_norm": 0.2609183192253113, "learning_rate": 9.896504875159122e-05, "loss": 0.0261, "step": 3350 }, { "epoch": 3.5555555555555554, "grad_norm": 0.17955462634563446, "learning_rate": 9.89538630727064e-05, "loss": 0.0223, "step": 3360 }, { "epoch": 3.566137566137566, "grad_norm": 0.26073962450027466, "learning_rate": 9.894261791026189e-05, "loss": 0.0297, "step": 3370 }, { "epoch": 3.5767195767195767, "grad_norm": 0.16885675489902496, "learning_rate": 9.893131327792165e-05, "loss": 0.0229, "step": 3380 }, { "epoch": 3.5873015873015874, "grad_norm": 0.16659343242645264, "learning_rate": 9.891994918942182e-05, "loss": 0.0267, "step": 3390 }, { "epoch": 3.597883597883598, "grad_norm": 0.2803594470024109, "learning_rate": 9.890852565857092e-05, "loss": 0.0266, "step": 3400 }, { "epoch": 3.6084656084656084, "grad_norm": 0.15117418766021729, "learning_rate": 9.889704269924954e-05, "loss": 0.0261, "step": 3410 }, { "epoch": 3.619047619047619, "grad_norm": 0.18954652547836304, "learning_rate": 9.888550032541059e-05, "loss": 0.0245, "step": 3420 }, { "epoch": 3.6296296296296298, "grad_norm": 0.2922452986240387, "learning_rate": 9.887389855107916e-05, "loss": 0.0251, "step": 3430 }, { "epoch": 3.64021164021164, "grad_norm": 0.24185380339622498, "learning_rate": 9.886223739035248e-05, "loss": 0.0277, "step": 3440 }, { "epoch": 3.6507936507936507, "grad_norm": 0.27489957213401794, "learning_rate": 9.885051685739997e-05, "loss": 0.0248, "step": 3450 }, { "epoch": 3.6613756613756614, "grad_norm": 0.23385775089263916, "learning_rate": 9.883873696646316e-05, "loss": 0.028, "step": 3460 }, { "epoch": 3.671957671957672, "grad_norm": 0.20762483775615692, "learning_rate": 9.882689773185575e-05, "loss": 0.0266, "step": 3470 }, { "epoch": 3.682539682539683, "grad_norm": 0.22717130184173584, "learning_rate": 9.881499916796353e-05, "loss": 0.0271, "step": 3480 }, { "epoch": 3.693121693121693, "grad_norm": 0.23919163644313812, "learning_rate": 9.880304128924434e-05, "loss": 0.028, "step": 3490 }, { "epoch": 3.7037037037037037, "grad_norm": 0.2206435650587082, "learning_rate": 9.879102411022817e-05, "loss": 0.0283, "step": 3500 }, { "epoch": 3.7142857142857144, "grad_norm": 0.2736937701702118, "learning_rate": 9.877894764551703e-05, "loss": 0.0261, "step": 3510 }, { "epoch": 3.7248677248677247, "grad_norm": 0.26780933141708374, "learning_rate": 9.876681190978494e-05, "loss": 0.0265, "step": 3520 }, { "epoch": 3.7354497354497354, "grad_norm": 0.26356884837150574, "learning_rate": 9.875461691777797e-05, "loss": 0.0274, "step": 3530 }, { "epoch": 3.746031746031746, "grad_norm": 0.19829510152339935, "learning_rate": 9.874236268431417e-05, "loss": 0.0249, "step": 3540 }, { "epoch": 3.7566137566137567, "grad_norm": 0.16109490394592285, "learning_rate": 9.873004922428361e-05, "loss": 0.0263, "step": 3550 }, { "epoch": 3.7671957671957674, "grad_norm": 0.2343643754720688, "learning_rate": 9.871767655264829e-05, "loss": 0.0269, "step": 3560 }, { "epoch": 3.7777777777777777, "grad_norm": 0.17887639999389648, "learning_rate": 9.87052446844422e-05, "loss": 0.0235, "step": 3570 }, { "epoch": 3.7883597883597884, "grad_norm": 0.154287189245224, "learning_rate": 9.869275363477122e-05, "loss": 0.0262, "step": 3580 }, { "epoch": 3.798941798941799, "grad_norm": 0.1983914077281952, "learning_rate": 9.868020341881312e-05, "loss": 0.0263, "step": 3590 }, { "epoch": 3.8095238095238093, "grad_norm": 0.20611201226711273, "learning_rate": 9.866759405181765e-05, "loss": 0.0317, "step": 3600 }, { "epoch": 3.82010582010582, "grad_norm": 0.26925981044769287, "learning_rate": 9.865492554910633e-05, "loss": 0.0244, "step": 3610 }, { "epoch": 3.8306878306878307, "grad_norm": 0.33553582429885864, "learning_rate": 9.864219792607262e-05, "loss": 0.0271, "step": 3620 }, { "epoch": 3.8412698412698414, "grad_norm": 0.2319973111152649, "learning_rate": 9.862941119818177e-05, "loss": 0.0232, "step": 3630 }, { "epoch": 3.851851851851852, "grad_norm": 0.31576570868492126, "learning_rate": 9.861656538097086e-05, "loss": 0.0241, "step": 3640 }, { "epoch": 3.8624338624338623, "grad_norm": 0.2165345996618271, "learning_rate": 9.860366049004877e-05, "loss": 0.0229, "step": 3650 }, { "epoch": 3.873015873015873, "grad_norm": 0.28317102789878845, "learning_rate": 9.859069654109615e-05, "loss": 0.0264, "step": 3660 }, { "epoch": 3.8835978835978837, "grad_norm": 0.23034577071666718, "learning_rate": 9.857767354986545e-05, "loss": 0.0285, "step": 3670 }, { "epoch": 3.894179894179894, "grad_norm": 0.19838079810142517, "learning_rate": 9.856459153218078e-05, "loss": 0.0237, "step": 3680 }, { "epoch": 3.9047619047619047, "grad_norm": 0.27764615416526794, "learning_rate": 9.855145050393808e-05, "loss": 0.0275, "step": 3690 }, { "epoch": 3.9153439153439153, "grad_norm": 0.18089258670806885, "learning_rate": 9.85382504811049e-05, "loss": 0.0281, "step": 3700 }, { "epoch": 3.925925925925926, "grad_norm": 0.2636431157588959, "learning_rate": 9.852499147972054e-05, "loss": 0.026, "step": 3710 }, { "epoch": 3.9365079365079367, "grad_norm": 0.1795637607574463, "learning_rate": 9.851167351589592e-05, "loss": 0.0242, "step": 3720 }, { "epoch": 3.947089947089947, "grad_norm": 0.21620658040046692, "learning_rate": 9.849829660581363e-05, "loss": 0.0227, "step": 3730 }, { "epoch": 3.9576719576719577, "grad_norm": 0.23454268276691437, "learning_rate": 9.848486076572787e-05, "loss": 0.0242, "step": 3740 }, { "epoch": 3.9682539682539684, "grad_norm": 0.24008338153362274, "learning_rate": 9.847136601196446e-05, "loss": 0.0269, "step": 3750 }, { "epoch": 3.9788359788359786, "grad_norm": 0.2582893967628479, "learning_rate": 9.845781236092078e-05, "loss": 0.0276, "step": 3760 }, { "epoch": 3.9894179894179893, "grad_norm": 0.15563596785068512, "learning_rate": 9.844419982906583e-05, "loss": 0.0227, "step": 3770 }, { "epoch": 4.0, "grad_norm": 0.24047264456748962, "learning_rate": 9.843052843294008e-05, "loss": 0.0216, "step": 3780 }, { "epoch": 4.01058201058201, "grad_norm": 0.12403402477502823, "learning_rate": 9.841679818915559e-05, "loss": 0.0254, "step": 3790 }, { "epoch": 4.021164021164021, "grad_norm": 0.20727738738059998, "learning_rate": 9.840300911439591e-05, "loss": 0.0267, "step": 3800 }, { "epoch": 4.031746031746032, "grad_norm": 0.19928860664367676, "learning_rate": 9.838916122541603e-05, "loss": 0.0262, "step": 3810 }, { "epoch": 4.042328042328043, "grad_norm": 0.16022944450378418, "learning_rate": 9.837525453904246e-05, "loss": 0.0256, "step": 3820 }, { "epoch": 4.052910052910053, "grad_norm": 0.22448217868804932, "learning_rate": 9.836128907217314e-05, "loss": 0.026, "step": 3830 }, { "epoch": 4.063492063492063, "grad_norm": 0.2020333707332611, "learning_rate": 9.834726484177743e-05, "loss": 0.0242, "step": 3840 }, { "epoch": 4.074074074074074, "grad_norm": 0.18751752376556396, "learning_rate": 9.833318186489609e-05, "loss": 0.0249, "step": 3850 }, { "epoch": 4.084656084656085, "grad_norm": 0.27519747614860535, "learning_rate": 9.831904015864126e-05, "loss": 0.0252, "step": 3860 }, { "epoch": 4.095238095238095, "grad_norm": 0.2303021252155304, "learning_rate": 9.830483974019645e-05, "loss": 0.0278, "step": 3870 }, { "epoch": 4.105820105820106, "grad_norm": 0.26601746678352356, "learning_rate": 9.82905806268165e-05, "loss": 0.0266, "step": 3880 }, { "epoch": 4.116402116402116, "grad_norm": 0.27241936326026917, "learning_rate": 9.82762628358276e-05, "loss": 0.0248, "step": 3890 }, { "epoch": 4.1269841269841265, "grad_norm": 0.16039694845676422, "learning_rate": 9.826188638462718e-05, "loss": 0.0281, "step": 3900 }, { "epoch": 4.137566137566138, "grad_norm": 0.22135676443576813, "learning_rate": 9.824745129068402e-05, "loss": 0.0232, "step": 3910 }, { "epoch": 4.148148148148148, "grad_norm": 0.1929844170808792, "learning_rate": 9.82329575715381e-05, "loss": 0.0297, "step": 3920 }, { "epoch": 4.158730158730159, "grad_norm": 0.23706193268299103, "learning_rate": 9.821840524480066e-05, "loss": 0.0233, "step": 3930 }, { "epoch": 4.169312169312169, "grad_norm": 0.2113131731748581, "learning_rate": 9.820379432815414e-05, "loss": 0.022, "step": 3940 }, { "epoch": 4.1798941798941796, "grad_norm": 0.2233104705810547, "learning_rate": 9.81891248393522e-05, "loss": 0.0239, "step": 3950 }, { "epoch": 4.190476190476191, "grad_norm": 0.2668057382106781, "learning_rate": 9.817439679621963e-05, "loss": 0.0268, "step": 3960 }, { "epoch": 4.201058201058201, "grad_norm": 0.2868625819683075, "learning_rate": 9.815961021665243e-05, "loss": 0.0237, "step": 3970 }, { "epoch": 4.211640211640212, "grad_norm": 0.1583409160375595, "learning_rate": 9.814476511861763e-05, "loss": 0.0299, "step": 3980 }, { "epoch": 4.222222222222222, "grad_norm": 0.25458574295043945, "learning_rate": 9.812986152015348e-05, "loss": 0.027, "step": 3990 }, { "epoch": 4.232804232804233, "grad_norm": 0.24500775337219238, "learning_rate": 9.811489943936922e-05, "loss": 0.0286, "step": 4000 }, { "epoch": 4.243386243386244, "grad_norm": 0.19651690125465393, "learning_rate": 9.809987889444522e-05, "loss": 0.0258, "step": 4010 }, { "epoch": 4.253968253968254, "grad_norm": 0.17092068493366241, "learning_rate": 9.808479990363282e-05, "loss": 0.025, "step": 4020 }, { "epoch": 4.264550264550264, "grad_norm": 0.19489115476608276, "learning_rate": 9.806966248525445e-05, "loss": 0.0239, "step": 4030 }, { "epoch": 4.275132275132275, "grad_norm": 0.3030531108379364, "learning_rate": 9.805446665770348e-05, "loss": 0.0232, "step": 4040 }, { "epoch": 4.285714285714286, "grad_norm": 0.24585126340389252, "learning_rate": 9.803921243944429e-05, "loss": 0.0256, "step": 4050 }, { "epoch": 4.296296296296296, "grad_norm": 0.18935532867908478, "learning_rate": 9.802389984901218e-05, "loss": 0.0265, "step": 4060 }, { "epoch": 4.306878306878307, "grad_norm": 0.1543596237897873, "learning_rate": 9.80085289050134e-05, "loss": 0.0228, "step": 4070 }, { "epoch": 4.317460317460317, "grad_norm": 0.21338380873203278, "learning_rate": 9.799309962612508e-05, "loss": 0.0251, "step": 4080 }, { "epoch": 4.328042328042328, "grad_norm": 0.15570683777332306, "learning_rate": 9.797761203109527e-05, "loss": 0.0262, "step": 4090 }, { "epoch": 4.338624338624339, "grad_norm": 0.2029392272233963, "learning_rate": 9.796206613874283e-05, "loss": 0.0281, "step": 4100 }, { "epoch": 4.349206349206349, "grad_norm": 0.3018639385700226, "learning_rate": 9.794646196795754e-05, "loss": 0.026, "step": 4110 }, { "epoch": 4.35978835978836, "grad_norm": 0.26904919743537903, "learning_rate": 9.793079953769987e-05, "loss": 0.0225, "step": 4120 }, { "epoch": 4.37037037037037, "grad_norm": 0.2630746066570282, "learning_rate": 9.79150788670012e-05, "loss": 0.0254, "step": 4130 }, { "epoch": 4.380952380952381, "grad_norm": 0.2744022011756897, "learning_rate": 9.78992999749636e-05, "loss": 0.0271, "step": 4140 }, { "epoch": 4.391534391534392, "grad_norm": 0.2544535994529724, "learning_rate": 9.788346288075994e-05, "loss": 0.0225, "step": 4150 }, { "epoch": 4.402116402116402, "grad_norm": 0.38601627945899963, "learning_rate": 9.786756760363373e-05, "loss": 0.0244, "step": 4160 }, { "epoch": 4.412698412698413, "grad_norm": 0.2449631243944168, "learning_rate": 9.78516141628993e-05, "loss": 0.0238, "step": 4170 }, { "epoch": 4.423280423280423, "grad_norm": 0.24148914217948914, "learning_rate": 9.783560257794154e-05, "loss": 0.0241, "step": 4180 }, { "epoch": 4.4338624338624335, "grad_norm": 0.3827368915081024, "learning_rate": 9.781953286821603e-05, "loss": 0.0217, "step": 4190 }, { "epoch": 4.444444444444445, "grad_norm": 0.18683764338493347, "learning_rate": 9.780340505324901e-05, "loss": 0.0232, "step": 4200 }, { "epoch": 4.455026455026455, "grad_norm": 0.26629316806793213, "learning_rate": 9.778721915263727e-05, "loss": 0.0247, "step": 4210 }, { "epoch": 4.465608465608465, "grad_norm": 0.2593284547328949, "learning_rate": 9.777097518604824e-05, "loss": 0.0281, "step": 4220 }, { "epoch": 4.476190476190476, "grad_norm": 0.2807139754295349, "learning_rate": 9.775467317321984e-05, "loss": 0.0257, "step": 4230 }, { "epoch": 4.4867724867724865, "grad_norm": 0.2142978310585022, "learning_rate": 9.773831313396055e-05, "loss": 0.0325, "step": 4240 }, { "epoch": 4.497354497354498, "grad_norm": 0.19354799389839172, "learning_rate": 9.77218950881494e-05, "loss": 0.0277, "step": 4250 }, { "epoch": 4.507936507936508, "grad_norm": 0.2503867447376251, "learning_rate": 9.770541905573583e-05, "loss": 0.0253, "step": 4260 }, { "epoch": 4.518518518518518, "grad_norm": 0.19918426871299744, "learning_rate": 9.768888505673976e-05, "loss": 0.022, "step": 4270 }, { "epoch": 4.529100529100529, "grad_norm": 0.19768327474594116, "learning_rate": 9.767229311125162e-05, "loss": 0.023, "step": 4280 }, { "epoch": 4.5396825396825395, "grad_norm": 0.18985965847969055, "learning_rate": 9.765564323943211e-05, "loss": 0.0244, "step": 4290 }, { "epoch": 4.550264550264551, "grad_norm": 0.16906580328941345, "learning_rate": 9.763893546151244e-05, "loss": 0.0238, "step": 4300 }, { "epoch": 4.560846560846561, "grad_norm": 0.13173073530197144, "learning_rate": 9.762216979779412e-05, "loss": 0.0221, "step": 4310 }, { "epoch": 4.571428571428571, "grad_norm": 0.19679220020771027, "learning_rate": 9.760534626864902e-05, "loss": 0.0243, "step": 4320 }, { "epoch": 4.582010582010582, "grad_norm": 0.26890161633491516, "learning_rate": 9.758846489451931e-05, "loss": 0.0238, "step": 4330 }, { "epoch": 4.592592592592593, "grad_norm": 0.17014014720916748, "learning_rate": 9.757152569591748e-05, "loss": 0.0228, "step": 4340 }, { "epoch": 4.603174603174603, "grad_norm": 0.15604127943515778, "learning_rate": 9.75545286934262e-05, "loss": 0.0202, "step": 4350 }, { "epoch": 4.613756613756614, "grad_norm": 0.19211073219776154, "learning_rate": 9.753747390769847e-05, "loss": 0.0178, "step": 4360 }, { "epoch": 4.624338624338624, "grad_norm": 0.21945540606975555, "learning_rate": 9.752036135945744e-05, "loss": 0.0249, "step": 4370 }, { "epoch": 4.634920634920634, "grad_norm": 0.21749217808246613, "learning_rate": 9.750319106949649e-05, "loss": 0.0245, "step": 4380 }, { "epoch": 4.645502645502646, "grad_norm": 0.17147471010684967, "learning_rate": 9.748596305867913e-05, "loss": 0.0249, "step": 4390 }, { "epoch": 4.656084656084656, "grad_norm": 0.17796573042869568, "learning_rate": 9.746867734793903e-05, "loss": 0.0245, "step": 4400 }, { "epoch": 4.666666666666667, "grad_norm": 0.19773082435131073, "learning_rate": 9.745133395827993e-05, "loss": 0.0207, "step": 4410 }, { "epoch": 4.677248677248677, "grad_norm": 0.15823116898536682, "learning_rate": 9.743393291077572e-05, "loss": 0.0219, "step": 4420 }, { "epoch": 4.6878306878306875, "grad_norm": 0.1216282993555069, "learning_rate": 9.741647422657028e-05, "loss": 0.0241, "step": 4430 }, { "epoch": 4.698412698412699, "grad_norm": 0.2054513543844223, "learning_rate": 9.739895792687758e-05, "loss": 0.0203, "step": 4440 }, { "epoch": 4.708994708994709, "grad_norm": 0.26556485891342163, "learning_rate": 9.738138403298157e-05, "loss": 0.0268, "step": 4450 }, { "epoch": 4.71957671957672, "grad_norm": 0.22782352566719055, "learning_rate": 9.736375256623619e-05, "loss": 0.0224, "step": 4460 }, { "epoch": 4.73015873015873, "grad_norm": 0.22724492847919464, "learning_rate": 9.734606354806533e-05, "loss": 0.0221, "step": 4470 }, { "epoch": 4.7407407407407405, "grad_norm": 0.14242666959762573, "learning_rate": 9.73283169999628e-05, "loss": 0.0225, "step": 4480 }, { "epoch": 4.751322751322752, "grad_norm": 0.22193555533885956, "learning_rate": 9.731051294349238e-05, "loss": 0.026, "step": 4490 }, { "epoch": 4.761904761904762, "grad_norm": 0.21143323183059692, "learning_rate": 9.729265140028762e-05, "loss": 0.0227, "step": 4500 }, { "epoch": 4.772486772486772, "grad_norm": 0.14150027930736542, "learning_rate": 9.727473239205201e-05, "loss": 0.0206, "step": 4510 }, { "epoch": 4.783068783068783, "grad_norm": 0.18081557750701904, "learning_rate": 9.725675594055883e-05, "loss": 0.026, "step": 4520 }, { "epoch": 4.7936507936507935, "grad_norm": 0.18688952922821045, "learning_rate": 9.723872206765116e-05, "loss": 0.0257, "step": 4530 }, { "epoch": 4.804232804232804, "grad_norm": 0.19352172315120697, "learning_rate": 9.722063079524185e-05, "loss": 0.0268, "step": 4540 }, { "epoch": 4.814814814814815, "grad_norm": 0.24126966297626495, "learning_rate": 9.720248214531351e-05, "loss": 0.0231, "step": 4550 }, { "epoch": 4.825396825396825, "grad_norm": 0.17925548553466797, "learning_rate": 9.718427613991848e-05, "loss": 0.0253, "step": 4560 }, { "epoch": 4.835978835978836, "grad_norm": 0.31540095806121826, "learning_rate": 9.716601280117873e-05, "loss": 0.0234, "step": 4570 }, { "epoch": 4.8465608465608465, "grad_norm": 0.26310083270072937, "learning_rate": 9.714769215128596e-05, "loss": 0.0258, "step": 4580 }, { "epoch": 4.857142857142857, "grad_norm": 0.28720715641975403, "learning_rate": 9.712931421250152e-05, "loss": 0.0285, "step": 4590 }, { "epoch": 4.867724867724868, "grad_norm": 0.24702413380146027, "learning_rate": 9.711087900715627e-05, "loss": 0.0236, "step": 4600 }, { "epoch": 4.878306878306878, "grad_norm": 0.20137597620487213, "learning_rate": 9.709238655765078e-05, "loss": 0.0255, "step": 4610 }, { "epoch": 4.888888888888889, "grad_norm": 0.20297038555145264, "learning_rate": 9.707383688645511e-05, "loss": 0.022, "step": 4620 }, { "epoch": 4.8994708994708995, "grad_norm": 0.15274037420749664, "learning_rate": 9.705523001610883e-05, "loss": 0.0224, "step": 4630 }, { "epoch": 4.91005291005291, "grad_norm": 0.23375476896762848, "learning_rate": 9.703656596922107e-05, "loss": 0.0261, "step": 4640 }, { "epoch": 4.920634920634921, "grad_norm": 0.21456469595432281, "learning_rate": 9.70178447684704e-05, "loss": 0.0278, "step": 4650 }, { "epoch": 4.931216931216931, "grad_norm": 0.17710421979427338, "learning_rate": 9.699906643660483e-05, "loss": 0.0232, "step": 4660 }, { "epoch": 4.941798941798941, "grad_norm": 0.1668485850095749, "learning_rate": 9.698023099644185e-05, "loss": 0.0231, "step": 4670 }, { "epoch": 4.9523809523809526, "grad_norm": 0.19019924104213715, "learning_rate": 9.696133847086823e-05, "loss": 0.0241, "step": 4680 }, { "epoch": 4.962962962962963, "grad_norm": 0.2823118269443512, "learning_rate": 9.694238888284022e-05, "loss": 0.0291, "step": 4690 }, { "epoch": 4.973544973544973, "grad_norm": 0.19920039176940918, "learning_rate": 9.692338225538333e-05, "loss": 0.0239, "step": 4700 }, { "epoch": 4.984126984126984, "grad_norm": 0.14869330823421478, "learning_rate": 9.690431861159241e-05, "loss": 0.0246, "step": 4710 }, { "epoch": 4.994708994708994, "grad_norm": 0.21263545751571655, "learning_rate": 9.688519797463161e-05, "loss": 0.0245, "step": 4720 }, { "epoch": 5.005291005291006, "grad_norm": 0.19977271556854248, "learning_rate": 9.686602036773426e-05, "loss": 0.0226, "step": 4730 }, { "epoch": 5.015873015873016, "grad_norm": 0.17780494689941406, "learning_rate": 9.684678581420302e-05, "loss": 0.021, "step": 4740 }, { "epoch": 5.026455026455026, "grad_norm": 0.16790705919265747, "learning_rate": 9.682749433740962e-05, "loss": 0.0217, "step": 4750 }, { "epoch": 5.037037037037037, "grad_norm": 0.1470905989408493, "learning_rate": 9.680814596079507e-05, "loss": 0.025, "step": 4760 }, { "epoch": 5.0476190476190474, "grad_norm": 0.1309415102005005, "learning_rate": 9.678874070786945e-05, "loss": 0.0259, "step": 4770 }, { "epoch": 5.058201058201059, "grad_norm": 0.1975143551826477, "learning_rate": 9.676927860221199e-05, "loss": 0.0217, "step": 4780 }, { "epoch": 5.068783068783069, "grad_norm": 0.19001717865467072, "learning_rate": 9.674975966747097e-05, "loss": 0.026, "step": 4790 }, { "epoch": 5.079365079365079, "grad_norm": 0.2403058558702469, "learning_rate": 9.673018392736374e-05, "loss": 0.0215, "step": 4800 }, { "epoch": 5.08994708994709, "grad_norm": 0.28260865807533264, "learning_rate": 9.671055140567667e-05, "loss": 0.0249, "step": 4810 }, { "epoch": 5.1005291005291005, "grad_norm": 0.28890368342399597, "learning_rate": 9.669086212626511e-05, "loss": 0.0234, "step": 4820 }, { "epoch": 5.111111111111111, "grad_norm": 0.1421542912721634, "learning_rate": 9.667111611305341e-05, "loss": 0.0222, "step": 4830 }, { "epoch": 5.121693121693122, "grad_norm": 0.18633832037448883, "learning_rate": 9.665131339003486e-05, "loss": 0.0204, "step": 4840 }, { "epoch": 5.132275132275132, "grad_norm": 0.16851188242435455, "learning_rate": 9.663145398127158e-05, "loss": 0.0225, "step": 4850 }, { "epoch": 5.142857142857143, "grad_norm": 0.15450173616409302, "learning_rate": 9.661153791089467e-05, "loss": 0.0217, "step": 4860 }, { "epoch": 5.1534391534391535, "grad_norm": 0.23905780911445618, "learning_rate": 9.659156520310402e-05, "loss": 0.0234, "step": 4870 }, { "epoch": 5.164021164021164, "grad_norm": 0.24496310949325562, "learning_rate": 9.657153588216835e-05, "loss": 0.0243, "step": 4880 }, { "epoch": 5.174603174603175, "grad_norm": 0.19452770054340363, "learning_rate": 9.655144997242516e-05, "loss": 0.024, "step": 4890 }, { "epoch": 5.185185185185185, "grad_norm": 0.3049011528491974, "learning_rate": 9.653130749828075e-05, "loss": 0.0236, "step": 4900 }, { "epoch": 5.195767195767195, "grad_norm": 0.2567329406738281, "learning_rate": 9.65111084842101e-05, "loss": 0.0261, "step": 4910 }, { "epoch": 5.2063492063492065, "grad_norm": 0.20184192061424255, "learning_rate": 9.649085295475695e-05, "loss": 0.0275, "step": 4920 }, { "epoch": 5.216931216931217, "grad_norm": 0.2786656618118286, "learning_rate": 9.647054093453365e-05, "loss": 0.0255, "step": 4930 }, { "epoch": 5.227513227513228, "grad_norm": 0.22429434955120087, "learning_rate": 9.645017244822123e-05, "loss": 0.0225, "step": 4940 }, { "epoch": 5.238095238095238, "grad_norm": 0.23075157403945923, "learning_rate": 9.642974752056931e-05, "loss": 0.0255, "step": 4950 }, { "epoch": 5.248677248677248, "grad_norm": 0.18942216038703918, "learning_rate": 9.640926617639613e-05, "loss": 0.0258, "step": 4960 }, { "epoch": 5.2592592592592595, "grad_norm": 0.1489534229040146, "learning_rate": 9.638872844058843e-05, "loss": 0.024, "step": 4970 }, { "epoch": 5.26984126984127, "grad_norm": 0.15265871584415436, "learning_rate": 9.63681343381015e-05, "loss": 0.024, "step": 4980 }, { "epoch": 5.28042328042328, "grad_norm": 0.16209790110588074, "learning_rate": 9.634748389395914e-05, "loss": 0.0247, "step": 4990 }, { "epoch": 5.291005291005291, "grad_norm": 0.1513448804616928, "learning_rate": 9.632677713325353e-05, "loss": 0.0243, "step": 5000 }, { "epoch": 5.301587301587301, "grad_norm": 0.14487916231155396, "learning_rate": 9.63060140811454e-05, "loss": 0.0234, "step": 5010 }, { "epoch": 5.3121693121693125, "grad_norm": 0.26723966002464294, "learning_rate": 9.628519476286379e-05, "loss": 0.0229, "step": 5020 }, { "epoch": 5.322751322751323, "grad_norm": 0.1334444284439087, "learning_rate": 9.626431920370612e-05, "loss": 0.0227, "step": 5030 }, { "epoch": 5.333333333333333, "grad_norm": 0.27420133352279663, "learning_rate": 9.624338742903819e-05, "loss": 0.0259, "step": 5040 }, { "epoch": 5.343915343915344, "grad_norm": 0.16367533802986145, "learning_rate": 9.622239946429406e-05, "loss": 0.0212, "step": 5050 }, { "epoch": 5.354497354497354, "grad_norm": 0.13817910850048065, "learning_rate": 9.620135533497609e-05, "loss": 0.0212, "step": 5060 }, { "epoch": 5.365079365079365, "grad_norm": 0.13842833042144775, "learning_rate": 9.61802550666549e-05, "loss": 0.0214, "step": 5070 }, { "epoch": 5.375661375661376, "grad_norm": 0.17538587749004364, "learning_rate": 9.615909868496928e-05, "loss": 0.0222, "step": 5080 }, { "epoch": 5.386243386243386, "grad_norm": 0.1784484088420868, "learning_rate": 9.613788621562622e-05, "loss": 0.0202, "step": 5090 }, { "epoch": 5.396825396825397, "grad_norm": 0.18745511770248413, "learning_rate": 9.611661768440093e-05, "loss": 0.0203, "step": 5100 }, { "epoch": 5.407407407407407, "grad_norm": 0.15349143743515015, "learning_rate": 9.609529311713661e-05, "loss": 0.0216, "step": 5110 }, { "epoch": 5.417989417989418, "grad_norm": 0.16199474036693573, "learning_rate": 9.607391253974466e-05, "loss": 0.0198, "step": 5120 }, { "epoch": 5.428571428571429, "grad_norm": 0.19510476291179657, "learning_rate": 9.605247597820448e-05, "loss": 0.0204, "step": 5130 }, { "epoch": 5.439153439153439, "grad_norm": 0.16607210040092468, "learning_rate": 9.603098345856354e-05, "loss": 0.0216, "step": 5140 }, { "epoch": 5.449735449735449, "grad_norm": 0.2188313752412796, "learning_rate": 9.600943500693725e-05, "loss": 0.0248, "step": 5150 }, { "epoch": 5.4603174603174605, "grad_norm": 0.2104046642780304, "learning_rate": 9.598783064950902e-05, "loss": 0.0233, "step": 5160 }, { "epoch": 5.470899470899471, "grad_norm": 0.12807565927505493, "learning_rate": 9.596617041253018e-05, "loss": 0.0208, "step": 5170 }, { "epoch": 5.481481481481482, "grad_norm": 0.28353264927864075, "learning_rate": 9.594445432231996e-05, "loss": 0.0208, "step": 5180 }, { "epoch": 5.492063492063492, "grad_norm": 0.18138745427131653, "learning_rate": 9.592268240526547e-05, "loss": 0.0243, "step": 5190 }, { "epoch": 5.502645502645502, "grad_norm": 0.1664595603942871, "learning_rate": 9.590085468782162e-05, "loss": 0.0233, "step": 5200 }, { "epoch": 5.5132275132275135, "grad_norm": 0.12447670847177505, "learning_rate": 9.587897119651116e-05, "loss": 0.0265, "step": 5210 }, { "epoch": 5.523809523809524, "grad_norm": 0.2000388652086258, "learning_rate": 9.585703195792459e-05, "loss": 0.0235, "step": 5220 }, { "epoch": 5.534391534391535, "grad_norm": 0.2775711417198181, "learning_rate": 9.583503699872016e-05, "loss": 0.0225, "step": 5230 }, { "epoch": 5.544973544973545, "grad_norm": 0.20965375006198883, "learning_rate": 9.581298634562381e-05, "loss": 0.0238, "step": 5240 }, { "epoch": 5.555555555555555, "grad_norm": 0.1587841808795929, "learning_rate": 9.579088002542917e-05, "loss": 0.0236, "step": 5250 }, { "epoch": 5.5661375661375665, "grad_norm": 0.1798291802406311, "learning_rate": 9.57687180649975e-05, "loss": 0.0247, "step": 5260 }, { "epoch": 5.576719576719577, "grad_norm": 0.19637994468212128, "learning_rate": 9.574650049125768e-05, "loss": 0.0205, "step": 5270 }, { "epoch": 5.587301587301587, "grad_norm": 0.3106609284877777, "learning_rate": 9.572422733120614e-05, "loss": 0.0245, "step": 5280 }, { "epoch": 5.597883597883598, "grad_norm": 0.1459248811006546, "learning_rate": 9.570189861190689e-05, "loss": 0.0227, "step": 5290 }, { "epoch": 5.608465608465608, "grad_norm": 0.15138126909732819, "learning_rate": 9.56795143604914e-05, "loss": 0.0226, "step": 5300 }, { "epoch": 5.619047619047619, "grad_norm": 0.17585909366607666, "learning_rate": 9.565707460415869e-05, "loss": 0.0185, "step": 5310 }, { "epoch": 5.62962962962963, "grad_norm": 0.1764351725578308, "learning_rate": 9.563457937017515e-05, "loss": 0.0217, "step": 5320 }, { "epoch": 5.64021164021164, "grad_norm": 0.2536950409412384, "learning_rate": 9.56120286858746e-05, "loss": 0.0226, "step": 5330 }, { "epoch": 5.650793650793651, "grad_norm": 0.17318256199359894, "learning_rate": 9.558942257865829e-05, "loss": 0.0265, "step": 5340 }, { "epoch": 5.661375661375661, "grad_norm": 0.15001510083675385, "learning_rate": 9.556676107599472e-05, "loss": 0.0231, "step": 5350 }, { "epoch": 5.671957671957672, "grad_norm": 0.18146498501300812, "learning_rate": 9.554404420541978e-05, "loss": 0.0236, "step": 5360 }, { "epoch": 5.682539682539683, "grad_norm": 0.14126279950141907, "learning_rate": 9.55212719945366e-05, "loss": 0.0209, "step": 5370 }, { "epoch": 5.693121693121693, "grad_norm": 0.16769146919250488, "learning_rate": 9.549844447101559e-05, "loss": 0.0212, "step": 5380 }, { "epoch": 5.703703703703704, "grad_norm": 0.22893056273460388, "learning_rate": 9.547556166259433e-05, "loss": 0.0207, "step": 5390 }, { "epoch": 5.714285714285714, "grad_norm": 0.16360661387443542, "learning_rate": 9.545262359707756e-05, "loss": 0.0226, "step": 5400 }, { "epoch": 5.724867724867725, "grad_norm": 0.18979816138744354, "learning_rate": 9.542963030233724e-05, "loss": 0.0229, "step": 5410 }, { "epoch": 5.735449735449736, "grad_norm": 0.1348208487033844, "learning_rate": 9.540658180631237e-05, "loss": 0.022, "step": 5420 }, { "epoch": 5.746031746031746, "grad_norm": 0.15333133935928345, "learning_rate": 9.538347813700904e-05, "loss": 0.0226, "step": 5430 }, { "epoch": 5.756613756613756, "grad_norm": 0.17689169943332672, "learning_rate": 9.536031932250036e-05, "loss": 0.0242, "step": 5440 }, { "epoch": 5.767195767195767, "grad_norm": 0.11487551778554916, "learning_rate": 9.533710539092653e-05, "loss": 0.0223, "step": 5450 }, { "epoch": 5.777777777777778, "grad_norm": 0.2047729343175888, "learning_rate": 9.531383637049464e-05, "loss": 0.0194, "step": 5460 }, { "epoch": 5.788359788359788, "grad_norm": 0.17957742512226105, "learning_rate": 9.529051228947875e-05, "loss": 0.0216, "step": 5470 }, { "epoch": 5.798941798941799, "grad_norm": 0.26392608880996704, "learning_rate": 9.52671331762198e-05, "loss": 0.026, "step": 5480 }, { "epoch": 5.809523809523809, "grad_norm": 0.1643499881029129, "learning_rate": 9.524369905912565e-05, "loss": 0.0221, "step": 5490 }, { "epoch": 5.8201058201058204, "grad_norm": 0.2017263025045395, "learning_rate": 9.522020996667092e-05, "loss": 0.0249, "step": 5500 }, { "epoch": 5.830687830687831, "grad_norm": 0.12917660176753998, "learning_rate": 9.519666592739709e-05, "loss": 0.0187, "step": 5510 }, { "epoch": 5.841269841269841, "grad_norm": 0.19572894275188446, "learning_rate": 9.517306696991241e-05, "loss": 0.0227, "step": 5520 }, { "epoch": 5.851851851851852, "grad_norm": 0.1744583696126938, "learning_rate": 9.51494131228918e-05, "loss": 0.0263, "step": 5530 }, { "epoch": 5.862433862433862, "grad_norm": 0.17951607704162598, "learning_rate": 9.512570441507695e-05, "loss": 0.0241, "step": 5540 }, { "epoch": 5.8730158730158735, "grad_norm": 0.21809431910514832, "learning_rate": 9.510194087527615e-05, "loss": 0.0205, "step": 5550 }, { "epoch": 5.883597883597884, "grad_norm": 0.18476571142673492, "learning_rate": 9.507812253236435e-05, "loss": 0.0203, "step": 5560 }, { "epoch": 5.894179894179894, "grad_norm": 0.18377670645713806, "learning_rate": 9.505424941528309e-05, "loss": 0.023, "step": 5570 }, { "epoch": 5.904761904761905, "grad_norm": 0.13379259407520294, "learning_rate": 9.503032155304046e-05, "loss": 0.0266, "step": 5580 }, { "epoch": 5.915343915343915, "grad_norm": 0.24041050672531128, "learning_rate": 9.500633897471106e-05, "loss": 0.0217, "step": 5590 }, { "epoch": 5.925925925925926, "grad_norm": 0.21654775738716125, "learning_rate": 9.498230170943596e-05, "loss": 0.0237, "step": 5600 }, { "epoch": 5.936507936507937, "grad_norm": 0.16069233417510986, "learning_rate": 9.495820978642275e-05, "loss": 0.0226, "step": 5610 }, { "epoch": 5.947089947089947, "grad_norm": 0.16103222966194153, "learning_rate": 9.493406323494535e-05, "loss": 0.0227, "step": 5620 }, { "epoch": 5.957671957671957, "grad_norm": 0.18171921372413635, "learning_rate": 9.490986208434413e-05, "loss": 0.0225, "step": 5630 }, { "epoch": 5.968253968253968, "grad_norm": 0.18027199804782867, "learning_rate": 9.488560636402577e-05, "loss": 0.0219, "step": 5640 }, { "epoch": 5.978835978835979, "grad_norm": 0.19509384036064148, "learning_rate": 9.486129610346321e-05, "loss": 0.0206, "step": 5650 }, { "epoch": 5.98941798941799, "grad_norm": 0.2014419436454773, "learning_rate": 9.483693133219576e-05, "loss": 0.0192, "step": 5660 }, { "epoch": 6.0, "grad_norm": 0.29371291399002075, "learning_rate": 9.481251207982888e-05, "loss": 0.0244, "step": 5670 }, { "epoch": 6.01058201058201, "grad_norm": 0.18492309749126434, "learning_rate": 9.47880383760343e-05, "loss": 0.0246, "step": 5680 }, { "epoch": 6.021164021164021, "grad_norm": 0.34969255328178406, "learning_rate": 9.476351025054983e-05, "loss": 0.0238, "step": 5690 }, { "epoch": 6.031746031746032, "grad_norm": 0.2165185809135437, "learning_rate": 9.473892773317952e-05, "loss": 0.0239, "step": 5700 }, { "epoch": 6.042328042328043, "grad_norm": 0.14802700281143188, "learning_rate": 9.471429085379338e-05, "loss": 0.0228, "step": 5710 }, { "epoch": 6.052910052910053, "grad_norm": 0.17832553386688232, "learning_rate": 9.468959964232757e-05, "loss": 0.0203, "step": 5720 }, { "epoch": 6.063492063492063, "grad_norm": 0.1818743646144867, "learning_rate": 9.466485412878425e-05, "loss": 0.0243, "step": 5730 }, { "epoch": 6.074074074074074, "grad_norm": 0.48946550488471985, "learning_rate": 9.464005434323154e-05, "loss": 0.0225, "step": 5740 }, { "epoch": 6.084656084656085, "grad_norm": 0.2275162935256958, "learning_rate": 9.461520031580352e-05, "loss": 0.0229, "step": 5750 }, { "epoch": 6.095238095238095, "grad_norm": 0.21593688428401947, "learning_rate": 9.459029207670019e-05, "loss": 0.0229, "step": 5760 }, { "epoch": 6.105820105820106, "grad_norm": 0.2040182501077652, "learning_rate": 9.456532965618737e-05, "loss": 0.0231, "step": 5770 }, { "epoch": 6.116402116402116, "grad_norm": 0.14748108386993408, "learning_rate": 9.454031308459681e-05, "loss": 0.0224, "step": 5780 }, { "epoch": 6.1269841269841265, "grad_norm": 0.18563517928123474, "learning_rate": 9.451524239232595e-05, "loss": 0.0209, "step": 5790 }, { "epoch": 6.137566137566138, "grad_norm": 0.3300509750843048, "learning_rate": 9.449011760983809e-05, "loss": 0.0208, "step": 5800 }, { "epoch": 6.148148148148148, "grad_norm": 0.2601502239704132, "learning_rate": 9.446493876766218e-05, "loss": 0.0198, "step": 5810 }, { "epoch": 6.158730158730159, "grad_norm": 0.1712155044078827, "learning_rate": 9.44397058963929e-05, "loss": 0.0242, "step": 5820 }, { "epoch": 6.169312169312169, "grad_norm": 0.12319833040237427, "learning_rate": 9.441441902669056e-05, "loss": 0.0208, "step": 5830 }, { "epoch": 6.1798941798941796, "grad_norm": 0.2489125281572342, "learning_rate": 9.43890781892811e-05, "loss": 0.0206, "step": 5840 }, { "epoch": 6.190476190476191, "grad_norm": 0.12324065715074539, "learning_rate": 9.436368341495603e-05, "loss": 0.022, "step": 5850 }, { "epoch": 6.201058201058201, "grad_norm": 0.2001962959766388, "learning_rate": 9.43382347345724e-05, "loss": 0.0258, "step": 5860 }, { "epoch": 6.211640211640212, "grad_norm": 0.2177634835243225, "learning_rate": 9.431273217905273e-05, "loss": 0.0248, "step": 5870 }, { "epoch": 6.222222222222222, "grad_norm": 0.25731775164604187, "learning_rate": 9.428717577938504e-05, "loss": 0.0232, "step": 5880 }, { "epoch": 6.232804232804233, "grad_norm": 0.2122044563293457, "learning_rate": 9.426156556662276e-05, "loss": 0.0249, "step": 5890 }, { "epoch": 6.243386243386244, "grad_norm": 0.19205014407634735, "learning_rate": 9.423590157188474e-05, "loss": 0.0224, "step": 5900 }, { "epoch": 6.253968253968254, "grad_norm": 0.16996918618679047, "learning_rate": 9.421018382635513e-05, "loss": 0.0206, "step": 5910 }, { "epoch": 6.264550264550264, "grad_norm": 0.13215231895446777, "learning_rate": 9.418441236128343e-05, "loss": 0.0222, "step": 5920 }, { "epoch": 6.275132275132275, "grad_norm": 0.14536046981811523, "learning_rate": 9.41585872079844e-05, "loss": 0.0218, "step": 5930 }, { "epoch": 6.285714285714286, "grad_norm": 0.1779017448425293, "learning_rate": 9.413270839783802e-05, "loss": 0.0224, "step": 5940 }, { "epoch": 6.296296296296296, "grad_norm": 0.19265064597129822, "learning_rate": 9.41067759622895e-05, "loss": 0.0233, "step": 5950 }, { "epoch": 6.306878306878307, "grad_norm": 0.18961210548877716, "learning_rate": 9.408078993284917e-05, "loss": 0.0256, "step": 5960 }, { "epoch": 6.317460317460317, "grad_norm": 0.1714249700307846, "learning_rate": 9.405475034109254e-05, "loss": 0.0233, "step": 5970 }, { "epoch": 6.328042328042328, "grad_norm": 0.18181532621383667, "learning_rate": 9.402865721866015e-05, "loss": 0.0252, "step": 5980 }, { "epoch": 6.338624338624339, "grad_norm": 0.21735495328903198, "learning_rate": 9.400251059725762e-05, "loss": 0.0254, "step": 5990 }, { "epoch": 6.349206349206349, "grad_norm": 0.2158617228269577, "learning_rate": 9.397631050865554e-05, "loss": 0.02, "step": 6000 }, { "epoch": 6.35978835978836, "grad_norm": 0.149308979511261, "learning_rate": 9.395005698468949e-05, "loss": 0.0215, "step": 6010 }, { "epoch": 6.37037037037037, "grad_norm": 0.17855429649353027, "learning_rate": 9.392375005726e-05, "loss": 0.0192, "step": 6020 }, { "epoch": 6.380952380952381, "grad_norm": 0.18988282978534698, "learning_rate": 9.389738975833242e-05, "loss": 0.0207, "step": 6030 }, { "epoch": 6.391534391534392, "grad_norm": 0.2102522850036621, "learning_rate": 9.387097611993707e-05, "loss": 0.0249, "step": 6040 }, { "epoch": 6.402116402116402, "grad_norm": 0.18764233589172363, "learning_rate": 9.384450917416894e-05, "loss": 0.0236, "step": 6050 }, { "epoch": 6.412698412698413, "grad_norm": 0.23040169477462769, "learning_rate": 9.381798895318792e-05, "loss": 0.0202, "step": 6060 }, { "epoch": 6.423280423280423, "grad_norm": 0.2280212938785553, "learning_rate": 9.379141548921853e-05, "loss": 0.0198, "step": 6070 }, { "epoch": 6.4338624338624335, "grad_norm": 0.12011121958494186, "learning_rate": 9.376478881455009e-05, "loss": 0.028, "step": 6080 }, { "epoch": 6.444444444444445, "grad_norm": 0.1720249205827713, "learning_rate": 9.373810896153647e-05, "loss": 0.0201, "step": 6090 }, { "epoch": 6.455026455026455, "grad_norm": 0.277042031288147, "learning_rate": 9.371137596259623e-05, "loss": 0.0212, "step": 6100 }, { "epoch": 6.465608465608465, "grad_norm": 0.1936749517917633, "learning_rate": 9.368458985021248e-05, "loss": 0.0221, "step": 6110 }, { "epoch": 6.476190476190476, "grad_norm": 0.1833265870809555, "learning_rate": 9.365775065693287e-05, "loss": 0.0177, "step": 6120 }, { "epoch": 6.4867724867724865, "grad_norm": 0.18313643336296082, "learning_rate": 9.363085841536957e-05, "loss": 0.0195, "step": 6130 }, { "epoch": 6.497354497354498, "grad_norm": 0.18789002299308777, "learning_rate": 9.360391315819917e-05, "loss": 0.0208, "step": 6140 }, { "epoch": 6.507936507936508, "grad_norm": 0.1849595010280609, "learning_rate": 9.357691491816269e-05, "loss": 0.0202, "step": 6150 }, { "epoch": 6.518518518518518, "grad_norm": 0.17876897752285004, "learning_rate": 9.354986372806557e-05, "loss": 0.019, "step": 6160 }, { "epoch": 6.529100529100529, "grad_norm": 0.20616750419139862, "learning_rate": 9.352275962077752e-05, "loss": 0.0206, "step": 6170 }, { "epoch": 6.5396825396825395, "grad_norm": 0.21335752308368683, "learning_rate": 9.349560262923262e-05, "loss": 0.0205, "step": 6180 }, { "epoch": 6.550264550264551, "grad_norm": 0.19443462789058685, "learning_rate": 9.346839278642913e-05, "loss": 0.0203, "step": 6190 }, { "epoch": 6.560846560846561, "grad_norm": 0.2122451514005661, "learning_rate": 9.344113012542963e-05, "loss": 0.0221, "step": 6200 }, { "epoch": 6.571428571428571, "grad_norm": 0.19128330051898956, "learning_rate": 9.341381467936079e-05, "loss": 0.0234, "step": 6210 }, { "epoch": 6.582010582010582, "grad_norm": 0.1612822413444519, "learning_rate": 9.338644648141346e-05, "loss": 0.0195, "step": 6220 }, { "epoch": 6.592592592592593, "grad_norm": 0.12845978140830994, "learning_rate": 9.335902556484257e-05, "loss": 0.0218, "step": 6230 }, { "epoch": 6.603174603174603, "grad_norm": 0.17286637425422668, "learning_rate": 9.333155196296712e-05, "loss": 0.0222, "step": 6240 }, { "epoch": 6.613756613756614, "grad_norm": 0.12904222309589386, "learning_rate": 9.330402570917016e-05, "loss": 0.02, "step": 6250 }, { "epoch": 6.624338624338624, "grad_norm": 0.18530046939849854, "learning_rate": 9.327644683689865e-05, "loss": 0.0186, "step": 6260 }, { "epoch": 6.634920634920634, "grad_norm": 0.12361671030521393, "learning_rate": 9.324881537966354e-05, "loss": 0.0219, "step": 6270 }, { "epoch": 6.645502645502646, "grad_norm": 0.1665956676006317, "learning_rate": 9.322113137103964e-05, "loss": 0.0233, "step": 6280 }, { "epoch": 6.656084656084656, "grad_norm": 0.19551175832748413, "learning_rate": 9.319339484466564e-05, "loss": 0.022, "step": 6290 }, { "epoch": 6.666666666666667, "grad_norm": 0.21934978663921356, "learning_rate": 9.316560583424403e-05, "loss": 0.0261, "step": 6300 }, { "epoch": 6.677248677248677, "grad_norm": 0.19010791182518005, "learning_rate": 9.313776437354108e-05, "loss": 0.0231, "step": 6310 }, { "epoch": 6.6878306878306875, "grad_norm": 0.1399822235107422, "learning_rate": 9.31098704963868e-05, "loss": 0.0217, "step": 6320 }, { "epoch": 6.698412698412699, "grad_norm": 0.23027560114860535, "learning_rate": 9.308192423667485e-05, "loss": 0.0205, "step": 6330 }, { "epoch": 6.708994708994709, "grad_norm": 0.24820154905319214, "learning_rate": 9.305392562836262e-05, "loss": 0.0206, "step": 6340 }, { "epoch": 6.71957671957672, "grad_norm": 0.1495678722858429, "learning_rate": 9.3025874705471e-05, "loss": 0.0209, "step": 6350 }, { "epoch": 6.73015873015873, "grad_norm": 0.20092251896858215, "learning_rate": 9.299777150208457e-05, "loss": 0.023, "step": 6360 }, { "epoch": 6.7407407407407405, "grad_norm": 0.1592237651348114, "learning_rate": 9.296961605235133e-05, "loss": 0.0197, "step": 6370 }, { "epoch": 6.751322751322752, "grad_norm": 0.19090069830417633, "learning_rate": 9.29414083904828e-05, "loss": 0.0181, "step": 6380 }, { "epoch": 6.761904761904762, "grad_norm": 0.1444035917520523, "learning_rate": 9.291314855075397e-05, "loss": 0.0226, "step": 6390 }, { "epoch": 6.772486772486772, "grad_norm": 0.15458105504512787, "learning_rate": 9.288483656750322e-05, "loss": 0.0229, "step": 6400 }, { "epoch": 6.783068783068783, "grad_norm": 0.15021175146102905, "learning_rate": 9.285647247513225e-05, "loss": 0.0219, "step": 6410 }, { "epoch": 6.7936507936507935, "grad_norm": 0.21341487765312195, "learning_rate": 9.282805630810614e-05, "loss": 0.0207, "step": 6420 }, { "epoch": 6.804232804232804, "grad_norm": 0.1554815173149109, "learning_rate": 9.279958810095317e-05, "loss": 0.0183, "step": 6430 }, { "epoch": 6.814814814814815, "grad_norm": 0.15357381105422974, "learning_rate": 9.277106788826494e-05, "loss": 0.021, "step": 6440 }, { "epoch": 6.825396825396825, "grad_norm": 0.17323346436023712, "learning_rate": 9.274249570469618e-05, "loss": 0.0199, "step": 6450 }, { "epoch": 6.835978835978836, "grad_norm": 0.11775460839271545, "learning_rate": 9.271387158496476e-05, "loss": 0.0215, "step": 6460 }, { "epoch": 6.8465608465608465, "grad_norm": 0.1533900797367096, "learning_rate": 9.268519556385173e-05, "loss": 0.0222, "step": 6470 }, { "epoch": 6.857142857142857, "grad_norm": 0.166974738240242, "learning_rate": 9.265646767620113e-05, "loss": 0.0183, "step": 6480 }, { "epoch": 6.867724867724868, "grad_norm": 0.1824696660041809, "learning_rate": 9.262768795692006e-05, "loss": 0.0225, "step": 6490 }, { "epoch": 6.878306878306878, "grad_norm": 0.1770448237657547, "learning_rate": 9.259885644097861e-05, "loss": 0.0207, "step": 6500 }, { "epoch": 6.888888888888889, "grad_norm": 0.2062506228685379, "learning_rate": 9.256997316340976e-05, "loss": 0.0212, "step": 6510 }, { "epoch": 6.8994708994708995, "grad_norm": 0.21468693017959595, "learning_rate": 9.254103815930943e-05, "loss": 0.0234, "step": 6520 }, { "epoch": 6.91005291005291, "grad_norm": 0.18053120374679565, "learning_rate": 9.251205146383636e-05, "loss": 0.0212, "step": 6530 }, { "epoch": 6.920634920634921, "grad_norm": 0.19176653027534485, "learning_rate": 9.248301311221216e-05, "loss": 0.0208, "step": 6540 }, { "epoch": 6.931216931216931, "grad_norm": 0.14799228310585022, "learning_rate": 9.245392313972115e-05, "loss": 0.0248, "step": 6550 }, { "epoch": 6.941798941798941, "grad_norm": 0.13719846308231354, "learning_rate": 9.242478158171037e-05, "loss": 0.0219, "step": 6560 }, { "epoch": 6.9523809523809526, "grad_norm": 0.1645166277885437, "learning_rate": 9.239558847358957e-05, "loss": 0.0231, "step": 6570 }, { "epoch": 6.962962962962963, "grad_norm": 0.19157449901103973, "learning_rate": 9.236634385083114e-05, "loss": 0.0206, "step": 6580 }, { "epoch": 6.973544973544973, "grad_norm": 0.1357564777135849, "learning_rate": 9.233704774897005e-05, "loss": 0.0215, "step": 6590 }, { "epoch": 6.984126984126984, "grad_norm": 0.16563691198825836, "learning_rate": 9.230770020360382e-05, "loss": 0.0226, "step": 6600 }, { "epoch": 6.994708994708994, "grad_norm": 0.1946519911289215, "learning_rate": 9.227830125039247e-05, "loss": 0.0211, "step": 6610 }, { "epoch": 7.005291005291006, "grad_norm": 0.21056078374385834, "learning_rate": 9.224885092505853e-05, "loss": 0.0237, "step": 6620 }, { "epoch": 7.015873015873016, "grad_norm": 0.17660044133663177, "learning_rate": 9.221934926338691e-05, "loss": 0.0196, "step": 6630 }, { "epoch": 7.026455026455026, "grad_norm": 0.21623964607715607, "learning_rate": 9.21897963012249e-05, "loss": 0.0238, "step": 6640 }, { "epoch": 7.037037037037037, "grad_norm": 0.2419058233499527, "learning_rate": 9.216019207448217e-05, "loss": 0.0193, "step": 6650 }, { "epoch": 7.0476190476190474, "grad_norm": 0.19709932804107666, "learning_rate": 9.21305366191306e-05, "loss": 0.0196, "step": 6660 }, { "epoch": 7.058201058201059, "grad_norm": 0.16408425569534302, "learning_rate": 9.21008299712044e-05, "loss": 0.0218, "step": 6670 }, { "epoch": 7.068783068783069, "grad_norm": 0.20442360639572144, "learning_rate": 9.207107216679993e-05, "loss": 0.0244, "step": 6680 }, { "epoch": 7.079365079365079, "grad_norm": 0.23607459664344788, "learning_rate": 9.204126324207574e-05, "loss": 0.0208, "step": 6690 }, { "epoch": 7.08994708994709, "grad_norm": 0.16632795333862305, "learning_rate": 9.201140323325248e-05, "loss": 0.0243, "step": 6700 }, { "epoch": 7.1005291005291005, "grad_norm": 0.14960983395576477, "learning_rate": 9.198149217661287e-05, "loss": 0.023, "step": 6710 }, { "epoch": 7.111111111111111, "grad_norm": 0.19850869476795197, "learning_rate": 9.195153010850167e-05, "loss": 0.0279, "step": 6720 }, { "epoch": 7.121693121693122, "grad_norm": 0.24772217869758606, "learning_rate": 9.192151706532562e-05, "loss": 0.025, "step": 6730 }, { "epoch": 7.132275132275132, "grad_norm": 0.225437730550766, "learning_rate": 9.189145308355337e-05, "loss": 0.0249, "step": 6740 }, { "epoch": 7.142857142857143, "grad_norm": 0.17270970344543457, "learning_rate": 9.186133819971555e-05, "loss": 0.0238, "step": 6750 }, { "epoch": 7.1534391534391535, "grad_norm": 0.16929887235164642, "learning_rate": 9.183117245040455e-05, "loss": 0.0221, "step": 6760 }, { "epoch": 7.164021164021164, "grad_norm": 0.16940592229366302, "learning_rate": 9.180095587227458e-05, "loss": 0.0196, "step": 6770 }, { "epoch": 7.174603174603175, "grad_norm": 0.2192145138978958, "learning_rate": 9.177068850204167e-05, "loss": 0.023, "step": 6780 }, { "epoch": 7.185185185185185, "grad_norm": 0.17907096445560455, "learning_rate": 9.174037037648351e-05, "loss": 0.0236, "step": 6790 }, { "epoch": 7.195767195767195, "grad_norm": 0.27532628178596497, "learning_rate": 9.171000153243947e-05, "loss": 0.0202, "step": 6800 }, { "epoch": 7.2063492063492065, "grad_norm": 0.15969857573509216, "learning_rate": 9.167958200681057e-05, "loss": 0.0194, "step": 6810 }, { "epoch": 7.216931216931217, "grad_norm": 0.17519356310367584, "learning_rate": 9.164911183655943e-05, "loss": 0.0222, "step": 6820 }, { "epoch": 7.227513227513228, "grad_norm": 0.16141709685325623, "learning_rate": 9.161859105871013e-05, "loss": 0.0198, "step": 6830 }, { "epoch": 7.238095238095238, "grad_norm": 0.14699578285217285, "learning_rate": 9.158801971034832e-05, "loss": 0.0219, "step": 6840 }, { "epoch": 7.248677248677248, "grad_norm": 0.23129884898662567, "learning_rate": 9.155739782862107e-05, "loss": 0.0193, "step": 6850 }, { "epoch": 7.2592592592592595, "grad_norm": 0.12972491979599, "learning_rate": 9.152672545073687e-05, "loss": 0.0228, "step": 6860 }, { "epoch": 7.26984126984127, "grad_norm": 0.17482340335845947, "learning_rate": 9.149600261396551e-05, "loss": 0.0222, "step": 6870 }, { "epoch": 7.28042328042328, "grad_norm": 0.1542358547449112, "learning_rate": 9.146522935563817e-05, "loss": 0.0227, "step": 6880 }, { "epoch": 7.291005291005291, "grad_norm": 0.18554989993572235, "learning_rate": 9.143440571314723e-05, "loss": 0.022, "step": 6890 }, { "epoch": 7.301587301587301, "grad_norm": 0.21276222169399261, "learning_rate": 9.140353172394636e-05, "loss": 0.0233, "step": 6900 }, { "epoch": 7.3121693121693125, "grad_norm": 0.17499566078186035, "learning_rate": 9.137260742555032e-05, "loss": 0.0224, "step": 6910 }, { "epoch": 7.322751322751323, "grad_norm": 0.1837032288312912, "learning_rate": 9.13416328555351e-05, "loss": 0.0213, "step": 6920 }, { "epoch": 7.333333333333333, "grad_norm": 0.17598700523376465, "learning_rate": 9.131060805153769e-05, "loss": 0.0239, "step": 6930 }, { "epoch": 7.343915343915344, "grad_norm": 0.1783585250377655, "learning_rate": 9.127953305125618e-05, "loss": 0.023, "step": 6940 }, { "epoch": 7.354497354497354, "grad_norm": 0.19481849670410156, "learning_rate": 9.124840789244958e-05, "loss": 0.0208, "step": 6950 }, { "epoch": 7.365079365079365, "grad_norm": 0.2063082456588745, "learning_rate": 9.121723261293794e-05, "loss": 0.0232, "step": 6960 }, { "epoch": 7.375661375661376, "grad_norm": 0.16276001930236816, "learning_rate": 9.118600725060214e-05, "loss": 0.0189, "step": 6970 }, { "epoch": 7.386243386243386, "grad_norm": 0.20677883923053741, "learning_rate": 9.115473184338392e-05, "loss": 0.0207, "step": 6980 }, { "epoch": 7.396825396825397, "grad_norm": 0.14885324239730835, "learning_rate": 9.11234064292859e-05, "loss": 0.0237, "step": 6990 }, { "epoch": 7.407407407407407, "grad_norm": 0.13029098510742188, "learning_rate": 9.109203104637137e-05, "loss": 0.0187, "step": 7000 }, { "epoch": 7.417989417989418, "grad_norm": 0.14754833281040192, "learning_rate": 9.10606057327644e-05, "loss": 0.0194, "step": 7010 }, { "epoch": 7.428571428571429, "grad_norm": 0.21491985023021698, "learning_rate": 9.10291305266497e-05, "loss": 0.0196, "step": 7020 }, { "epoch": 7.439153439153439, "grad_norm": 0.24787749350070953, "learning_rate": 9.099760546627261e-05, "loss": 0.0196, "step": 7030 }, { "epoch": 7.449735449735449, "grad_norm": 0.19525426626205444, "learning_rate": 9.096603058993907e-05, "loss": 0.0191, "step": 7040 }, { "epoch": 7.4603174603174605, "grad_norm": 0.23715920746326447, "learning_rate": 9.093440593601552e-05, "loss": 0.0191, "step": 7050 }, { "epoch": 7.470899470899471, "grad_norm": 0.17566142976284027, "learning_rate": 9.090273154292889e-05, "loss": 0.0178, "step": 7060 }, { "epoch": 7.481481481481482, "grad_norm": 0.12365799397230148, "learning_rate": 9.087100744916656e-05, "loss": 0.0216, "step": 7070 }, { "epoch": 7.492063492063492, "grad_norm": 0.1712784618139267, "learning_rate": 9.08392336932763e-05, "loss": 0.0202, "step": 7080 }, { "epoch": 7.502645502645502, "grad_norm": 0.19723421335220337, "learning_rate": 9.080741031386619e-05, "loss": 0.0198, "step": 7090 }, { "epoch": 7.5132275132275135, "grad_norm": 0.17695248126983643, "learning_rate": 9.077553734960468e-05, "loss": 0.0196, "step": 7100 }, { "epoch": 7.523809523809524, "grad_norm": 0.30005213618278503, "learning_rate": 9.074361483922041e-05, "loss": 0.022, "step": 7110 }, { "epoch": 7.534391534391535, "grad_norm": 0.1541295349597931, "learning_rate": 9.071164282150224e-05, "loss": 0.0222, "step": 7120 }, { "epoch": 7.544973544973545, "grad_norm": 0.2410442978143692, "learning_rate": 9.067962133529919e-05, "loss": 0.0213, "step": 7130 }, { "epoch": 7.555555555555555, "grad_norm": 0.16902434825897217, "learning_rate": 9.064755041952036e-05, "loss": 0.0182, "step": 7140 }, { "epoch": 7.5661375661375665, "grad_norm": 0.18500594794750214, "learning_rate": 9.061543011313497e-05, "loss": 0.0176, "step": 7150 }, { "epoch": 7.576719576719577, "grad_norm": 0.1659843623638153, "learning_rate": 9.05832604551722e-05, "loss": 0.0202, "step": 7160 }, { "epoch": 7.587301587301587, "grad_norm": 0.1577642858028412, "learning_rate": 9.055104148472124e-05, "loss": 0.0191, "step": 7170 }, { "epoch": 7.597883597883598, "grad_norm": 0.1464722752571106, "learning_rate": 9.051877324093114e-05, "loss": 0.0184, "step": 7180 }, { "epoch": 7.608465608465608, "grad_norm": 0.22034798562526703, "learning_rate": 9.04864557630109e-05, "loss": 0.021, "step": 7190 }, { "epoch": 7.619047619047619, "grad_norm": 0.16102257370948792, "learning_rate": 9.045408909022928e-05, "loss": 0.0206, "step": 7200 }, { "epoch": 7.62962962962963, "grad_norm": 0.14642134308815002, "learning_rate": 9.042167326191484e-05, "loss": 0.0193, "step": 7210 }, { "epoch": 7.64021164021164, "grad_norm": 0.1869398057460785, "learning_rate": 9.038920831745587e-05, "loss": 0.0206, "step": 7220 }, { "epoch": 7.650793650793651, "grad_norm": 0.1317805051803589, "learning_rate": 9.035669429630036e-05, "loss": 0.0224, "step": 7230 }, { "epoch": 7.661375661375661, "grad_norm": 0.18073664605617523, "learning_rate": 9.032413123795588e-05, "loss": 0.0219, "step": 7240 }, { "epoch": 7.671957671957672, "grad_norm": 0.15350483357906342, "learning_rate": 9.029151918198962e-05, "loss": 0.0232, "step": 7250 }, { "epoch": 7.682539682539683, "grad_norm": 0.1687629669904709, "learning_rate": 9.025885816802833e-05, "loss": 0.0183, "step": 7260 }, { "epoch": 7.693121693121693, "grad_norm": 0.16351835429668427, "learning_rate": 9.02261482357582e-05, "loss": 0.0193, "step": 7270 }, { "epoch": 7.703703703703704, "grad_norm": 0.168185755610466, "learning_rate": 9.019338942492485e-05, "loss": 0.0188, "step": 7280 }, { "epoch": 7.714285714285714, "grad_norm": 0.16276027262210846, "learning_rate": 9.01605817753334e-05, "loss": 0.0208, "step": 7290 }, { "epoch": 7.724867724867725, "grad_norm": 0.17533668875694275, "learning_rate": 9.012772532684818e-05, "loss": 0.0183, "step": 7300 }, { "epoch": 7.735449735449736, "grad_norm": 0.10035920888185501, "learning_rate": 9.00948201193929e-05, "loss": 0.018, "step": 7310 }, { "epoch": 7.746031746031746, "grad_norm": 0.23366333544254303, "learning_rate": 9.006186619295047e-05, "loss": 0.0224, "step": 7320 }, { "epoch": 7.756613756613756, "grad_norm": 0.22357039153575897, "learning_rate": 9.002886358756305e-05, "loss": 0.0175, "step": 7330 }, { "epoch": 7.767195767195767, "grad_norm": 0.19233427941799164, "learning_rate": 8.999581234333188e-05, "loss": 0.0186, "step": 7340 }, { "epoch": 7.777777777777778, "grad_norm": 0.231095090508461, "learning_rate": 8.996271250041734e-05, "loss": 0.0207, "step": 7350 }, { "epoch": 7.788359788359788, "grad_norm": 0.19041655957698822, "learning_rate": 8.99295640990389e-05, "loss": 0.0241, "step": 7360 }, { "epoch": 7.798941798941799, "grad_norm": 0.2818748354911804, "learning_rate": 8.989636717947494e-05, "loss": 0.0214, "step": 7370 }, { "epoch": 7.809523809523809, "grad_norm": 0.21214599907398224, "learning_rate": 8.986312178206291e-05, "loss": 0.0231, "step": 7380 }, { "epoch": 7.8201058201058204, "grad_norm": 0.2486923336982727, "learning_rate": 8.982982794719904e-05, "loss": 0.0213, "step": 7390 }, { "epoch": 7.830687830687831, "grad_norm": 0.18815404176712036, "learning_rate": 8.979648571533852e-05, "loss": 0.0204, "step": 7400 }, { "epoch": 7.841269841269841, "grad_norm": 0.25723105669021606, "learning_rate": 8.976309512699528e-05, "loss": 0.0191, "step": 7410 }, { "epoch": 7.851851851851852, "grad_norm": 0.2217625379562378, "learning_rate": 8.972965622274205e-05, "loss": 0.0187, "step": 7420 }, { "epoch": 7.862433862433862, "grad_norm": 0.20707690715789795, "learning_rate": 8.969616904321026e-05, "loss": 0.0225, "step": 7430 }, { "epoch": 7.8730158730158735, "grad_norm": 0.1773764193058014, "learning_rate": 8.966263362908998e-05, "loss": 0.0181, "step": 7440 }, { "epoch": 7.883597883597884, "grad_norm": 0.13469870388507843, "learning_rate": 8.962905002112989e-05, "loss": 0.0217, "step": 7450 }, { "epoch": 7.894179894179894, "grad_norm": 0.11114483326673508, "learning_rate": 8.959541826013724e-05, "loss": 0.0208, "step": 7460 }, { "epoch": 7.904761904761905, "grad_norm": 0.2000305950641632, "learning_rate": 8.956173838697781e-05, "loss": 0.0197, "step": 7470 }, { "epoch": 7.915343915343915, "grad_norm": 0.2265944927930832, "learning_rate": 8.952801044257581e-05, "loss": 0.0232, "step": 7480 }, { "epoch": 7.925925925925926, "grad_norm": 0.17437851428985596, "learning_rate": 8.949423446791387e-05, "loss": 0.0194, "step": 7490 }, { "epoch": 7.936507936507937, "grad_norm": 0.1410323977470398, "learning_rate": 8.9460410504033e-05, "loss": 0.0209, "step": 7500 }, { "epoch": 7.947089947089947, "grad_norm": 0.21191424131393433, "learning_rate": 8.942653859203248e-05, "loss": 0.0167, "step": 7510 }, { "epoch": 7.957671957671957, "grad_norm": 0.1518864780664444, "learning_rate": 8.93926187730699e-05, "loss": 0.0209, "step": 7520 }, { "epoch": 7.968253968253968, "grad_norm": 0.19442668557167053, "learning_rate": 8.935865108836102e-05, "loss": 0.0191, "step": 7530 }, { "epoch": 7.978835978835979, "grad_norm": 0.1614450216293335, "learning_rate": 8.932463557917981e-05, "loss": 0.0207, "step": 7540 }, { "epoch": 7.98941798941799, "grad_norm": 0.11601876467466354, "learning_rate": 8.929057228685829e-05, "loss": 0.0213, "step": 7550 }, { "epoch": 8.0, "grad_norm": 0.15299370884895325, "learning_rate": 8.925646125278658e-05, "loss": 0.0201, "step": 7560 }, { "epoch": 8.010582010582011, "grad_norm": 0.19948652386665344, "learning_rate": 8.92223025184128e-05, "loss": 0.0197, "step": 7570 }, { "epoch": 8.02116402116402, "grad_norm": 0.20237910747528076, "learning_rate": 8.918809612524303e-05, "loss": 0.0214, "step": 7580 }, { "epoch": 8.031746031746032, "grad_norm": 0.206132173538208, "learning_rate": 8.91538421148413e-05, "loss": 0.0176, "step": 7590 }, { "epoch": 8.042328042328043, "grad_norm": 0.22186441719532013, "learning_rate": 8.91195405288294e-05, "loss": 0.0211, "step": 7600 }, { "epoch": 8.052910052910052, "grad_norm": 0.16562454402446747, "learning_rate": 8.908519140888705e-05, "loss": 0.0154, "step": 7610 }, { "epoch": 8.063492063492063, "grad_norm": 0.17328223586082458, "learning_rate": 8.905079479675164e-05, "loss": 0.0188, "step": 7620 }, { "epoch": 8.074074074074074, "grad_norm": 0.20946140587329865, "learning_rate": 8.901635073421831e-05, "loss": 0.0203, "step": 7630 }, { "epoch": 8.084656084656086, "grad_norm": 0.17676621675491333, "learning_rate": 8.898185926313983e-05, "loss": 0.0214, "step": 7640 }, { "epoch": 8.095238095238095, "grad_norm": 0.15856830775737762, "learning_rate": 8.894732042542659e-05, "loss": 0.0203, "step": 7650 }, { "epoch": 8.105820105820106, "grad_norm": 0.19148455560207367, "learning_rate": 8.891273426304654e-05, "loss": 0.0215, "step": 7660 }, { "epoch": 8.116402116402117, "grad_norm": 0.17020326852798462, "learning_rate": 8.887810081802515e-05, "loss": 0.0196, "step": 7670 }, { "epoch": 8.126984126984127, "grad_norm": 0.17844848334789276, "learning_rate": 8.884342013244529e-05, "loss": 0.0189, "step": 7680 }, { "epoch": 8.137566137566138, "grad_norm": 0.11302601546049118, "learning_rate": 8.880869224844725e-05, "loss": 0.0173, "step": 7690 }, { "epoch": 8.148148148148149, "grad_norm": 0.1568141132593155, "learning_rate": 8.877391720822875e-05, "loss": 0.0208, "step": 7700 }, { "epoch": 8.158730158730158, "grad_norm": 0.18944302201271057, "learning_rate": 8.873909505404468e-05, "loss": 0.0199, "step": 7710 }, { "epoch": 8.16931216931217, "grad_norm": 0.1576729714870453, "learning_rate": 8.870422582820726e-05, "loss": 0.0149, "step": 7720 }, { "epoch": 8.17989417989418, "grad_norm": 0.17133435606956482, "learning_rate": 8.866930957308588e-05, "loss": 0.0197, "step": 7730 }, { "epoch": 8.19047619047619, "grad_norm": 0.13454996049404144, "learning_rate": 8.86343463311071e-05, "loss": 0.0175, "step": 7740 }, { "epoch": 8.201058201058201, "grad_norm": 0.17119957506656647, "learning_rate": 8.859933614475452e-05, "loss": 0.019, "step": 7750 }, { "epoch": 8.211640211640212, "grad_norm": 0.23377977311611176, "learning_rate": 8.85642790565689e-05, "loss": 0.019, "step": 7760 }, { "epoch": 8.222222222222221, "grad_norm": 0.15414002537727356, "learning_rate": 8.852917510914782e-05, "loss": 0.0212, "step": 7770 }, { "epoch": 8.232804232804233, "grad_norm": 0.21779370307922363, "learning_rate": 8.84940243451459e-05, "loss": 0.0217, "step": 7780 }, { "epoch": 8.243386243386244, "grad_norm": 0.1511513590812683, "learning_rate": 8.845882680727469e-05, "loss": 0.0215, "step": 7790 }, { "epoch": 8.253968253968253, "grad_norm": 0.11904202401638031, "learning_rate": 8.842358253830246e-05, "loss": 0.0191, "step": 7800 }, { "epoch": 8.264550264550264, "grad_norm": 0.19639311730861664, "learning_rate": 8.838829158105434e-05, "loss": 0.0196, "step": 7810 }, { "epoch": 8.275132275132275, "grad_norm": 0.18062737584114075, "learning_rate": 8.835295397841217e-05, "loss": 0.0161, "step": 7820 }, { "epoch": 8.285714285714286, "grad_norm": 0.14609721302986145, "learning_rate": 8.831756977331447e-05, "loss": 0.0197, "step": 7830 }, { "epoch": 8.296296296296296, "grad_norm": 0.10294722765684128, "learning_rate": 8.828213900875638e-05, "loss": 0.0198, "step": 7840 }, { "epoch": 8.306878306878307, "grad_norm": 0.11115147173404694, "learning_rate": 8.824666172778964e-05, "loss": 0.0162, "step": 7850 }, { "epoch": 8.317460317460318, "grad_norm": 0.21246322989463806, "learning_rate": 8.821113797352245e-05, "loss": 0.0227, "step": 7860 }, { "epoch": 8.328042328042327, "grad_norm": 0.17382094264030457, "learning_rate": 8.817556778911956e-05, "loss": 0.0209, "step": 7870 }, { "epoch": 8.338624338624339, "grad_norm": 0.22226396203041077, "learning_rate": 8.81399512178021e-05, "loss": 0.0186, "step": 7880 }, { "epoch": 8.34920634920635, "grad_norm": 0.11061115562915802, "learning_rate": 8.810428830284752e-05, "loss": 0.0195, "step": 7890 }, { "epoch": 8.359788359788359, "grad_norm": 0.19772595167160034, "learning_rate": 8.806857908758967e-05, "loss": 0.0182, "step": 7900 }, { "epoch": 8.37037037037037, "grad_norm": 0.17727534472942352, "learning_rate": 8.80328236154186e-05, "loss": 0.0203, "step": 7910 }, { "epoch": 8.380952380952381, "grad_norm": 0.23033872246742249, "learning_rate": 8.799702192978056e-05, "loss": 0.0196, "step": 7920 }, { "epoch": 8.39153439153439, "grad_norm": 0.15787623822689056, "learning_rate": 8.7961174074178e-05, "loss": 0.0212, "step": 7930 }, { "epoch": 8.402116402116402, "grad_norm": 0.13897103071212769, "learning_rate": 8.792528009216942e-05, "loss": 0.0205, "step": 7940 }, { "epoch": 8.412698412698413, "grad_norm": 0.3023090958595276, "learning_rate": 8.788934002736943e-05, "loss": 0.019, "step": 7950 }, { "epoch": 8.423280423280424, "grad_norm": 0.30803853273391724, "learning_rate": 8.785335392344857e-05, "loss": 0.0207, "step": 7960 }, { "epoch": 8.433862433862434, "grad_norm": 0.19411146640777588, "learning_rate": 8.781732182413336e-05, "loss": 0.0187, "step": 7970 }, { "epoch": 8.444444444444445, "grad_norm": 0.28386545181274414, "learning_rate": 8.778124377320618e-05, "loss": 0.0175, "step": 7980 }, { "epoch": 8.455026455026456, "grad_norm": 0.1937234252691269, "learning_rate": 8.774511981450529e-05, "loss": 0.0182, "step": 7990 }, { "epoch": 8.465608465608465, "grad_norm": 0.1713896244764328, "learning_rate": 8.770894999192467e-05, "loss": 0.0209, "step": 8000 }, { "epoch": 8.476190476190476, "grad_norm": 0.20320789515972137, "learning_rate": 8.767273434941412e-05, "loss": 0.0179, "step": 8010 }, { "epoch": 8.486772486772487, "grad_norm": 0.22029393911361694, "learning_rate": 8.7636472930979e-05, "loss": 0.0181, "step": 8020 }, { "epoch": 8.497354497354497, "grad_norm": 0.19536320865154266, "learning_rate": 8.76001657806804e-05, "loss": 0.0178, "step": 8030 }, { "epoch": 8.507936507936508, "grad_norm": 0.18808218836784363, "learning_rate": 8.756381294263491e-05, "loss": 0.0192, "step": 8040 }, { "epoch": 8.518518518518519, "grad_norm": 0.16115792095661163, "learning_rate": 8.752741446101464e-05, "loss": 0.0171, "step": 8050 }, { "epoch": 8.529100529100528, "grad_norm": 0.1421302855014801, "learning_rate": 8.749097038004722e-05, "loss": 0.0182, "step": 8060 }, { "epoch": 8.53968253968254, "grad_norm": 0.20594890415668488, "learning_rate": 8.745448074401561e-05, "loss": 0.0185, "step": 8070 }, { "epoch": 8.55026455026455, "grad_norm": 0.12282946705818176, "learning_rate": 8.741794559725818e-05, "loss": 0.023, "step": 8080 }, { "epoch": 8.56084656084656, "grad_norm": 0.1461414247751236, "learning_rate": 8.738136498416856e-05, "loss": 0.0202, "step": 8090 }, { "epoch": 8.571428571428571, "grad_norm": 0.139751136302948, "learning_rate": 8.734473894919564e-05, "loss": 0.0209, "step": 8100 }, { "epoch": 8.582010582010582, "grad_norm": 0.15774254500865936, "learning_rate": 8.730806753684353e-05, "loss": 0.0189, "step": 8110 }, { "epoch": 8.592592592592592, "grad_norm": 0.15614144504070282, "learning_rate": 8.727135079167144e-05, "loss": 0.0186, "step": 8120 }, { "epoch": 8.603174603174603, "grad_norm": 0.15441524982452393, "learning_rate": 8.723458875829368e-05, "loss": 0.0181, "step": 8130 }, { "epoch": 8.613756613756614, "grad_norm": 0.14589627087116241, "learning_rate": 8.719778148137958e-05, "loss": 0.0189, "step": 8140 }, { "epoch": 8.624338624338625, "grad_norm": 0.20918488502502441, "learning_rate": 8.716092900565346e-05, "loss": 0.0211, "step": 8150 }, { "epoch": 8.634920634920634, "grad_norm": 0.1761360615491867, "learning_rate": 8.712403137589456e-05, "loss": 0.018, "step": 8160 }, { "epoch": 8.645502645502646, "grad_norm": 0.22873260080814362, "learning_rate": 8.708708863693697e-05, "loss": 0.0195, "step": 8170 }, { "epoch": 8.656084656084657, "grad_norm": 0.20595593750476837, "learning_rate": 8.70501008336696e-05, "loss": 0.0224, "step": 8180 }, { "epoch": 8.666666666666666, "grad_norm": 0.12372051179409027, "learning_rate": 8.701306801103611e-05, "loss": 0.0176, "step": 8190 }, { "epoch": 8.677248677248677, "grad_norm": 0.16812418401241302, "learning_rate": 8.697599021403489e-05, "loss": 0.019, "step": 8200 }, { "epoch": 8.687830687830688, "grad_norm": 0.11865603923797607, "learning_rate": 8.693886748771896e-05, "loss": 0.0162, "step": 8210 }, { "epoch": 8.698412698412698, "grad_norm": 0.09152933210134506, "learning_rate": 8.690169987719593e-05, "loss": 0.0182, "step": 8220 }, { "epoch": 8.708994708994709, "grad_norm": 0.10140620917081833, "learning_rate": 8.686448742762791e-05, "loss": 0.0166, "step": 8230 }, { "epoch": 8.71957671957672, "grad_norm": 0.14081087708473206, "learning_rate": 8.682723018423161e-05, "loss": 0.0171, "step": 8240 }, { "epoch": 8.73015873015873, "grad_norm": 0.13910022377967834, "learning_rate": 8.678992819227804e-05, "loss": 0.0163, "step": 8250 }, { "epoch": 8.74074074074074, "grad_norm": 0.1576622575521469, "learning_rate": 8.675258149709265e-05, "loss": 0.0168, "step": 8260 }, { "epoch": 8.751322751322752, "grad_norm": 0.14109571278095245, "learning_rate": 8.67151901440552e-05, "loss": 0.0181, "step": 8270 }, { "epoch": 8.761904761904763, "grad_norm": 0.16854515671730042, "learning_rate": 8.667775417859971e-05, "loss": 0.0172, "step": 8280 }, { "epoch": 8.772486772486772, "grad_norm": 0.11551665514707565, "learning_rate": 8.664027364621441e-05, "loss": 0.0198, "step": 8290 }, { "epoch": 8.783068783068783, "grad_norm": 0.14371244609355927, "learning_rate": 8.660274859244168e-05, "loss": 0.0169, "step": 8300 }, { "epoch": 8.793650793650794, "grad_norm": 0.13205094635486603, "learning_rate": 8.656517906287799e-05, "loss": 0.0201, "step": 8310 }, { "epoch": 8.804232804232804, "grad_norm": 0.21439173817634583, "learning_rate": 8.652756510317387e-05, "loss": 0.0186, "step": 8320 }, { "epoch": 8.814814814814815, "grad_norm": 0.1649792492389679, "learning_rate": 8.648990675903382e-05, "loss": 0.0209, "step": 8330 }, { "epoch": 8.825396825396826, "grad_norm": 0.14583049714565277, "learning_rate": 8.64522040762163e-05, "loss": 0.0186, "step": 8340 }, { "epoch": 8.835978835978835, "grad_norm": 0.20688718557357788, "learning_rate": 8.64144571005336e-05, "loss": 0.0189, "step": 8350 }, { "epoch": 8.846560846560847, "grad_norm": 0.20854610204696655, "learning_rate": 8.637666587785184e-05, "loss": 0.0192, "step": 8360 }, { "epoch": 8.857142857142858, "grad_norm": 0.1529761105775833, "learning_rate": 8.633883045409096e-05, "loss": 0.0211, "step": 8370 }, { "epoch": 8.867724867724867, "grad_norm": 0.18162932991981506, "learning_rate": 8.630095087522457e-05, "loss": 0.0186, "step": 8380 }, { "epoch": 8.878306878306878, "grad_norm": 0.18541498482227325, "learning_rate": 8.62630271872799e-05, "loss": 0.0228, "step": 8390 }, { "epoch": 8.88888888888889, "grad_norm": 0.128691628575325, "learning_rate": 8.62250594363378e-05, "loss": 0.019, "step": 8400 }, { "epoch": 8.899470899470899, "grad_norm": 0.13798856735229492, "learning_rate": 8.61870476685327e-05, "loss": 0.0189, "step": 8410 }, { "epoch": 8.91005291005291, "grad_norm": 0.16443189978599548, "learning_rate": 8.614899193005248e-05, "loss": 0.0175, "step": 8420 }, { "epoch": 8.920634920634921, "grad_norm": 0.16403548419475555, "learning_rate": 8.611089226713842e-05, "loss": 0.0172, "step": 8430 }, { "epoch": 8.93121693121693, "grad_norm": 0.17452533543109894, "learning_rate": 8.607274872608521e-05, "loss": 0.0152, "step": 8440 }, { "epoch": 8.941798941798941, "grad_norm": 0.22807064652442932, "learning_rate": 8.603456135324088e-05, "loss": 0.0231, "step": 8450 }, { "epoch": 8.952380952380953, "grad_norm": 0.18264548480510712, "learning_rate": 8.599633019500665e-05, "loss": 0.0201, "step": 8460 }, { "epoch": 8.962962962962964, "grad_norm": 0.17378154397010803, "learning_rate": 8.595805529783702e-05, "loss": 0.0189, "step": 8470 }, { "epoch": 8.973544973544973, "grad_norm": 0.15510252118110657, "learning_rate": 8.591973670823958e-05, "loss": 0.018, "step": 8480 }, { "epoch": 8.984126984126984, "grad_norm": 0.17981910705566406, "learning_rate": 8.588137447277503e-05, "loss": 0.0239, "step": 8490 }, { "epoch": 8.994708994708995, "grad_norm": 0.20087961852550507, "learning_rate": 8.584296863805709e-05, "loss": 0.0207, "step": 8500 }, { "epoch": 9.005291005291005, "grad_norm": 0.1004253625869751, "learning_rate": 8.580451925075248e-05, "loss": 0.0195, "step": 8510 }, { "epoch": 9.015873015873016, "grad_norm": 0.17029958963394165, "learning_rate": 8.576602635758085e-05, "loss": 0.019, "step": 8520 }, { "epoch": 9.026455026455027, "grad_norm": 0.14118917286396027, "learning_rate": 8.572749000531469e-05, "loss": 0.0174, "step": 8530 }, { "epoch": 9.037037037037036, "grad_norm": 0.187066450715065, "learning_rate": 8.568891024077926e-05, "loss": 0.0214, "step": 8540 }, { "epoch": 9.047619047619047, "grad_norm": 0.1376589685678482, "learning_rate": 8.565028711085265e-05, "loss": 0.0151, "step": 8550 }, { "epoch": 9.058201058201059, "grad_norm": 0.18463920056819916, "learning_rate": 8.561162066246561e-05, "loss": 0.0157, "step": 8560 }, { "epoch": 9.068783068783068, "grad_norm": 0.22746042907238007, "learning_rate": 8.557291094260152e-05, "loss": 0.0231, "step": 8570 }, { "epoch": 9.079365079365079, "grad_norm": 0.2107260376214981, "learning_rate": 8.553415799829629e-05, "loss": 0.0192, "step": 8580 }, { "epoch": 9.08994708994709, "grad_norm": 0.12579496204853058, "learning_rate": 8.549536187663847e-05, "loss": 0.0195, "step": 8590 }, { "epoch": 9.100529100529101, "grad_norm": 0.22728829085826874, "learning_rate": 8.545652262476897e-05, "loss": 0.0154, "step": 8600 }, { "epoch": 9.11111111111111, "grad_norm": 0.14726810157299042, "learning_rate": 8.541764028988115e-05, "loss": 0.017, "step": 8610 }, { "epoch": 9.121693121693122, "grad_norm": 0.2168988138437271, "learning_rate": 8.537871491922071e-05, "loss": 0.0158, "step": 8620 }, { "epoch": 9.132275132275133, "grad_norm": 0.1797293871641159, "learning_rate": 8.533974656008566e-05, "loss": 0.0165, "step": 8630 }, { "epoch": 9.142857142857142, "grad_norm": 0.17334169149398804, "learning_rate": 8.530073525982621e-05, "loss": 0.0167, "step": 8640 }, { "epoch": 9.153439153439153, "grad_norm": 0.1863032579421997, "learning_rate": 8.526168106584476e-05, "loss": 0.0226, "step": 8650 }, { "epoch": 9.164021164021165, "grad_norm": 0.15072186291217804, "learning_rate": 8.522258402559586e-05, "loss": 0.0208, "step": 8660 }, { "epoch": 9.174603174603174, "grad_norm": 0.2747143805027008, "learning_rate": 8.518344418658613e-05, "loss": 0.0199, "step": 8670 }, { "epoch": 9.185185185185185, "grad_norm": 0.254176527261734, "learning_rate": 8.51442615963741e-05, "loss": 0.0172, "step": 8680 }, { "epoch": 9.195767195767196, "grad_norm": 0.17608124017715454, "learning_rate": 8.510503630257034e-05, "loss": 0.0174, "step": 8690 }, { "epoch": 9.206349206349206, "grad_norm": 0.14772461354732513, "learning_rate": 8.50657683528373e-05, "loss": 0.0178, "step": 8700 }, { "epoch": 9.216931216931217, "grad_norm": 0.16262388229370117, "learning_rate": 8.502645779488922e-05, "loss": 0.0174, "step": 8710 }, { "epoch": 9.227513227513228, "grad_norm": 0.17153163254261017, "learning_rate": 8.498710467649213e-05, "loss": 0.0216, "step": 8720 }, { "epoch": 9.238095238095237, "grad_norm": 0.15243101119995117, "learning_rate": 8.494770904546381e-05, "loss": 0.0185, "step": 8730 }, { "epoch": 9.248677248677248, "grad_norm": 0.12915921211242676, "learning_rate": 8.490827094967363e-05, "loss": 0.0178, "step": 8740 }, { "epoch": 9.25925925925926, "grad_norm": 0.2095094621181488, "learning_rate": 8.486879043704263e-05, "loss": 0.0167, "step": 8750 }, { "epoch": 9.26984126984127, "grad_norm": 0.08836907148361206, "learning_rate": 8.482926755554332e-05, "loss": 0.0172, "step": 8760 }, { "epoch": 9.28042328042328, "grad_norm": 0.1824076920747757, "learning_rate": 8.478970235319975e-05, "loss": 0.0212, "step": 8770 }, { "epoch": 9.291005291005291, "grad_norm": 0.1631210595369339, "learning_rate": 8.475009487808737e-05, "loss": 0.0198, "step": 8780 }, { "epoch": 9.301587301587302, "grad_norm": 0.16248774528503418, "learning_rate": 8.471044517833298e-05, "loss": 0.0189, "step": 8790 }, { "epoch": 9.312169312169312, "grad_norm": 0.14825062453746796, "learning_rate": 8.467075330211474e-05, "loss": 0.018, "step": 8800 }, { "epoch": 9.322751322751323, "grad_norm": 0.14928293228149414, "learning_rate": 8.463101929766198e-05, "loss": 0.0197, "step": 8810 }, { "epoch": 9.333333333333334, "grad_norm": 0.16277897357940674, "learning_rate": 8.459124321325528e-05, "loss": 0.0193, "step": 8820 }, { "epoch": 9.343915343915343, "grad_norm": 0.12936939299106598, "learning_rate": 8.455142509722634e-05, "loss": 0.0167, "step": 8830 }, { "epoch": 9.354497354497354, "grad_norm": 0.1465858519077301, "learning_rate": 8.451156499795792e-05, "loss": 0.0175, "step": 8840 }, { "epoch": 9.365079365079366, "grad_norm": 0.19495722651481628, "learning_rate": 8.447166296388381e-05, "loss": 0.0187, "step": 8850 }, { "epoch": 9.375661375661375, "grad_norm": 0.1735941618680954, "learning_rate": 8.443171904348872e-05, "loss": 0.0163, "step": 8860 }, { "epoch": 9.386243386243386, "grad_norm": 0.15923917293548584, "learning_rate": 8.43917332853083e-05, "loss": 0.0179, "step": 8870 }, { "epoch": 9.396825396825397, "grad_norm": 0.13281619548797607, "learning_rate": 8.435170573792901e-05, "loss": 0.018, "step": 8880 }, { "epoch": 9.407407407407407, "grad_norm": 0.1861899197101593, "learning_rate": 8.431163644998808e-05, "loss": 0.0202, "step": 8890 }, { "epoch": 9.417989417989418, "grad_norm": 0.14845024049282074, "learning_rate": 8.42715254701735e-05, "loss": 0.0184, "step": 8900 }, { "epoch": 9.428571428571429, "grad_norm": 0.14154121279716492, "learning_rate": 8.42313728472239e-05, "loss": 0.0164, "step": 8910 }, { "epoch": 9.43915343915344, "grad_norm": 0.21073301136493683, "learning_rate": 8.419117862992844e-05, "loss": 0.0183, "step": 8920 }, { "epoch": 9.44973544973545, "grad_norm": 0.1156744509935379, "learning_rate": 8.415094286712694e-05, "loss": 0.0217, "step": 8930 }, { "epoch": 9.46031746031746, "grad_norm": 0.1587861031293869, "learning_rate": 8.411066560770965e-05, "loss": 0.0173, "step": 8940 }, { "epoch": 9.470899470899472, "grad_norm": 0.12085498869419098, "learning_rate": 8.407034690061721e-05, "loss": 0.0184, "step": 8950 }, { "epoch": 9.481481481481481, "grad_norm": 0.12061701714992523, "learning_rate": 8.402998679484067e-05, "loss": 0.0196, "step": 8960 }, { "epoch": 9.492063492063492, "grad_norm": 0.11670458316802979, "learning_rate": 8.398958533942135e-05, "loss": 0.0179, "step": 8970 }, { "epoch": 9.502645502645503, "grad_norm": 0.1845267117023468, "learning_rate": 8.394914258345083e-05, "loss": 0.0208, "step": 8980 }, { "epoch": 9.513227513227513, "grad_norm": 0.1711067408323288, "learning_rate": 8.390865857607089e-05, "loss": 0.0245, "step": 8990 }, { "epoch": 9.523809523809524, "grad_norm": 0.16920337080955505, "learning_rate": 8.386813336647339e-05, "loss": 0.0214, "step": 9000 }, { "epoch": 9.534391534391535, "grad_norm": 0.2599749267101288, "learning_rate": 8.38275670039003e-05, "loss": 0.0217, "step": 9010 }, { "epoch": 9.544973544973544, "grad_norm": 0.1122012734413147, "learning_rate": 8.378695953764356e-05, "loss": 0.0208, "step": 9020 }, { "epoch": 9.555555555555555, "grad_norm": 0.25079116225242615, "learning_rate": 8.374631101704509e-05, "loss": 0.0202, "step": 9030 }, { "epoch": 9.566137566137566, "grad_norm": 0.21227101981639862, "learning_rate": 8.370562149149665e-05, "loss": 0.0195, "step": 9040 }, { "epoch": 9.576719576719576, "grad_norm": 0.16131591796875, "learning_rate": 8.366489101043989e-05, "loss": 0.0205, "step": 9050 }, { "epoch": 9.587301587301587, "grad_norm": 0.19881194829940796, "learning_rate": 8.362411962336614e-05, "loss": 0.0169, "step": 9060 }, { "epoch": 9.597883597883598, "grad_norm": 0.18571533262729645, "learning_rate": 8.358330737981651e-05, "loss": 0.0184, "step": 9070 }, { "epoch": 9.60846560846561, "grad_norm": 0.26290106773376465, "learning_rate": 8.35424543293817e-05, "loss": 0.0195, "step": 9080 }, { "epoch": 9.619047619047619, "grad_norm": 0.26029184460639954, "learning_rate": 8.350156052170206e-05, "loss": 0.0185, "step": 9090 }, { "epoch": 9.62962962962963, "grad_norm": 0.15810027718544006, "learning_rate": 8.346062600646738e-05, "loss": 0.0147, "step": 9100 }, { "epoch": 9.640211640211641, "grad_norm": 0.2012079954147339, "learning_rate": 8.341965083341697e-05, "loss": 0.0157, "step": 9110 }, { "epoch": 9.65079365079365, "grad_norm": 0.31453773379325867, "learning_rate": 8.337863505233953e-05, "loss": 0.0199, "step": 9120 }, { "epoch": 9.661375661375661, "grad_norm": 0.26651105284690857, "learning_rate": 8.33375787130731e-05, "loss": 0.0188, "step": 9130 }, { "epoch": 9.671957671957673, "grad_norm": 0.24122586846351624, "learning_rate": 8.3296481865505e-05, "loss": 0.0172, "step": 9140 }, { "epoch": 9.682539682539682, "grad_norm": 0.15171270072460175, "learning_rate": 8.325534455957181e-05, "loss": 0.0195, "step": 9150 }, { "epoch": 9.693121693121693, "grad_norm": 0.19530822336673737, "learning_rate": 8.321416684525917e-05, "loss": 0.0184, "step": 9160 }, { "epoch": 9.703703703703704, "grad_norm": 0.12821416556835175, "learning_rate": 8.317294877260193e-05, "loss": 0.0188, "step": 9170 }, { "epoch": 9.714285714285714, "grad_norm": 0.14520268142223358, "learning_rate": 8.313169039168394e-05, "loss": 0.0205, "step": 9180 }, { "epoch": 9.724867724867725, "grad_norm": 0.17838351428508759, "learning_rate": 8.3090391752638e-05, "loss": 0.0179, "step": 9190 }, { "epoch": 9.735449735449736, "grad_norm": 0.2162761241197586, "learning_rate": 8.304905290564586e-05, "loss": 0.0207, "step": 9200 }, { "epoch": 9.746031746031747, "grad_norm": 0.2094467282295227, "learning_rate": 8.300767390093813e-05, "loss": 0.0195, "step": 9210 }, { "epoch": 9.756613756613756, "grad_norm": 0.2108340561389923, "learning_rate": 8.296625478879417e-05, "loss": 0.0174, "step": 9220 }, { "epoch": 9.767195767195767, "grad_norm": 0.2150123119354248, "learning_rate": 8.292479561954214e-05, "loss": 0.0185, "step": 9230 }, { "epoch": 9.777777777777779, "grad_norm": 0.14891047775745392, "learning_rate": 8.288329644355883e-05, "loss": 0.0177, "step": 9240 }, { "epoch": 9.788359788359788, "grad_norm": 0.18539343774318695, "learning_rate": 8.284175731126965e-05, "loss": 0.0196, "step": 9250 }, { "epoch": 9.798941798941799, "grad_norm": 0.1496395766735077, "learning_rate": 8.280017827314854e-05, "loss": 0.0185, "step": 9260 }, { "epoch": 9.80952380952381, "grad_norm": 0.15835033357143402, "learning_rate": 8.2758559379718e-05, "loss": 0.0196, "step": 9270 }, { "epoch": 9.82010582010582, "grad_norm": 0.14678263664245605, "learning_rate": 8.271690068154887e-05, "loss": 0.019, "step": 9280 }, { "epoch": 9.83068783068783, "grad_norm": 0.16715601086616516, "learning_rate": 8.26752022292604e-05, "loss": 0.0151, "step": 9290 }, { "epoch": 9.841269841269842, "grad_norm": 0.16175886988639832, "learning_rate": 8.263346407352018e-05, "loss": 0.0201, "step": 9300 }, { "epoch": 9.851851851851851, "grad_norm": 0.14697131514549255, "learning_rate": 8.259168626504395e-05, "loss": 0.0185, "step": 9310 }, { "epoch": 9.862433862433862, "grad_norm": 0.1414712816476822, "learning_rate": 8.254986885459572e-05, "loss": 0.017, "step": 9320 }, { "epoch": 9.873015873015873, "grad_norm": 0.24626120924949646, "learning_rate": 8.250801189298757e-05, "loss": 0.0174, "step": 9330 }, { "epoch": 9.883597883597883, "grad_norm": 0.1411730945110321, "learning_rate": 8.246611543107967e-05, "loss": 0.0221, "step": 9340 }, { "epoch": 9.894179894179894, "grad_norm": 0.1356646567583084, "learning_rate": 8.242417951978018e-05, "loss": 0.0197, "step": 9350 }, { "epoch": 9.904761904761905, "grad_norm": 0.13897746801376343, "learning_rate": 8.238220421004517e-05, "loss": 0.0202, "step": 9360 }, { "epoch": 9.915343915343914, "grad_norm": 0.17540132999420166, "learning_rate": 8.234018955287859e-05, "loss": 0.0165, "step": 9370 }, { "epoch": 9.925925925925926, "grad_norm": 0.10786277055740356, "learning_rate": 8.229813559933224e-05, "loss": 0.0151, "step": 9380 }, { "epoch": 9.936507936507937, "grad_norm": 0.1266981065273285, "learning_rate": 8.22560424005056e-05, "loss": 0.0173, "step": 9390 }, { "epoch": 9.947089947089948, "grad_norm": 0.17386619746685028, "learning_rate": 8.22139100075459e-05, "loss": 0.0211, "step": 9400 }, { "epoch": 9.957671957671957, "grad_norm": 0.10173843055963516, "learning_rate": 8.217173847164799e-05, "loss": 0.0183, "step": 9410 }, { "epoch": 9.968253968253968, "grad_norm": 0.22005802392959595, "learning_rate": 8.212952784405422e-05, "loss": 0.0179, "step": 9420 }, { "epoch": 9.97883597883598, "grad_norm": 0.23021216690540314, "learning_rate": 8.208727817605451e-05, "loss": 0.0194, "step": 9430 }, { "epoch": 9.989417989417989, "grad_norm": 0.22162282466888428, "learning_rate": 8.204498951898618e-05, "loss": 0.017, "step": 9440 }, { "epoch": 10.0, "grad_norm": 0.16684693098068237, "learning_rate": 8.200266192423395e-05, "loss": 0.0171, "step": 9450 }, { "epoch": 10.010582010582011, "grad_norm": 0.14423096179962158, "learning_rate": 8.196029544322981e-05, "loss": 0.018, "step": 9460 }, { "epoch": 10.02116402116402, "grad_norm": 0.11663028597831726, "learning_rate": 8.191789012745305e-05, "loss": 0.0176, "step": 9470 }, { "epoch": 10.031746031746032, "grad_norm": 0.16955547034740448, "learning_rate": 8.187544602843014e-05, "loss": 0.0166, "step": 9480 }, { "epoch": 10.042328042328043, "grad_norm": 0.13802854716777802, "learning_rate": 8.183296319773466e-05, "loss": 0.0141, "step": 9490 }, { "epoch": 10.052910052910052, "grad_norm": 0.17989321053028107, "learning_rate": 8.179044168698721e-05, "loss": 0.0139, "step": 9500 }, { "epoch": 10.063492063492063, "grad_norm": 0.111312136054039, "learning_rate": 8.174788154785548e-05, "loss": 0.0149, "step": 9510 }, { "epoch": 10.074074074074074, "grad_norm": 0.17340274155139923, "learning_rate": 8.170528283205403e-05, "loss": 0.0172, "step": 9520 }, { "epoch": 10.084656084656086, "grad_norm": 0.18068848550319672, "learning_rate": 8.166264559134435e-05, "loss": 0.0158, "step": 9530 }, { "epoch": 10.095238095238095, "grad_norm": 0.1529015153646469, "learning_rate": 8.161996987753467e-05, "loss": 0.0198, "step": 9540 }, { "epoch": 10.105820105820106, "grad_norm": 0.14757715165615082, "learning_rate": 8.157725574248e-05, "loss": 0.0155, "step": 9550 }, { "epoch": 10.116402116402117, "grad_norm": 0.17389406263828278, "learning_rate": 8.153450323808209e-05, "loss": 0.0163, "step": 9560 }, { "epoch": 10.126984126984127, "grad_norm": 0.13249878585338593, "learning_rate": 8.149171241628923e-05, "loss": 0.0162, "step": 9570 }, { "epoch": 10.137566137566138, "grad_norm": 0.10693816095590591, "learning_rate": 8.144888332909631e-05, "loss": 0.0157, "step": 9580 }, { "epoch": 10.148148148148149, "grad_norm": 0.13686880469322205, "learning_rate": 8.14060160285447e-05, "loss": 0.0177, "step": 9590 }, { "epoch": 10.158730158730158, "grad_norm": 0.14110872149467468, "learning_rate": 8.136311056672223e-05, "loss": 0.0176, "step": 9600 }, { "epoch": 10.16931216931217, "grad_norm": 0.11020515114068985, "learning_rate": 8.132016699576308e-05, "loss": 0.0145, "step": 9610 }, { "epoch": 10.17989417989418, "grad_norm": 0.18996462225914001, "learning_rate": 8.127718536784771e-05, "loss": 0.0142, "step": 9620 }, { "epoch": 10.19047619047619, "grad_norm": 0.15018007159233093, "learning_rate": 8.12341657352029e-05, "loss": 0.0167, "step": 9630 }, { "epoch": 10.201058201058201, "grad_norm": 0.1434657871723175, "learning_rate": 8.119110815010151e-05, "loss": 0.0134, "step": 9640 }, { "epoch": 10.211640211640212, "grad_norm": 0.14664538204669952, "learning_rate": 8.114801266486261e-05, "loss": 0.0167, "step": 9650 }, { "epoch": 10.222222222222221, "grad_norm": 0.11212164908647537, "learning_rate": 8.110487933185122e-05, "loss": 0.0155, "step": 9660 }, { "epoch": 10.232804232804233, "grad_norm": 0.1455252766609192, "learning_rate": 8.106170820347847e-05, "loss": 0.0144, "step": 9670 }, { "epoch": 10.243386243386244, "grad_norm": 0.13999071717262268, "learning_rate": 8.101849933220132e-05, "loss": 0.0179, "step": 9680 }, { "epoch": 10.253968253968253, "grad_norm": 0.14936228096485138, "learning_rate": 8.097525277052264e-05, "loss": 0.0162, "step": 9690 }, { "epoch": 10.264550264550264, "grad_norm": 0.13775281608104706, "learning_rate": 8.093196857099106e-05, "loss": 0.0158, "step": 9700 }, { "epoch": 10.275132275132275, "grad_norm": 0.1742825210094452, "learning_rate": 8.088864678620096e-05, "loss": 0.0174, "step": 9710 }, { "epoch": 10.285714285714286, "grad_norm": 0.15209874510765076, "learning_rate": 8.084528746879243e-05, "loss": 0.0176, "step": 9720 }, { "epoch": 10.296296296296296, "grad_norm": 0.20809268951416016, "learning_rate": 8.080189067145108e-05, "loss": 0.0169, "step": 9730 }, { "epoch": 10.306878306878307, "grad_norm": 0.1349266916513443, "learning_rate": 8.075845644690813e-05, "loss": 0.0193, "step": 9740 }, { "epoch": 10.317460317460318, "grad_norm": 0.16458472609519958, "learning_rate": 8.07149848479403e-05, "loss": 0.0159, "step": 9750 }, { "epoch": 10.328042328042327, "grad_norm": 0.16413038969039917, "learning_rate": 8.067147592736962e-05, "loss": 0.0161, "step": 9760 }, { "epoch": 10.338624338624339, "grad_norm": 0.21694044768810272, "learning_rate": 8.062792973806357e-05, "loss": 0.0162, "step": 9770 }, { "epoch": 10.34920634920635, "grad_norm": 0.1779213398694992, "learning_rate": 8.058434633293485e-05, "loss": 0.0201, "step": 9780 }, { "epoch": 10.359788359788359, "grad_norm": 0.17265577614307404, "learning_rate": 8.054072576494143e-05, "loss": 0.022, "step": 9790 }, { "epoch": 10.37037037037037, "grad_norm": 0.21412432193756104, "learning_rate": 8.04970680870864e-05, "loss": 0.0162, "step": 9800 }, { "epoch": 10.380952380952381, "grad_norm": 0.1997523307800293, "learning_rate": 8.045337335241793e-05, "loss": 0.0176, "step": 9810 }, { "epoch": 10.39153439153439, "grad_norm": 0.13308224081993103, "learning_rate": 8.040964161402932e-05, "loss": 0.02, "step": 9820 }, { "epoch": 10.402116402116402, "grad_norm": 0.17819887399673462, "learning_rate": 8.036587292505869e-05, "loss": 0.0164, "step": 9830 }, { "epoch": 10.412698412698413, "grad_norm": 0.2626210153102875, "learning_rate": 8.032206733868912e-05, "loss": 0.015, "step": 9840 }, { "epoch": 10.423280423280424, "grad_norm": 0.155051127076149, "learning_rate": 8.027822490814858e-05, "loss": 0.0166, "step": 9850 }, { "epoch": 10.433862433862434, "grad_norm": 0.16080161929130554, "learning_rate": 8.023434568670972e-05, "loss": 0.0194, "step": 9860 }, { "epoch": 10.444444444444445, "grad_norm": 0.15135325491428375, "learning_rate": 8.019042972768992e-05, "loss": 0.0172, "step": 9870 }, { "epoch": 10.455026455026456, "grad_norm": 0.18332719802856445, "learning_rate": 8.014647708445124e-05, "loss": 0.0189, "step": 9880 }, { "epoch": 10.465608465608465, "grad_norm": 0.16505780816078186, "learning_rate": 8.010248781040027e-05, "loss": 0.017, "step": 9890 }, { "epoch": 10.476190476190476, "grad_norm": 0.12192723900079727, "learning_rate": 8.005846195898814e-05, "loss": 0.0201, "step": 9900 }, { "epoch": 10.486772486772487, "grad_norm": 0.17065304517745972, "learning_rate": 8.00143995837104e-05, "loss": 0.0153, "step": 9910 }, { "epoch": 10.497354497354497, "grad_norm": 0.09667833894491196, "learning_rate": 7.997030073810698e-05, "loss": 0.0199, "step": 9920 }, { "epoch": 10.507936507936508, "grad_norm": 0.18551033735275269, "learning_rate": 7.992616547576217e-05, "loss": 0.0203, "step": 9930 }, { "epoch": 10.518518518518519, "grad_norm": 0.15998470783233643, "learning_rate": 7.988199385030445e-05, "loss": 0.0167, "step": 9940 }, { "epoch": 10.529100529100528, "grad_norm": 0.19083727896213531, "learning_rate": 7.98377859154065e-05, "loss": 0.0172, "step": 9950 }, { "epoch": 10.53968253968254, "grad_norm": 0.17820106446743011, "learning_rate": 7.979354172478516e-05, "loss": 0.016, "step": 9960 }, { "epoch": 10.55026455026455, "grad_norm": 0.1565254032611847, "learning_rate": 7.974926133220127e-05, "loss": 0.0172, "step": 9970 }, { "epoch": 10.56084656084656, "grad_norm": 0.15900857746601105, "learning_rate": 7.970494479145967e-05, "loss": 0.0161, "step": 9980 }, { "epoch": 10.571428571428571, "grad_norm": 0.18809965252876282, "learning_rate": 7.966059215640919e-05, "loss": 0.0182, "step": 9990 }, { "epoch": 10.582010582010582, "grad_norm": 0.15251828730106354, "learning_rate": 7.961620348094241e-05, "loss": 0.0167, "step": 10000 }, { "epoch": 10.592592592592592, "grad_norm": 0.11248621344566345, "learning_rate": 7.957177881899577e-05, "loss": 0.0168, "step": 10010 }, { "epoch": 10.603174603174603, "grad_norm": 0.19336508214473724, "learning_rate": 7.952731822454944e-05, "loss": 0.0172, "step": 10020 }, { "epoch": 10.613756613756614, "grad_norm": 0.1566939353942871, "learning_rate": 7.948282175162722e-05, "loss": 0.0199, "step": 10030 }, { "epoch": 10.624338624338625, "grad_norm": 0.14574860036373138, "learning_rate": 7.943828945429652e-05, "loss": 0.0159, "step": 10040 }, { "epoch": 10.634920634920634, "grad_norm": 0.126283198595047, "learning_rate": 7.939372138666827e-05, "loss": 0.0184, "step": 10050 }, { "epoch": 10.645502645502646, "grad_norm": 0.15500576794147491, "learning_rate": 7.934911760289692e-05, "loss": 0.0163, "step": 10060 }, { "epoch": 10.656084656084657, "grad_norm": 0.17155727744102478, "learning_rate": 7.930447815718022e-05, "loss": 0.0175, "step": 10070 }, { "epoch": 10.666666666666666, "grad_norm": 0.18272952735424042, "learning_rate": 7.925980310375934e-05, "loss": 0.0182, "step": 10080 }, { "epoch": 10.677248677248677, "grad_norm": 0.16555188596248627, "learning_rate": 7.921509249691865e-05, "loss": 0.0173, "step": 10090 }, { "epoch": 10.687830687830688, "grad_norm": 0.15025760233402252, "learning_rate": 7.917034639098579e-05, "loss": 0.0194, "step": 10100 }, { "epoch": 10.698412698412698, "grad_norm": 0.09903678297996521, "learning_rate": 7.912556484033146e-05, "loss": 0.0157, "step": 10110 }, { "epoch": 10.708994708994709, "grad_norm": 0.19934149086475372, "learning_rate": 7.908074789936952e-05, "loss": 0.0171, "step": 10120 }, { "epoch": 10.71957671957672, "grad_norm": 0.13772308826446533, "learning_rate": 7.903589562255673e-05, "loss": 0.0183, "step": 10130 }, { "epoch": 10.73015873015873, "grad_norm": 0.10359026491641998, "learning_rate": 7.899100806439285e-05, "loss": 0.0145, "step": 10140 }, { "epoch": 10.74074074074074, "grad_norm": 0.1603563278913498, "learning_rate": 7.894608527942049e-05, "loss": 0.0165, "step": 10150 }, { "epoch": 10.751322751322752, "grad_norm": 0.1190338209271431, "learning_rate": 7.89011273222251e-05, "loss": 0.0158, "step": 10160 }, { "epoch": 10.761904761904763, "grad_norm": 0.14533643424510956, "learning_rate": 7.885613424743482e-05, "loss": 0.0178, "step": 10170 }, { "epoch": 10.772486772486772, "grad_norm": 0.1315339207649231, "learning_rate": 7.881110610972044e-05, "loss": 0.017, "step": 10180 }, { "epoch": 10.783068783068783, "grad_norm": 0.1711287945508957, "learning_rate": 7.876604296379544e-05, "loss": 0.0168, "step": 10190 }, { "epoch": 10.793650793650794, "grad_norm": 0.21337328851222992, "learning_rate": 7.872094486441579e-05, "loss": 0.0165, "step": 10200 }, { "epoch": 10.804232804232804, "grad_norm": 0.2290569692850113, "learning_rate": 7.867581186637991e-05, "loss": 0.0162, "step": 10210 }, { "epoch": 10.814814814814815, "grad_norm": 0.14865747094154358, "learning_rate": 7.863064402452866e-05, "loss": 0.0177, "step": 10220 }, { "epoch": 10.825396825396826, "grad_norm": 0.17122487723827362, "learning_rate": 7.858544139374524e-05, "loss": 0.016, "step": 10230 }, { "epoch": 10.835978835978835, "grad_norm": 0.13022202253341675, "learning_rate": 7.854020402895508e-05, "loss": 0.0155, "step": 10240 }, { "epoch": 10.846560846560847, "grad_norm": 0.10258594900369644, "learning_rate": 7.849493198512587e-05, "loss": 0.0172, "step": 10250 }, { "epoch": 10.857142857142858, "grad_norm": 0.16295641660690308, "learning_rate": 7.84496253172674e-05, "loss": 0.0166, "step": 10260 }, { "epoch": 10.867724867724867, "grad_norm": 0.1116187646985054, "learning_rate": 7.840428408043155e-05, "loss": 0.0174, "step": 10270 }, { "epoch": 10.878306878306878, "grad_norm": 0.16047242283821106, "learning_rate": 7.835890832971218e-05, "loss": 0.0144, "step": 10280 }, { "epoch": 10.88888888888889, "grad_norm": 0.10510637611150742, "learning_rate": 7.831349812024512e-05, "loss": 0.0185, "step": 10290 }, { "epoch": 10.899470899470899, "grad_norm": 0.1260281652212143, "learning_rate": 7.826805350720807e-05, "loss": 0.0172, "step": 10300 }, { "epoch": 10.91005291005291, "grad_norm": 0.20648322999477386, "learning_rate": 7.822257454582049e-05, "loss": 0.0171, "step": 10310 }, { "epoch": 10.920634920634921, "grad_norm": 0.13615769147872925, "learning_rate": 7.817706129134363e-05, "loss": 0.0147, "step": 10320 }, { "epoch": 10.93121693121693, "grad_norm": 0.16917085647583008, "learning_rate": 7.813151379908036e-05, "loss": 0.0167, "step": 10330 }, { "epoch": 10.941798941798941, "grad_norm": 0.2059391885995865, "learning_rate": 7.808593212437523e-05, "loss": 0.0171, "step": 10340 }, { "epoch": 10.952380952380953, "grad_norm": 0.14925649762153625, "learning_rate": 7.804031632261421e-05, "loss": 0.0158, "step": 10350 }, { "epoch": 10.962962962962964, "grad_norm": 0.17624665796756744, "learning_rate": 7.799466644922484e-05, "loss": 0.0184, "step": 10360 }, { "epoch": 10.973544973544973, "grad_norm": 0.18466569483280182, "learning_rate": 7.794898255967602e-05, "loss": 0.0182, "step": 10370 }, { "epoch": 10.984126984126984, "grad_norm": 0.1740439087152481, "learning_rate": 7.790326470947795e-05, "loss": 0.0169, "step": 10380 }, { "epoch": 10.994708994708995, "grad_norm": 0.15253259241580963, "learning_rate": 7.785751295418217e-05, "loss": 0.0158, "step": 10390 }, { "epoch": 11.005291005291005, "grad_norm": 0.167111337184906, "learning_rate": 7.781172734938136e-05, "loss": 0.0169, "step": 10400 }, { "epoch": 11.015873015873016, "grad_norm": 0.14555642008781433, "learning_rate": 7.776590795070933e-05, "loss": 0.016, "step": 10410 }, { "epoch": 11.026455026455027, "grad_norm": 0.19797445833683014, "learning_rate": 7.772005481384099e-05, "loss": 0.0144, "step": 10420 }, { "epoch": 11.037037037037036, "grad_norm": 0.1657400131225586, "learning_rate": 7.767416799449223e-05, "loss": 0.0143, "step": 10430 }, { "epoch": 11.047619047619047, "grad_norm": 0.2392760068178177, "learning_rate": 7.762824754841984e-05, "loss": 0.0141, "step": 10440 }, { "epoch": 11.058201058201059, "grad_norm": 0.1851608008146286, "learning_rate": 7.758229353142152e-05, "loss": 0.0194, "step": 10450 }, { "epoch": 11.068783068783068, "grad_norm": 0.19199596345424652, "learning_rate": 7.753630599933572e-05, "loss": 0.0158, "step": 10460 }, { "epoch": 11.079365079365079, "grad_norm": 0.15884645283222198, "learning_rate": 7.749028500804161e-05, "loss": 0.0189, "step": 10470 }, { "epoch": 11.08994708994709, "grad_norm": 0.17603681981563568, "learning_rate": 7.744423061345906e-05, "loss": 0.0162, "step": 10480 }, { "epoch": 11.100529100529101, "grad_norm": 0.14138297736644745, "learning_rate": 7.73981428715485e-05, "loss": 0.0193, "step": 10490 }, { "epoch": 11.11111111111111, "grad_norm": 0.2183818221092224, "learning_rate": 7.735202183831085e-05, "loss": 0.0198, "step": 10500 }, { "epoch": 11.121693121693122, "grad_norm": 0.1251598298549652, "learning_rate": 7.730586756978757e-05, "loss": 0.0134, "step": 10510 }, { "epoch": 11.132275132275133, "grad_norm": 0.15142373740673065, "learning_rate": 7.72596801220604e-05, "loss": 0.0171, "step": 10520 }, { "epoch": 11.142857142857142, "grad_norm": 0.10665736347436905, "learning_rate": 7.721345955125147e-05, "loss": 0.0153, "step": 10530 }, { "epoch": 11.153439153439153, "grad_norm": 0.1457769274711609, "learning_rate": 7.71672059135231e-05, "loss": 0.0174, "step": 10540 }, { "epoch": 11.164021164021165, "grad_norm": 0.13509933650493622, "learning_rate": 7.712091926507787e-05, "loss": 0.0168, "step": 10550 }, { "epoch": 11.174603174603174, "grad_norm": 0.16695331037044525, "learning_rate": 7.70745996621584e-05, "loss": 0.0169, "step": 10560 }, { "epoch": 11.185185185185185, "grad_norm": 0.2290833741426468, "learning_rate": 7.702824716104735e-05, "loss": 0.0171, "step": 10570 }, { "epoch": 11.195767195767196, "grad_norm": 0.11694402247667313, "learning_rate": 7.698186181806743e-05, "loss": 0.0151, "step": 10580 }, { "epoch": 11.206349206349206, "grad_norm": 0.18576139211654663, "learning_rate": 7.693544368958116e-05, "loss": 0.0172, "step": 10590 }, { "epoch": 11.216931216931217, "grad_norm": 0.11731576919555664, "learning_rate": 7.688899283199096e-05, "loss": 0.0139, "step": 10600 }, { "epoch": 11.227513227513228, "grad_norm": 0.2339949905872345, "learning_rate": 7.684250930173901e-05, "loss": 0.0182, "step": 10610 }, { "epoch": 11.238095238095237, "grad_norm": 0.14200828969478607, "learning_rate": 7.679599315530718e-05, "loss": 0.0187, "step": 10620 }, { "epoch": 11.248677248677248, "grad_norm": 0.12492359429597855, "learning_rate": 7.674944444921695e-05, "loss": 0.0188, "step": 10630 }, { "epoch": 11.25925925925926, "grad_norm": 0.19339804351329803, "learning_rate": 7.670286324002944e-05, "loss": 0.0203, "step": 10640 }, { "epoch": 11.26984126984127, "grad_norm": 0.14723435044288635, "learning_rate": 7.665624958434514e-05, "loss": 0.0162, "step": 10650 }, { "epoch": 11.28042328042328, "grad_norm": 0.15799805521965027, "learning_rate": 7.66096035388041e-05, "loss": 0.016, "step": 10660 }, { "epoch": 11.291005291005291, "grad_norm": 0.14902062714099884, "learning_rate": 7.656292516008563e-05, "loss": 0.0193, "step": 10670 }, { "epoch": 11.301587301587302, "grad_norm": 0.20896035432815552, "learning_rate": 7.651621450490837e-05, "loss": 0.0201, "step": 10680 }, { "epoch": 11.312169312169312, "grad_norm": 0.1430826336145401, "learning_rate": 7.646947163003017e-05, "loss": 0.0152, "step": 10690 }, { "epoch": 11.322751322751323, "grad_norm": 0.18388979136943817, "learning_rate": 7.642269659224803e-05, "loss": 0.0166, "step": 10700 }, { "epoch": 11.333333333333334, "grad_norm": 0.16811971366405487, "learning_rate": 7.637588944839803e-05, "loss": 0.0161, "step": 10710 }, { "epoch": 11.343915343915343, "grad_norm": 0.16978034377098083, "learning_rate": 7.632905025535528e-05, "loss": 0.0171, "step": 10720 }, { "epoch": 11.354497354497354, "grad_norm": 0.1480986773967743, "learning_rate": 7.628217907003378e-05, "loss": 0.0188, "step": 10730 }, { "epoch": 11.365079365079366, "grad_norm": 0.17400865256786346, "learning_rate": 7.623527594938649e-05, "loss": 0.0149, "step": 10740 }, { "epoch": 11.375661375661375, "grad_norm": 0.1433708667755127, "learning_rate": 7.618834095040509e-05, "loss": 0.0165, "step": 10750 }, { "epoch": 11.386243386243386, "grad_norm": 0.14022809267044067, "learning_rate": 7.614137413012e-05, "loss": 0.0147, "step": 10760 }, { "epoch": 11.396825396825397, "grad_norm": 0.11850980669260025, "learning_rate": 7.609437554560042e-05, "loss": 0.0144, "step": 10770 }, { "epoch": 11.407407407407407, "grad_norm": 0.20763088762760162, "learning_rate": 7.604734525395398e-05, "loss": 0.0152, "step": 10780 }, { "epoch": 11.417989417989418, "grad_norm": 0.1510779857635498, "learning_rate": 7.600028331232699e-05, "loss": 0.0159, "step": 10790 }, { "epoch": 11.428571428571429, "grad_norm": 0.11851391941308975, "learning_rate": 7.595318977790408e-05, "loss": 0.0168, "step": 10800 }, { "epoch": 11.43915343915344, "grad_norm": 0.16385264694690704, "learning_rate": 7.590606470790836e-05, "loss": 0.0136, "step": 10810 }, { "epoch": 11.44973544973545, "grad_norm": 0.15224173665046692, "learning_rate": 7.585890815960125e-05, "loss": 0.0152, "step": 10820 }, { "epoch": 11.46031746031746, "grad_norm": 0.1445734053850174, "learning_rate": 7.581172019028238e-05, "loss": 0.0153, "step": 10830 }, { "epoch": 11.470899470899472, "grad_norm": 0.14458367228507996, "learning_rate": 7.576450085728959e-05, "loss": 0.0154, "step": 10840 }, { "epoch": 11.481481481481481, "grad_norm": 0.12058544158935547, "learning_rate": 7.571725021799885e-05, "loss": 0.0161, "step": 10850 }, { "epoch": 11.492063492063492, "grad_norm": 0.15639238059520721, "learning_rate": 7.566996832982409e-05, "loss": 0.0143, "step": 10860 }, { "epoch": 11.502645502645503, "grad_norm": 0.17140445113182068, "learning_rate": 7.56226552502173e-05, "loss": 0.0159, "step": 10870 }, { "epoch": 11.513227513227513, "grad_norm": 0.12499185651540756, "learning_rate": 7.557531103666833e-05, "loss": 0.0141, "step": 10880 }, { "epoch": 11.523809523809524, "grad_norm": 0.13725444674491882, "learning_rate": 7.552793574670485e-05, "loss": 0.0138, "step": 10890 }, { "epoch": 11.534391534391535, "grad_norm": 0.14149929583072662, "learning_rate": 7.54805294378923e-05, "loss": 0.0165, "step": 10900 }, { "epoch": 11.544973544973544, "grad_norm": 0.1427084356546402, "learning_rate": 7.543309216783384e-05, "loss": 0.0169, "step": 10910 }, { "epoch": 11.555555555555555, "grad_norm": 0.2062721848487854, "learning_rate": 7.53856239941702e-05, "loss": 0.0154, "step": 10920 }, { "epoch": 11.566137566137566, "grad_norm": 0.1481231153011322, "learning_rate": 7.533812497457972e-05, "loss": 0.0185, "step": 10930 }, { "epoch": 11.576719576719576, "grad_norm": 0.1372426152229309, "learning_rate": 7.529059516677814e-05, "loss": 0.0171, "step": 10940 }, { "epoch": 11.587301587301587, "grad_norm": 0.11712191998958588, "learning_rate": 7.524303462851872e-05, "loss": 0.017, "step": 10950 }, { "epoch": 11.597883597883598, "grad_norm": 0.1446828544139862, "learning_rate": 7.519544341759192e-05, "loss": 0.0128, "step": 10960 }, { "epoch": 11.60846560846561, "grad_norm": 0.13651500642299652, "learning_rate": 7.514782159182562e-05, "loss": 0.0177, "step": 10970 }, { "epoch": 11.619047619047619, "grad_norm": 0.17607565224170685, "learning_rate": 7.510016920908481e-05, "loss": 0.0138, "step": 10980 }, { "epoch": 11.62962962962963, "grad_norm": 0.24730385839939117, "learning_rate": 7.505248632727158e-05, "loss": 0.0133, "step": 10990 }, { "epoch": 11.640211640211641, "grad_norm": 0.11319776624441147, "learning_rate": 7.50047730043252e-05, "loss": 0.0165, "step": 11000 }, { "epoch": 11.65079365079365, "grad_norm": 0.19627943634986877, "learning_rate": 7.495702929822183e-05, "loss": 0.0145, "step": 11010 }, { "epoch": 11.661375661375661, "grad_norm": 0.1494280993938446, "learning_rate": 7.490925526697455e-05, "loss": 0.0154, "step": 11020 }, { "epoch": 11.671957671957673, "grad_norm": 0.09604530036449432, "learning_rate": 7.486145096863334e-05, "loss": 0.0156, "step": 11030 }, { "epoch": 11.682539682539682, "grad_norm": 0.12281036376953125, "learning_rate": 7.481361646128491e-05, "loss": 0.0147, "step": 11040 }, { "epoch": 11.693121693121693, "grad_norm": 0.12713780999183655, "learning_rate": 7.476575180305271e-05, "loss": 0.0171, "step": 11050 }, { "epoch": 11.703703703703704, "grad_norm": 0.10401991754770279, "learning_rate": 7.471785705209682e-05, "loss": 0.0149, "step": 11060 }, { "epoch": 11.714285714285714, "grad_norm": 0.14850853383541107, "learning_rate": 7.466993226661387e-05, "loss": 0.0156, "step": 11070 }, { "epoch": 11.724867724867725, "grad_norm": 0.13182906806468964, "learning_rate": 7.462197750483699e-05, "loss": 0.0137, "step": 11080 }, { "epoch": 11.735449735449736, "grad_norm": 0.18154430389404297, "learning_rate": 7.457399282503574e-05, "loss": 0.0139, "step": 11090 }, { "epoch": 11.746031746031747, "grad_norm": 0.16830036044120789, "learning_rate": 7.452597828551604e-05, "loss": 0.0151, "step": 11100 }, { "epoch": 11.756613756613756, "grad_norm": 0.15579013526439667, "learning_rate": 7.447793394462006e-05, "loss": 0.0145, "step": 11110 }, { "epoch": 11.767195767195767, "grad_norm": 0.1327953189611435, "learning_rate": 7.442985986072624e-05, "loss": 0.0164, "step": 11120 }, { "epoch": 11.777777777777779, "grad_norm": 0.17196395993232727, "learning_rate": 7.438175609224908e-05, "loss": 0.0161, "step": 11130 }, { "epoch": 11.788359788359788, "grad_norm": 0.1955474466085434, "learning_rate": 7.433362269763924e-05, "loss": 0.0146, "step": 11140 }, { "epoch": 11.798941798941799, "grad_norm": 0.14997707307338715, "learning_rate": 7.428545973538329e-05, "loss": 0.0125, "step": 11150 }, { "epoch": 11.80952380952381, "grad_norm": 0.15259280800819397, "learning_rate": 7.42372672640038e-05, "loss": 0.0181, "step": 11160 }, { "epoch": 11.82010582010582, "grad_norm": 0.1911485642194748, "learning_rate": 7.418904534205917e-05, "loss": 0.0176, "step": 11170 }, { "epoch": 11.83068783068783, "grad_norm": 0.11616220325231552, "learning_rate": 7.414079402814356e-05, "loss": 0.0153, "step": 11180 }, { "epoch": 11.841269841269842, "grad_norm": 0.15144428610801697, "learning_rate": 7.409251338088687e-05, "loss": 0.0175, "step": 11190 }, { "epoch": 11.851851851851851, "grad_norm": 0.18782606720924377, "learning_rate": 7.404420345895467e-05, "loss": 0.0161, "step": 11200 }, { "epoch": 11.862433862433862, "grad_norm": 0.14810536801815033, "learning_rate": 7.399586432104804e-05, "loss": 0.0142, "step": 11210 }, { "epoch": 11.873015873015873, "grad_norm": 0.15593817830085754, "learning_rate": 7.394749602590358e-05, "loss": 0.018, "step": 11220 }, { "epoch": 11.883597883597883, "grad_norm": 0.10093612968921661, "learning_rate": 7.389909863229336e-05, "loss": 0.0153, "step": 11230 }, { "epoch": 11.894179894179894, "grad_norm": 0.20632988214492798, "learning_rate": 7.385067219902477e-05, "loss": 0.016, "step": 11240 }, { "epoch": 11.904761904761905, "grad_norm": 0.1748553365468979, "learning_rate": 7.380221678494049e-05, "loss": 0.0148, "step": 11250 }, { "epoch": 11.915343915343914, "grad_norm": 0.16098956763744354, "learning_rate": 7.37537324489184e-05, "loss": 0.0164, "step": 11260 }, { "epoch": 11.925925925925926, "grad_norm": 0.1394336223602295, "learning_rate": 7.370521924987155e-05, "loss": 0.0147, "step": 11270 }, { "epoch": 11.936507936507937, "grad_norm": 0.11299649626016617, "learning_rate": 7.365667724674806e-05, "loss": 0.0171, "step": 11280 }, { "epoch": 11.947089947089948, "grad_norm": 0.16665229201316833, "learning_rate": 7.360810649853105e-05, "loss": 0.0157, "step": 11290 }, { "epoch": 11.957671957671957, "grad_norm": 0.16673704981803894, "learning_rate": 7.355950706423853e-05, "loss": 0.0155, "step": 11300 }, { "epoch": 11.968253968253968, "grad_norm": 0.15061406791210175, "learning_rate": 7.351087900292342e-05, "loss": 0.0178, "step": 11310 }, { "epoch": 11.97883597883598, "grad_norm": 0.1501716673374176, "learning_rate": 7.346222237367339e-05, "loss": 0.0136, "step": 11320 }, { "epoch": 11.989417989417989, "grad_norm": 0.12501667439937592, "learning_rate": 7.341353723561081e-05, "loss": 0.0209, "step": 11330 }, { "epoch": 12.0, "grad_norm": 0.17312663793563843, "learning_rate": 7.336482364789277e-05, "loss": 0.0156, "step": 11340 }, { "epoch": 12.010582010582011, "grad_norm": 0.17601194977760315, "learning_rate": 7.331608166971082e-05, "loss": 0.0189, "step": 11350 }, { "epoch": 12.02116402116402, "grad_norm": 0.22953324019908905, "learning_rate": 7.326731136029108e-05, "loss": 0.019, "step": 11360 }, { "epoch": 12.031746031746032, "grad_norm": 0.1512768715620041, "learning_rate": 7.321851277889408e-05, "loss": 0.0154, "step": 11370 }, { "epoch": 12.042328042328043, "grad_norm": 0.20997072756290436, "learning_rate": 7.316968598481469e-05, "loss": 0.0164, "step": 11380 }, { "epoch": 12.052910052910052, "grad_norm": 0.17648394405841827, "learning_rate": 7.312083103738206e-05, "loss": 0.0141, "step": 11390 }, { "epoch": 12.063492063492063, "grad_norm": 0.189301535487175, "learning_rate": 7.307194799595958e-05, "loss": 0.0171, "step": 11400 }, { "epoch": 12.074074074074074, "grad_norm": 0.18631279468536377, "learning_rate": 7.302303691994473e-05, "loss": 0.017, "step": 11410 }, { "epoch": 12.084656084656086, "grad_norm": 0.1839209944009781, "learning_rate": 7.29740978687691e-05, "loss": 0.0141, "step": 11420 }, { "epoch": 12.095238095238095, "grad_norm": 0.31311294436454773, "learning_rate": 7.292513090189825e-05, "loss": 0.0163, "step": 11430 }, { "epoch": 12.105820105820106, "grad_norm": 0.17783492803573608, "learning_rate": 7.287613607883163e-05, "loss": 0.0169, "step": 11440 }, { "epoch": 12.116402116402117, "grad_norm": 0.2347680628299713, "learning_rate": 7.282711345910263e-05, "loss": 0.0156, "step": 11450 }, { "epoch": 12.126984126984127, "grad_norm": 0.15422728657722473, "learning_rate": 7.27780631022783e-05, "loss": 0.0186, "step": 11460 }, { "epoch": 12.137566137566138, "grad_norm": 0.1331804245710373, "learning_rate": 7.272898506795948e-05, "loss": 0.0172, "step": 11470 }, { "epoch": 12.148148148148149, "grad_norm": 0.1318727731704712, "learning_rate": 7.267987941578058e-05, "loss": 0.0173, "step": 11480 }, { "epoch": 12.158730158730158, "grad_norm": 0.10192675143480301, "learning_rate": 7.263074620540962e-05, "loss": 0.0161, "step": 11490 }, { "epoch": 12.16931216931217, "grad_norm": 0.17894276976585388, "learning_rate": 7.25815854965481e-05, "loss": 0.0187, "step": 11500 }, { "epoch": 12.17989417989418, "grad_norm": 0.12232589721679688, "learning_rate": 7.253239734893089e-05, "loss": 0.0178, "step": 11510 }, { "epoch": 12.19047619047619, "grad_norm": 0.17203617095947266, "learning_rate": 7.248318182232623e-05, "loss": 0.0142, "step": 11520 }, { "epoch": 12.201058201058201, "grad_norm": 0.15718448162078857, "learning_rate": 7.243393897653565e-05, "loss": 0.0188, "step": 11530 }, { "epoch": 12.211640211640212, "grad_norm": 0.18686896562576294, "learning_rate": 7.23846688713938e-05, "loss": 0.0185, "step": 11540 }, { "epoch": 12.222222222222221, "grad_norm": 0.21081963181495667, "learning_rate": 7.233537156676854e-05, "loss": 0.0161, "step": 11550 }, { "epoch": 12.232804232804233, "grad_norm": 0.15197180211544037, "learning_rate": 7.228604712256076e-05, "loss": 0.0174, "step": 11560 }, { "epoch": 12.243386243386244, "grad_norm": 0.12725025415420532, "learning_rate": 7.223669559870426e-05, "loss": 0.0149, "step": 11570 }, { "epoch": 12.253968253968253, "grad_norm": 0.15071865916252136, "learning_rate": 7.218731705516585e-05, "loss": 0.0177, "step": 11580 }, { "epoch": 12.264550264550264, "grad_norm": 0.10861744731664658, "learning_rate": 7.21379115519451e-05, "loss": 0.0136, "step": 11590 }, { "epoch": 12.275132275132275, "grad_norm": 0.11628078669309616, "learning_rate": 7.20884791490743e-05, "loss": 0.0153, "step": 11600 }, { "epoch": 12.285714285714286, "grad_norm": 0.12958702445030212, "learning_rate": 7.203901990661857e-05, "loss": 0.0161, "step": 11610 }, { "epoch": 12.296296296296296, "grad_norm": 0.1676003485918045, "learning_rate": 7.198953388467549e-05, "loss": 0.0174, "step": 11620 }, { "epoch": 12.306878306878307, "grad_norm": 0.15146680176258087, "learning_rate": 7.194002114337526e-05, "loss": 0.0155, "step": 11630 }, { "epoch": 12.317460317460318, "grad_norm": 0.18644975125789642, "learning_rate": 7.189048174288054e-05, "loss": 0.0172, "step": 11640 }, { "epoch": 12.328042328042327, "grad_norm": 0.1326739490032196, "learning_rate": 7.184091574338636e-05, "loss": 0.0146, "step": 11650 }, { "epoch": 12.338624338624339, "grad_norm": 0.1964118629693985, "learning_rate": 7.179132320512009e-05, "loss": 0.0134, "step": 11660 }, { "epoch": 12.34920634920635, "grad_norm": 0.19681821763515472, "learning_rate": 7.174170418834134e-05, "loss": 0.0149, "step": 11670 }, { "epoch": 12.359788359788359, "grad_norm": 0.17235447466373444, "learning_rate": 7.169205875334189e-05, "loss": 0.0149, "step": 11680 }, { "epoch": 12.37037037037037, "grad_norm": 0.1271544247865677, "learning_rate": 7.164238696044562e-05, "loss": 0.015, "step": 11690 }, { "epoch": 12.380952380952381, "grad_norm": 0.12907624244689941, "learning_rate": 7.159268887000846e-05, "loss": 0.017, "step": 11700 }, { "epoch": 12.39153439153439, "grad_norm": 0.1762828230857849, "learning_rate": 7.154296454241827e-05, "loss": 0.0156, "step": 11710 }, { "epoch": 12.402116402116402, "grad_norm": 0.1813759058713913, "learning_rate": 7.149321403809479e-05, "loss": 0.0148, "step": 11720 }, { "epoch": 12.412698412698413, "grad_norm": 0.1316537857055664, "learning_rate": 7.144343741748956e-05, "loss": 0.0132, "step": 11730 }, { "epoch": 12.423280423280424, "grad_norm": 0.14475040137767792, "learning_rate": 7.139363474108589e-05, "loss": 0.0135, "step": 11740 }, { "epoch": 12.433862433862434, "grad_norm": 0.1190594807267189, "learning_rate": 7.134380606939874e-05, "loss": 0.0143, "step": 11750 }, { "epoch": 12.444444444444445, "grad_norm": 0.16127800941467285, "learning_rate": 7.12939514629746e-05, "loss": 0.0159, "step": 11760 }, { "epoch": 12.455026455026456, "grad_norm": 0.16400882601737976, "learning_rate": 7.124407098239155e-05, "loss": 0.0147, "step": 11770 }, { "epoch": 12.465608465608465, "grad_norm": 0.12457337230443954, "learning_rate": 7.119416468825908e-05, "loss": 0.0154, "step": 11780 }, { "epoch": 12.476190476190476, "grad_norm": 0.111515112221241, "learning_rate": 7.114423264121804e-05, "loss": 0.0204, "step": 11790 }, { "epoch": 12.486772486772487, "grad_norm": 0.15820318460464478, "learning_rate": 7.109427490194056e-05, "loss": 0.0155, "step": 11800 }, { "epoch": 12.497354497354497, "grad_norm": 0.10965754091739655, "learning_rate": 7.104429153113001e-05, "loss": 0.0126, "step": 11810 }, { "epoch": 12.507936507936508, "grad_norm": 0.1204577311873436, "learning_rate": 7.099428258952092e-05, "loss": 0.0168, "step": 11820 }, { "epoch": 12.518518518518519, "grad_norm": 0.21079012751579285, "learning_rate": 7.094424813787883e-05, "loss": 0.017, "step": 11830 }, { "epoch": 12.529100529100528, "grad_norm": 0.20222780108451843, "learning_rate": 7.089418823700034e-05, "loss": 0.0157, "step": 11840 }, { "epoch": 12.53968253968254, "grad_norm": 0.1827419549226761, "learning_rate": 7.084410294771298e-05, "loss": 0.0147, "step": 11850 }, { "epoch": 12.55026455026455, "grad_norm": 0.11591069400310516, "learning_rate": 7.079399233087504e-05, "loss": 0.0125, "step": 11860 }, { "epoch": 12.56084656084656, "grad_norm": 0.131125345826149, "learning_rate": 7.074385644737568e-05, "loss": 0.0163, "step": 11870 }, { "epoch": 12.571428571428571, "grad_norm": 0.14558573067188263, "learning_rate": 7.069369535813473e-05, "loss": 0.016, "step": 11880 }, { "epoch": 12.582010582010582, "grad_norm": 0.23214906454086304, "learning_rate": 7.06435091241026e-05, "loss": 0.0172, "step": 11890 }, { "epoch": 12.592592592592592, "grad_norm": 0.1298687756061554, "learning_rate": 7.059329780626034e-05, "loss": 0.0192, "step": 11900 }, { "epoch": 12.603174603174603, "grad_norm": 0.17411567270755768, "learning_rate": 7.054306146561944e-05, "loss": 0.0177, "step": 11910 }, { "epoch": 12.613756613756614, "grad_norm": 0.12121574580669403, "learning_rate": 7.049280016322178e-05, "loss": 0.0157, "step": 11920 }, { "epoch": 12.624338624338625, "grad_norm": 0.1883932650089264, "learning_rate": 7.044251396013957e-05, "loss": 0.0174, "step": 11930 }, { "epoch": 12.634920634920634, "grad_norm": 0.10054516792297363, "learning_rate": 7.039220291747528e-05, "loss": 0.0146, "step": 11940 }, { "epoch": 12.645502645502646, "grad_norm": 0.1843080371618271, "learning_rate": 7.034186709636159e-05, "loss": 0.0151, "step": 11950 }, { "epoch": 12.656084656084657, "grad_norm": 0.12823595106601715, "learning_rate": 7.029150655796129e-05, "loss": 0.0172, "step": 11960 }, { "epoch": 12.666666666666666, "grad_norm": 0.19059891998767853, "learning_rate": 7.024112136346712e-05, "loss": 0.015, "step": 11970 }, { "epoch": 12.677248677248677, "grad_norm": 0.12430836260318756, "learning_rate": 7.01907115741019e-05, "loss": 0.0144, "step": 11980 }, { "epoch": 12.687830687830688, "grad_norm": 0.10923315584659576, "learning_rate": 7.014027725111825e-05, "loss": 0.0157, "step": 11990 }, { "epoch": 12.698412698412698, "grad_norm": 0.1637873500585556, "learning_rate": 7.008981845579864e-05, "loss": 0.0174, "step": 12000 }, { "epoch": 12.708994708994709, "grad_norm": 0.1612076759338379, "learning_rate": 7.003933524945528e-05, "loss": 0.0174, "step": 12010 }, { "epoch": 12.71957671957672, "grad_norm": 0.09894504398107529, "learning_rate": 6.998882769342998e-05, "loss": 0.0151, "step": 12020 }, { "epoch": 12.73015873015873, "grad_norm": 0.1415034383535385, "learning_rate": 6.993829584909423e-05, "loss": 0.0149, "step": 12030 }, { "epoch": 12.74074074074074, "grad_norm": 0.14391416311264038, "learning_rate": 6.988773977784895e-05, "loss": 0.0164, "step": 12040 }, { "epoch": 12.751322751322752, "grad_norm": 0.09954191744327545, "learning_rate": 6.983715954112454e-05, "loss": 0.0133, "step": 12050 }, { "epoch": 12.761904761904763, "grad_norm": 0.15104250609874725, "learning_rate": 6.978655520038079e-05, "loss": 0.0149, "step": 12060 }, { "epoch": 12.772486772486772, "grad_norm": 0.17496533691883087, "learning_rate": 6.97359268171067e-05, "loss": 0.0153, "step": 12070 }, { "epoch": 12.783068783068783, "grad_norm": 0.15728351473808289, "learning_rate": 6.968527445282056e-05, "loss": 0.013, "step": 12080 }, { "epoch": 12.793650793650794, "grad_norm": 0.12831172347068787, "learning_rate": 6.963459816906977e-05, "loss": 0.0145, "step": 12090 }, { "epoch": 12.804232804232804, "grad_norm": 0.167995423078537, "learning_rate": 6.958389802743077e-05, "loss": 0.0159, "step": 12100 }, { "epoch": 12.814814814814815, "grad_norm": 0.14579463005065918, "learning_rate": 6.953317408950903e-05, "loss": 0.014, "step": 12110 }, { "epoch": 12.825396825396826, "grad_norm": 0.15403074026107788, "learning_rate": 6.94824264169389e-05, "loss": 0.0156, "step": 12120 }, { "epoch": 12.835978835978835, "grad_norm": 0.11602116376161575, "learning_rate": 6.94316550713836e-05, "loss": 0.0164, "step": 12130 }, { "epoch": 12.846560846560847, "grad_norm": 0.18342183530330658, "learning_rate": 6.938086011453513e-05, "loss": 0.016, "step": 12140 }, { "epoch": 12.857142857142858, "grad_norm": 0.17057187855243683, "learning_rate": 6.93300416081141e-05, "loss": 0.0151, "step": 12150 }, { "epoch": 12.867724867724867, "grad_norm": 0.13185717165470123, "learning_rate": 6.927919961386984e-05, "loss": 0.0148, "step": 12160 }, { "epoch": 12.878306878306878, "grad_norm": 0.1539977639913559, "learning_rate": 6.922833419358013e-05, "loss": 0.0146, "step": 12170 }, { "epoch": 12.88888888888889, "grad_norm": 0.11095532774925232, "learning_rate": 6.917744540905124e-05, "loss": 0.0144, "step": 12180 }, { "epoch": 12.899470899470899, "grad_norm": 0.12635090947151184, "learning_rate": 6.912653332211787e-05, "loss": 0.0145, "step": 12190 }, { "epoch": 12.91005291005291, "grad_norm": 0.18247438967227936, "learning_rate": 6.9075597994643e-05, "loss": 0.0155, "step": 12200 }, { "epoch": 12.920634920634921, "grad_norm": 0.19986937940120697, "learning_rate": 6.902463948851787e-05, "loss": 0.0181, "step": 12210 }, { "epoch": 12.93121693121693, "grad_norm": 0.18702639639377594, "learning_rate": 6.897365786566184e-05, "loss": 0.016, "step": 12220 }, { "epoch": 12.941798941798941, "grad_norm": 0.1503743827342987, "learning_rate": 6.892265318802242e-05, "loss": 0.0156, "step": 12230 }, { "epoch": 12.952380952380953, "grad_norm": 0.1441185176372528, "learning_rate": 6.887162551757507e-05, "loss": 0.0137, "step": 12240 }, { "epoch": 12.962962962962964, "grad_norm": 0.12353788316249847, "learning_rate": 6.882057491632326e-05, "loss": 0.013, "step": 12250 }, { "epoch": 12.973544973544973, "grad_norm": 0.10600680112838745, "learning_rate": 6.876950144629824e-05, "loss": 0.0142, "step": 12260 }, { "epoch": 12.984126984126984, "grad_norm": 0.1747460514307022, "learning_rate": 6.87184051695591e-05, "loss": 0.0145, "step": 12270 }, { "epoch": 12.994708994708995, "grad_norm": 0.12039068341255188, "learning_rate": 6.866728614819268e-05, "loss": 0.0121, "step": 12280 }, { "epoch": 13.005291005291005, "grad_norm": 0.16734932363033295, "learning_rate": 6.861614444431337e-05, "loss": 0.0132, "step": 12290 }, { "epoch": 13.015873015873016, "grad_norm": 0.18904908001422882, "learning_rate": 6.856498012006318e-05, "loss": 0.0139, "step": 12300 }, { "epoch": 13.026455026455027, "grad_norm": 0.09490162879228592, "learning_rate": 6.851379323761157e-05, "loss": 0.0168, "step": 12310 }, { "epoch": 13.037037037037036, "grad_norm": 0.15239815413951874, "learning_rate": 6.846258385915545e-05, "loss": 0.0156, "step": 12320 }, { "epoch": 13.047619047619047, "grad_norm": 0.09891900420188904, "learning_rate": 6.841135204691902e-05, "loss": 0.0143, "step": 12330 }, { "epoch": 13.058201058201059, "grad_norm": 0.12097975611686707, "learning_rate": 6.836009786315377e-05, "loss": 0.0149, "step": 12340 }, { "epoch": 13.068783068783068, "grad_norm": 0.15538673102855682, "learning_rate": 6.830882137013839e-05, "loss": 0.0156, "step": 12350 }, { "epoch": 13.079365079365079, "grad_norm": 0.11900186538696289, "learning_rate": 6.825752263017863e-05, "loss": 0.0133, "step": 12360 }, { "epoch": 13.08994708994709, "grad_norm": 0.21833842992782593, "learning_rate": 6.820620170560731e-05, "loss": 0.0134, "step": 12370 }, { "epoch": 13.100529100529101, "grad_norm": 0.1082443818449974, "learning_rate": 6.815485865878418e-05, "loss": 0.0173, "step": 12380 }, { "epoch": 13.11111111111111, "grad_norm": 0.16492408514022827, "learning_rate": 6.810349355209588e-05, "loss": 0.0139, "step": 12390 }, { "epoch": 13.121693121693122, "grad_norm": 0.11105608195066452, "learning_rate": 6.805210644795588e-05, "loss": 0.016, "step": 12400 }, { "epoch": 13.132275132275133, "grad_norm": 0.1286090761423111, "learning_rate": 6.800069740880433e-05, "loss": 0.0147, "step": 12410 }, { "epoch": 13.142857142857142, "grad_norm": 0.14115813374519348, "learning_rate": 6.794926649710807e-05, "loss": 0.0162, "step": 12420 }, { "epoch": 13.153439153439153, "grad_norm": 0.14346003532409668, "learning_rate": 6.789781377536052e-05, "loss": 0.0174, "step": 12430 }, { "epoch": 13.164021164021165, "grad_norm": 0.15987585484981537, "learning_rate": 6.784633930608158e-05, "loss": 0.0159, "step": 12440 }, { "epoch": 13.174603174603174, "grad_norm": 0.17939028143882751, "learning_rate": 6.779484315181759e-05, "loss": 0.018, "step": 12450 }, { "epoch": 13.185185185185185, "grad_norm": 0.09884550422430038, "learning_rate": 6.774332537514122e-05, "loss": 0.013, "step": 12460 }, { "epoch": 13.195767195767196, "grad_norm": 0.2312924712896347, "learning_rate": 6.769178603865143e-05, "loss": 0.0158, "step": 12470 }, { "epoch": 13.206349206349206, "grad_norm": 0.1497785449028015, "learning_rate": 6.764022520497337e-05, "loss": 0.0148, "step": 12480 }, { "epoch": 13.216931216931217, "grad_norm": 0.12755294144153595, "learning_rate": 6.758864293675833e-05, "loss": 0.0176, "step": 12490 }, { "epoch": 13.227513227513228, "grad_norm": 0.1514863818883896, "learning_rate": 6.753703929668362e-05, "loss": 0.0128, "step": 12500 }, { "epoch": 13.238095238095237, "grad_norm": 0.14714230597019196, "learning_rate": 6.748541434745253e-05, "loss": 0.0156, "step": 12510 }, { "epoch": 13.248677248677248, "grad_norm": 0.1621895432472229, "learning_rate": 6.743376815179424e-05, "loss": 0.0146, "step": 12520 }, { "epoch": 13.25925925925926, "grad_norm": 0.1630212664604187, "learning_rate": 6.738210077246376e-05, "loss": 0.0125, "step": 12530 }, { "epoch": 13.26984126984127, "grad_norm": 0.14706124365329742, "learning_rate": 6.733041227224181e-05, "loss": 0.016, "step": 12540 }, { "epoch": 13.28042328042328, "grad_norm": 0.24231070280075073, "learning_rate": 6.72787027139348e-05, "loss": 0.0139, "step": 12550 }, { "epoch": 13.291005291005291, "grad_norm": 0.27410683035850525, "learning_rate": 6.72269721603747e-05, "loss": 0.0168, "step": 12560 }, { "epoch": 13.301587301587302, "grad_norm": 0.1507665514945984, "learning_rate": 6.717522067441904e-05, "loss": 0.0145, "step": 12570 }, { "epoch": 13.312169312169312, "grad_norm": 0.20570971071720123, "learning_rate": 6.712344831895074e-05, "loss": 0.018, "step": 12580 }, { "epoch": 13.322751322751323, "grad_norm": 0.11561724543571472, "learning_rate": 6.707165515687811e-05, "loss": 0.0174, "step": 12590 }, { "epoch": 13.333333333333334, "grad_norm": 0.1797116994857788, "learning_rate": 6.70198412511347e-05, "loss": 0.0163, "step": 12600 }, { "epoch": 13.343915343915343, "grad_norm": 0.11711863428354263, "learning_rate": 6.69680066646793e-05, "loss": 0.0149, "step": 12610 }, { "epoch": 13.354497354497354, "grad_norm": 0.11805655807256699, "learning_rate": 6.691615146049583e-05, "loss": 0.0131, "step": 12620 }, { "epoch": 13.365079365079366, "grad_norm": 0.1391814947128296, "learning_rate": 6.686427570159324e-05, "loss": 0.0164, "step": 12630 }, { "epoch": 13.375661375661375, "grad_norm": 0.16913428902626038, "learning_rate": 6.681237945100548e-05, "loss": 0.0152, "step": 12640 }, { "epoch": 13.386243386243386, "grad_norm": 0.16021254658699036, "learning_rate": 6.676046277179139e-05, "loss": 0.0134, "step": 12650 }, { "epoch": 13.396825396825397, "grad_norm": 0.11889854073524475, "learning_rate": 6.670852572703462e-05, "loss": 0.0145, "step": 12660 }, { "epoch": 13.407407407407407, "grad_norm": 0.16511516273021698, "learning_rate": 6.665656837984359e-05, "loss": 0.0152, "step": 12670 }, { "epoch": 13.417989417989418, "grad_norm": 0.10683462768793106, "learning_rate": 6.660459079335135e-05, "loss": 0.0147, "step": 12680 }, { "epoch": 13.428571428571429, "grad_norm": 0.181754469871521, "learning_rate": 6.655259303071558e-05, "loss": 0.0154, "step": 12690 }, { "epoch": 13.43915343915344, "grad_norm": 0.10041944682598114, "learning_rate": 6.650057515511848e-05, "loss": 0.0168, "step": 12700 }, { "epoch": 13.44973544973545, "grad_norm": 0.1711355447769165, "learning_rate": 6.644853722976667e-05, "loss": 0.0143, "step": 12710 }, { "epoch": 13.46031746031746, "grad_norm": 0.12357668578624725, "learning_rate": 6.639647931789113e-05, "loss": 0.0133, "step": 12720 }, { "epoch": 13.470899470899472, "grad_norm": 0.18307052552700043, "learning_rate": 6.634440148274713e-05, "loss": 0.0159, "step": 12730 }, { "epoch": 13.481481481481481, "grad_norm": 0.1368793100118637, "learning_rate": 6.629230378761415e-05, "loss": 0.0134, "step": 12740 }, { "epoch": 13.492063492063492, "grad_norm": 0.1849122941493988, "learning_rate": 6.624018629579581e-05, "loss": 0.0143, "step": 12750 }, { "epoch": 13.502645502645503, "grad_norm": 0.15390101075172424, "learning_rate": 6.618804907061976e-05, "loss": 0.0186, "step": 12760 }, { "epoch": 13.513227513227513, "grad_norm": 0.13289803266525269, "learning_rate": 6.613589217543766e-05, "loss": 0.0141, "step": 12770 }, { "epoch": 13.523809523809524, "grad_norm": 0.11651697754859924, "learning_rate": 6.608371567362504e-05, "loss": 0.0128, "step": 12780 }, { "epoch": 13.534391534391535, "grad_norm": 0.12643297016620636, "learning_rate": 6.60315196285813e-05, "loss": 0.0157, "step": 12790 }, { "epoch": 13.544973544973544, "grad_norm": 0.1516922563314438, "learning_rate": 6.597930410372954e-05, "loss": 0.0184, "step": 12800 }, { "epoch": 13.555555555555555, "grad_norm": 0.1625143587589264, "learning_rate": 6.592706916251653e-05, "loss": 0.0169, "step": 12810 }, { "epoch": 13.566137566137566, "grad_norm": 0.17178773880004883, "learning_rate": 6.587481486841267e-05, "loss": 0.0166, "step": 12820 }, { "epoch": 13.576719576719576, "grad_norm": 0.18780039250850677, "learning_rate": 6.582254128491184e-05, "loss": 0.0116, "step": 12830 }, { "epoch": 13.587301587301587, "grad_norm": 0.13586503267288208, "learning_rate": 6.577024847553139e-05, "loss": 0.0143, "step": 12840 }, { "epoch": 13.597883597883598, "grad_norm": 0.1254788041114807, "learning_rate": 6.571793650381201e-05, "loss": 0.0181, "step": 12850 }, { "epoch": 13.60846560846561, "grad_norm": 0.12519437074661255, "learning_rate": 6.566560543331766e-05, "loss": 0.0143, "step": 12860 }, { "epoch": 13.619047619047619, "grad_norm": 0.14802025258541107, "learning_rate": 6.561325532763554e-05, "loss": 0.018, "step": 12870 }, { "epoch": 13.62962962962963, "grad_norm": 0.17349623143672943, "learning_rate": 6.556088625037597e-05, "loss": 0.0171, "step": 12880 }, { "epoch": 13.640211640211641, "grad_norm": 0.12251042574644089, "learning_rate": 6.55084982651723e-05, "loss": 0.0132, "step": 12890 }, { "epoch": 13.65079365079365, "grad_norm": 0.15643128752708435, "learning_rate": 6.54560914356809e-05, "loss": 0.0154, "step": 12900 }, { "epoch": 13.661375661375661, "grad_norm": 0.1359013020992279, "learning_rate": 6.540366582558097e-05, "loss": 0.012, "step": 12910 }, { "epoch": 13.671957671957673, "grad_norm": 0.10443221032619476, "learning_rate": 6.53512214985746e-05, "loss": 0.0161, "step": 12920 }, { "epoch": 13.682539682539682, "grad_norm": 0.21298764646053314, "learning_rate": 6.529875851838658e-05, "loss": 0.0179, "step": 12930 }, { "epoch": 13.693121693121693, "grad_norm": 0.07328639924526215, "learning_rate": 6.52462769487644e-05, "loss": 0.0143, "step": 12940 }, { "epoch": 13.703703703703704, "grad_norm": 0.13829566538333893, "learning_rate": 6.519377685347807e-05, "loss": 0.014, "step": 12950 }, { "epoch": 13.714285714285714, "grad_norm": 0.17588874697685242, "learning_rate": 6.514125829632021e-05, "loss": 0.0153, "step": 12960 }, { "epoch": 13.724867724867725, "grad_norm": 0.11814041435718536, "learning_rate": 6.508872134110577e-05, "loss": 0.0126, "step": 12970 }, { "epoch": 13.735449735449736, "grad_norm": 0.11782871186733246, "learning_rate": 6.503616605167213e-05, "loss": 0.0135, "step": 12980 }, { "epoch": 13.746031746031747, "grad_norm": 0.148656964302063, "learning_rate": 6.498359249187893e-05, "loss": 0.0137, "step": 12990 }, { "epoch": 13.756613756613756, "grad_norm": 0.15478958189487457, "learning_rate": 6.493100072560799e-05, "loss": 0.0131, "step": 13000 }, { "epoch": 13.767195767195767, "grad_norm": 0.20939433574676514, "learning_rate": 6.487839081676327e-05, "loss": 0.0165, "step": 13010 }, { "epoch": 13.777777777777779, "grad_norm": 0.16892413794994354, "learning_rate": 6.482576282927076e-05, "loss": 0.0162, "step": 13020 }, { "epoch": 13.788359788359788, "grad_norm": 0.14126059412956238, "learning_rate": 6.477311682707845e-05, "loss": 0.0135, "step": 13030 }, { "epoch": 13.798941798941799, "grad_norm": 0.10185698419809341, "learning_rate": 6.472045287415616e-05, "loss": 0.0139, "step": 13040 }, { "epoch": 13.80952380952381, "grad_norm": 0.19186753034591675, "learning_rate": 6.466777103449558e-05, "loss": 0.0149, "step": 13050 }, { "epoch": 13.82010582010582, "grad_norm": 0.13678810000419617, "learning_rate": 6.461507137211012e-05, "loss": 0.0148, "step": 13060 }, { "epoch": 13.83068783068783, "grad_norm": 0.11247415095567703, "learning_rate": 6.456235395103483e-05, "loss": 0.0129, "step": 13070 }, { "epoch": 13.841269841269842, "grad_norm": 0.12719009816646576, "learning_rate": 6.450961883532634e-05, "loss": 0.014, "step": 13080 }, { "epoch": 13.851851851851851, "grad_norm": 0.1357097178697586, "learning_rate": 6.445686608906283e-05, "loss": 0.015, "step": 13090 }, { "epoch": 13.862433862433862, "grad_norm": 0.11440759152173996, "learning_rate": 6.44040957763438e-05, "loss": 0.0133, "step": 13100 }, { "epoch": 13.873015873015873, "grad_norm": 0.15497325360774994, "learning_rate": 6.435130796129018e-05, "loss": 0.0159, "step": 13110 }, { "epoch": 13.883597883597883, "grad_norm": 0.18392284214496613, "learning_rate": 6.429850270804416e-05, "loss": 0.0139, "step": 13120 }, { "epoch": 13.894179894179894, "grad_norm": 0.16896787285804749, "learning_rate": 6.424568008076909e-05, "loss": 0.014, "step": 13130 }, { "epoch": 13.904761904761905, "grad_norm": 0.15702573955059052, "learning_rate": 6.419284014364943e-05, "loss": 0.0125, "step": 13140 }, { "epoch": 13.915343915343914, "grad_norm": 0.11793252825737, "learning_rate": 6.41399829608907e-05, "loss": 0.0128, "step": 13150 }, { "epoch": 13.925925925925926, "grad_norm": 0.10486666858196259, "learning_rate": 6.408710859671938e-05, "loss": 0.0136, "step": 13160 }, { "epoch": 13.936507936507937, "grad_norm": 0.18865247070789337, "learning_rate": 6.403421711538278e-05, "loss": 0.0116, "step": 13170 }, { "epoch": 13.947089947089948, "grad_norm": 0.2084113359451294, "learning_rate": 6.398130858114903e-05, "loss": 0.013, "step": 13180 }, { "epoch": 13.957671957671957, "grad_norm": 0.13861629366874695, "learning_rate": 6.392838305830701e-05, "loss": 0.0142, "step": 13190 }, { "epoch": 13.968253968253968, "grad_norm": 0.12937027215957642, "learning_rate": 6.387544061116621e-05, "loss": 0.0106, "step": 13200 }, { "epoch": 13.97883597883598, "grad_norm": 0.16809387505054474, "learning_rate": 6.382248130405671e-05, "loss": 0.0143, "step": 13210 }, { "epoch": 13.989417989417989, "grad_norm": 0.16509059071540833, "learning_rate": 6.376950520132905e-05, "loss": 0.0148, "step": 13220 }, { "epoch": 14.0, "grad_norm": 0.140420064330101, "learning_rate": 6.371651236735418e-05, "loss": 0.0154, "step": 13230 }, { "epoch": 14.010582010582011, "grad_norm": 0.14782917499542236, "learning_rate": 6.366350286652341e-05, "loss": 0.0133, "step": 13240 }, { "epoch": 14.02116402116402, "grad_norm": 0.1444079577922821, "learning_rate": 6.361047676324827e-05, "loss": 0.0127, "step": 13250 }, { "epoch": 14.031746031746032, "grad_norm": 0.1562630534172058, "learning_rate": 6.355743412196047e-05, "loss": 0.0135, "step": 13260 }, { "epoch": 14.042328042328043, "grad_norm": 0.15513615310192108, "learning_rate": 6.350437500711184e-05, "loss": 0.0141, "step": 13270 }, { "epoch": 14.052910052910052, "grad_norm": 0.19360290467739105, "learning_rate": 6.345129948317418e-05, "loss": 0.0133, "step": 13280 }, { "epoch": 14.063492063492063, "grad_norm": 0.1325603723526001, "learning_rate": 6.33982076146393e-05, "loss": 0.0166, "step": 13290 }, { "epoch": 14.074074074074074, "grad_norm": 0.11801972985267639, "learning_rate": 6.334509946601879e-05, "loss": 0.0118, "step": 13300 }, { "epoch": 14.084656084656086, "grad_norm": 0.1639968752861023, "learning_rate": 6.329197510184406e-05, "loss": 0.0135, "step": 13310 }, { "epoch": 14.095238095238095, "grad_norm": 0.12149186432361603, "learning_rate": 6.323883458666624e-05, "loss": 0.0136, "step": 13320 }, { "epoch": 14.105820105820106, "grad_norm": 0.116496242582798, "learning_rate": 6.318567798505605e-05, "loss": 0.0181, "step": 13330 }, { "epoch": 14.116402116402117, "grad_norm": 0.10622723400592804, "learning_rate": 6.313250536160378e-05, "loss": 0.0112, "step": 13340 }, { "epoch": 14.126984126984127, "grad_norm": 0.13304218649864197, "learning_rate": 6.307931678091917e-05, "loss": 0.0167, "step": 13350 }, { "epoch": 14.137566137566138, "grad_norm": 0.16003704071044922, "learning_rate": 6.302611230763138e-05, "loss": 0.0161, "step": 13360 }, { "epoch": 14.148148148148149, "grad_norm": 0.133531391620636, "learning_rate": 6.297289200638887e-05, "loss": 0.0146, "step": 13370 }, { "epoch": 14.158730158730158, "grad_norm": 0.10696054250001907, "learning_rate": 6.29196559418593e-05, "loss": 0.0123, "step": 13380 }, { "epoch": 14.16931216931217, "grad_norm": 0.21820519864559174, "learning_rate": 6.286640417872951e-05, "loss": 0.012, "step": 13390 }, { "epoch": 14.17989417989418, "grad_norm": 0.14090077579021454, "learning_rate": 6.281313678170542e-05, "loss": 0.0137, "step": 13400 }, { "epoch": 14.19047619047619, "grad_norm": 0.12835563719272614, "learning_rate": 6.275985381551194e-05, "loss": 0.0118, "step": 13410 }, { "epoch": 14.201058201058201, "grad_norm": 0.15023797750473022, "learning_rate": 6.270655534489292e-05, "loss": 0.0158, "step": 13420 }, { "epoch": 14.211640211640212, "grad_norm": 0.2615918517112732, "learning_rate": 6.265324143461098e-05, "loss": 0.0136, "step": 13430 }, { "epoch": 14.222222222222221, "grad_norm": 0.1317720115184784, "learning_rate": 6.259991214944758e-05, "loss": 0.014, "step": 13440 }, { "epoch": 14.232804232804233, "grad_norm": 0.1570160835981369, "learning_rate": 6.254656755420283e-05, "loss": 0.0145, "step": 13450 }, { "epoch": 14.243386243386244, "grad_norm": 0.1638309359550476, "learning_rate": 6.249320771369545e-05, "loss": 0.0139, "step": 13460 }, { "epoch": 14.253968253968253, "grad_norm": 0.13432513177394867, "learning_rate": 6.243983269276263e-05, "loss": 0.0124, "step": 13470 }, { "epoch": 14.264550264550264, "grad_norm": 0.1425156593322754, "learning_rate": 6.238644255626012e-05, "loss": 0.0133, "step": 13480 }, { "epoch": 14.275132275132275, "grad_norm": 0.168029323220253, "learning_rate": 6.233303736906193e-05, "loss": 0.0148, "step": 13490 }, { "epoch": 14.285714285714286, "grad_norm": 0.16404187679290771, "learning_rate": 6.227961719606039e-05, "loss": 0.0186, "step": 13500 }, { "epoch": 14.296296296296296, "grad_norm": 0.19229665398597717, "learning_rate": 6.222618210216608e-05, "loss": 0.0171, "step": 13510 }, { "epoch": 14.306878306878307, "grad_norm": 0.13926075398921967, "learning_rate": 6.217273215230767e-05, "loss": 0.0139, "step": 13520 }, { "epoch": 14.317460317460318, "grad_norm": 0.11396767944097519, "learning_rate": 6.211926741143188e-05, "loss": 0.0141, "step": 13530 }, { "epoch": 14.328042328042327, "grad_norm": 0.17979037761688232, "learning_rate": 6.20657879445034e-05, "loss": 0.0142, "step": 13540 }, { "epoch": 14.338624338624339, "grad_norm": 0.14275245368480682, "learning_rate": 6.201229381650485e-05, "loss": 0.0152, "step": 13550 }, { "epoch": 14.34920634920635, "grad_norm": 0.1622018814086914, "learning_rate": 6.195878509243661e-05, "loss": 0.0155, "step": 13560 }, { "epoch": 14.359788359788359, "grad_norm": 0.15363255143165588, "learning_rate": 6.190526183731685e-05, "loss": 0.0148, "step": 13570 }, { "epoch": 14.37037037037037, "grad_norm": 0.18609574437141418, "learning_rate": 6.185172411618138e-05, "loss": 0.015, "step": 13580 }, { "epoch": 14.380952380952381, "grad_norm": 0.1521168351173401, "learning_rate": 6.179817199408355e-05, "loss": 0.0159, "step": 13590 }, { "epoch": 14.39153439153439, "grad_norm": 0.1034182533621788, "learning_rate": 6.174460553609426e-05, "loss": 0.0149, "step": 13600 }, { "epoch": 14.402116402116402, "grad_norm": 0.18168137967586517, "learning_rate": 6.16910248073018e-05, "loss": 0.0149, "step": 13610 }, { "epoch": 14.412698412698413, "grad_norm": 0.14350634813308716, "learning_rate": 6.16374298728118e-05, "loss": 0.0149, "step": 13620 }, { "epoch": 14.423280423280424, "grad_norm": 0.1944950371980667, "learning_rate": 6.158382079774716e-05, "loss": 0.0144, "step": 13630 }, { "epoch": 14.433862433862434, "grad_norm": 0.23631899058818817, "learning_rate": 6.153019764724799e-05, "loss": 0.0149, "step": 13640 }, { "epoch": 14.444444444444445, "grad_norm": 0.11710357666015625, "learning_rate": 6.147656048647144e-05, "loss": 0.0143, "step": 13650 }, { "epoch": 14.455026455026456, "grad_norm": 0.08130736649036407, "learning_rate": 6.142290938059173e-05, "loss": 0.0121, "step": 13660 }, { "epoch": 14.465608465608465, "grad_norm": 0.15861205756664276, "learning_rate": 6.136924439480001e-05, "loss": 0.0142, "step": 13670 }, { "epoch": 14.476190476190476, "grad_norm": 0.1289794147014618, "learning_rate": 6.13155655943043e-05, "loss": 0.0152, "step": 13680 }, { "epoch": 14.486772486772487, "grad_norm": 0.13518886268138885, "learning_rate": 6.126187304432941e-05, "loss": 0.0139, "step": 13690 }, { "epoch": 14.497354497354497, "grad_norm": 0.1775902956724167, "learning_rate": 6.120816681011684e-05, "loss": 0.0155, "step": 13700 }, { "epoch": 14.507936507936508, "grad_norm": 0.13354086875915527, "learning_rate": 6.115444695692474e-05, "loss": 0.0144, "step": 13710 }, { "epoch": 14.518518518518519, "grad_norm": 0.1303110122680664, "learning_rate": 6.110071355002779e-05, "loss": 0.0145, "step": 13720 }, { "epoch": 14.529100529100528, "grad_norm": 0.10445213317871094, "learning_rate": 6.104696665471714e-05, "loss": 0.0129, "step": 13730 }, { "epoch": 14.53968253968254, "grad_norm": 0.1792774349451065, "learning_rate": 6.099320633630036e-05, "loss": 0.0157, "step": 13740 }, { "epoch": 14.55026455026455, "grad_norm": 0.11070969700813293, "learning_rate": 6.093943266010128e-05, "loss": 0.0149, "step": 13750 }, { "epoch": 14.56084656084656, "grad_norm": 0.20601093769073486, "learning_rate": 6.088564569146e-05, "loss": 0.0141, "step": 13760 }, { "epoch": 14.571428571428571, "grad_norm": 0.06934300065040588, "learning_rate": 6.083184549573274e-05, "loss": 0.0149, "step": 13770 }, { "epoch": 14.582010582010582, "grad_norm": 0.16573341190814972, "learning_rate": 6.077803213829184e-05, "loss": 0.0132, "step": 13780 }, { "epoch": 14.592592592592592, "grad_norm": 0.09251513332128525, "learning_rate": 6.0724205684525594e-05, "loss": 0.0121, "step": 13790 }, { "epoch": 14.603174603174603, "grad_norm": 0.08905959874391556, "learning_rate": 6.0670366199838215e-05, "loss": 0.0126, "step": 13800 }, { "epoch": 14.613756613756614, "grad_norm": 0.0826789140701294, "learning_rate": 6.061651374964974e-05, "loss": 0.011, "step": 13810 }, { "epoch": 14.624338624338625, "grad_norm": 0.09131411463022232, "learning_rate": 6.0562648399396004e-05, "loss": 0.0125, "step": 13820 }, { "epoch": 14.634920634920634, "grad_norm": 0.09689934551715851, "learning_rate": 6.050877021452845e-05, "loss": 0.0123, "step": 13830 }, { "epoch": 14.645502645502646, "grad_norm": 0.07754877209663391, "learning_rate": 6.04548792605142e-05, "loss": 0.0116, "step": 13840 }, { "epoch": 14.656084656084657, "grad_norm": 0.1814226508140564, "learning_rate": 6.04009756028358e-05, "loss": 0.0152, "step": 13850 }, { "epoch": 14.666666666666666, "grad_norm": 0.12367632240056992, "learning_rate": 6.0347059306991294e-05, "loss": 0.012, "step": 13860 }, { "epoch": 14.677248677248677, "grad_norm": 0.12674853205680847, "learning_rate": 6.029313043849407e-05, "loss": 0.0134, "step": 13870 }, { "epoch": 14.687830687830688, "grad_norm": 0.1660003960132599, "learning_rate": 6.023918906287279e-05, "loss": 0.0143, "step": 13880 }, { "epoch": 14.698412698412698, "grad_norm": 0.14694061875343323, "learning_rate": 6.01852352456713e-05, "loss": 0.0139, "step": 13890 }, { "epoch": 14.708994708994709, "grad_norm": 0.1308487057685852, "learning_rate": 6.013126905244858e-05, "loss": 0.0145, "step": 13900 }, { "epoch": 14.71957671957672, "grad_norm": 0.20952732861042023, "learning_rate": 6.007729054877864e-05, "loss": 0.0152, "step": 13910 }, { "epoch": 14.73015873015873, "grad_norm": 0.11215442419052124, "learning_rate": 6.0023299800250466e-05, "loss": 0.0129, "step": 13920 }, { "epoch": 14.74074074074074, "grad_norm": 0.13220550119876862, "learning_rate": 5.9969296872467894e-05, "loss": 0.0145, "step": 13930 }, { "epoch": 14.751322751322752, "grad_norm": 0.1845506876707077, "learning_rate": 5.9915281831049585e-05, "loss": 0.014, "step": 13940 }, { "epoch": 14.761904761904763, "grad_norm": 0.15848499536514282, "learning_rate": 5.98612547416289e-05, "loss": 0.0146, "step": 13950 }, { "epoch": 14.772486772486772, "grad_norm": 0.11135569959878922, "learning_rate": 5.9807215669853855e-05, "loss": 0.0149, "step": 13960 }, { "epoch": 14.783068783068783, "grad_norm": 0.09606221318244934, "learning_rate": 5.9753164681387e-05, "loss": 0.013, "step": 13970 }, { "epoch": 14.793650793650794, "grad_norm": 0.14873123168945312, "learning_rate": 5.969910184190539e-05, "loss": 0.0156, "step": 13980 }, { "epoch": 14.804232804232804, "grad_norm": 0.14512431621551514, "learning_rate": 5.9645027217100477e-05, "loss": 0.0129, "step": 13990 }, { "epoch": 14.814814814814815, "grad_norm": 0.13126476109027863, "learning_rate": 5.959094087267804e-05, "loss": 0.0129, "step": 14000 }, { "epoch": 14.825396825396826, "grad_norm": 0.14063474535942078, "learning_rate": 5.9536842874358065e-05, "loss": 0.0099, "step": 14010 }, { "epoch": 14.835978835978835, "grad_norm": 0.14216510951519012, "learning_rate": 5.9482733287874734e-05, "loss": 0.0121, "step": 14020 }, { "epoch": 14.846560846560847, "grad_norm": 0.13525789976119995, "learning_rate": 5.942861217897631e-05, "loss": 0.0122, "step": 14030 }, { "epoch": 14.857142857142858, "grad_norm": 0.11080493777990341, "learning_rate": 5.9374479613425e-05, "loss": 0.0137, "step": 14040 }, { "epoch": 14.867724867724867, "grad_norm": 0.1599884033203125, "learning_rate": 5.932033565699704e-05, "loss": 0.0128, "step": 14050 }, { "epoch": 14.878306878306878, "grad_norm": 0.15055221319198608, "learning_rate": 5.926618037548237e-05, "loss": 0.0122, "step": 14060 }, { "epoch": 14.88888888888889, "grad_norm": 0.14180968701839447, "learning_rate": 5.9212013834684824e-05, "loss": 0.0131, "step": 14070 }, { "epoch": 14.899470899470899, "grad_norm": 0.10864979773759842, "learning_rate": 5.9157836100421824e-05, "loss": 0.012, "step": 14080 }, { "epoch": 14.91005291005291, "grad_norm": 0.12265057861804962, "learning_rate": 5.910364723852444e-05, "loss": 0.0114, "step": 14090 }, { "epoch": 14.920634920634921, "grad_norm": 0.18045081198215485, "learning_rate": 5.904944731483723e-05, "loss": 0.0129, "step": 14100 }, { "epoch": 14.93121693121693, "grad_norm": 0.10174672305583954, "learning_rate": 5.899523639521825e-05, "loss": 0.0152, "step": 14110 }, { "epoch": 14.941798941798941, "grad_norm": 0.10803879797458649, "learning_rate": 5.894101454553883e-05, "loss": 0.0144, "step": 14120 }, { "epoch": 14.952380952380953, "grad_norm": 0.3464270234107971, "learning_rate": 5.888678183168368e-05, "loss": 0.0132, "step": 14130 }, { "epoch": 14.962962962962964, "grad_norm": 0.1068209856748581, "learning_rate": 5.883253831955061e-05, "loss": 0.0148, "step": 14140 }, { "epoch": 14.973544973544973, "grad_norm": 0.13364610075950623, "learning_rate": 5.8778284075050625e-05, "loss": 0.0149, "step": 14150 }, { "epoch": 14.984126984126984, "grad_norm": 0.15609954297542572, "learning_rate": 5.872401916410777e-05, "loss": 0.0131, "step": 14160 }, { "epoch": 14.994708994708995, "grad_norm": 0.16066236793994904, "learning_rate": 5.8669743652659014e-05, "loss": 0.013, "step": 14170 }, { "epoch": 15.005291005291005, "grad_norm": 0.17526495456695557, "learning_rate": 5.86154576066542e-05, "loss": 0.0122, "step": 14180 }, { "epoch": 15.015873015873016, "grad_norm": 0.12220416963100433, "learning_rate": 5.856116109205602e-05, "loss": 0.0142, "step": 14190 }, { "epoch": 15.026455026455027, "grad_norm": 0.21337449550628662, "learning_rate": 5.850685417483983e-05, "loss": 0.015, "step": 14200 }, { "epoch": 15.037037037037036, "grad_norm": 0.1017012968659401, "learning_rate": 5.8452536920993684e-05, "loss": 0.0131, "step": 14210 }, { "epoch": 15.047619047619047, "grad_norm": 0.13810822367668152, "learning_rate": 5.839820939651817e-05, "loss": 0.0156, "step": 14220 }, { "epoch": 15.058201058201059, "grad_norm": 0.18169832229614258, "learning_rate": 5.834387166742632e-05, "loss": 0.0134, "step": 14230 }, { "epoch": 15.068783068783068, "grad_norm": 0.1079118549823761, "learning_rate": 5.828952379974364e-05, "loss": 0.013, "step": 14240 }, { "epoch": 15.079365079365079, "grad_norm": 0.1597425937652588, "learning_rate": 5.8235165859507864e-05, "loss": 0.0143, "step": 14250 }, { "epoch": 15.08994708994709, "grad_norm": 0.10484649240970612, "learning_rate": 5.818079791276907e-05, "loss": 0.0128, "step": 14260 }, { "epoch": 15.100529100529101, "grad_norm": 0.17379750311374664, "learning_rate": 5.8126420025589415e-05, "loss": 0.0138, "step": 14270 }, { "epoch": 15.11111111111111, "grad_norm": 0.11727312952280045, "learning_rate": 5.807203226404313e-05, "loss": 0.0159, "step": 14280 }, { "epoch": 15.121693121693122, "grad_norm": 0.2043287605047226, "learning_rate": 5.801763469421652e-05, "loss": 0.0136, "step": 14290 }, { "epoch": 15.132275132275133, "grad_norm": 0.14796437323093414, "learning_rate": 5.796322738220774e-05, "loss": 0.014, "step": 14300 }, { "epoch": 15.142857142857142, "grad_norm": 0.18091650307178497, "learning_rate": 5.79088103941268e-05, "loss": 0.0137, "step": 14310 }, { "epoch": 15.153439153439153, "grad_norm": 0.16605959832668304, "learning_rate": 5.785438379609549e-05, "loss": 0.0109, "step": 14320 }, { "epoch": 15.164021164021165, "grad_norm": 0.15680406987667084, "learning_rate": 5.779994765424724e-05, "loss": 0.0126, "step": 14330 }, { "epoch": 15.174603174603174, "grad_norm": 0.1223442479968071, "learning_rate": 5.7745502034727114e-05, "loss": 0.0125, "step": 14340 }, { "epoch": 15.185185185185185, "grad_norm": 0.14641691744327545, "learning_rate": 5.769104700369164e-05, "loss": 0.0122, "step": 14350 }, { "epoch": 15.195767195767196, "grad_norm": 0.18282079696655273, "learning_rate": 5.7636582627308854e-05, "loss": 0.014, "step": 14360 }, { "epoch": 15.206349206349206, "grad_norm": 0.1440655142068863, "learning_rate": 5.7582108971758095e-05, "loss": 0.0122, "step": 14370 }, { "epoch": 15.216931216931217, "grad_norm": 0.1486009955406189, "learning_rate": 5.752762610322999e-05, "loss": 0.0124, "step": 14380 }, { "epoch": 15.227513227513228, "grad_norm": 0.14444106817245483, "learning_rate": 5.747313408792636e-05, "loss": 0.0122, "step": 14390 }, { "epoch": 15.238095238095237, "grad_norm": 0.20641116797924042, "learning_rate": 5.741863299206014e-05, "loss": 0.0108, "step": 14400 }, { "epoch": 15.248677248677248, "grad_norm": 0.1504615992307663, "learning_rate": 5.73641228818553e-05, "loss": 0.0133, "step": 14410 }, { "epoch": 15.25925925925926, "grad_norm": 0.15840892493724823, "learning_rate": 5.730960382354676e-05, "loss": 0.0132, "step": 14420 }, { "epoch": 15.26984126984127, "grad_norm": 0.12382443249225616, "learning_rate": 5.725507588338035e-05, "loss": 0.0136, "step": 14430 }, { "epoch": 15.28042328042328, "grad_norm": 0.1549975723028183, "learning_rate": 5.7200539127612604e-05, "loss": 0.0152, "step": 14440 }, { "epoch": 15.291005291005291, "grad_norm": 0.14088425040245056, "learning_rate": 5.714599362251087e-05, "loss": 0.0107, "step": 14450 }, { "epoch": 15.301587301587302, "grad_norm": 0.14599187672138214, "learning_rate": 5.709143943435307e-05, "loss": 0.0139, "step": 14460 }, { "epoch": 15.312169312169312, "grad_norm": 0.15660905838012695, "learning_rate": 5.7036876629427646e-05, "loss": 0.0144, "step": 14470 }, { "epoch": 15.322751322751323, "grad_norm": 0.0941755622625351, "learning_rate": 5.698230527403361e-05, "loss": 0.0109, "step": 14480 }, { "epoch": 15.333333333333334, "grad_norm": 0.1022164523601532, "learning_rate": 5.6927725434480274e-05, "loss": 0.0119, "step": 14490 }, { "epoch": 15.343915343915343, "grad_norm": 0.12540951371192932, "learning_rate": 5.687313717708728e-05, "loss": 0.0115, "step": 14500 }, { "epoch": 15.354497354497354, "grad_norm": 0.1474149078130722, "learning_rate": 5.681854056818453e-05, "loss": 0.0124, "step": 14510 }, { "epoch": 15.365079365079366, "grad_norm": 0.1355440467596054, "learning_rate": 5.676393567411205e-05, "loss": 0.0122, "step": 14520 }, { "epoch": 15.375661375661375, "grad_norm": 0.12605991959571838, "learning_rate": 5.670932256121991e-05, "loss": 0.0124, "step": 14530 }, { "epoch": 15.386243386243386, "grad_norm": 0.12386475503444672, "learning_rate": 5.6654701295868216e-05, "loss": 0.0124, "step": 14540 }, { "epoch": 15.396825396825397, "grad_norm": 0.12370302528142929, "learning_rate": 5.660007194442697e-05, "loss": 0.0125, "step": 14550 }, { "epoch": 15.407407407407407, "grad_norm": 0.20037788152694702, "learning_rate": 5.6545434573275946e-05, "loss": 0.0132, "step": 14560 }, { "epoch": 15.417989417989418, "grad_norm": 0.18843266367912292, "learning_rate": 5.649078924880472e-05, "loss": 0.0125, "step": 14570 }, { "epoch": 15.428571428571429, "grad_norm": 0.11902154237031937, "learning_rate": 5.643613603741252e-05, "loss": 0.011, "step": 14580 }, { "epoch": 15.43915343915344, "grad_norm": 0.21317724883556366, "learning_rate": 5.6381475005508154e-05, "loss": 0.0145, "step": 14590 }, { "epoch": 15.44973544973545, "grad_norm": 0.1254330575466156, "learning_rate": 5.632680621950992e-05, "loss": 0.0145, "step": 14600 }, { "epoch": 15.46031746031746, "grad_norm": 0.1650930792093277, "learning_rate": 5.627212974584555e-05, "loss": 0.0137, "step": 14610 }, { "epoch": 15.470899470899472, "grad_norm": 0.12016919255256653, "learning_rate": 5.62174456509521e-05, "loss": 0.0124, "step": 14620 }, { "epoch": 15.481481481481481, "grad_norm": 0.09299294650554657, "learning_rate": 5.616275400127594e-05, "loss": 0.012, "step": 14630 }, { "epoch": 15.492063492063492, "grad_norm": 0.1702631115913391, "learning_rate": 5.610805486327254e-05, "loss": 0.0144, "step": 14640 }, { "epoch": 15.502645502645503, "grad_norm": 0.08499672263860703, "learning_rate": 5.6053348303406536e-05, "loss": 0.0138, "step": 14650 }, { "epoch": 15.513227513227513, "grad_norm": 0.10546469688415527, "learning_rate": 5.599863438815156e-05, "loss": 0.0107, "step": 14660 }, { "epoch": 15.523809523809524, "grad_norm": 0.1281641721725464, "learning_rate": 5.594391318399017e-05, "loss": 0.0108, "step": 14670 }, { "epoch": 15.534391534391535, "grad_norm": 0.18578265607357025, "learning_rate": 5.588918475741377e-05, "loss": 0.0114, "step": 14680 }, { "epoch": 15.544973544973544, "grad_norm": 0.09975679963827133, "learning_rate": 5.5834449174922586e-05, "loss": 0.0126, "step": 14690 }, { "epoch": 15.555555555555555, "grad_norm": 0.11396104842424393, "learning_rate": 5.57797065030255e-05, "loss": 0.0152, "step": 14700 }, { "epoch": 15.566137566137566, "grad_norm": 0.0760219469666481, "learning_rate": 5.5724956808240016e-05, "loss": 0.016, "step": 14710 }, { "epoch": 15.576719576719576, "grad_norm": 0.14015363156795502, "learning_rate": 5.567020015709219e-05, "loss": 0.0121, "step": 14720 }, { "epoch": 15.587301587301587, "grad_norm": 0.14406350255012512, "learning_rate": 5.561543661611649e-05, "loss": 0.0145, "step": 14730 }, { "epoch": 15.597883597883598, "grad_norm": 0.15124252438545227, "learning_rate": 5.556066625185583e-05, "loss": 0.0123, "step": 14740 }, { "epoch": 15.60846560846561, "grad_norm": 0.09609399735927582, "learning_rate": 5.550588913086131e-05, "loss": 0.0129, "step": 14750 }, { "epoch": 15.619047619047619, "grad_norm": 0.09960085153579712, "learning_rate": 5.545110531969234e-05, "loss": 0.012, "step": 14760 }, { "epoch": 15.62962962962963, "grad_norm": 0.09883574396371841, "learning_rate": 5.539631488491641e-05, "loss": 0.0133, "step": 14770 }, { "epoch": 15.640211640211641, "grad_norm": 0.12396075576543808, "learning_rate": 5.534151789310904e-05, "loss": 0.0138, "step": 14780 }, { "epoch": 15.65079365079365, "grad_norm": 0.15025092661380768, "learning_rate": 5.528671441085376e-05, "loss": 0.0139, "step": 14790 }, { "epoch": 15.661375661375661, "grad_norm": 0.12128642946481705, "learning_rate": 5.523190450474197e-05, "loss": 0.014, "step": 14800 }, { "epoch": 15.671957671957673, "grad_norm": 0.16446490585803986, "learning_rate": 5.5177088241372865e-05, "loss": 0.0122, "step": 14810 }, { "epoch": 15.682539682539682, "grad_norm": 0.13195759057998657, "learning_rate": 5.512226568735338e-05, "loss": 0.0152, "step": 14820 }, { "epoch": 15.693121693121693, "grad_norm": 0.1274782121181488, "learning_rate": 5.506743690929809e-05, "loss": 0.0131, "step": 14830 }, { "epoch": 15.703703703703704, "grad_norm": 0.14987578988075256, "learning_rate": 5.501260197382912e-05, "loss": 0.0146, "step": 14840 }, { "epoch": 15.714285714285714, "grad_norm": 0.09525519609451294, "learning_rate": 5.49577609475761e-05, "loss": 0.0125, "step": 14850 }, { "epoch": 15.724867724867725, "grad_norm": 0.11937165260314941, "learning_rate": 5.490291389717603e-05, "loss": 0.0111, "step": 14860 }, { "epoch": 15.735449735449736, "grad_norm": 0.11569222807884216, "learning_rate": 5.484806088927329e-05, "loss": 0.0128, "step": 14870 }, { "epoch": 15.746031746031747, "grad_norm": 0.20481474697589874, "learning_rate": 5.479320199051942e-05, "loss": 0.0128, "step": 14880 }, { "epoch": 15.756613756613756, "grad_norm": 0.20565685629844666, "learning_rate": 5.473833726757314e-05, "loss": 0.0113, "step": 14890 }, { "epoch": 15.767195767195767, "grad_norm": 0.08181557059288025, "learning_rate": 5.4683466787100304e-05, "loss": 0.0117, "step": 14900 }, { "epoch": 15.777777777777779, "grad_norm": 0.1824617087841034, "learning_rate": 5.462859061577369e-05, "loss": 0.0136, "step": 14910 }, { "epoch": 15.788359788359788, "grad_norm": 0.1542206108570099, "learning_rate": 5.4573708820273026e-05, "loss": 0.0138, "step": 14920 }, { "epoch": 15.798941798941799, "grad_norm": 0.15037013590335846, "learning_rate": 5.451882146728488e-05, "loss": 0.0133, "step": 14930 }, { "epoch": 15.80952380952381, "grad_norm": 0.1486506313085556, "learning_rate": 5.446392862350255e-05, "loss": 0.0138, "step": 14940 }, { "epoch": 15.82010582010582, "grad_norm": 0.18442335724830627, "learning_rate": 5.440903035562603e-05, "loss": 0.0133, "step": 14950 }, { "epoch": 15.83068783068783, "grad_norm": 0.14541484415531158, "learning_rate": 5.435412673036188e-05, "loss": 0.0121, "step": 14960 }, { "epoch": 15.841269841269842, "grad_norm": 0.16735409200191498, "learning_rate": 5.429921781442318e-05, "loss": 0.0133, "step": 14970 }, { "epoch": 15.851851851851851, "grad_norm": 0.1530480682849884, "learning_rate": 5.424430367452945e-05, "loss": 0.0114, "step": 14980 }, { "epoch": 15.862433862433862, "grad_norm": 0.11313065141439438, "learning_rate": 5.418938437740655e-05, "loss": 0.015, "step": 14990 }, { "epoch": 15.873015873015873, "grad_norm": 0.12056147307157516, "learning_rate": 5.413445998978658e-05, "loss": 0.0128, "step": 15000 }, { "epoch": 15.883597883597883, "grad_norm": 0.1433691680431366, "learning_rate": 5.407953057840789e-05, "loss": 0.017, "step": 15010 }, { "epoch": 15.894179894179894, "grad_norm": 0.11511925607919693, "learning_rate": 5.4024596210014853e-05, "loss": 0.0133, "step": 15020 }, { "epoch": 15.904761904761905, "grad_norm": 0.13236697018146515, "learning_rate": 5.3969656951357935e-05, "loss": 0.0121, "step": 15030 }, { "epoch": 15.915343915343914, "grad_norm": 0.14408859610557556, "learning_rate": 5.3914712869193504e-05, "loss": 0.014, "step": 15040 }, { "epoch": 15.925925925925926, "grad_norm": 0.15878137946128845, "learning_rate": 5.385976403028381e-05, "loss": 0.0133, "step": 15050 }, { "epoch": 15.936507936507937, "grad_norm": 0.19343966245651245, "learning_rate": 5.3804810501396875e-05, "loss": 0.0125, "step": 15060 }, { "epoch": 15.947089947089948, "grad_norm": 0.12353609502315521, "learning_rate": 5.374985234930639e-05, "loss": 0.0142, "step": 15070 }, { "epoch": 15.957671957671957, "grad_norm": 0.12138605862855911, "learning_rate": 5.3694889640791724e-05, "loss": 0.0161, "step": 15080 }, { "epoch": 15.968253968253968, "grad_norm": 0.13985870778560638, "learning_rate": 5.3639922442637736e-05, "loss": 0.0144, "step": 15090 }, { "epoch": 15.97883597883598, "grad_norm": 0.1042596697807312, "learning_rate": 5.358495082163476e-05, "loss": 0.0159, "step": 15100 }, { "epoch": 15.989417989417989, "grad_norm": 0.23243264853954315, "learning_rate": 5.35299748445785e-05, "loss": 0.0141, "step": 15110 }, { "epoch": 16.0, "grad_norm": 0.12005287408828735, "learning_rate": 5.3474994578269944e-05, "loss": 0.0131, "step": 15120 }, { "epoch": 16.01058201058201, "grad_norm": 0.2069077342748642, "learning_rate": 5.34200100895153e-05, "loss": 0.0127, "step": 15130 }, { "epoch": 16.021164021164022, "grad_norm": 0.1489700824022293, "learning_rate": 5.3365021445125916e-05, "loss": 0.0141, "step": 15140 }, { "epoch": 16.03174603174603, "grad_norm": 0.14887060225009918, "learning_rate": 5.331002871191817e-05, "loss": 0.0134, "step": 15150 }, { "epoch": 16.04232804232804, "grad_norm": 0.10854610800743103, "learning_rate": 5.3255031956713443e-05, "loss": 0.0134, "step": 15160 }, { "epoch": 16.052910052910054, "grad_norm": 0.11911124736070633, "learning_rate": 5.320003124633795e-05, "loss": 0.0138, "step": 15170 }, { "epoch": 16.063492063492063, "grad_norm": 0.23017702996730804, "learning_rate": 5.314502664762274e-05, "loss": 0.0134, "step": 15180 }, { "epoch": 16.074074074074073, "grad_norm": 0.2118794173002243, "learning_rate": 5.3090018227403604e-05, "loss": 0.0143, "step": 15190 }, { "epoch": 16.084656084656086, "grad_norm": 0.19661158323287964, "learning_rate": 5.3035006052520955e-05, "loss": 0.0144, "step": 15200 }, { "epoch": 16.095238095238095, "grad_norm": 0.17967084050178528, "learning_rate": 5.297999018981976e-05, "loss": 0.0121, "step": 15210 }, { "epoch": 16.105820105820104, "grad_norm": 0.15680159628391266, "learning_rate": 5.2924970706149505e-05, "loss": 0.0115, "step": 15220 }, { "epoch": 16.116402116402117, "grad_norm": 0.15728023648262024, "learning_rate": 5.2869947668364014e-05, "loss": 0.0116, "step": 15230 }, { "epoch": 16.126984126984127, "grad_norm": 0.14713099598884583, "learning_rate": 5.2814921143321504e-05, "loss": 0.0129, "step": 15240 }, { "epoch": 16.137566137566136, "grad_norm": 0.14196540415287018, "learning_rate": 5.275989119788436e-05, "loss": 0.0128, "step": 15250 }, { "epoch": 16.14814814814815, "grad_norm": 0.13068674504756927, "learning_rate": 5.270485789891919e-05, "loss": 0.0106, "step": 15260 }, { "epoch": 16.158730158730158, "grad_norm": 0.11639229208230972, "learning_rate": 5.264982131329661e-05, "loss": 0.0126, "step": 15270 }, { "epoch": 16.16931216931217, "grad_norm": 0.1115134209394455, "learning_rate": 5.259478150789128e-05, "loss": 0.0129, "step": 15280 }, { "epoch": 16.17989417989418, "grad_norm": 0.12103778123855591, "learning_rate": 5.253973854958173e-05, "loss": 0.0143, "step": 15290 }, { "epoch": 16.19047619047619, "grad_norm": 0.09492682665586472, "learning_rate": 5.2484692505250376e-05, "loss": 0.0151, "step": 15300 }, { "epoch": 16.201058201058203, "grad_norm": 0.11672979593276978, "learning_rate": 5.242964344178333e-05, "loss": 0.0154, "step": 15310 }, { "epoch": 16.211640211640212, "grad_norm": 0.08805330097675323, "learning_rate": 5.237459142607041e-05, "loss": 0.0101, "step": 15320 }, { "epoch": 16.22222222222222, "grad_norm": 0.13827240467071533, "learning_rate": 5.231953652500498e-05, "loss": 0.0107, "step": 15330 }, { "epoch": 16.232804232804234, "grad_norm": 0.2025768905878067, "learning_rate": 5.2264478805483976e-05, "loss": 0.0103, "step": 15340 }, { "epoch": 16.243386243386244, "grad_norm": 0.12457545846700668, "learning_rate": 5.220941833440768e-05, "loss": 0.0129, "step": 15350 }, { "epoch": 16.253968253968253, "grad_norm": 0.10426756739616394, "learning_rate": 5.215435517867977e-05, "loss": 0.0134, "step": 15360 }, { "epoch": 16.264550264550266, "grad_norm": 0.1090281680226326, "learning_rate": 5.209928940520719e-05, "loss": 0.0127, "step": 15370 }, { "epoch": 16.275132275132275, "grad_norm": 0.13949702680110931, "learning_rate": 5.204422108090004e-05, "loss": 0.0125, "step": 15380 }, { "epoch": 16.285714285714285, "grad_norm": 0.13257016241550446, "learning_rate": 5.19891502726715e-05, "loss": 0.0129, "step": 15390 }, { "epoch": 16.296296296296298, "grad_norm": 0.13893498480319977, "learning_rate": 5.193407704743782e-05, "loss": 0.0139, "step": 15400 }, { "epoch": 16.306878306878307, "grad_norm": 0.09524160623550415, "learning_rate": 5.187900147211815e-05, "loss": 0.0117, "step": 15410 }, { "epoch": 16.317460317460316, "grad_norm": 0.1115308627486229, "learning_rate": 5.182392361363453e-05, "loss": 0.0139, "step": 15420 }, { "epoch": 16.32804232804233, "grad_norm": 0.11841810494661331, "learning_rate": 5.176884353891172e-05, "loss": 0.0157, "step": 15430 }, { "epoch": 16.33862433862434, "grad_norm": 0.21566203236579895, "learning_rate": 5.171376131487722e-05, "loss": 0.0158, "step": 15440 }, { "epoch": 16.349206349206348, "grad_norm": 0.11472677439451218, "learning_rate": 5.165867700846113e-05, "loss": 0.0103, "step": 15450 }, { "epoch": 16.35978835978836, "grad_norm": 0.14214973151683807, "learning_rate": 5.160359068659606e-05, "loss": 0.0144, "step": 15460 }, { "epoch": 16.37037037037037, "grad_norm": 0.14412276446819305, "learning_rate": 5.154850241621712e-05, "loss": 0.0109, "step": 15470 }, { "epoch": 16.38095238095238, "grad_norm": 0.1790272444486618, "learning_rate": 5.149341226426172e-05, "loss": 0.0162, "step": 15480 }, { "epoch": 16.391534391534393, "grad_norm": 0.11517620831727982, "learning_rate": 5.14383202976696e-05, "loss": 0.0117, "step": 15490 }, { "epoch": 16.402116402116402, "grad_norm": 0.21316467225551605, "learning_rate": 5.1383226583382684e-05, "loss": 0.0176, "step": 15500 }, { "epoch": 16.41269841269841, "grad_norm": 0.19823101162910461, "learning_rate": 5.132813118834504e-05, "loss": 0.0131, "step": 15510 }, { "epoch": 16.423280423280424, "grad_norm": 0.14319349825382233, "learning_rate": 5.1273034179502777e-05, "loss": 0.0114, "step": 15520 }, { "epoch": 16.433862433862434, "grad_norm": 0.1323503702878952, "learning_rate": 5.1217935623803945e-05, "loss": 0.0139, "step": 15530 }, { "epoch": 16.444444444444443, "grad_norm": 0.12821529805660248, "learning_rate": 5.1162835588198476e-05, "loss": 0.0124, "step": 15540 }, { "epoch": 16.455026455026456, "grad_norm": 0.1400139033794403, "learning_rate": 5.110773413963813e-05, "loss": 0.0165, "step": 15550 }, { "epoch": 16.465608465608465, "grad_norm": 0.10050814598798752, "learning_rate": 5.105263134507636e-05, "loss": 0.0114, "step": 15560 }, { "epoch": 16.476190476190474, "grad_norm": 0.12286437302827835, "learning_rate": 5.099752727146824e-05, "loss": 0.0109, "step": 15570 }, { "epoch": 16.486772486772487, "grad_norm": 0.15674610435962677, "learning_rate": 5.094242198577042e-05, "loss": 0.0114, "step": 15580 }, { "epoch": 16.497354497354497, "grad_norm": 0.1369965523481369, "learning_rate": 5.088731555494102e-05, "loss": 0.0125, "step": 15590 }, { "epoch": 16.507936507936506, "grad_norm": 0.13084781169891357, "learning_rate": 5.0832208045939556e-05, "loss": 0.0109, "step": 15600 }, { "epoch": 16.51851851851852, "grad_norm": 0.13803225755691528, "learning_rate": 5.0777099525726844e-05, "loss": 0.0103, "step": 15610 }, { "epoch": 16.52910052910053, "grad_norm": 0.10484116524457932, "learning_rate": 5.072199006126494e-05, "loss": 0.0109, "step": 15620 }, { "epoch": 16.53968253968254, "grad_norm": 0.12603643536567688, "learning_rate": 5.0666879719517026e-05, "loss": 0.0124, "step": 15630 }, { "epoch": 16.55026455026455, "grad_norm": 0.12131494283676147, "learning_rate": 5.061176856744737e-05, "loss": 0.0113, "step": 15640 }, { "epoch": 16.56084656084656, "grad_norm": 0.12310589104890823, "learning_rate": 5.0556656672021205e-05, "loss": 0.0123, "step": 15650 }, { "epoch": 16.571428571428573, "grad_norm": 0.09095972031354904, "learning_rate": 5.050154410020472e-05, "loss": 0.0113, "step": 15660 }, { "epoch": 16.582010582010582, "grad_norm": 0.15033262968063354, "learning_rate": 5.044643091896485e-05, "loss": 0.0115, "step": 15670 }, { "epoch": 16.59259259259259, "grad_norm": 0.12995164096355438, "learning_rate": 5.039131719526932e-05, "loss": 0.0123, "step": 15680 }, { "epoch": 16.603174603174605, "grad_norm": 0.16051678359508514, "learning_rate": 5.03362029960865e-05, "loss": 0.0138, "step": 15690 }, { "epoch": 16.613756613756614, "grad_norm": 0.17235465347766876, "learning_rate": 5.028108838838533e-05, "loss": 0.014, "step": 15700 }, { "epoch": 16.624338624338623, "grad_norm": 0.19204668700695038, "learning_rate": 5.0225973439135276e-05, "loss": 0.0127, "step": 15710 }, { "epoch": 16.634920634920636, "grad_norm": 0.15669910609722137, "learning_rate": 5.017085821530617e-05, "loss": 0.0101, "step": 15720 }, { "epoch": 16.645502645502646, "grad_norm": 0.15236160159111023, "learning_rate": 5.011574278386822e-05, "loss": 0.0123, "step": 15730 }, { "epoch": 16.656084656084655, "grad_norm": 0.17700032889842987, "learning_rate": 5.006062721179189e-05, "loss": 0.0137, "step": 15740 }, { "epoch": 16.666666666666668, "grad_norm": 0.12598039209842682, "learning_rate": 5.000551156604777e-05, "loss": 0.0113, "step": 15750 }, { "epoch": 16.677248677248677, "grad_norm": 0.15742485225200653, "learning_rate": 4.9950395913606594e-05, "loss": 0.0137, "step": 15760 }, { "epoch": 16.687830687830687, "grad_norm": 0.15420593321323395, "learning_rate": 4.989528032143903e-05, "loss": 0.0121, "step": 15770 }, { "epoch": 16.6984126984127, "grad_norm": 0.1750151664018631, "learning_rate": 4.984016485651578e-05, "loss": 0.0145, "step": 15780 }, { "epoch": 16.70899470899471, "grad_norm": 0.182233527302742, "learning_rate": 4.9785049585807274e-05, "loss": 0.0139, "step": 15790 }, { "epoch": 16.719576719576718, "grad_norm": 0.16587334871292114, "learning_rate": 4.972993457628381e-05, "loss": 0.0132, "step": 15800 }, { "epoch": 16.73015873015873, "grad_norm": 0.1621350347995758, "learning_rate": 4.9674819894915306e-05, "loss": 0.0131, "step": 15810 }, { "epoch": 16.74074074074074, "grad_norm": 0.17688465118408203, "learning_rate": 4.9619705608671265e-05, "loss": 0.012, "step": 15820 }, { "epoch": 16.75132275132275, "grad_norm": 0.16085602343082428, "learning_rate": 4.956459178452079e-05, "loss": 0.0125, "step": 15830 }, { "epoch": 16.761904761904763, "grad_norm": 0.09965850412845612, "learning_rate": 4.950947848943235e-05, "loss": 0.0138, "step": 15840 }, { "epoch": 16.772486772486772, "grad_norm": 0.18176238238811493, "learning_rate": 4.9454365790373805e-05, "loss": 0.0141, "step": 15850 }, { "epoch": 16.78306878306878, "grad_norm": 0.11114046722650528, "learning_rate": 4.939925375431226e-05, "loss": 0.0109, "step": 15860 }, { "epoch": 16.793650793650794, "grad_norm": 0.18989549577236176, "learning_rate": 4.9344142448214046e-05, "loss": 0.0117, "step": 15870 }, { "epoch": 16.804232804232804, "grad_norm": 0.1346413642168045, "learning_rate": 4.928903193904461e-05, "loss": 0.0134, "step": 15880 }, { "epoch": 16.814814814814813, "grad_norm": 0.16107094287872314, "learning_rate": 4.923392229376842e-05, "loss": 0.0128, "step": 15890 }, { "epoch": 16.825396825396826, "grad_norm": 0.1393662691116333, "learning_rate": 4.9178813579348865e-05, "loss": 0.0126, "step": 15900 }, { "epoch": 16.835978835978835, "grad_norm": 0.16980494558811188, "learning_rate": 4.9123705862748253e-05, "loss": 0.0124, "step": 15910 }, { "epoch": 16.84656084656085, "grad_norm": 0.284553587436676, "learning_rate": 4.9068599210927627e-05, "loss": 0.0132, "step": 15920 }, { "epoch": 16.857142857142858, "grad_norm": 0.2642212510108948, "learning_rate": 4.901349369084681e-05, "loss": 0.0137, "step": 15930 }, { "epoch": 16.867724867724867, "grad_norm": 0.12066797912120819, "learning_rate": 4.8958389369464156e-05, "loss": 0.013, "step": 15940 }, { "epoch": 16.87830687830688, "grad_norm": 0.1565287709236145, "learning_rate": 4.890328631373666e-05, "loss": 0.013, "step": 15950 }, { "epoch": 16.88888888888889, "grad_norm": 0.15178310871124268, "learning_rate": 4.88481845906197e-05, "loss": 0.0105, "step": 15960 }, { "epoch": 16.8994708994709, "grad_norm": 0.16108259558677673, "learning_rate": 4.879308426706707e-05, "loss": 0.0135, "step": 15970 }, { "epoch": 16.91005291005291, "grad_norm": 0.16068682074546814, "learning_rate": 4.873798541003084e-05, "loss": 0.0135, "step": 15980 }, { "epoch": 16.92063492063492, "grad_norm": 0.09219949692487717, "learning_rate": 4.868288808646136e-05, "loss": 0.0112, "step": 15990 }, { "epoch": 16.93121693121693, "grad_norm": 0.14106321334838867, "learning_rate": 4.862779236330705e-05, "loss": 0.0119, "step": 16000 }, { "epoch": 16.941798941798943, "grad_norm": 0.13332432508468628, "learning_rate": 4.8572698307514395e-05, "loss": 0.0108, "step": 16010 }, { "epoch": 16.952380952380953, "grad_norm": 0.11475053429603577, "learning_rate": 4.85176059860279e-05, "loss": 0.0109, "step": 16020 }, { "epoch": 16.962962962962962, "grad_norm": 0.132324680685997, "learning_rate": 4.846251546578989e-05, "loss": 0.0106, "step": 16030 }, { "epoch": 16.973544973544975, "grad_norm": 0.22547008097171783, "learning_rate": 4.840742681374058e-05, "loss": 0.0127, "step": 16040 }, { "epoch": 16.984126984126984, "grad_norm": 0.1103038564324379, "learning_rate": 4.835234009681787e-05, "loss": 0.0132, "step": 16050 }, { "epoch": 16.994708994708994, "grad_norm": 0.12924544513225555, "learning_rate": 4.8297255381957296e-05, "loss": 0.013, "step": 16060 }, { "epoch": 17.005291005291006, "grad_norm": 0.24991540610790253, "learning_rate": 4.824217273609198e-05, "loss": 0.0118, "step": 16070 }, { "epoch": 17.015873015873016, "grad_norm": 0.10248234868049622, "learning_rate": 4.8187092226152555e-05, "loss": 0.0121, "step": 16080 }, { "epoch": 17.026455026455025, "grad_norm": 0.2319203019142151, "learning_rate": 4.8132013919067016e-05, "loss": 0.0149, "step": 16090 }, { "epoch": 17.037037037037038, "grad_norm": 0.16658233106136322, "learning_rate": 4.807693788176071e-05, "loss": 0.0134, "step": 16100 }, { "epoch": 17.047619047619047, "grad_norm": 0.07044043391942978, "learning_rate": 4.8021864181156214e-05, "loss": 0.0111, "step": 16110 }, { "epoch": 17.058201058201057, "grad_norm": 0.13738176226615906, "learning_rate": 4.796679288417326e-05, "loss": 0.0135, "step": 16120 }, { "epoch": 17.06878306878307, "grad_norm": 0.11629265546798706, "learning_rate": 4.791172405772866e-05, "loss": 0.0133, "step": 16130 }, { "epoch": 17.07936507936508, "grad_norm": 0.09280512481927872, "learning_rate": 4.785665776873625e-05, "loss": 0.0101, "step": 16140 }, { "epoch": 17.08994708994709, "grad_norm": 0.14393217861652374, "learning_rate": 4.7801594084106763e-05, "loss": 0.0135, "step": 16150 }, { "epoch": 17.1005291005291, "grad_norm": 0.1491629034280777, "learning_rate": 4.774653307074775e-05, "loss": 0.0114, "step": 16160 }, { "epoch": 17.11111111111111, "grad_norm": 0.15208977460861206, "learning_rate": 4.769147479556355e-05, "loss": 0.011, "step": 16170 }, { "epoch": 17.12169312169312, "grad_norm": 0.10190711170434952, "learning_rate": 4.763641932545515e-05, "loss": 0.0128, "step": 16180 }, { "epoch": 17.132275132275133, "grad_norm": 0.09540244191884995, "learning_rate": 4.758136672732013e-05, "loss": 0.0121, "step": 16190 }, { "epoch": 17.142857142857142, "grad_norm": 0.25540104508399963, "learning_rate": 4.752631706805261e-05, "loss": 0.0147, "step": 16200 }, { "epoch": 17.15343915343915, "grad_norm": 0.16341018676757812, "learning_rate": 4.7471270414543114e-05, "loss": 0.0123, "step": 16210 }, { "epoch": 17.164021164021165, "grad_norm": 0.11476954072713852, "learning_rate": 4.741622683367849e-05, "loss": 0.0115, "step": 16220 }, { "epoch": 17.174603174603174, "grad_norm": 0.19264909625053406, "learning_rate": 4.736118639234191e-05, "loss": 0.0133, "step": 16230 }, { "epoch": 17.185185185185187, "grad_norm": 0.34855011105537415, "learning_rate": 4.730614915741267e-05, "loss": 0.0119, "step": 16240 }, { "epoch": 17.195767195767196, "grad_norm": 0.16099834442138672, "learning_rate": 4.7251115195766234e-05, "loss": 0.014, "step": 16250 }, { "epoch": 17.206349206349206, "grad_norm": 0.11796627193689346, "learning_rate": 4.719608457427404e-05, "loss": 0.0119, "step": 16260 }, { "epoch": 17.21693121693122, "grad_norm": 0.16548281908035278, "learning_rate": 4.714105735980347e-05, "loss": 0.0129, "step": 16270 }, { "epoch": 17.227513227513228, "grad_norm": 0.13052500784397125, "learning_rate": 4.70860336192178e-05, "loss": 0.0117, "step": 16280 }, { "epoch": 17.238095238095237, "grad_norm": 0.16293227672576904, "learning_rate": 4.7031013419376034e-05, "loss": 0.0121, "step": 16290 }, { "epoch": 17.24867724867725, "grad_norm": 0.13072042167186737, "learning_rate": 4.6975996827132926e-05, "loss": 0.0137, "step": 16300 }, { "epoch": 17.25925925925926, "grad_norm": 0.08539000153541565, "learning_rate": 4.692098390933883e-05, "loss": 0.0104, "step": 16310 }, { "epoch": 17.26984126984127, "grad_norm": 0.10265880823135376, "learning_rate": 4.6865974732839615e-05, "loss": 0.0116, "step": 16320 }, { "epoch": 17.280423280423282, "grad_norm": 0.10935094207525253, "learning_rate": 4.6810969364476615e-05, "loss": 0.0136, "step": 16330 }, { "epoch": 17.29100529100529, "grad_norm": 0.15497995913028717, "learning_rate": 4.675596787108653e-05, "loss": 0.014, "step": 16340 }, { "epoch": 17.3015873015873, "grad_norm": 0.14884671568870544, "learning_rate": 4.670097031950139e-05, "loss": 0.0115, "step": 16350 }, { "epoch": 17.312169312169313, "grad_norm": 0.12306711077690125, "learning_rate": 4.664597677654838e-05, "loss": 0.0108, "step": 16360 }, { "epoch": 17.322751322751323, "grad_norm": 0.17426109313964844, "learning_rate": 4.659098730904986e-05, "loss": 0.0122, "step": 16370 }, { "epoch": 17.333333333333332, "grad_norm": 0.17204515635967255, "learning_rate": 4.6536001983823204e-05, "loss": 0.0134, "step": 16380 }, { "epoch": 17.343915343915345, "grad_norm": 0.1908477395772934, "learning_rate": 4.648102086768078e-05, "loss": 0.0143, "step": 16390 }, { "epoch": 17.354497354497354, "grad_norm": 0.11623547971248627, "learning_rate": 4.6426044027429786e-05, "loss": 0.0144, "step": 16400 }, { "epoch": 17.365079365079364, "grad_norm": 0.10874585062265396, "learning_rate": 4.637107152987234e-05, "loss": 0.0128, "step": 16410 }, { "epoch": 17.375661375661377, "grad_norm": 0.09987933188676834, "learning_rate": 4.631610344180515e-05, "loss": 0.0114, "step": 16420 }, { "epoch": 17.386243386243386, "grad_norm": 0.15605445206165314, "learning_rate": 4.626113983001965e-05, "loss": 0.0144, "step": 16430 }, { "epoch": 17.396825396825395, "grad_norm": 0.12376736849546432, "learning_rate": 4.620618076130182e-05, "loss": 0.0111, "step": 16440 }, { "epoch": 17.40740740740741, "grad_norm": 0.10166148841381073, "learning_rate": 4.6151226302432074e-05, "loss": 0.0134, "step": 16450 }, { "epoch": 17.417989417989418, "grad_norm": 0.14688198268413544, "learning_rate": 4.6096276520185304e-05, "loss": 0.012, "step": 16460 }, { "epoch": 17.428571428571427, "grad_norm": 0.15138088166713715, "learning_rate": 4.604133148133066e-05, "loss": 0.0092, "step": 16470 }, { "epoch": 17.43915343915344, "grad_norm": 0.142354816198349, "learning_rate": 4.598639125263155e-05, "loss": 0.0121, "step": 16480 }, { "epoch": 17.44973544973545, "grad_norm": 0.10321366041898727, "learning_rate": 4.593145590084553e-05, "loss": 0.0112, "step": 16490 }, { "epoch": 17.46031746031746, "grad_norm": 0.142492413520813, "learning_rate": 4.58765254927242e-05, "loss": 0.0108, "step": 16500 }, { "epoch": 17.47089947089947, "grad_norm": 0.16489724814891815, "learning_rate": 4.582160009501323e-05, "loss": 0.0104, "step": 16510 }, { "epoch": 17.48148148148148, "grad_norm": 0.08787307143211365, "learning_rate": 4.5766679774452143e-05, "loss": 0.0132, "step": 16520 }, { "epoch": 17.49206349206349, "grad_norm": 0.13492070138454437, "learning_rate": 4.571176459777431e-05, "loss": 0.012, "step": 16530 }, { "epoch": 17.502645502645503, "grad_norm": 0.1358214169740677, "learning_rate": 4.5656854631706846e-05, "loss": 0.0118, "step": 16540 }, { "epoch": 17.513227513227513, "grad_norm": 0.14134474098682404, "learning_rate": 4.560194994297054e-05, "loss": 0.0111, "step": 16550 }, { "epoch": 17.523809523809526, "grad_norm": 0.09904133528470993, "learning_rate": 4.554705059827974e-05, "loss": 0.0137, "step": 16560 }, { "epoch": 17.534391534391535, "grad_norm": 0.14029008150100708, "learning_rate": 4.5492156664342365e-05, "loss": 0.0127, "step": 16570 }, { "epoch": 17.544973544973544, "grad_norm": 0.11488819122314453, "learning_rate": 4.5437268207859695e-05, "loss": 0.0133, "step": 16580 }, { "epoch": 17.555555555555557, "grad_norm": 0.15699215233325958, "learning_rate": 4.538238529552641e-05, "loss": 0.0131, "step": 16590 }, { "epoch": 17.566137566137566, "grad_norm": 0.13457639515399933, "learning_rate": 4.5327507994030394e-05, "loss": 0.0122, "step": 16600 }, { "epoch": 17.576719576719576, "grad_norm": 0.11712028086185455, "learning_rate": 4.527263637005273e-05, "loss": 0.0121, "step": 16610 }, { "epoch": 17.58730158730159, "grad_norm": 0.13563096523284912, "learning_rate": 4.521777049026766e-05, "loss": 0.0119, "step": 16620 }, { "epoch": 17.597883597883598, "grad_norm": 0.14823375642299652, "learning_rate": 4.5162910421342386e-05, "loss": 0.0124, "step": 16630 }, { "epoch": 17.608465608465607, "grad_norm": 0.13292834162712097, "learning_rate": 4.510805622993706e-05, "loss": 0.0128, "step": 16640 }, { "epoch": 17.61904761904762, "grad_norm": 0.14627130329608917, "learning_rate": 4.505320798270467e-05, "loss": 0.0096, "step": 16650 }, { "epoch": 17.62962962962963, "grad_norm": 0.1317969411611557, "learning_rate": 4.499836574629105e-05, "loss": 0.0102, "step": 16660 }, { "epoch": 17.64021164021164, "grad_norm": 0.20883065462112427, "learning_rate": 4.494352958733465e-05, "loss": 0.0137, "step": 16670 }, { "epoch": 17.650793650793652, "grad_norm": 0.1528601050376892, "learning_rate": 4.488869957246663e-05, "loss": 0.0106, "step": 16680 }, { "epoch": 17.66137566137566, "grad_norm": 0.22863399982452393, "learning_rate": 4.4833875768310575e-05, "loss": 0.0133, "step": 16690 }, { "epoch": 17.67195767195767, "grad_norm": 0.1322150081396103, "learning_rate": 4.47790582414826e-05, "loss": 0.0125, "step": 16700 }, { "epoch": 17.682539682539684, "grad_norm": 0.21058320999145508, "learning_rate": 4.472424705859114e-05, "loss": 0.0096, "step": 16710 }, { "epoch": 17.693121693121693, "grad_norm": 0.10872579365968704, "learning_rate": 4.466944228623701e-05, "loss": 0.0094, "step": 16720 }, { "epoch": 17.703703703703702, "grad_norm": 0.15141329169273376, "learning_rate": 4.461464399101312e-05, "loss": 0.0098, "step": 16730 }, { "epoch": 17.714285714285715, "grad_norm": 0.12332247942686081, "learning_rate": 4.45598522395046e-05, "loss": 0.0114, "step": 16740 }, { "epoch": 17.724867724867725, "grad_norm": 0.14661310613155365, "learning_rate": 4.4505067098288577e-05, "loss": 0.015, "step": 16750 }, { "epoch": 17.735449735449734, "grad_norm": 0.11157266050577164, "learning_rate": 4.445028863393417e-05, "loss": 0.0092, "step": 16760 }, { "epoch": 17.746031746031747, "grad_norm": 0.12840837240219116, "learning_rate": 4.439551691300236e-05, "loss": 0.0102, "step": 16770 }, { "epoch": 17.756613756613756, "grad_norm": 0.1547306627035141, "learning_rate": 4.434075200204599e-05, "loss": 0.0116, "step": 16780 }, { "epoch": 17.767195767195766, "grad_norm": 0.10567419975996017, "learning_rate": 4.428599396760956e-05, "loss": 0.0089, "step": 16790 }, { "epoch": 17.77777777777778, "grad_norm": 0.15894128382205963, "learning_rate": 4.423124287622926e-05, "loss": 0.0124, "step": 16800 }, { "epoch": 17.788359788359788, "grad_norm": 0.10719239711761475, "learning_rate": 4.417649879443282e-05, "loss": 0.0132, "step": 16810 }, { "epoch": 17.798941798941797, "grad_norm": 0.11281711608171463, "learning_rate": 4.4121761788739444e-05, "loss": 0.0118, "step": 16820 }, { "epoch": 17.80952380952381, "grad_norm": 0.1199617087841034, "learning_rate": 4.4067031925659805e-05, "loss": 0.0145, "step": 16830 }, { "epoch": 17.82010582010582, "grad_norm": 0.1318206787109375, "learning_rate": 4.4012309271695816e-05, "loss": 0.0113, "step": 16840 }, { "epoch": 17.83068783068783, "grad_norm": 0.15757715702056885, "learning_rate": 4.395759389334067e-05, "loss": 0.0113, "step": 16850 }, { "epoch": 17.841269841269842, "grad_norm": 0.10120819509029388, "learning_rate": 4.3902885857078685e-05, "loss": 0.0123, "step": 16860 }, { "epoch": 17.85185185185185, "grad_norm": 0.0981786698102951, "learning_rate": 4.384818522938531e-05, "loss": 0.0118, "step": 16870 }, { "epoch": 17.862433862433864, "grad_norm": 0.1622476726770401, "learning_rate": 4.379349207672695e-05, "loss": 0.013, "step": 16880 }, { "epoch": 17.873015873015873, "grad_norm": 0.19655190408229828, "learning_rate": 4.373880646556098e-05, "loss": 0.0134, "step": 16890 }, { "epoch": 17.883597883597883, "grad_norm": 0.11728501319885254, "learning_rate": 4.3684128462335533e-05, "loss": 0.0112, "step": 16900 }, { "epoch": 17.894179894179896, "grad_norm": 0.1142304316163063, "learning_rate": 4.362945813348955e-05, "loss": 0.0127, "step": 16910 }, { "epoch": 17.904761904761905, "grad_norm": 0.11513261497020721, "learning_rate": 4.357479554545262e-05, "loss": 0.0108, "step": 16920 }, { "epoch": 17.915343915343914, "grad_norm": 0.10578437894582748, "learning_rate": 4.352014076464499e-05, "loss": 0.0106, "step": 16930 }, { "epoch": 17.925925925925927, "grad_norm": 0.0868181586265564, "learning_rate": 4.346549385747734e-05, "loss": 0.0111, "step": 16940 }, { "epoch": 17.936507936507937, "grad_norm": 0.15072223544120789, "learning_rate": 4.34108548903508e-05, "loss": 0.0114, "step": 16950 }, { "epoch": 17.947089947089946, "grad_norm": 0.07590598613023758, "learning_rate": 4.335622392965689e-05, "loss": 0.0102, "step": 16960 }, { "epoch": 17.95767195767196, "grad_norm": 0.07557179778814316, "learning_rate": 4.330160104177738e-05, "loss": 0.0104, "step": 16970 }, { "epoch": 17.96825396825397, "grad_norm": 0.1088862270116806, "learning_rate": 4.32469862930842e-05, "loss": 0.0105, "step": 16980 }, { "epoch": 17.978835978835978, "grad_norm": 0.11173544079065323, "learning_rate": 4.319237974993946e-05, "loss": 0.0132, "step": 16990 }, { "epoch": 17.98941798941799, "grad_norm": 0.15895436704158783, "learning_rate": 4.3137781478695236e-05, "loss": 0.0127, "step": 17000 }, { "epoch": 18.0, "grad_norm": 0.241832435131073, "learning_rate": 4.308319154569358e-05, "loss": 0.0143, "step": 17010 }, { "epoch": 18.01058201058201, "grad_norm": 0.13925471901893616, "learning_rate": 4.302861001726641e-05, "loss": 0.0115, "step": 17020 }, { "epoch": 18.021164021164022, "grad_norm": 0.10167950391769409, "learning_rate": 4.2974036959735415e-05, "loss": 0.0121, "step": 17030 }, { "epoch": 18.03174603174603, "grad_norm": 0.09198401868343353, "learning_rate": 4.2919472439412034e-05, "loss": 0.0144, "step": 17040 }, { "epoch": 18.04232804232804, "grad_norm": 0.17663513123989105, "learning_rate": 4.286491652259729e-05, "loss": 0.0128, "step": 17050 }, { "epoch": 18.052910052910054, "grad_norm": 0.14790396392345428, "learning_rate": 4.281036927558178e-05, "loss": 0.0143, "step": 17060 }, { "epoch": 18.063492063492063, "grad_norm": 0.11602305620908737, "learning_rate": 4.275583076464552e-05, "loss": 0.0114, "step": 17070 }, { "epoch": 18.074074074074073, "grad_norm": 0.12833355367183685, "learning_rate": 4.270130105605794e-05, "loss": 0.0118, "step": 17080 }, { "epoch": 18.084656084656086, "grad_norm": 0.10998509079217911, "learning_rate": 4.264678021607782e-05, "loss": 0.0128, "step": 17090 }, { "epoch": 18.095238095238095, "grad_norm": 0.16495171189308167, "learning_rate": 4.25922683109531e-05, "loss": 0.0111, "step": 17100 }, { "epoch": 18.105820105820104, "grad_norm": 0.13065031170845032, "learning_rate": 4.25377654069209e-05, "loss": 0.0138, "step": 17110 }, { "epoch": 18.116402116402117, "grad_norm": 0.11184810847043991, "learning_rate": 4.248327157020737e-05, "loss": 0.0101, "step": 17120 }, { "epoch": 18.126984126984127, "grad_norm": 0.1155150905251503, "learning_rate": 4.242878686702763e-05, "loss": 0.0145, "step": 17130 }, { "epoch": 18.137566137566136, "grad_norm": 0.13337776064872742, "learning_rate": 4.23743113635858e-05, "loss": 0.0118, "step": 17140 }, { "epoch": 18.14814814814815, "grad_norm": 0.09556908160448074, "learning_rate": 4.2319845126074705e-05, "loss": 0.0108, "step": 17150 }, { "epoch": 18.158730158730158, "grad_norm": 0.1377558559179306, "learning_rate": 4.226538822067598e-05, "loss": 0.0121, "step": 17160 }, { "epoch": 18.16931216931217, "grad_norm": 0.187893807888031, "learning_rate": 4.2210940713559896e-05, "loss": 0.0155, "step": 17170 }, { "epoch": 18.17989417989418, "grad_norm": 0.16292668879032135, "learning_rate": 4.21565026708853e-05, "loss": 0.0116, "step": 17180 }, { "epoch": 18.19047619047619, "grad_norm": 0.30474749207496643, "learning_rate": 4.2102074158799535e-05, "loss": 0.0134, "step": 17190 }, { "epoch": 18.201058201058203, "grad_norm": 0.12391939759254456, "learning_rate": 4.204765524343841e-05, "loss": 0.0125, "step": 17200 }, { "epoch": 18.211640211640212, "grad_norm": 0.16122379899024963, "learning_rate": 4.1993245990926036e-05, "loss": 0.0131, "step": 17210 }, { "epoch": 18.22222222222222, "grad_norm": 0.10511612147092819, "learning_rate": 4.193884646737474e-05, "loss": 0.0132, "step": 17220 }, { "epoch": 18.232804232804234, "grad_norm": 0.14097073674201965, "learning_rate": 4.1884456738885124e-05, "loss": 0.0109, "step": 17230 }, { "epoch": 18.243386243386244, "grad_norm": 0.11918596178293228, "learning_rate": 4.1830076871545795e-05, "loss": 0.0134, "step": 17240 }, { "epoch": 18.253968253968253, "grad_norm": 0.187265083193779, "learning_rate": 4.177570693143346e-05, "loss": 0.0118, "step": 17250 }, { "epoch": 18.264550264550266, "grad_norm": 0.15288445353507996, "learning_rate": 4.172134698461271e-05, "loss": 0.0117, "step": 17260 }, { "epoch": 18.275132275132275, "grad_norm": 0.1368255615234375, "learning_rate": 4.166699709713599e-05, "loss": 0.0123, "step": 17270 }, { "epoch": 18.285714285714285, "grad_norm": 0.12482428550720215, "learning_rate": 4.1612657335043555e-05, "loss": 0.0099, "step": 17280 }, { "epoch": 18.296296296296298, "grad_norm": 0.13972212374210358, "learning_rate": 4.155832776436331e-05, "loss": 0.009, "step": 17290 }, { "epoch": 18.306878306878307, "grad_norm": 0.24268753826618195, "learning_rate": 4.1504008451110845e-05, "loss": 0.0131, "step": 17300 }, { "epoch": 18.317460317460316, "grad_norm": 0.17904719710350037, "learning_rate": 4.1449699461289225e-05, "loss": 0.0119, "step": 17310 }, { "epoch": 18.32804232804233, "grad_norm": 0.168814554810524, "learning_rate": 4.139540086088901e-05, "loss": 0.0141, "step": 17320 }, { "epoch": 18.33862433862434, "grad_norm": 0.10924948006868362, "learning_rate": 4.1341112715888106e-05, "loss": 0.0128, "step": 17330 }, { "epoch": 18.349206349206348, "grad_norm": 0.15132278203964233, "learning_rate": 4.128683509225172e-05, "loss": 0.0095, "step": 17340 }, { "epoch": 18.35978835978836, "grad_norm": 0.16796784102916718, "learning_rate": 4.123256805593231e-05, "loss": 0.0112, "step": 17350 }, { "epoch": 18.37037037037037, "grad_norm": 0.12661173939704895, "learning_rate": 4.117831167286943e-05, "loss": 0.0116, "step": 17360 }, { "epoch": 18.38095238095238, "grad_norm": 0.12492911517620087, "learning_rate": 4.1124066008989685e-05, "loss": 0.0121, "step": 17370 }, { "epoch": 18.391534391534393, "grad_norm": 0.1077297031879425, "learning_rate": 4.106983113020669e-05, "loss": 0.0108, "step": 17380 }, { "epoch": 18.402116402116402, "grad_norm": 0.11634361743927002, "learning_rate": 4.101560710242094e-05, "loss": 0.0112, "step": 17390 }, { "epoch": 18.41269841269841, "grad_norm": 0.1905323714017868, "learning_rate": 4.096139399151971e-05, "loss": 0.0121, "step": 17400 }, { "epoch": 18.423280423280424, "grad_norm": 0.1585444211959839, "learning_rate": 4.090719186337709e-05, "loss": 0.0125, "step": 17410 }, { "epoch": 18.433862433862434, "grad_norm": 0.10167360305786133, "learning_rate": 4.0853000783853746e-05, "loss": 0.0099, "step": 17420 }, { "epoch": 18.444444444444443, "grad_norm": 0.11832509934902191, "learning_rate": 4.0798820818796956e-05, "loss": 0.0104, "step": 17430 }, { "epoch": 18.455026455026456, "grad_norm": 0.13395990431308746, "learning_rate": 4.074465203404048e-05, "loss": 0.0111, "step": 17440 }, { "epoch": 18.465608465608465, "grad_norm": 0.10107868164777756, "learning_rate": 4.0690494495404495e-05, "loss": 0.011, "step": 17450 }, { "epoch": 18.476190476190474, "grad_norm": 0.15975722670555115, "learning_rate": 4.0636348268695535e-05, "loss": 0.0103, "step": 17460 }, { "epoch": 18.486772486772487, "grad_norm": 0.09330674260854721, "learning_rate": 4.058221341970637e-05, "loss": 0.0112, "step": 17470 }, { "epoch": 18.497354497354497, "grad_norm": 0.14239293336868286, "learning_rate": 4.0528090014215945e-05, "loss": 0.0088, "step": 17480 }, { "epoch": 18.507936507936506, "grad_norm": 0.16151666641235352, "learning_rate": 4.047397811798929e-05, "loss": 0.0122, "step": 17490 }, { "epoch": 18.51851851851852, "grad_norm": 0.1272761970758438, "learning_rate": 4.041987779677745e-05, "loss": 0.0136, "step": 17500 }, { "epoch": 18.52910052910053, "grad_norm": 0.11411380022764206, "learning_rate": 4.036578911631745e-05, "loss": 0.0123, "step": 17510 }, { "epoch": 18.53968253968254, "grad_norm": 0.15396955609321594, "learning_rate": 4.031171214233211e-05, "loss": 0.0134, "step": 17520 }, { "epoch": 18.55026455026455, "grad_norm": 0.11523213982582092, "learning_rate": 4.025764694053008e-05, "loss": 0.0128, "step": 17530 }, { "epoch": 18.56084656084656, "grad_norm": 0.36107462644577026, "learning_rate": 4.020359357660566e-05, "loss": 0.0115, "step": 17540 }, { "epoch": 18.571428571428573, "grad_norm": 0.15083995461463928, "learning_rate": 4.014955211623875e-05, "loss": 0.0101, "step": 17550 }, { "epoch": 18.582010582010582, "grad_norm": 0.13413527607917786, "learning_rate": 4.0095522625094895e-05, "loss": 0.0142, "step": 17560 }, { "epoch": 18.59259259259259, "grad_norm": 0.14417193830013275, "learning_rate": 4.004150516882497e-05, "loss": 0.0132, "step": 17570 }, { "epoch": 18.603174603174605, "grad_norm": 0.1219739094376564, "learning_rate": 3.99874998130653e-05, "loss": 0.012, "step": 17580 }, { "epoch": 18.613756613756614, "grad_norm": 0.11161097884178162, "learning_rate": 3.9933506623437455e-05, "loss": 0.0146, "step": 17590 }, { "epoch": 18.624338624338623, "grad_norm": 0.12853464484214783, "learning_rate": 3.9879525665548276e-05, "loss": 0.0127, "step": 17600 }, { "epoch": 18.634920634920636, "grad_norm": 0.19022315740585327, "learning_rate": 3.9825557004989704e-05, "loss": 0.0112, "step": 17610 }, { "epoch": 18.645502645502646, "grad_norm": 0.08590932935476303, "learning_rate": 3.977160070733877e-05, "loss": 0.0095, "step": 17620 }, { "epoch": 18.656084656084655, "grad_norm": 0.12077876925468445, "learning_rate": 3.971765683815746e-05, "loss": 0.0114, "step": 17630 }, { "epoch": 18.666666666666668, "grad_norm": 0.14142611622810364, "learning_rate": 3.966372546299265e-05, "loss": 0.0117, "step": 17640 }, { "epoch": 18.677248677248677, "grad_norm": 0.10499543696641922, "learning_rate": 3.960980664737604e-05, "loss": 0.013, "step": 17650 }, { "epoch": 18.687830687830687, "grad_norm": 0.08761482685804367, "learning_rate": 3.955590045682408e-05, "loss": 0.0132, "step": 17660 }, { "epoch": 18.6984126984127, "grad_norm": 0.134656623005867, "learning_rate": 3.950200695683788e-05, "loss": 0.0099, "step": 17670 }, { "epoch": 18.70899470899471, "grad_norm": 0.09696627408266068, "learning_rate": 3.944812621290314e-05, "loss": 0.0088, "step": 17680 }, { "epoch": 18.719576719576718, "grad_norm": 0.07050564140081406, "learning_rate": 3.939425829049002e-05, "loss": 0.0137, "step": 17690 }, { "epoch": 18.73015873015873, "grad_norm": 0.07239237427711487, "learning_rate": 3.9340403255053135e-05, "loss": 0.0105, "step": 17700 }, { "epoch": 18.74074074074074, "grad_norm": 0.11770341545343399, "learning_rate": 3.9286561172031407e-05, "loss": 0.0104, "step": 17710 }, { "epoch": 18.75132275132275, "grad_norm": 0.12790606915950775, "learning_rate": 3.923273210684809e-05, "loss": 0.0113, "step": 17720 }, { "epoch": 18.761904761904763, "grad_norm": 0.11828199028968811, "learning_rate": 3.9178916124910555e-05, "loss": 0.0092, "step": 17730 }, { "epoch": 18.772486772486772, "grad_norm": 0.09813781082630157, "learning_rate": 3.9125113291610274e-05, "loss": 0.0103, "step": 17740 }, { "epoch": 18.78306878306878, "grad_norm": 0.16003599762916565, "learning_rate": 3.907132367232278e-05, "loss": 0.0123, "step": 17750 }, { "epoch": 18.793650793650794, "grad_norm": 0.10179164260625839, "learning_rate": 3.901754733240752e-05, "loss": 0.0099, "step": 17760 }, { "epoch": 18.804232804232804, "grad_norm": 0.14451727271080017, "learning_rate": 3.896378433720785e-05, "loss": 0.0119, "step": 17770 }, { "epoch": 18.814814814814813, "grad_norm": 0.1568106710910797, "learning_rate": 3.891003475205086e-05, "loss": 0.0135, "step": 17780 }, { "epoch": 18.825396825396826, "grad_norm": 0.06209408864378929, "learning_rate": 3.885629864224736e-05, "loss": 0.0095, "step": 17790 }, { "epoch": 18.835978835978835, "grad_norm": 0.10210586339235306, "learning_rate": 3.880257607309178e-05, "loss": 0.0086, "step": 17800 }, { "epoch": 18.84656084656085, "grad_norm": 0.10427426546812057, "learning_rate": 3.874886710986213e-05, "loss": 0.0143, "step": 17810 }, { "epoch": 18.857142857142858, "grad_norm": 0.13116273283958435, "learning_rate": 3.8695171817819834e-05, "loss": 0.012, "step": 17820 }, { "epoch": 18.867724867724867, "grad_norm": 0.13553255796432495, "learning_rate": 3.8641490262209774e-05, "loss": 0.0126, "step": 17830 }, { "epoch": 18.87830687830688, "grad_norm": 0.10491863638162613, "learning_rate": 3.8587822508260084e-05, "loss": 0.013, "step": 17840 }, { "epoch": 18.88888888888889, "grad_norm": 0.12552405893802643, "learning_rate": 3.853416862118214e-05, "loss": 0.0111, "step": 17850 }, { "epoch": 18.8994708994709, "grad_norm": 0.17929816246032715, "learning_rate": 3.848052866617049e-05, "loss": 0.0104, "step": 17860 }, { "epoch": 18.91005291005291, "grad_norm": 0.20010153949260712, "learning_rate": 3.84269027084027e-05, "loss": 0.0126, "step": 17870 }, { "epoch": 18.92063492063492, "grad_norm": 0.14885267615318298, "learning_rate": 3.83732908130394e-05, "loss": 0.0114, "step": 17880 }, { "epoch": 18.93121693121693, "grad_norm": 0.10470334440469742, "learning_rate": 3.831969304522409e-05, "loss": 0.0117, "step": 17890 }, { "epoch": 18.941798941798943, "grad_norm": 0.10660981386899948, "learning_rate": 3.826610947008312e-05, "loss": 0.0092, "step": 17900 }, { "epoch": 18.952380952380953, "grad_norm": 0.11861076951026917, "learning_rate": 3.821254015272559e-05, "loss": 0.0105, "step": 17910 }, { "epoch": 18.962962962962962, "grad_norm": 0.14086972177028656, "learning_rate": 3.8158985158243214e-05, "loss": 0.0085, "step": 17920 }, { "epoch": 18.973544973544975, "grad_norm": 0.15626157820224762, "learning_rate": 3.810544455171044e-05, "loss": 0.0091, "step": 17930 }, { "epoch": 18.984126984126984, "grad_norm": 0.11125937849283218, "learning_rate": 3.8051918398184114e-05, "loss": 0.0097, "step": 17940 }, { "epoch": 18.994708994708994, "grad_norm": 0.16443435847759247, "learning_rate": 3.799840676270356e-05, "loss": 0.0086, "step": 17950 }, { "epoch": 19.005291005291006, "grad_norm": 0.10707315802574158, "learning_rate": 3.794490971029048e-05, "loss": 0.0114, "step": 17960 }, { "epoch": 19.015873015873016, "grad_norm": 0.11608882248401642, "learning_rate": 3.789142730594881e-05, "loss": 0.0089, "step": 17970 }, { "epoch": 19.026455026455025, "grad_norm": 0.17285925149917603, "learning_rate": 3.783795961466471e-05, "loss": 0.0119, "step": 17980 }, { "epoch": 19.037037037037038, "grad_norm": 0.12061625719070435, "learning_rate": 3.778450670140651e-05, "loss": 0.0098, "step": 17990 }, { "epoch": 19.047619047619047, "grad_norm": 0.12696552276611328, "learning_rate": 3.773106863112451e-05, "loss": 0.0107, "step": 18000 }, { "epoch": 19.058201058201057, "grad_norm": 0.11412665992975235, "learning_rate": 3.7677645468751e-05, "loss": 0.0102, "step": 18010 }, { "epoch": 19.06878306878307, "grad_norm": 0.18210124969482422, "learning_rate": 3.762423727920018e-05, "loss": 0.0126, "step": 18020 }, { "epoch": 19.07936507936508, "grad_norm": 0.13909119367599487, "learning_rate": 3.7570844127367995e-05, "loss": 0.0089, "step": 18030 }, { "epoch": 19.08994708994709, "grad_norm": 0.16318362951278687, "learning_rate": 3.751746607813221e-05, "loss": 0.0123, "step": 18040 }, { "epoch": 19.1005291005291, "grad_norm": 0.26445385813713074, "learning_rate": 3.746410319635217e-05, "loss": 0.0117, "step": 18050 }, { "epoch": 19.11111111111111, "grad_norm": 0.18694403767585754, "learning_rate": 3.7410755546868806e-05, "loss": 0.0116, "step": 18060 }, { "epoch": 19.12169312169312, "grad_norm": 0.1151684820652008, "learning_rate": 3.735742319450454e-05, "loss": 0.0121, "step": 18070 }, { "epoch": 19.132275132275133, "grad_norm": 0.13094723224639893, "learning_rate": 3.7304106204063184e-05, "loss": 0.0113, "step": 18080 }, { "epoch": 19.142857142857142, "grad_norm": 0.11046901345252991, "learning_rate": 3.725080464032996e-05, "loss": 0.0092, "step": 18090 }, { "epoch": 19.15343915343915, "grad_norm": 0.13211804628372192, "learning_rate": 3.719751856807125e-05, "loss": 0.0108, "step": 18100 }, { "epoch": 19.164021164021165, "grad_norm": 0.12612521648406982, "learning_rate": 3.71442480520347e-05, "loss": 0.0106, "step": 18110 }, { "epoch": 19.174603174603174, "grad_norm": 0.1444542557001114, "learning_rate": 3.709099315694897e-05, "loss": 0.0113, "step": 18120 }, { "epoch": 19.185185185185187, "grad_norm": 0.11307837069034576, "learning_rate": 3.703775394752378e-05, "loss": 0.0094, "step": 18130 }, { "epoch": 19.195767195767196, "grad_norm": 0.07534153759479523, "learning_rate": 3.698453048844983e-05, "loss": 0.0113, "step": 18140 }, { "epoch": 19.206349206349206, "grad_norm": 0.12158763408660889, "learning_rate": 3.693132284439861e-05, "loss": 0.0106, "step": 18150 }, { "epoch": 19.21693121693122, "grad_norm": 0.10051573812961578, "learning_rate": 3.6878131080022414e-05, "loss": 0.01, "step": 18160 }, { "epoch": 19.227513227513228, "grad_norm": 0.10679542273283005, "learning_rate": 3.682495525995429e-05, "loss": 0.0138, "step": 18170 }, { "epoch": 19.238095238095237, "grad_norm": 0.09660738706588745, "learning_rate": 3.6771795448807846e-05, "loss": 0.0098, "step": 18180 }, { "epoch": 19.24867724867725, "grad_norm": 0.08982638269662857, "learning_rate": 3.671865171117724e-05, "loss": 0.0104, "step": 18190 }, { "epoch": 19.25925925925926, "grad_norm": 0.14974597096443176, "learning_rate": 3.666552411163718e-05, "loss": 0.011, "step": 18200 }, { "epoch": 19.26984126984127, "grad_norm": 0.11013387143611908, "learning_rate": 3.661241271474269e-05, "loss": 0.01, "step": 18210 }, { "epoch": 19.280423280423282, "grad_norm": 0.18227191269397736, "learning_rate": 3.6559317585029116e-05, "loss": 0.0126, "step": 18220 }, { "epoch": 19.29100529100529, "grad_norm": 0.14674632251262665, "learning_rate": 3.650623878701204e-05, "loss": 0.01, "step": 18230 }, { "epoch": 19.3015873015873, "grad_norm": 0.15215007960796356, "learning_rate": 3.645317638518721e-05, "loss": 0.0092, "step": 18240 }, { "epoch": 19.312169312169313, "grad_norm": 0.09930659085512161, "learning_rate": 3.640013044403046e-05, "loss": 0.013, "step": 18250 }, { "epoch": 19.322751322751323, "grad_norm": 0.09409113973379135, "learning_rate": 3.634710102799761e-05, "loss": 0.009, "step": 18260 }, { "epoch": 19.333333333333332, "grad_norm": 0.16964192688465118, "learning_rate": 3.6294088201524394e-05, "loss": 0.0118, "step": 18270 }, { "epoch": 19.343915343915345, "grad_norm": 0.10171884298324585, "learning_rate": 3.62410920290264e-05, "loss": 0.0113, "step": 18280 }, { "epoch": 19.354497354497354, "grad_norm": 0.12822724878787994, "learning_rate": 3.6188112574898954e-05, "loss": 0.0097, "step": 18290 }, { "epoch": 19.365079365079364, "grad_norm": 0.12050475925207138, "learning_rate": 3.6135149903517115e-05, "loss": 0.0111, "step": 18300 }, { "epoch": 19.375661375661377, "grad_norm": 0.0957622304558754, "learning_rate": 3.608220407923552e-05, "loss": 0.012, "step": 18310 }, { "epoch": 19.386243386243386, "grad_norm": 0.18321381509304047, "learning_rate": 3.602927516638833e-05, "loss": 0.0123, "step": 18320 }, { "epoch": 19.396825396825395, "grad_norm": 0.15325745940208435, "learning_rate": 3.5976363229289165e-05, "loss": 0.0106, "step": 18330 }, { "epoch": 19.40740740740741, "grad_norm": 0.15346136689186096, "learning_rate": 3.5923468332231e-05, "loss": 0.01, "step": 18340 }, { "epoch": 19.417989417989418, "grad_norm": 0.10610250383615494, "learning_rate": 3.587059053948616e-05, "loss": 0.0096, "step": 18350 }, { "epoch": 19.428571428571427, "grad_norm": 0.12737813591957092, "learning_rate": 3.5817729915306134e-05, "loss": 0.0112, "step": 18360 }, { "epoch": 19.43915343915344, "grad_norm": 0.18025659024715424, "learning_rate": 3.5764886523921566e-05, "loss": 0.01, "step": 18370 }, { "epoch": 19.44973544973545, "grad_norm": 0.14126789569854736, "learning_rate": 3.571206042954214e-05, "loss": 0.0145, "step": 18380 }, { "epoch": 19.46031746031746, "grad_norm": 0.13601629436016083, "learning_rate": 3.5659251696356566e-05, "loss": 0.0105, "step": 18390 }, { "epoch": 19.47089947089947, "grad_norm": 0.12209698557853699, "learning_rate": 3.56064603885324e-05, "loss": 0.0118, "step": 18400 }, { "epoch": 19.48148148148148, "grad_norm": 0.10047946870326996, "learning_rate": 3.5553686570216115e-05, "loss": 0.0121, "step": 18410 }, { "epoch": 19.49206349206349, "grad_norm": 0.15315504372119904, "learning_rate": 3.550093030553284e-05, "loss": 0.0133, "step": 18420 }, { "epoch": 19.502645502645503, "grad_norm": 0.11987663805484772, "learning_rate": 3.544819165858642e-05, "loss": 0.0117, "step": 18430 }, { "epoch": 19.513227513227513, "grad_norm": 0.1650518923997879, "learning_rate": 3.539547069345926e-05, "loss": 0.0106, "step": 18440 }, { "epoch": 19.523809523809526, "grad_norm": 0.09210915118455887, "learning_rate": 3.5342767474212344e-05, "loss": 0.0108, "step": 18450 }, { "epoch": 19.534391534391535, "grad_norm": 0.16526804864406586, "learning_rate": 3.529008206488502e-05, "loss": 0.0104, "step": 18460 }, { "epoch": 19.544973544973544, "grad_norm": 0.09174142777919769, "learning_rate": 3.5237414529495055e-05, "loss": 0.0119, "step": 18470 }, { "epoch": 19.555555555555557, "grad_norm": 0.0938543826341629, "learning_rate": 3.5184764932038454e-05, "loss": 0.009, "step": 18480 }, { "epoch": 19.566137566137566, "grad_norm": 0.1532650887966156, "learning_rate": 3.513213333648945e-05, "loss": 0.0117, "step": 18490 }, { "epoch": 19.576719576719576, "grad_norm": 0.1503310650587082, "learning_rate": 3.5079519806800374e-05, "loss": 0.011, "step": 18500 }, { "epoch": 19.58730158730159, "grad_norm": 0.08003376424312592, "learning_rate": 3.502692440690165e-05, "loss": 0.0114, "step": 18510 }, { "epoch": 19.597883597883598, "grad_norm": 0.14804178476333618, "learning_rate": 3.497434720070165e-05, "loss": 0.011, "step": 18520 }, { "epoch": 19.608465608465607, "grad_norm": 0.11549719423055649, "learning_rate": 3.4921788252086616e-05, "loss": 0.0106, "step": 18530 }, { "epoch": 19.61904761904762, "grad_norm": 0.10830007493495941, "learning_rate": 3.486924762492065e-05, "loss": 0.0101, "step": 18540 }, { "epoch": 19.62962962962963, "grad_norm": 0.12377385795116425, "learning_rate": 3.4816725383045534e-05, "loss": 0.0102, "step": 18550 }, { "epoch": 19.64021164021164, "grad_norm": 0.11085963994264603, "learning_rate": 3.476422159028079e-05, "loss": 0.0114, "step": 18560 }, { "epoch": 19.650793650793652, "grad_norm": 0.11240428686141968, "learning_rate": 3.471173631042345e-05, "loss": 0.0106, "step": 18570 }, { "epoch": 19.66137566137566, "grad_norm": 0.15735560655593872, "learning_rate": 3.465926960724808e-05, "loss": 0.0127, "step": 18580 }, { "epoch": 19.67195767195767, "grad_norm": 0.14589513838291168, "learning_rate": 3.460682154450666e-05, "loss": 0.0101, "step": 18590 }, { "epoch": 19.682539682539684, "grad_norm": 0.11781347543001175, "learning_rate": 3.4554392185928564e-05, "loss": 0.011, "step": 18600 }, { "epoch": 19.693121693121693, "grad_norm": 0.10659408569335938, "learning_rate": 3.450198159522037e-05, "loss": 0.0101, "step": 18610 }, { "epoch": 19.703703703703702, "grad_norm": 0.09280318766832352, "learning_rate": 3.444958983606592e-05, "loss": 0.011, "step": 18620 }, { "epoch": 19.714285714285715, "grad_norm": 0.1497066468000412, "learning_rate": 3.439721697212612e-05, "loss": 0.011, "step": 18630 }, { "epoch": 19.724867724867725, "grad_norm": 0.08597881346940994, "learning_rate": 3.4344863067038954e-05, "loss": 0.0111, "step": 18640 }, { "epoch": 19.735449735449734, "grad_norm": 0.10138877481222153, "learning_rate": 3.429252818441935e-05, "loss": 0.0071, "step": 18650 }, { "epoch": 19.746031746031747, "grad_norm": 0.10505537688732147, "learning_rate": 3.4240212387859094e-05, "loss": 0.0101, "step": 18660 }, { "epoch": 19.756613756613756, "grad_norm": 0.08446280658245087, "learning_rate": 3.418791574092686e-05, "loss": 0.0076, "step": 18670 }, { "epoch": 19.767195767195766, "grad_norm": 0.11867457628250122, "learning_rate": 3.413563830716796e-05, "loss": 0.0106, "step": 18680 }, { "epoch": 19.77777777777778, "grad_norm": 0.12245366722345352, "learning_rate": 3.408338015010445e-05, "loss": 0.009, "step": 18690 }, { "epoch": 19.788359788359788, "grad_norm": 0.1448907107114792, "learning_rate": 3.40311413332349e-05, "loss": 0.0076, "step": 18700 }, { "epoch": 19.798941798941797, "grad_norm": 0.1489969938993454, "learning_rate": 3.3978921920034365e-05, "loss": 0.0093, "step": 18710 }, { "epoch": 19.80952380952381, "grad_norm": 0.16815413534641266, "learning_rate": 3.392672197395441e-05, "loss": 0.0101, "step": 18720 }, { "epoch": 19.82010582010582, "grad_norm": 0.14315995573997498, "learning_rate": 3.387454155842287e-05, "loss": 0.0108, "step": 18730 }, { "epoch": 19.83068783068783, "grad_norm": 0.14304515719413757, "learning_rate": 3.382238073684386e-05, "loss": 0.0103, "step": 18740 }, { "epoch": 19.841269841269842, "grad_norm": 0.12437903136014938, "learning_rate": 3.377023957259771e-05, "loss": 0.0098, "step": 18750 }, { "epoch": 19.85185185185185, "grad_norm": 0.10345996171236038, "learning_rate": 3.3718118129040835e-05, "loss": 0.0095, "step": 18760 }, { "epoch": 19.862433862433864, "grad_norm": 0.16433851420879364, "learning_rate": 3.3666016469505724e-05, "loss": 0.0105, "step": 18770 }, { "epoch": 19.873015873015873, "grad_norm": 0.12119382619857788, "learning_rate": 3.361393465730079e-05, "loss": 0.0084, "step": 18780 }, { "epoch": 19.883597883597883, "grad_norm": 0.22213156521320343, "learning_rate": 3.3561872755710366e-05, "loss": 0.0116, "step": 18790 }, { "epoch": 19.894179894179896, "grad_norm": 0.09494311362504959, "learning_rate": 3.350983082799456e-05, "loss": 0.01, "step": 18800 }, { "epoch": 19.904761904761905, "grad_norm": 0.1667143553495407, "learning_rate": 3.34578089373892e-05, "loss": 0.0093, "step": 18810 }, { "epoch": 19.915343915343914, "grad_norm": 0.19188018143177032, "learning_rate": 3.340580714710581e-05, "loss": 0.0101, "step": 18820 }, { "epoch": 19.925925925925927, "grad_norm": 0.17429398000240326, "learning_rate": 3.3353825520331466e-05, "loss": 0.0083, "step": 18830 }, { "epoch": 19.936507936507937, "grad_norm": 0.11172422766685486, "learning_rate": 3.330186412022876e-05, "loss": 0.0094, "step": 18840 }, { "epoch": 19.947089947089946, "grad_norm": 0.1589718908071518, "learning_rate": 3.324992300993568e-05, "loss": 0.01, "step": 18850 }, { "epoch": 19.95767195767196, "grad_norm": 0.10947707295417786, "learning_rate": 3.319800225256556e-05, "loss": 0.0099, "step": 18860 }, { "epoch": 19.96825396825397, "grad_norm": 0.13887353241443634, "learning_rate": 3.314610191120702e-05, "loss": 0.0099, "step": 18870 }, { "epoch": 19.978835978835978, "grad_norm": 0.08827392011880875, "learning_rate": 3.30942220489239e-05, "loss": 0.01, "step": 18880 }, { "epoch": 19.98941798941799, "grad_norm": 0.13502193987369537, "learning_rate": 3.3042362728755086e-05, "loss": 0.0093, "step": 18890 }, { "epoch": 20.0, "grad_norm": 0.35230809450149536, "learning_rate": 3.299052401371456e-05, "loss": 0.0106, "step": 18900 }, { "epoch": 20.01058201058201, "grad_norm": 0.15252447128295898, "learning_rate": 3.293870596679125e-05, "loss": 0.0077, "step": 18910 }, { "epoch": 20.021164021164022, "grad_norm": 0.11773746460676193, "learning_rate": 3.288690865094895e-05, "loss": 0.0087, "step": 18920 }, { "epoch": 20.03174603174603, "grad_norm": 0.08998213708400726, "learning_rate": 3.283513212912632e-05, "loss": 0.011, "step": 18930 }, { "epoch": 20.04232804232804, "grad_norm": 0.10814781486988068, "learning_rate": 3.2783376464236684e-05, "loss": 0.0095, "step": 18940 }, { "epoch": 20.052910052910054, "grad_norm": 0.09632468968629837, "learning_rate": 3.273164171916806e-05, "loss": 0.012, "step": 18950 }, { "epoch": 20.063492063492063, "grad_norm": 0.12214640527963638, "learning_rate": 3.267992795678306e-05, "loss": 0.0117, "step": 18960 }, { "epoch": 20.074074074074073, "grad_norm": 0.17100465297698975, "learning_rate": 3.2628235239918744e-05, "loss": 0.0123, "step": 18970 }, { "epoch": 20.084656084656086, "grad_norm": 0.13817349076271057, "learning_rate": 3.2576563631386695e-05, "loss": 0.0121, "step": 18980 }, { "epoch": 20.095238095238095, "grad_norm": 0.21214257180690765, "learning_rate": 3.252491319397275e-05, "loss": 0.0127, "step": 18990 }, { "epoch": 20.105820105820104, "grad_norm": 0.12267030030488968, "learning_rate": 3.247328399043706e-05, "loss": 0.0115, "step": 19000 }, { "epoch": 20.116402116402117, "grad_norm": 0.09730885922908783, "learning_rate": 3.242167608351399e-05, "loss": 0.0079, "step": 19010 }, { "epoch": 20.126984126984127, "grad_norm": 0.07766564935445786, "learning_rate": 3.2370089535911986e-05, "loss": 0.0096, "step": 19020 }, { "epoch": 20.137566137566136, "grad_norm": 0.09058134257793427, "learning_rate": 3.23185244103136e-05, "loss": 0.0083, "step": 19030 }, { "epoch": 20.14814814814815, "grad_norm": 0.11574947088956833, "learning_rate": 3.22669807693753e-05, "loss": 0.0097, "step": 19040 }, { "epoch": 20.158730158730158, "grad_norm": 0.12481769174337387, "learning_rate": 3.2215458675727495e-05, "loss": 0.0111, "step": 19050 }, { "epoch": 20.16931216931217, "grad_norm": 0.06211547926068306, "learning_rate": 3.216395819197438e-05, "loss": 0.0099, "step": 19060 }, { "epoch": 20.17989417989418, "grad_norm": 0.1373850256204605, "learning_rate": 3.211247938069387e-05, "loss": 0.0113, "step": 19070 }, { "epoch": 20.19047619047619, "grad_norm": 0.16939745843410492, "learning_rate": 3.206102230443759e-05, "loss": 0.011, "step": 19080 }, { "epoch": 20.201058201058203, "grad_norm": 0.2178727686405182, "learning_rate": 3.2009587025730764e-05, "loss": 0.0102, "step": 19090 }, { "epoch": 20.211640211640212, "grad_norm": 0.23394757509231567, "learning_rate": 3.195817360707207e-05, "loss": 0.0118, "step": 19100 }, { "epoch": 20.22222222222222, "grad_norm": 0.22716914117336273, "learning_rate": 3.19067821109337e-05, "loss": 0.0115, "step": 19110 }, { "epoch": 20.232804232804234, "grad_norm": 0.12821584939956665, "learning_rate": 3.185541259976114e-05, "loss": 0.0127, "step": 19120 }, { "epoch": 20.243386243386244, "grad_norm": 0.15461547672748566, "learning_rate": 3.180406513597316e-05, "loss": 0.011, "step": 19130 }, { "epoch": 20.253968253968253, "grad_norm": 0.1283264309167862, "learning_rate": 3.1752739781961835e-05, "loss": 0.0103, "step": 19140 }, { "epoch": 20.264550264550266, "grad_norm": 0.10056287795305252, "learning_rate": 3.170143660009228e-05, "loss": 0.0128, "step": 19150 }, { "epoch": 20.275132275132275, "grad_norm": 0.12536995112895966, "learning_rate": 3.16501556527027e-05, "loss": 0.0112, "step": 19160 }, { "epoch": 20.285714285714285, "grad_norm": 0.12990149855613708, "learning_rate": 3.1598897002104265e-05, "loss": 0.0113, "step": 19170 }, { "epoch": 20.296296296296298, "grad_norm": 0.10574468225240707, "learning_rate": 3.154766071058108e-05, "loss": 0.0101, "step": 19180 }, { "epoch": 20.306878306878307, "grad_norm": 0.1820424497127533, "learning_rate": 3.149644684039008e-05, "loss": 0.0098, "step": 19190 }, { "epoch": 20.317460317460316, "grad_norm": 0.19389313459396362, "learning_rate": 3.144525545376095e-05, "loss": 0.0127, "step": 19200 }, { "epoch": 20.32804232804233, "grad_norm": 0.10064177960157394, "learning_rate": 3.139408661289603e-05, "loss": 0.0113, "step": 19210 }, { "epoch": 20.33862433862434, "grad_norm": 0.11108633875846863, "learning_rate": 3.134294037997032e-05, "loss": 0.0105, "step": 19220 }, { "epoch": 20.349206349206348, "grad_norm": 0.11144047975540161, "learning_rate": 3.129181681713127e-05, "loss": 0.0115, "step": 19230 }, { "epoch": 20.35978835978836, "grad_norm": 0.09952244162559509, "learning_rate": 3.1240715986498855e-05, "loss": 0.0095, "step": 19240 }, { "epoch": 20.37037037037037, "grad_norm": 0.11222370713949203, "learning_rate": 3.1189637950165394e-05, "loss": 0.0096, "step": 19250 }, { "epoch": 20.38095238095238, "grad_norm": 0.09948199987411499, "learning_rate": 3.1138582770195544e-05, "loss": 0.0112, "step": 19260 }, { "epoch": 20.391534391534393, "grad_norm": 0.13574299216270447, "learning_rate": 3.108755050862615e-05, "loss": 0.0098, "step": 19270 }, { "epoch": 20.402116402116402, "grad_norm": 0.11953767389059067, "learning_rate": 3.1036541227466204e-05, "loss": 0.0091, "step": 19280 }, { "epoch": 20.41269841269841, "grad_norm": 0.13352979719638824, "learning_rate": 3.0985554988696786e-05, "loss": 0.0089, "step": 19290 }, { "epoch": 20.423280423280424, "grad_norm": 0.2121303230524063, "learning_rate": 3.093459185427102e-05, "loss": 0.0124, "step": 19300 }, { "epoch": 20.433862433862434, "grad_norm": 0.07770460844039917, "learning_rate": 3.0883651886113905e-05, "loss": 0.0103, "step": 19310 }, { "epoch": 20.444444444444443, "grad_norm": 0.1435854285955429, "learning_rate": 3.0832735146122294e-05, "loss": 0.0101, "step": 19320 }, { "epoch": 20.455026455026456, "grad_norm": 0.12199913710355759, "learning_rate": 3.078184169616485e-05, "loss": 0.0103, "step": 19330 }, { "epoch": 20.465608465608465, "grad_norm": 0.13125786185264587, "learning_rate": 3.073097159808187e-05, "loss": 0.0119, "step": 19340 }, { "epoch": 20.476190476190474, "grad_norm": 0.24557147920131683, "learning_rate": 3.068012491368537e-05, "loss": 0.012, "step": 19350 }, { "epoch": 20.486772486772487, "grad_norm": 0.10075777769088745, "learning_rate": 3.062930170475885e-05, "loss": 0.0099, "step": 19360 }, { "epoch": 20.497354497354497, "grad_norm": 0.23499180376529694, "learning_rate": 3.057850203305729e-05, "loss": 0.0105, "step": 19370 }, { "epoch": 20.507936507936506, "grad_norm": 0.10097205638885498, "learning_rate": 3.052772596030708e-05, "loss": 0.0103, "step": 19380 }, { "epoch": 20.51851851851852, "grad_norm": 0.1041564792394638, "learning_rate": 3.0476973548205943e-05, "loss": 0.0115, "step": 19390 }, { "epoch": 20.52910052910053, "grad_norm": 0.09410128742456436, "learning_rate": 3.0426244858422847e-05, "loss": 0.01, "step": 19400 }, { "epoch": 20.53968253968254, "grad_norm": 0.15246930718421936, "learning_rate": 3.0375539952597943e-05, "loss": 0.0109, "step": 19410 }, { "epoch": 20.55026455026455, "grad_norm": 0.2567597031593323, "learning_rate": 3.0324858892342468e-05, "loss": 0.0126, "step": 19420 }, { "epoch": 20.56084656084656, "grad_norm": 0.12323429435491562, "learning_rate": 3.0274201739238672e-05, "loss": 0.01, "step": 19430 }, { "epoch": 20.571428571428573, "grad_norm": 0.15022434294223785, "learning_rate": 3.0223568554839786e-05, "loss": 0.0115, "step": 19440 }, { "epoch": 20.582010582010582, "grad_norm": 0.1112704947590828, "learning_rate": 3.0172959400669886e-05, "loss": 0.0106, "step": 19450 }, { "epoch": 20.59259259259259, "grad_norm": 0.1272101253271103, "learning_rate": 3.0122374338223902e-05, "loss": 0.0086, "step": 19460 }, { "epoch": 20.603174603174605, "grad_norm": 0.07355689257383347, "learning_rate": 3.0071813428967427e-05, "loss": 0.0096, "step": 19470 }, { "epoch": 20.613756613756614, "grad_norm": 0.1463806927204132, "learning_rate": 3.0021276734336746e-05, "loss": 0.0102, "step": 19480 }, { "epoch": 20.624338624338623, "grad_norm": 0.10659120231866837, "learning_rate": 2.997076431573871e-05, "loss": 0.0082, "step": 19490 }, { "epoch": 20.634920634920636, "grad_norm": 0.059247903525829315, "learning_rate": 2.9920276234550636e-05, "loss": 0.0092, "step": 19500 }, { "epoch": 20.645502645502646, "grad_norm": 0.15052996575832367, "learning_rate": 2.986981255212035e-05, "loss": 0.0097, "step": 19510 }, { "epoch": 20.656084656084655, "grad_norm": 0.08663109689950943, "learning_rate": 2.9819373329765977e-05, "loss": 0.0077, "step": 19520 }, { "epoch": 20.666666666666668, "grad_norm": 0.19728557765483856, "learning_rate": 2.9768958628775902e-05, "loss": 0.01, "step": 19530 }, { "epoch": 20.677248677248677, "grad_norm": 0.3534969687461853, "learning_rate": 2.9718568510408762e-05, "loss": 0.009, "step": 19540 }, { "epoch": 20.687830687830687, "grad_norm": 0.07529741525650024, "learning_rate": 2.9668203035893272e-05, "loss": 0.008, "step": 19550 }, { "epoch": 20.6984126984127, "grad_norm": 0.10689077526330948, "learning_rate": 2.9617862266428287e-05, "loss": 0.0114, "step": 19560 }, { "epoch": 20.70899470899471, "grad_norm": 0.1323467493057251, "learning_rate": 2.9567546263182556e-05, "loss": 0.0093, "step": 19570 }, { "epoch": 20.719576719576718, "grad_norm": 0.16686071455478668, "learning_rate": 2.951725508729476e-05, "loss": 0.0102, "step": 19580 }, { "epoch": 20.73015873015873, "grad_norm": 0.13897252082824707, "learning_rate": 2.946698879987344e-05, "loss": 0.0076, "step": 19590 }, { "epoch": 20.74074074074074, "grad_norm": 0.28644633293151855, "learning_rate": 2.9416747461996853e-05, "loss": 0.0107, "step": 19600 }, { "epoch": 20.75132275132275, "grad_norm": 0.10213685780763626, "learning_rate": 2.9366531134712973e-05, "loss": 0.0098, "step": 19610 }, { "epoch": 20.761904761904763, "grad_norm": 0.13905030488967896, "learning_rate": 2.9316339879039367e-05, "loss": 0.0096, "step": 19620 }, { "epoch": 20.772486772486772, "grad_norm": 0.19534912705421448, "learning_rate": 2.9266173755963167e-05, "loss": 0.0108, "step": 19630 }, { "epoch": 20.78306878306878, "grad_norm": 0.2797314524650574, "learning_rate": 2.9216032826440926e-05, "loss": 0.0105, "step": 19640 }, { "epoch": 20.793650793650794, "grad_norm": 0.12469014525413513, "learning_rate": 2.9165917151398592e-05, "loss": 0.0108, "step": 19650 }, { "epoch": 20.804232804232804, "grad_norm": 0.17910653352737427, "learning_rate": 2.9115826791731428e-05, "loss": 0.0123, "step": 19660 }, { "epoch": 20.814814814814813, "grad_norm": 0.1755513846874237, "learning_rate": 2.906576180830398e-05, "loss": 0.0112, "step": 19670 }, { "epoch": 20.825396825396826, "grad_norm": 0.11084353178739548, "learning_rate": 2.9015722261949917e-05, "loss": 0.0102, "step": 19680 }, { "epoch": 20.835978835978835, "grad_norm": 0.15334568917751312, "learning_rate": 2.8965708213471986e-05, "loss": 0.0091, "step": 19690 }, { "epoch": 20.84656084656085, "grad_norm": 0.15367184579372406, "learning_rate": 2.8915719723641975e-05, "loss": 0.0103, "step": 19700 }, { "epoch": 20.857142857142858, "grad_norm": 0.13970662653446198, "learning_rate": 2.8865756853200604e-05, "loss": 0.0086, "step": 19710 }, { "epoch": 20.867724867724867, "grad_norm": 0.10002655535936356, "learning_rate": 2.8815819662857503e-05, "loss": 0.008, "step": 19720 }, { "epoch": 20.87830687830688, "grad_norm": 0.08838411420583725, "learning_rate": 2.876590821329105e-05, "loss": 0.0094, "step": 19730 }, { "epoch": 20.88888888888889, "grad_norm": 0.1677333116531372, "learning_rate": 2.871602256514836e-05, "loss": 0.0102, "step": 19740 }, { "epoch": 20.8994708994709, "grad_norm": 0.14699795842170715, "learning_rate": 2.8666162779045203e-05, "loss": 0.0097, "step": 19750 }, { "epoch": 20.91005291005291, "grad_norm": 0.07280182093381882, "learning_rate": 2.8616328915565904e-05, "loss": 0.0108, "step": 19760 }, { "epoch": 20.92063492063492, "grad_norm": 0.12061376124620438, "learning_rate": 2.856652103526334e-05, "loss": 0.0103, "step": 19770 }, { "epoch": 20.93121693121693, "grad_norm": 0.12445225566625595, "learning_rate": 2.8516739198658755e-05, "loss": 0.0096, "step": 19780 }, { "epoch": 20.941798941798943, "grad_norm": 0.10751165449619293, "learning_rate": 2.846698346624177e-05, "loss": 0.0099, "step": 19790 }, { "epoch": 20.952380952380953, "grad_norm": 0.15082231163978577, "learning_rate": 2.841725389847032e-05, "loss": 0.0114, "step": 19800 }, { "epoch": 20.962962962962962, "grad_norm": 0.12126300483942032, "learning_rate": 2.8367550555770506e-05, "loss": 0.0092, "step": 19810 }, { "epoch": 20.973544973544975, "grad_norm": 0.16907508671283722, "learning_rate": 2.831787349853655e-05, "loss": 0.0107, "step": 19820 }, { "epoch": 20.984126984126984, "grad_norm": 0.10987025499343872, "learning_rate": 2.8268222787130806e-05, "loss": 0.0092, "step": 19830 }, { "epoch": 20.994708994708994, "grad_norm": 0.1288667917251587, "learning_rate": 2.821859848188355e-05, "loss": 0.0108, "step": 19840 }, { "epoch": 21.005291005291006, "grad_norm": 0.14735166728496552, "learning_rate": 2.816900064309299e-05, "loss": 0.011, "step": 19850 }, { "epoch": 21.015873015873016, "grad_norm": 0.1978432685136795, "learning_rate": 2.811942933102517e-05, "loss": 0.0089, "step": 19860 }, { "epoch": 21.026455026455025, "grad_norm": 0.13127590715885162, "learning_rate": 2.806988460591391e-05, "loss": 0.0102, "step": 19870 }, { "epoch": 21.037037037037038, "grad_norm": 0.12720897793769836, "learning_rate": 2.802036652796074e-05, "loss": 0.0095, "step": 19880 }, { "epoch": 21.047619047619047, "grad_norm": 0.14872203767299652, "learning_rate": 2.797087515733478e-05, "loss": 0.0099, "step": 19890 }, { "epoch": 21.058201058201057, "grad_norm": 0.08955220133066177, "learning_rate": 2.7921410554172723e-05, "loss": 0.0095, "step": 19900 }, { "epoch": 21.06878306878307, "grad_norm": 0.09200132638216019, "learning_rate": 2.787197277857871e-05, "loss": 0.0114, "step": 19910 }, { "epoch": 21.07936507936508, "grad_norm": 0.11567212641239166, "learning_rate": 2.782256189062429e-05, "loss": 0.0093, "step": 19920 }, { "epoch": 21.08994708994709, "grad_norm": 0.1280301958322525, "learning_rate": 2.777317795034839e-05, "loss": 0.0106, "step": 19930 }, { "epoch": 21.1005291005291, "grad_norm": 0.07983856648206711, "learning_rate": 2.7723821017757112e-05, "loss": 0.0094, "step": 19940 }, { "epoch": 21.11111111111111, "grad_norm": 0.1315968632698059, "learning_rate": 2.7674491152823822e-05, "loss": 0.0078, "step": 19950 }, { "epoch": 21.12169312169312, "grad_norm": 0.10763780027627945, "learning_rate": 2.7625188415488944e-05, "loss": 0.0119, "step": 19960 }, { "epoch": 21.132275132275133, "grad_norm": 0.0669177770614624, "learning_rate": 2.7575912865659924e-05, "loss": 0.009, "step": 19970 }, { "epoch": 21.142857142857142, "grad_norm": 0.12197723984718323, "learning_rate": 2.7526664563211245e-05, "loss": 0.0087, "step": 19980 }, { "epoch": 21.15343915343915, "grad_norm": 0.13749268651008606, "learning_rate": 2.7477443567984224e-05, "loss": 0.0106, "step": 19990 }, { "epoch": 21.164021164021165, "grad_norm": 0.10203485190868378, "learning_rate": 2.7428249939787e-05, "loss": 0.0099, "step": 20000 }, { "epoch": 21.174603174603174, "grad_norm": 0.1775227040052414, "learning_rate": 2.7379083738394484e-05, "loss": 0.0094, "step": 20010 }, { "epoch": 21.185185185185187, "grad_norm": 0.07413356006145477, "learning_rate": 2.732994502354823e-05, "loss": 0.0094, "step": 20020 }, { "epoch": 21.195767195767196, "grad_norm": 0.11867181956768036, "learning_rate": 2.72808338549564e-05, "loss": 0.0094, "step": 20030 }, { "epoch": 21.206349206349206, "grad_norm": 0.09745877236127853, "learning_rate": 2.723175029229374e-05, "loss": 0.01, "step": 20040 }, { "epoch": 21.21693121693122, "grad_norm": 0.1452837437391281, "learning_rate": 2.718269439520138e-05, "loss": 0.0098, "step": 20050 }, { "epoch": 21.227513227513228, "grad_norm": 0.11404120177030563, "learning_rate": 2.7133666223286858e-05, "loss": 0.0092, "step": 20060 }, { "epoch": 21.238095238095237, "grad_norm": 0.11668221652507782, "learning_rate": 2.7084665836124006e-05, "loss": 0.0096, "step": 20070 }, { "epoch": 21.24867724867725, "grad_norm": 0.11886418610811234, "learning_rate": 2.703569329325296e-05, "loss": 0.0106, "step": 20080 }, { "epoch": 21.25925925925926, "grad_norm": 0.14442318677902222, "learning_rate": 2.698674865417994e-05, "loss": 0.0109, "step": 20090 }, { "epoch": 21.26984126984127, "grad_norm": 0.23178520798683167, "learning_rate": 2.6937831978377325e-05, "loss": 0.0091, "step": 20100 }, { "epoch": 21.280423280423282, "grad_norm": 0.11694367974996567, "learning_rate": 2.6888943325283485e-05, "loss": 0.0123, "step": 20110 }, { "epoch": 21.29100529100529, "grad_norm": 0.08318165689706802, "learning_rate": 2.6840082754302733e-05, "loss": 0.0077, "step": 20120 }, { "epoch": 21.3015873015873, "grad_norm": 0.17472611367702484, "learning_rate": 2.679125032480525e-05, "loss": 0.0112, "step": 20130 }, { "epoch": 21.312169312169313, "grad_norm": 0.10344789922237396, "learning_rate": 2.674244609612708e-05, "loss": 0.0095, "step": 20140 }, { "epoch": 21.322751322751323, "grad_norm": 0.19993729889392853, "learning_rate": 2.6693670127569958e-05, "loss": 0.0093, "step": 20150 }, { "epoch": 21.333333333333332, "grad_norm": 0.1659214198589325, "learning_rate": 2.6644922478401268e-05, "loss": 0.0115, "step": 20160 }, { "epoch": 21.343915343915345, "grad_norm": 0.1809556484222412, "learning_rate": 2.6596203207854004e-05, "loss": 0.0115, "step": 20170 }, { "epoch": 21.354497354497354, "grad_norm": 0.09492571651935577, "learning_rate": 2.654751237512666e-05, "loss": 0.0129, "step": 20180 }, { "epoch": 21.365079365079364, "grad_norm": 0.1270444095134735, "learning_rate": 2.649885003938323e-05, "loss": 0.0101, "step": 20190 }, { "epoch": 21.375661375661377, "grad_norm": 0.1896924078464508, "learning_rate": 2.6450216259753e-05, "loss": 0.0127, "step": 20200 }, { "epoch": 21.386243386243386, "grad_norm": 0.1278003603219986, "learning_rate": 2.640161109533063e-05, "loss": 0.0096, "step": 20210 }, { "epoch": 21.396825396825395, "grad_norm": 0.06848739832639694, "learning_rate": 2.6353034605175936e-05, "loss": 0.0087, "step": 20220 }, { "epoch": 21.40740740740741, "grad_norm": 0.16232430934906006, "learning_rate": 2.630448684831398e-05, "loss": 0.0089, "step": 20230 }, { "epoch": 21.417989417989418, "grad_norm": 0.15507405996322632, "learning_rate": 2.625596788373482e-05, "loss": 0.0102, "step": 20240 }, { "epoch": 21.428571428571427, "grad_norm": 0.24265287816524506, "learning_rate": 2.620747777039363e-05, "loss": 0.0106, "step": 20250 }, { "epoch": 21.43915343915344, "grad_norm": 0.1667763739824295, "learning_rate": 2.6159016567210424e-05, "loss": 0.0098, "step": 20260 }, { "epoch": 21.44973544973545, "grad_norm": 0.1723683476448059, "learning_rate": 2.6110584333070153e-05, "loss": 0.0109, "step": 20270 }, { "epoch": 21.46031746031746, "grad_norm": 0.15367862582206726, "learning_rate": 2.6062181126822537e-05, "loss": 0.0079, "step": 20280 }, { "epoch": 21.47089947089947, "grad_norm": 0.10590901970863342, "learning_rate": 2.601380700728203e-05, "loss": 0.0086, "step": 20290 }, { "epoch": 21.48148148148148, "grad_norm": 0.15949983894824982, "learning_rate": 2.596546203322777e-05, "loss": 0.0098, "step": 20300 }, { "epoch": 21.49206349206349, "grad_norm": 0.09199567884206772, "learning_rate": 2.5917146263403457e-05, "loss": 0.0092, "step": 20310 }, { "epoch": 21.502645502645503, "grad_norm": 0.07088278979063034, "learning_rate": 2.5868859756517296e-05, "loss": 0.01, "step": 20320 }, { "epoch": 21.513227513227513, "grad_norm": 0.09439944475889206, "learning_rate": 2.5820602571241947e-05, "loss": 0.0104, "step": 20330 }, { "epoch": 21.523809523809526, "grad_norm": 0.07300359010696411, "learning_rate": 2.577237476621442e-05, "loss": 0.008, "step": 20340 }, { "epoch": 21.534391534391535, "grad_norm": 0.12904563546180725, "learning_rate": 2.572417640003609e-05, "loss": 0.0088, "step": 20350 }, { "epoch": 21.544973544973544, "grad_norm": 0.14128194749355316, "learning_rate": 2.5676007531272473e-05, "loss": 0.0087, "step": 20360 }, { "epoch": 21.555555555555557, "grad_norm": 0.11816711723804474, "learning_rate": 2.5627868218453332e-05, "loss": 0.0088, "step": 20370 }, { "epoch": 21.566137566137566, "grad_norm": 0.18274378776550293, "learning_rate": 2.5579758520072443e-05, "loss": 0.0089, "step": 20380 }, { "epoch": 21.576719576719576, "grad_norm": 0.15150316059589386, "learning_rate": 2.5531678494587615e-05, "loss": 0.0091, "step": 20390 }, { "epoch": 21.58730158730159, "grad_norm": 0.24176836013793945, "learning_rate": 2.5483628200420644e-05, "loss": 0.0104, "step": 20400 }, { "epoch": 21.597883597883598, "grad_norm": 0.11769750714302063, "learning_rate": 2.5435607695957154e-05, "loss": 0.0098, "step": 20410 }, { "epoch": 21.608465608465607, "grad_norm": 0.07923336327075958, "learning_rate": 2.5387617039546586e-05, "loss": 0.0088, "step": 20420 }, { "epoch": 21.61904761904762, "grad_norm": 0.16593855619430542, "learning_rate": 2.5339656289502106e-05, "loss": 0.0104, "step": 20430 }, { "epoch": 21.62962962962963, "grad_norm": 0.09871983528137207, "learning_rate": 2.529172550410056e-05, "loss": 0.0097, "step": 20440 }, { "epoch": 21.64021164021164, "grad_norm": 0.10243546217679977, "learning_rate": 2.524382474158234e-05, "loss": 0.0087, "step": 20450 }, { "epoch": 21.650793650793652, "grad_norm": 0.12927091121673584, "learning_rate": 2.5195954060151433e-05, "loss": 0.0105, "step": 20460 }, { "epoch": 21.66137566137566, "grad_norm": 0.1241525188088417, "learning_rate": 2.5148113517975212e-05, "loss": 0.0095, "step": 20470 }, { "epoch": 21.67195767195767, "grad_norm": 0.1771206259727478, "learning_rate": 2.5100303173184447e-05, "loss": 0.0096, "step": 20480 }, { "epoch": 21.682539682539684, "grad_norm": 0.1374325305223465, "learning_rate": 2.5052523083873196e-05, "loss": 0.0094, "step": 20490 }, { "epoch": 21.693121693121693, "grad_norm": 0.1578689068555832, "learning_rate": 2.500477330809881e-05, "loss": 0.0106, "step": 20500 }, { "epoch": 21.703703703703702, "grad_norm": 0.1008077934384346, "learning_rate": 2.4957053903881733e-05, "loss": 0.0092, "step": 20510 }, { "epoch": 21.714285714285715, "grad_norm": 0.0627187192440033, "learning_rate": 2.4909364929205576e-05, "loss": 0.0091, "step": 20520 }, { "epoch": 21.724867724867725, "grad_norm": 0.1170390397310257, "learning_rate": 2.4861706442016923e-05, "loss": 0.0088, "step": 20530 }, { "epoch": 21.735449735449734, "grad_norm": 0.15348224341869354, "learning_rate": 2.481407850022533e-05, "loss": 0.0107, "step": 20540 }, { "epoch": 21.746031746031747, "grad_norm": 0.1362651288509369, "learning_rate": 2.4766481161703213e-05, "loss": 0.0119, "step": 20550 }, { "epoch": 21.756613756613756, "grad_norm": 0.08150946348905563, "learning_rate": 2.4718914484285876e-05, "loss": 0.0081, "step": 20560 }, { "epoch": 21.767195767195766, "grad_norm": 0.0716332420706749, "learning_rate": 2.467137852577129e-05, "loss": 0.0092, "step": 20570 }, { "epoch": 21.77777777777778, "grad_norm": 0.11009880155324936, "learning_rate": 2.4623873343920123e-05, "loss": 0.0103, "step": 20580 }, { "epoch": 21.788359788359788, "grad_norm": 0.10981529951095581, "learning_rate": 2.4576398996455656e-05, "loss": 0.008, "step": 20590 }, { "epoch": 21.798941798941797, "grad_norm": 0.0882052555680275, "learning_rate": 2.452895554106368e-05, "loss": 0.0097, "step": 20600 }, { "epoch": 21.80952380952381, "grad_norm": 0.1249731183052063, "learning_rate": 2.4481543035392506e-05, "loss": 0.0113, "step": 20610 }, { "epoch": 21.82010582010582, "grad_norm": 0.09011606872081757, "learning_rate": 2.4434161537052774e-05, "loss": 0.0077, "step": 20620 }, { "epoch": 21.83068783068783, "grad_norm": 0.09948776662349701, "learning_rate": 2.4386811103617474e-05, "loss": 0.0114, "step": 20630 }, { "epoch": 21.841269841269842, "grad_norm": 0.09204112738370895, "learning_rate": 2.4339491792621833e-05, "loss": 0.0084, "step": 20640 }, { "epoch": 21.85185185185185, "grad_norm": 0.11551312357187271, "learning_rate": 2.4292203661563312e-05, "loss": 0.0072, "step": 20650 }, { "epoch": 21.862433862433864, "grad_norm": 0.18026410043239594, "learning_rate": 2.424494676790141e-05, "loss": 0.0097, "step": 20660 }, { "epoch": 21.873015873015873, "grad_norm": 0.14167621731758118, "learning_rate": 2.4197721169057745e-05, "loss": 0.0095, "step": 20670 }, { "epoch": 21.883597883597883, "grad_norm": 0.16148096323013306, "learning_rate": 2.4150526922415857e-05, "loss": 0.0107, "step": 20680 }, { "epoch": 21.894179894179896, "grad_norm": 0.2657772898674011, "learning_rate": 2.41033640853212e-05, "loss": 0.0097, "step": 20690 }, { "epoch": 21.904761904761905, "grad_norm": 0.1630149930715561, "learning_rate": 2.405623271508108e-05, "loss": 0.0114, "step": 20700 }, { "epoch": 21.915343915343914, "grad_norm": 0.09956002980470657, "learning_rate": 2.4009132868964522e-05, "loss": 0.0093, "step": 20710 }, { "epoch": 21.925925925925927, "grad_norm": 0.10128141194581985, "learning_rate": 2.3962064604202327e-05, "loss": 0.0103, "step": 20720 }, { "epoch": 21.936507936507937, "grad_norm": 0.11446001380681992, "learning_rate": 2.3915027977986858e-05, "loss": 0.0128, "step": 20730 }, { "epoch": 21.947089947089946, "grad_norm": 0.13990065455436707, "learning_rate": 2.386802304747205e-05, "loss": 0.0096, "step": 20740 }, { "epoch": 21.95767195767196, "grad_norm": 0.31593018770217896, "learning_rate": 2.382104986977332e-05, "loss": 0.0104, "step": 20750 }, { "epoch": 21.96825396825397, "grad_norm": 0.07805515080690384, "learning_rate": 2.377410850196749e-05, "loss": 0.0084, "step": 20760 }, { "epoch": 21.978835978835978, "grad_norm": 0.1930316984653473, "learning_rate": 2.3727199001092797e-05, "loss": 0.0091, "step": 20770 }, { "epoch": 21.98941798941799, "grad_norm": 0.15703563392162323, "learning_rate": 2.3680321424148677e-05, "loss": 0.0087, "step": 20780 }, { "epoch": 22.0, "grad_norm": 0.10788397490978241, "learning_rate": 2.3633475828095793e-05, "loss": 0.0082, "step": 20790 }, { "epoch": 22.01058201058201, "grad_norm": 0.09723556041717529, "learning_rate": 2.358666226985599e-05, "loss": 0.0097, "step": 20800 }, { "epoch": 22.021164021164022, "grad_norm": 0.09826261550188065, "learning_rate": 2.353988080631213e-05, "loss": 0.0089, "step": 20810 }, { "epoch": 22.03174603174603, "grad_norm": 0.11993920803070068, "learning_rate": 2.3493131494308142e-05, "loss": 0.0085, "step": 20820 }, { "epoch": 22.04232804232804, "grad_norm": 0.10577547550201416, "learning_rate": 2.344641439064881e-05, "loss": 0.0096, "step": 20830 }, { "epoch": 22.052910052910054, "grad_norm": 0.2759973108768463, "learning_rate": 2.339972955209984e-05, "loss": 0.0093, "step": 20840 }, { "epoch": 22.063492063492063, "grad_norm": 0.143357515335083, "learning_rate": 2.335307703538771e-05, "loss": 0.0092, "step": 20850 }, { "epoch": 22.074074074074073, "grad_norm": 0.08359059691429138, "learning_rate": 2.330645689719962e-05, "loss": 0.0069, "step": 20860 }, { "epoch": 22.084656084656086, "grad_norm": 0.08679341524839401, "learning_rate": 2.3259869194183414e-05, "loss": 0.0074, "step": 20870 }, { "epoch": 22.095238095238095, "grad_norm": 0.1527824103832245, "learning_rate": 2.321331398294759e-05, "loss": 0.0099, "step": 20880 }, { "epoch": 22.105820105820104, "grad_norm": 0.1431329846382141, "learning_rate": 2.3166791320061094e-05, "loss": 0.0089, "step": 20890 }, { "epoch": 22.116402116402117, "grad_norm": 0.07321906834840775, "learning_rate": 2.312030126205335e-05, "loss": 0.0087, "step": 20900 }, { "epoch": 22.126984126984127, "grad_norm": 0.08120706677436829, "learning_rate": 2.3073843865414162e-05, "loss": 0.0092, "step": 20910 }, { "epoch": 22.137566137566136, "grad_norm": 0.12388573586940765, "learning_rate": 2.302741918659363e-05, "loss": 0.0114, "step": 20920 }, { "epoch": 22.14814814814815, "grad_norm": 0.10927873104810715, "learning_rate": 2.2981027282002154e-05, "loss": 0.0104, "step": 20930 }, { "epoch": 22.158730158730158, "grad_norm": 0.18908600509166718, "learning_rate": 2.2934668208010236e-05, "loss": 0.0113, "step": 20940 }, { "epoch": 22.16931216931217, "grad_norm": 0.10497846454381943, "learning_rate": 2.2888342020948555e-05, "loss": 0.0086, "step": 20950 }, { "epoch": 22.17989417989418, "grad_norm": 0.1394333690404892, "learning_rate": 2.2842048777107783e-05, "loss": 0.0112, "step": 20960 }, { "epoch": 22.19047619047619, "grad_norm": 0.1475355476140976, "learning_rate": 2.2795788532738553e-05, "loss": 0.0116, "step": 20970 }, { "epoch": 22.201058201058203, "grad_norm": 0.15635202825069427, "learning_rate": 2.2749561344051468e-05, "loss": 0.0087, "step": 20980 }, { "epoch": 22.211640211640212, "grad_norm": 0.10121671855449677, "learning_rate": 2.2703367267216896e-05, "loss": 0.0091, "step": 20990 }, { "epoch": 22.22222222222222, "grad_norm": 0.16963836550712585, "learning_rate": 2.2657206358364997e-05, "loss": 0.0107, "step": 21000 }, { "epoch": 22.232804232804234, "grad_norm": 0.10785111039876938, "learning_rate": 2.261107867358563e-05, "loss": 0.0101, "step": 21010 }, { "epoch": 22.243386243386244, "grad_norm": 0.07860156148672104, "learning_rate": 2.2564984268928263e-05, "loss": 0.0094, "step": 21020 }, { "epoch": 22.253968253968253, "grad_norm": 0.10885778069496155, "learning_rate": 2.251892320040198e-05, "loss": 0.0081, "step": 21030 }, { "epoch": 22.264550264550266, "grad_norm": 0.10444996505975723, "learning_rate": 2.2472895523975312e-05, "loss": 0.0091, "step": 21040 }, { "epoch": 22.275132275132275, "grad_norm": 0.12384411692619324, "learning_rate": 2.2426901295576212e-05, "loss": 0.0086, "step": 21050 }, { "epoch": 22.285714285714285, "grad_norm": 0.1072298064827919, "learning_rate": 2.2380940571092013e-05, "loss": 0.0082, "step": 21060 }, { "epoch": 22.296296296296298, "grad_norm": 0.1694258600473404, "learning_rate": 2.2335013406369302e-05, "loss": 0.0121, "step": 21070 }, { "epoch": 22.306878306878307, "grad_norm": 0.10739649832248688, "learning_rate": 2.228911985721397e-05, "loss": 0.0082, "step": 21080 }, { "epoch": 22.317460317460316, "grad_norm": 0.11542218923568726, "learning_rate": 2.224325997939095e-05, "loss": 0.0087, "step": 21090 }, { "epoch": 22.32804232804233, "grad_norm": 0.12737688422203064, "learning_rate": 2.219743382862437e-05, "loss": 0.0103, "step": 21100 }, { "epoch": 22.33862433862434, "grad_norm": 0.09368492662906647, "learning_rate": 2.2151641460597294e-05, "loss": 0.0079, "step": 21110 }, { "epoch": 22.349206349206348, "grad_norm": 0.13804003596305847, "learning_rate": 2.210588293095177e-05, "loss": 0.008, "step": 21120 }, { "epoch": 22.35978835978836, "grad_norm": 0.10431548207998276, "learning_rate": 2.2060158295288714e-05, "loss": 0.0094, "step": 21130 }, { "epoch": 22.37037037037037, "grad_norm": 0.16672368347644806, "learning_rate": 2.2014467609167905e-05, "loss": 0.0106, "step": 21140 }, { "epoch": 22.38095238095238, "grad_norm": 0.10247121751308441, "learning_rate": 2.1968810928107806e-05, "loss": 0.0068, "step": 21150 }, { "epoch": 22.391534391534393, "grad_norm": 0.13052937388420105, "learning_rate": 2.1923188307585606e-05, "loss": 0.0094, "step": 21160 }, { "epoch": 22.402116402116402, "grad_norm": 0.15001769363880157, "learning_rate": 2.187759980303708e-05, "loss": 0.0104, "step": 21170 }, { "epoch": 22.41269841269841, "grad_norm": 0.10364121943712234, "learning_rate": 2.1832045469856545e-05, "loss": 0.0085, "step": 21180 }, { "epoch": 22.423280423280424, "grad_norm": 0.12187139689922333, "learning_rate": 2.1786525363396844e-05, "loss": 0.0092, "step": 21190 }, { "epoch": 22.433862433862434, "grad_norm": 0.09680399298667908, "learning_rate": 2.1741039538969183e-05, "loss": 0.0098, "step": 21200 }, { "epoch": 22.444444444444443, "grad_norm": 0.15853388607501984, "learning_rate": 2.169558805184313e-05, "loss": 0.0083, "step": 21210 }, { "epoch": 22.455026455026456, "grad_norm": 0.14181563258171082, "learning_rate": 2.165017095724651e-05, "loss": 0.0105, "step": 21220 }, { "epoch": 22.465608465608465, "grad_norm": 0.14148105680942535, "learning_rate": 2.1604788310365405e-05, "loss": 0.0103, "step": 21230 }, { "epoch": 22.476190476190474, "grad_norm": 0.08489071577787399, "learning_rate": 2.155944016634398e-05, "loss": 0.0086, "step": 21240 }, { "epoch": 22.486772486772487, "grad_norm": 0.08881475031375885, "learning_rate": 2.1514126580284538e-05, "loss": 0.0079, "step": 21250 }, { "epoch": 22.497354497354497, "grad_norm": 0.11917892098426819, "learning_rate": 2.1468847607247344e-05, "loss": 0.0094, "step": 21260 }, { "epoch": 22.507936507936506, "grad_norm": 0.11167936772108078, "learning_rate": 2.1423603302250623e-05, "loss": 0.0086, "step": 21270 }, { "epoch": 22.51851851851852, "grad_norm": 0.09827908128499985, "learning_rate": 2.1378393720270468e-05, "loss": 0.0084, "step": 21280 }, { "epoch": 22.52910052910053, "grad_norm": 0.23491092026233673, "learning_rate": 2.133321891624076e-05, "loss": 0.0099, "step": 21290 }, { "epoch": 22.53968253968254, "grad_norm": 0.09350389987230301, "learning_rate": 2.1288078945053193e-05, "loss": 0.0093, "step": 21300 }, { "epoch": 22.55026455026455, "grad_norm": 0.12732429802417755, "learning_rate": 2.1242973861557064e-05, "loss": 0.0098, "step": 21310 }, { "epoch": 22.56084656084656, "grad_norm": 0.08573411405086517, "learning_rate": 2.1197903720559304e-05, "loss": 0.0098, "step": 21320 }, { "epoch": 22.571428571428573, "grad_norm": 0.08740609139204025, "learning_rate": 2.115286857682438e-05, "loss": 0.0095, "step": 21330 }, { "epoch": 22.582010582010582, "grad_norm": 0.1599167436361313, "learning_rate": 2.1107868485074228e-05, "loss": 0.0088, "step": 21340 }, { "epoch": 22.59259259259259, "grad_norm": 0.18112237751483917, "learning_rate": 2.1062903499988234e-05, "loss": 0.0099, "step": 21350 }, { "epoch": 22.603174603174605, "grad_norm": 0.11496388167142868, "learning_rate": 2.101797367620308e-05, "loss": 0.0095, "step": 21360 }, { "epoch": 22.613756613756614, "grad_norm": 0.1418973058462143, "learning_rate": 2.0973079068312713e-05, "loss": 0.0066, "step": 21370 }, { "epoch": 22.624338624338623, "grad_norm": 0.14564676582813263, "learning_rate": 2.0928219730868358e-05, "loss": 0.0104, "step": 21380 }, { "epoch": 22.634920634920636, "grad_norm": 0.08877326548099518, "learning_rate": 2.0883395718378302e-05, "loss": 0.0088, "step": 21390 }, { "epoch": 22.645502645502646, "grad_norm": 0.12689390778541565, "learning_rate": 2.083860708530798e-05, "loss": 0.0071, "step": 21400 }, { "epoch": 22.656084656084655, "grad_norm": 0.12637445330619812, "learning_rate": 2.0793853886079796e-05, "loss": 0.0101, "step": 21410 }, { "epoch": 22.666666666666668, "grad_norm": 0.1492360532283783, "learning_rate": 2.074913617507309e-05, "loss": 0.0098, "step": 21420 }, { "epoch": 22.677248677248677, "grad_norm": 0.1092238649725914, "learning_rate": 2.070445400662411e-05, "loss": 0.0091, "step": 21430 }, { "epoch": 22.687830687830687, "grad_norm": 0.1604773998260498, "learning_rate": 2.0659807435025907e-05, "loss": 0.0078, "step": 21440 }, { "epoch": 22.6984126984127, "grad_norm": 0.11973849684000015, "learning_rate": 2.061519651452825e-05, "loss": 0.0077, "step": 21450 }, { "epoch": 22.70899470899471, "grad_norm": 0.09573972970247269, "learning_rate": 2.0570621299337656e-05, "loss": 0.0087, "step": 21460 }, { "epoch": 22.719576719576718, "grad_norm": 0.12329800426959991, "learning_rate": 2.0526081843617183e-05, "loss": 0.0097, "step": 21470 }, { "epoch": 22.73015873015873, "grad_norm": 0.12980236113071442, "learning_rate": 2.0481578201486485e-05, "loss": 0.0072, "step": 21480 }, { "epoch": 22.74074074074074, "grad_norm": 0.16027289628982544, "learning_rate": 2.0437110427021677e-05, "loss": 0.0084, "step": 21490 }, { "epoch": 22.75132275132275, "grad_norm": 0.17094872891902924, "learning_rate": 2.0392678574255282e-05, "loss": 0.0076, "step": 21500 }, { "epoch": 22.761904761904763, "grad_norm": 0.1305486261844635, "learning_rate": 2.0348282697176217e-05, "loss": 0.0089, "step": 21510 }, { "epoch": 22.772486772486772, "grad_norm": 0.13248306512832642, "learning_rate": 2.030392284972964e-05, "loss": 0.009, "step": 21520 }, { "epoch": 22.78306878306878, "grad_norm": 0.09475421905517578, "learning_rate": 2.0259599085816973e-05, "loss": 0.0072, "step": 21530 }, { "epoch": 22.793650793650794, "grad_norm": 0.1976487785577774, "learning_rate": 2.0215311459295755e-05, "loss": 0.0086, "step": 21540 }, { "epoch": 22.804232804232804, "grad_norm": 0.12566205859184265, "learning_rate": 2.0171060023979606e-05, "loss": 0.0081, "step": 21550 }, { "epoch": 22.814814814814813, "grad_norm": 0.13032564520835876, "learning_rate": 2.012684483363823e-05, "loss": 0.0111, "step": 21560 }, { "epoch": 22.825396825396826, "grad_norm": 0.07762400805950165, "learning_rate": 2.0082665941997237e-05, "loss": 0.0078, "step": 21570 }, { "epoch": 22.835978835978835, "grad_norm": 0.26834315061569214, "learning_rate": 2.0038523402738146e-05, "loss": 0.0074, "step": 21580 }, { "epoch": 22.84656084656085, "grad_norm": 0.14009875059127808, "learning_rate": 1.99944172694983e-05, "loss": 0.0095, "step": 21590 }, { "epoch": 22.857142857142858, "grad_norm": 0.1445290595293045, "learning_rate": 1.99503475958708e-05, "loss": 0.0087, "step": 21600 }, { "epoch": 22.867724867724867, "grad_norm": 0.1425415277481079, "learning_rate": 1.9906314435404484e-05, "loss": 0.0083, "step": 21610 }, { "epoch": 22.87830687830688, "grad_norm": 0.07200033217668533, "learning_rate": 1.986231784160378e-05, "loss": 0.0092, "step": 21620 }, { "epoch": 22.88888888888889, "grad_norm": 0.13191533088684082, "learning_rate": 1.9818357867928693e-05, "loss": 0.0097, "step": 21630 }, { "epoch": 22.8994708994709, "grad_norm": 0.09943963587284088, "learning_rate": 1.977443456779474e-05, "loss": 0.008, "step": 21640 }, { "epoch": 22.91005291005291, "grad_norm": 0.1524483561515808, "learning_rate": 1.9730547994572858e-05, "loss": 0.01, "step": 21650 }, { "epoch": 22.92063492063492, "grad_norm": 0.10916678607463837, "learning_rate": 1.9686698201589393e-05, "loss": 0.009, "step": 21660 }, { "epoch": 22.93121693121693, "grad_norm": 0.09720636159181595, "learning_rate": 1.9642885242125964e-05, "loss": 0.0094, "step": 21670 }, { "epoch": 22.941798941798943, "grad_norm": 0.07514113932847977, "learning_rate": 1.9599109169419466e-05, "loss": 0.0079, "step": 21680 }, { "epoch": 22.952380952380953, "grad_norm": 0.11142101883888245, "learning_rate": 1.9555370036661948e-05, "loss": 0.0093, "step": 21690 }, { "epoch": 22.962962962962962, "grad_norm": 0.10101354122161865, "learning_rate": 1.9511667897000574e-05, "loss": 0.0082, "step": 21700 }, { "epoch": 22.973544973544975, "grad_norm": 0.17620056867599487, "learning_rate": 1.946800280353755e-05, "loss": 0.0075, "step": 21710 }, { "epoch": 22.984126984126984, "grad_norm": 0.17811189591884613, "learning_rate": 1.9424374809330114e-05, "loss": 0.0089, "step": 21720 }, { "epoch": 22.994708994708994, "grad_norm": 0.2679801881313324, "learning_rate": 1.9380783967390382e-05, "loss": 0.0082, "step": 21730 }, { "epoch": 23.005291005291006, "grad_norm": 0.09341602027416229, "learning_rate": 1.933723033068533e-05, "loss": 0.0083, "step": 21740 }, { "epoch": 23.015873015873016, "grad_norm": 0.08189374953508377, "learning_rate": 1.929371395213674e-05, "loss": 0.0099, "step": 21750 }, { "epoch": 23.026455026455025, "grad_norm": 0.10591569542884827, "learning_rate": 1.9250234884621092e-05, "loss": 0.0103, "step": 21760 }, { "epoch": 23.037037037037038, "grad_norm": 0.16754278540611267, "learning_rate": 1.920679318096959e-05, "loss": 0.0086, "step": 21770 }, { "epoch": 23.047619047619047, "grad_norm": 0.0771554633975029, "learning_rate": 1.9163388893967982e-05, "loss": 0.0106, "step": 21780 }, { "epoch": 23.058201058201057, "grad_norm": 0.08094627410173416, "learning_rate": 1.9120022076356575e-05, "loss": 0.0097, "step": 21790 }, { "epoch": 23.06878306878307, "grad_norm": 0.13192610442638397, "learning_rate": 1.9076692780830114e-05, "loss": 0.0073, "step": 21800 }, { "epoch": 23.07936507936508, "grad_norm": 0.18959186971187592, "learning_rate": 1.903340106003782e-05, "loss": 0.0099, "step": 21810 }, { "epoch": 23.08994708994709, "grad_norm": 0.12445251643657684, "learning_rate": 1.8990146966583183e-05, "loss": 0.0071, "step": 21820 }, { "epoch": 23.1005291005291, "grad_norm": 0.21443238854408264, "learning_rate": 1.8946930553024035e-05, "loss": 0.0102, "step": 21830 }, { "epoch": 23.11111111111111, "grad_norm": 0.10146408528089523, "learning_rate": 1.8903751871872378e-05, "loss": 0.0078, "step": 21840 }, { "epoch": 23.12169312169312, "grad_norm": 0.07795804738998413, "learning_rate": 1.8860610975594382e-05, "loss": 0.0111, "step": 21850 }, { "epoch": 23.132275132275133, "grad_norm": 0.12569619715213776, "learning_rate": 1.8817507916610307e-05, "loss": 0.0128, "step": 21860 }, { "epoch": 23.142857142857142, "grad_norm": 0.08426173776388168, "learning_rate": 1.8774442747294407e-05, "loss": 0.009, "step": 21870 }, { "epoch": 23.15343915343915, "grad_norm": 0.12394174933433533, "learning_rate": 1.8731415519974965e-05, "loss": 0.0078, "step": 21880 }, { "epoch": 23.164021164021165, "grad_norm": 0.09202061593532562, "learning_rate": 1.86884262869341e-05, "loss": 0.0085, "step": 21890 }, { "epoch": 23.174603174603174, "grad_norm": 0.12944495677947998, "learning_rate": 1.8645475100407788e-05, "loss": 0.0105, "step": 21900 }, { "epoch": 23.185185185185187, "grad_norm": 0.08850299566984177, "learning_rate": 1.8602562012585768e-05, "loss": 0.0084, "step": 21910 }, { "epoch": 23.195767195767196, "grad_norm": 0.12853819131851196, "learning_rate": 1.8559687075611466e-05, "loss": 0.0092, "step": 21920 }, { "epoch": 23.206349206349206, "grad_norm": 0.13262562453746796, "learning_rate": 1.8516850341582017e-05, "loss": 0.0102, "step": 21930 }, { "epoch": 23.21693121693122, "grad_norm": 0.10895363241434097, "learning_rate": 1.847405186254807e-05, "loss": 0.0089, "step": 21940 }, { "epoch": 23.227513227513228, "grad_norm": 0.19268310070037842, "learning_rate": 1.843129169051379e-05, "loss": 0.0078, "step": 21950 }, { "epoch": 23.238095238095237, "grad_norm": 0.09230136126279831, "learning_rate": 1.8388569877436863e-05, "loss": 0.0092, "step": 21960 }, { "epoch": 23.24867724867725, "grad_norm": 0.1629510372877121, "learning_rate": 1.834588647522828e-05, "loss": 0.0089, "step": 21970 }, { "epoch": 23.25925925925926, "grad_norm": 0.10990961641073227, "learning_rate": 1.8303241535752435e-05, "loss": 0.0095, "step": 21980 }, { "epoch": 23.26984126984127, "grad_norm": 0.14406901597976685, "learning_rate": 1.8260635110826934e-05, "loss": 0.0103, "step": 21990 }, { "epoch": 23.280423280423282, "grad_norm": 0.1172713115811348, "learning_rate": 1.8218067252222597e-05, "loss": 0.0092, "step": 22000 }, { "epoch": 23.29100529100529, "grad_norm": 0.20602647960186005, "learning_rate": 1.817553801166339e-05, "loss": 0.0075, "step": 22010 }, { "epoch": 23.3015873015873, "grad_norm": 0.1463833600282669, "learning_rate": 1.8133047440826333e-05, "loss": 0.0106, "step": 22020 }, { "epoch": 23.312169312169313, "grad_norm": 0.10298803448677063, "learning_rate": 1.809059559134151e-05, "loss": 0.0087, "step": 22030 }, { "epoch": 23.322751322751323, "grad_norm": 0.10650307685136795, "learning_rate": 1.80481825147919e-05, "loss": 0.0082, "step": 22040 }, { "epoch": 23.333333333333332, "grad_norm": 0.11293243616819382, "learning_rate": 1.8005808262713396e-05, "loss": 0.0082, "step": 22050 }, { "epoch": 23.343915343915345, "grad_norm": 0.17427295446395874, "learning_rate": 1.7963472886594714e-05, "loss": 0.0102, "step": 22060 }, { "epoch": 23.354497354497354, "grad_norm": 0.1659996658563614, "learning_rate": 1.7921176437877302e-05, "loss": 0.0072, "step": 22070 }, { "epoch": 23.365079365079364, "grad_norm": 0.09650042653083801, "learning_rate": 1.7878918967955366e-05, "loss": 0.0086, "step": 22080 }, { "epoch": 23.375661375661377, "grad_norm": 0.18384364247322083, "learning_rate": 1.7836700528175693e-05, "loss": 0.0084, "step": 22090 }, { "epoch": 23.386243386243386, "grad_norm": 0.15955223143100739, "learning_rate": 1.779452116983769e-05, "loss": 0.009, "step": 22100 }, { "epoch": 23.396825396825395, "grad_norm": 0.1251775473356247, "learning_rate": 1.775238094419325e-05, "loss": 0.0095, "step": 22110 }, { "epoch": 23.40740740740741, "grad_norm": 0.10237053036689758, "learning_rate": 1.7710279902446708e-05, "loss": 0.0072, "step": 22120 }, { "epoch": 23.417989417989418, "grad_norm": 0.10019965469837189, "learning_rate": 1.7668218095754795e-05, "loss": 0.0105, "step": 22130 }, { "epoch": 23.428571428571427, "grad_norm": 0.1537276655435562, "learning_rate": 1.7626195575226594e-05, "loss": 0.0093, "step": 22140 }, { "epoch": 23.43915343915344, "grad_norm": 0.1856215000152588, "learning_rate": 1.758421239192343e-05, "loss": 0.0096, "step": 22150 }, { "epoch": 23.44973544973545, "grad_norm": 0.08900201320648193, "learning_rate": 1.7542268596858814e-05, "loss": 0.0112, "step": 22160 }, { "epoch": 23.46031746031746, "grad_norm": 0.1528773009777069, "learning_rate": 1.750036424099841e-05, "loss": 0.0091, "step": 22170 }, { "epoch": 23.47089947089947, "grad_norm": 0.09785521030426025, "learning_rate": 1.7458499375259956e-05, "loss": 0.0079, "step": 22180 }, { "epoch": 23.48148148148148, "grad_norm": 0.0643005222082138, "learning_rate": 1.7416674050513245e-05, "loss": 0.0087, "step": 22190 }, { "epoch": 23.49206349206349, "grad_norm": 0.08323578536510468, "learning_rate": 1.737488831757997e-05, "loss": 0.009, "step": 22200 }, { "epoch": 23.502645502645503, "grad_norm": 0.1483549326658249, "learning_rate": 1.733314222723373e-05, "loss": 0.0096, "step": 22210 }, { "epoch": 23.513227513227513, "grad_norm": 0.19128504395484924, "learning_rate": 1.7291435830199954e-05, "loss": 0.009, "step": 22220 }, { "epoch": 23.523809523809526, "grad_norm": 0.08231422305107117, "learning_rate": 1.7249769177155878e-05, "loss": 0.0078, "step": 22230 }, { "epoch": 23.534391534391535, "grad_norm": 0.1427789032459259, "learning_rate": 1.720814231873038e-05, "loss": 0.008, "step": 22240 }, { "epoch": 23.544973544973544, "grad_norm": 0.1707647740840912, "learning_rate": 1.716655530550405e-05, "loss": 0.0078, "step": 22250 }, { "epoch": 23.555555555555557, "grad_norm": 0.1429567039012909, "learning_rate": 1.7125008188009017e-05, "loss": 0.0086, "step": 22260 }, { "epoch": 23.566137566137566, "grad_norm": 0.06432969868183136, "learning_rate": 1.708350101672894e-05, "loss": 0.0072, "step": 22270 }, { "epoch": 23.576719576719576, "grad_norm": 0.16359278559684753, "learning_rate": 1.7042033842098957e-05, "loss": 0.0071, "step": 22280 }, { "epoch": 23.58730158730159, "grad_norm": 0.08053523302078247, "learning_rate": 1.7000606714505567e-05, "loss": 0.0085, "step": 22290 }, { "epoch": 23.597883597883598, "grad_norm": 0.09239017963409424, "learning_rate": 1.695921968428668e-05, "loss": 0.0101, "step": 22300 }, { "epoch": 23.608465608465607, "grad_norm": 0.11347002536058426, "learning_rate": 1.6917872801731416e-05, "loss": 0.009, "step": 22310 }, { "epoch": 23.61904761904762, "grad_norm": 0.09693965315818787, "learning_rate": 1.687656611708014e-05, "loss": 0.008, "step": 22320 }, { "epoch": 23.62962962962963, "grad_norm": 0.10878345370292664, "learning_rate": 1.6835299680524368e-05, "loss": 0.0095, "step": 22330 }, { "epoch": 23.64021164021164, "grad_norm": 0.10494758188724518, "learning_rate": 1.679407354220669e-05, "loss": 0.0087, "step": 22340 }, { "epoch": 23.650793650793652, "grad_norm": 0.11821960657835007, "learning_rate": 1.6752887752220793e-05, "loss": 0.0071, "step": 22350 }, { "epoch": 23.66137566137566, "grad_norm": 0.3137259781360626, "learning_rate": 1.6711742360611277e-05, "loss": 0.0091, "step": 22360 }, { "epoch": 23.67195767195767, "grad_norm": 0.131147101521492, "learning_rate": 1.6670637417373652e-05, "loss": 0.0092, "step": 22370 }, { "epoch": 23.682539682539684, "grad_norm": 0.18390606343746185, "learning_rate": 1.6629572972454333e-05, "loss": 0.0097, "step": 22380 }, { "epoch": 23.693121693121693, "grad_norm": 0.11296270042657852, "learning_rate": 1.6588549075750464e-05, "loss": 0.0069, "step": 22390 }, { "epoch": 23.703703703703702, "grad_norm": 0.12147419154644012, "learning_rate": 1.6547565777109976e-05, "loss": 0.0098, "step": 22400 }, { "epoch": 23.714285714285715, "grad_norm": 0.06756311655044556, "learning_rate": 1.650662312633143e-05, "loss": 0.0076, "step": 22410 }, { "epoch": 23.724867724867725, "grad_norm": 0.11824812740087509, "learning_rate": 1.6465721173164002e-05, "loss": 0.0092, "step": 22420 }, { "epoch": 23.735449735449734, "grad_norm": 0.16397972404956818, "learning_rate": 1.6424859967307428e-05, "loss": 0.0083, "step": 22430 }, { "epoch": 23.746031746031747, "grad_norm": 0.10996085405349731, "learning_rate": 1.6384039558411903e-05, "loss": 0.0097, "step": 22440 }, { "epoch": 23.756613756613756, "grad_norm": 0.11501927673816681, "learning_rate": 1.634325999607811e-05, "loss": 0.0077, "step": 22450 }, { "epoch": 23.767195767195766, "grad_norm": 0.07987785339355469, "learning_rate": 1.6302521329857045e-05, "loss": 0.0086, "step": 22460 }, { "epoch": 23.77777777777778, "grad_norm": 0.15657928586006165, "learning_rate": 1.6261823609250027e-05, "loss": 0.0068, "step": 22470 }, { "epoch": 23.788359788359788, "grad_norm": 0.19225220382213593, "learning_rate": 1.6221166883708627e-05, "loss": 0.0092, "step": 22480 }, { "epoch": 23.798941798941797, "grad_norm": 0.11999720335006714, "learning_rate": 1.6180551202634603e-05, "loss": 0.0077, "step": 22490 }, { "epoch": 23.80952380952381, "grad_norm": 0.19638560712337494, "learning_rate": 1.613997661537981e-05, "loss": 0.0095, "step": 22500 }, { "epoch": 23.82010582010582, "grad_norm": 0.08586866408586502, "learning_rate": 1.609944317124624e-05, "loss": 0.0082, "step": 22510 }, { "epoch": 23.83068783068783, "grad_norm": 0.10672365874052048, "learning_rate": 1.605895091948582e-05, "loss": 0.008, "step": 22520 }, { "epoch": 23.841269841269842, "grad_norm": 0.13844750821590424, "learning_rate": 1.6018499909300477e-05, "loss": 0.0092, "step": 22530 }, { "epoch": 23.85185185185185, "grad_norm": 0.11444062739610672, "learning_rate": 1.597809018984199e-05, "loss": 0.009, "step": 22540 }, { "epoch": 23.862433862433864, "grad_norm": 0.13176864385604858, "learning_rate": 1.593772181021196e-05, "loss": 0.0089, "step": 22550 }, { "epoch": 23.873015873015873, "grad_norm": 0.08721797913312912, "learning_rate": 1.5897394819461813e-05, "loss": 0.0062, "step": 22560 }, { "epoch": 23.883597883597883, "grad_norm": 0.09307023137807846, "learning_rate": 1.5857109266592625e-05, "loss": 0.0099, "step": 22570 }, { "epoch": 23.894179894179896, "grad_norm": 0.1920469105243683, "learning_rate": 1.581686520055514e-05, "loss": 0.009, "step": 22580 }, { "epoch": 23.904761904761905, "grad_norm": 0.40649351477622986, "learning_rate": 1.5776662670249704e-05, "loss": 0.0099, "step": 22590 }, { "epoch": 23.915343915343914, "grad_norm": 0.2220049649477005, "learning_rate": 1.573650172452615e-05, "loss": 0.0089, "step": 22600 }, { "epoch": 23.925925925925927, "grad_norm": 0.06975656002759933, "learning_rate": 1.5696382412183852e-05, "loss": 0.0098, "step": 22610 }, { "epoch": 23.936507936507937, "grad_norm": 0.11213213950395584, "learning_rate": 1.565630478197155e-05, "loss": 0.0082, "step": 22620 }, { "epoch": 23.947089947089946, "grad_norm": 0.13478964567184448, "learning_rate": 1.5616268882587332e-05, "loss": 0.0097, "step": 22630 }, { "epoch": 23.95767195767196, "grad_norm": 0.07301664352416992, "learning_rate": 1.5576274762678594e-05, "loss": 0.0072, "step": 22640 }, { "epoch": 23.96825396825397, "grad_norm": 0.10476706922054291, "learning_rate": 1.5536322470841952e-05, "loss": 0.0077, "step": 22650 }, { "epoch": 23.978835978835978, "grad_norm": 0.08319173753261566, "learning_rate": 1.549641205562324e-05, "loss": 0.0084, "step": 22660 }, { "epoch": 23.98941798941799, "grad_norm": 0.15328668057918549, "learning_rate": 1.5456543565517334e-05, "loss": 0.0108, "step": 22670 }, { "epoch": 24.0, "grad_norm": 0.07879733294248581, "learning_rate": 1.541671704896825e-05, "loss": 0.007, "step": 22680 }, { "epoch": 24.01058201058201, "grad_norm": 0.09151367098093033, "learning_rate": 1.5376932554368938e-05, "loss": 0.0087, "step": 22690 }, { "epoch": 24.021164021164022, "grad_norm": 0.07005073130130768, "learning_rate": 1.5337190130061318e-05, "loss": 0.0101, "step": 22700 }, { "epoch": 24.03174603174603, "grad_norm": 0.15707571804523468, "learning_rate": 1.5297489824336165e-05, "loss": 0.0094, "step": 22710 }, { "epoch": 24.04232804232804, "grad_norm": 0.14298252761363983, "learning_rate": 1.5257831685433121e-05, "loss": 0.0089, "step": 22720 }, { "epoch": 24.052910052910054, "grad_norm": 0.1329488754272461, "learning_rate": 1.521821576154055e-05, "loss": 0.0083, "step": 22730 }, { "epoch": 24.063492063492063, "grad_norm": 0.09882708638906479, "learning_rate": 1.5178642100795542e-05, "loss": 0.0079, "step": 22740 }, { "epoch": 24.074074074074073, "grad_norm": 0.23429663479328156, "learning_rate": 1.5139110751283819e-05, "loss": 0.0097, "step": 22750 }, { "epoch": 24.084656084656086, "grad_norm": 0.13513536751270294, "learning_rate": 1.5099621761039683e-05, "loss": 0.0081, "step": 22760 }, { "epoch": 24.095238095238095, "grad_norm": 0.0948987826704979, "learning_rate": 1.5060175178046016e-05, "loss": 0.0086, "step": 22770 }, { "epoch": 24.105820105820104, "grad_norm": 0.05198455974459648, "learning_rate": 1.5020771050234117e-05, "loss": 0.0091, "step": 22780 }, { "epoch": 24.116402116402117, "grad_norm": 0.14139533042907715, "learning_rate": 1.4981409425483716e-05, "loss": 0.0076, "step": 22790 }, { "epoch": 24.126984126984127, "grad_norm": 0.12896257638931274, "learning_rate": 1.4942090351622883e-05, "loss": 0.0073, "step": 22800 }, { "epoch": 24.137566137566136, "grad_norm": 0.3182849884033203, "learning_rate": 1.490281387642804e-05, "loss": 0.0071, "step": 22810 }, { "epoch": 24.14814814814815, "grad_norm": 0.10398029536008835, "learning_rate": 1.4863580047623765e-05, "loss": 0.0081, "step": 22820 }, { "epoch": 24.158730158730158, "grad_norm": 0.18296407163143158, "learning_rate": 1.4824388912882897e-05, "loss": 0.0063, "step": 22830 }, { "epoch": 24.16931216931217, "grad_norm": 0.12393161654472351, "learning_rate": 1.4785240519826344e-05, "loss": 0.0061, "step": 22840 }, { "epoch": 24.17989417989418, "grad_norm": 0.10889433324337006, "learning_rate": 1.4746134916023097e-05, "loss": 0.0075, "step": 22850 }, { "epoch": 24.19047619047619, "grad_norm": 0.14776410162448883, "learning_rate": 1.4707072148990141e-05, "loss": 0.0083, "step": 22860 }, { "epoch": 24.201058201058203, "grad_norm": 0.13818365335464478, "learning_rate": 1.4668052266192423e-05, "loss": 0.0086, "step": 22870 }, { "epoch": 24.211640211640212, "grad_norm": 0.12552866339683533, "learning_rate": 1.4629075315042795e-05, "loss": 0.009, "step": 22880 }, { "epoch": 24.22222222222222, "grad_norm": 0.1002889946103096, "learning_rate": 1.4590141342901925e-05, "loss": 0.0079, "step": 22890 }, { "epoch": 24.232804232804234, "grad_norm": 0.11359041184186935, "learning_rate": 1.4551250397078253e-05, "loss": 0.009, "step": 22900 }, { "epoch": 24.243386243386244, "grad_norm": 0.15767860412597656, "learning_rate": 1.4512402524827945e-05, "loss": 0.0077, "step": 22910 }, { "epoch": 24.253968253968253, "grad_norm": 0.1263275444507599, "learning_rate": 1.447359777335482e-05, "loss": 0.0095, "step": 22920 }, { "epoch": 24.264550264550266, "grad_norm": 0.0836203396320343, "learning_rate": 1.4434836189810335e-05, "loss": 0.0074, "step": 22930 }, { "epoch": 24.275132275132275, "grad_norm": 0.21757566928863525, "learning_rate": 1.4396117821293453e-05, "loss": 0.0085, "step": 22940 }, { "epoch": 24.285714285714285, "grad_norm": 0.10634062439203262, "learning_rate": 1.4357442714850633e-05, "loss": 0.0051, "step": 22950 }, { "epoch": 24.296296296296298, "grad_norm": 0.15618447959423065, "learning_rate": 1.43188109174758e-05, "loss": 0.009, "step": 22960 }, { "epoch": 24.306878306878307, "grad_norm": 0.12954957783222198, "learning_rate": 1.4280222476110206e-05, "loss": 0.0084, "step": 22970 }, { "epoch": 24.317460317460316, "grad_norm": 0.11990432441234589, "learning_rate": 1.4241677437642469e-05, "loss": 0.0091, "step": 22980 }, { "epoch": 24.32804232804233, "grad_norm": 0.09490177035331726, "learning_rate": 1.420317584890844e-05, "loss": 0.0077, "step": 22990 }, { "epoch": 24.33862433862434, "grad_norm": 0.07976845651865005, "learning_rate": 1.4164717756691176e-05, "loss": 0.0081, "step": 23000 }, { "epoch": 24.349206349206348, "grad_norm": 0.3149688243865967, "learning_rate": 1.4126303207720882e-05, "loss": 0.0069, "step": 23010 }, { "epoch": 24.35978835978836, "grad_norm": 0.148951917886734, "learning_rate": 1.408793224867484e-05, "loss": 0.0077, "step": 23020 }, { "epoch": 24.37037037037037, "grad_norm": 0.10653408616781235, "learning_rate": 1.4049604926177423e-05, "loss": 0.0083, "step": 23030 }, { "epoch": 24.38095238095238, "grad_norm": 0.09416577965021133, "learning_rate": 1.4011321286799916e-05, "loss": 0.0079, "step": 23040 }, { "epoch": 24.391534391534393, "grad_norm": 0.11519728600978851, "learning_rate": 1.3973081377060565e-05, "loss": 0.0069, "step": 23050 }, { "epoch": 24.402116402116402, "grad_norm": 0.11025872826576233, "learning_rate": 1.393488524342445e-05, "loss": 0.0107, "step": 23060 }, { "epoch": 24.41269841269841, "grad_norm": 0.09284371882677078, "learning_rate": 1.3896732932303486e-05, "loss": 0.0084, "step": 23070 }, { "epoch": 24.423280423280424, "grad_norm": 0.1493648737668991, "learning_rate": 1.3858624490056304e-05, "loss": 0.0078, "step": 23080 }, { "epoch": 24.433862433862434, "grad_norm": 0.13188910484313965, "learning_rate": 1.3820559962988299e-05, "loss": 0.0073, "step": 23090 }, { "epoch": 24.444444444444443, "grad_norm": 0.15770234167575836, "learning_rate": 1.3782539397351418e-05, "loss": 0.0097, "step": 23100 }, { "epoch": 24.455026455026456, "grad_norm": 0.1350797712802887, "learning_rate": 1.3744562839344266e-05, "loss": 0.0086, "step": 23110 }, { "epoch": 24.465608465608465, "grad_norm": 0.12009082734584808, "learning_rate": 1.370663033511193e-05, "loss": 0.0089, "step": 23120 }, { "epoch": 24.476190476190474, "grad_norm": 0.10749568790197372, "learning_rate": 1.3668741930745964e-05, "loss": 0.0098, "step": 23130 }, { "epoch": 24.486772486772487, "grad_norm": 0.2077810913324356, "learning_rate": 1.3630897672284382e-05, "loss": 0.0084, "step": 23140 }, { "epoch": 24.497354497354497, "grad_norm": 0.19502514600753784, "learning_rate": 1.3593097605711508e-05, "loss": 0.0081, "step": 23150 }, { "epoch": 24.507936507936506, "grad_norm": 0.09282442927360535, "learning_rate": 1.3555341776957992e-05, "loss": 0.0077, "step": 23160 }, { "epoch": 24.51851851851852, "grad_norm": 0.1204688623547554, "learning_rate": 1.3517630231900724e-05, "loss": 0.0087, "step": 23170 }, { "epoch": 24.52910052910053, "grad_norm": 0.1875302791595459, "learning_rate": 1.3479963016362767e-05, "loss": 0.0103, "step": 23180 }, { "epoch": 24.53968253968254, "grad_norm": 0.15209771692752838, "learning_rate": 1.3442340176113377e-05, "loss": 0.0084, "step": 23190 }, { "epoch": 24.55026455026455, "grad_norm": 0.08733559399843216, "learning_rate": 1.340476175686784e-05, "loss": 0.0089, "step": 23200 }, { "epoch": 24.56084656084656, "grad_norm": 0.11664913594722748, "learning_rate": 1.3367227804287469e-05, "loss": 0.0062, "step": 23210 }, { "epoch": 24.571428571428573, "grad_norm": 0.13095727562904358, "learning_rate": 1.332973836397956e-05, "loss": 0.0096, "step": 23220 }, { "epoch": 24.582010582010582, "grad_norm": 0.12970200181007385, "learning_rate": 1.329229348149731e-05, "loss": 0.0083, "step": 23230 }, { "epoch": 24.59259259259259, "grad_norm": 0.10731661319732666, "learning_rate": 1.3254893202339796e-05, "loss": 0.0073, "step": 23240 }, { "epoch": 24.603174603174605, "grad_norm": 0.21344631910324097, "learning_rate": 1.3217537571951872e-05, "loss": 0.0089, "step": 23250 }, { "epoch": 24.613756613756614, "grad_norm": 0.2856588363647461, "learning_rate": 1.3180226635724169e-05, "loss": 0.0086, "step": 23260 }, { "epoch": 24.624338624338623, "grad_norm": 0.0931103527545929, "learning_rate": 1.314296043899298e-05, "loss": 0.01, "step": 23270 }, { "epoch": 24.634920634920636, "grad_norm": 0.12638084590435028, "learning_rate": 1.3105739027040248e-05, "loss": 0.0106, "step": 23280 }, { "epoch": 24.645502645502646, "grad_norm": 0.14470450580120087, "learning_rate": 1.3068562445093491e-05, "loss": 0.009, "step": 23290 }, { "epoch": 24.656084656084655, "grad_norm": 0.12969708442687988, "learning_rate": 1.3031430738325778e-05, "loss": 0.0062, "step": 23300 }, { "epoch": 24.666666666666668, "grad_norm": 0.10755819082260132, "learning_rate": 1.299434395185563e-05, "loss": 0.0088, "step": 23310 }, { "epoch": 24.677248677248677, "grad_norm": 0.3030003607273102, "learning_rate": 1.295730213074699e-05, "loss": 0.0078, "step": 23320 }, { "epoch": 24.687830687830687, "grad_norm": 0.11896342039108276, "learning_rate": 1.2920305320009152e-05, "loss": 0.0086, "step": 23330 }, { "epoch": 24.6984126984127, "grad_norm": 0.0928819477558136, "learning_rate": 1.2883353564596729e-05, "loss": 0.0073, "step": 23340 }, { "epoch": 24.70899470899471, "grad_norm": 0.1564657837152481, "learning_rate": 1.2846446909409604e-05, "loss": 0.0069, "step": 23350 }, { "epoch": 24.719576719576718, "grad_norm": 0.15377895534038544, "learning_rate": 1.280958539929284e-05, "loss": 0.0078, "step": 23360 }, { "epoch": 24.73015873015873, "grad_norm": 0.11717280000448227, "learning_rate": 1.277276907903664e-05, "loss": 0.0099, "step": 23370 }, { "epoch": 24.74074074074074, "grad_norm": 0.1957828849554062, "learning_rate": 1.2735997993376298e-05, "loss": 0.0079, "step": 23380 }, { "epoch": 24.75132275132275, "grad_norm": 0.15485277771949768, "learning_rate": 1.2699272186992166e-05, "loss": 0.0091, "step": 23390 }, { "epoch": 24.761904761904763, "grad_norm": 0.13470767438411713, "learning_rate": 1.2662591704509547e-05, "loss": 0.0053, "step": 23400 }, { "epoch": 24.772486772486772, "grad_norm": 0.10606607049703598, "learning_rate": 1.2625956590498711e-05, "loss": 0.0071, "step": 23410 }, { "epoch": 24.78306878306878, "grad_norm": 0.15617190301418304, "learning_rate": 1.2589366889474757e-05, "loss": 0.0084, "step": 23420 }, { "epoch": 24.793650793650794, "grad_norm": 0.11464614421129227, "learning_rate": 1.2552822645897621e-05, "loss": 0.0086, "step": 23430 }, { "epoch": 24.804232804232804, "grad_norm": 0.07316923141479492, "learning_rate": 1.2516323904172e-05, "loss": 0.0062, "step": 23440 }, { "epoch": 24.814814814814813, "grad_norm": 0.08291611075401306, "learning_rate": 1.2479870708647324e-05, "loss": 0.0078, "step": 23450 }, { "epoch": 24.825396825396826, "grad_norm": 0.09908267855644226, "learning_rate": 1.2443463103617658e-05, "loss": 0.0097, "step": 23460 }, { "epoch": 24.835978835978835, "grad_norm": 0.11848761886358261, "learning_rate": 1.240710113332167e-05, "loss": 0.0098, "step": 23470 }, { "epoch": 24.84656084656085, "grad_norm": 0.2525632083415985, "learning_rate": 1.2370784841942584e-05, "loss": 0.0092, "step": 23480 }, { "epoch": 24.857142857142858, "grad_norm": 0.10893790423870087, "learning_rate": 1.2334514273608117e-05, "loss": 0.0076, "step": 23490 }, { "epoch": 24.867724867724867, "grad_norm": 0.1123809888958931, "learning_rate": 1.2298289472390417e-05, "loss": 0.0074, "step": 23500 }, { "epoch": 24.87830687830688, "grad_norm": 0.17489147186279297, "learning_rate": 1.226211048230606e-05, "loss": 0.0083, "step": 23510 }, { "epoch": 24.88888888888889, "grad_norm": 0.08032778650522232, "learning_rate": 1.2225977347315921e-05, "loss": 0.0066, "step": 23520 }, { "epoch": 24.8994708994709, "grad_norm": 0.1482192426919937, "learning_rate": 1.2189890111325148e-05, "loss": 0.0072, "step": 23530 }, { "epoch": 24.91005291005291, "grad_norm": 0.099896140396595, "learning_rate": 1.2153848818183161e-05, "loss": 0.0086, "step": 23540 }, { "epoch": 24.92063492063492, "grad_norm": 0.4076886475086212, "learning_rate": 1.2117853511683507e-05, "loss": 0.0077, "step": 23550 }, { "epoch": 24.93121693121693, "grad_norm": 0.13023807108402252, "learning_rate": 1.2081904235563906e-05, "loss": 0.0079, "step": 23560 }, { "epoch": 24.941798941798943, "grad_norm": 0.22397413849830627, "learning_rate": 1.20460010335061e-05, "loss": 0.0087, "step": 23570 }, { "epoch": 24.952380952380953, "grad_norm": 0.11843026429414749, "learning_rate": 1.2010143949135865e-05, "loss": 0.0089, "step": 23580 }, { "epoch": 24.962962962962962, "grad_norm": 0.0916251540184021, "learning_rate": 1.1974333026022938e-05, "loss": 0.0096, "step": 23590 }, { "epoch": 24.973544973544975, "grad_norm": 0.11127282679080963, "learning_rate": 1.1938568307680964e-05, "loss": 0.0076, "step": 23600 }, { "epoch": 24.984126984126984, "grad_norm": 0.07343287020921707, "learning_rate": 1.1902849837567464e-05, "loss": 0.0095, "step": 23610 }, { "epoch": 24.994708994708994, "grad_norm": 0.1815868765115738, "learning_rate": 1.1867177659083739e-05, "loss": 0.0079, "step": 23620 }, { "epoch": 25.005291005291006, "grad_norm": 0.07687604427337646, "learning_rate": 1.1831551815574848e-05, "loss": 0.0077, "step": 23630 }, { "epoch": 25.015873015873016, "grad_norm": 0.10051486641168594, "learning_rate": 1.1795972350329554e-05, "loss": 0.0072, "step": 23640 }, { "epoch": 25.026455026455025, "grad_norm": 0.10156408697366714, "learning_rate": 1.176043930658024e-05, "loss": 0.0064, "step": 23650 }, { "epoch": 25.037037037037038, "grad_norm": 0.08264391869306564, "learning_rate": 1.1724952727502942e-05, "loss": 0.0079, "step": 23660 }, { "epoch": 25.047619047619047, "grad_norm": 0.14298294484615326, "learning_rate": 1.1689512656217178e-05, "loss": 0.0093, "step": 23670 }, { "epoch": 25.058201058201057, "grad_norm": 0.130076602101326, "learning_rate": 1.1654119135785962e-05, "loss": 0.0091, "step": 23680 }, { "epoch": 25.06878306878307, "grad_norm": 0.08435402810573578, "learning_rate": 1.1618772209215794e-05, "loss": 0.0082, "step": 23690 }, { "epoch": 25.07936507936508, "grad_norm": 0.10305050015449524, "learning_rate": 1.1583471919456506e-05, "loss": 0.0075, "step": 23700 }, { "epoch": 25.08994708994709, "grad_norm": 0.21026301383972168, "learning_rate": 1.1548218309401266e-05, "loss": 0.0079, "step": 23710 }, { "epoch": 25.1005291005291, "grad_norm": 0.11457672715187073, "learning_rate": 1.1513011421886555e-05, "loss": 0.0109, "step": 23720 }, { "epoch": 25.11111111111111, "grad_norm": 0.1184048131108284, "learning_rate": 1.1477851299692056e-05, "loss": 0.0069, "step": 23730 }, { "epoch": 25.12169312169312, "grad_norm": 0.09511419385671616, "learning_rate": 1.144273798554063e-05, "loss": 0.0076, "step": 23740 }, { "epoch": 25.132275132275133, "grad_norm": 0.13938738405704498, "learning_rate": 1.1407671522098262e-05, "loss": 0.0099, "step": 23750 }, { "epoch": 25.142857142857142, "grad_norm": 0.09536655992269516, "learning_rate": 1.1372651951974e-05, "loss": 0.0078, "step": 23760 }, { "epoch": 25.15343915343915, "grad_norm": 0.14232014119625092, "learning_rate": 1.1337679317719952e-05, "loss": 0.0075, "step": 23770 }, { "epoch": 25.164021164021165, "grad_norm": 0.1213424950838089, "learning_rate": 1.1302753661831151e-05, "loss": 0.0115, "step": 23780 }, { "epoch": 25.174603174603174, "grad_norm": 0.06804095953702927, "learning_rate": 1.1267875026745561e-05, "loss": 0.0077, "step": 23790 }, { "epoch": 25.185185185185187, "grad_norm": 0.06279613077640533, "learning_rate": 1.1233043454844016e-05, "loss": 0.0058, "step": 23800 }, { "epoch": 25.195767195767196, "grad_norm": 0.17191492021083832, "learning_rate": 1.1198258988450144e-05, "loss": 0.0086, "step": 23810 }, { "epoch": 25.206349206349206, "grad_norm": 0.20736509561538696, "learning_rate": 1.1163521669830368e-05, "loss": 0.0085, "step": 23820 }, { "epoch": 25.21693121693122, "grad_norm": 0.13537192344665527, "learning_rate": 1.1128831541193818e-05, "loss": 0.0085, "step": 23830 }, { "epoch": 25.227513227513228, "grad_norm": 0.2603134512901306, "learning_rate": 1.1094188644692255e-05, "loss": 0.0078, "step": 23840 }, { "epoch": 25.238095238095237, "grad_norm": 0.1426069587469101, "learning_rate": 1.1059593022420067e-05, "loss": 0.0084, "step": 23850 }, { "epoch": 25.24867724867725, "grad_norm": 0.10020212084054947, "learning_rate": 1.1025044716414184e-05, "loss": 0.0073, "step": 23860 }, { "epoch": 25.25925925925926, "grad_norm": 0.14106258749961853, "learning_rate": 1.0990543768654082e-05, "loss": 0.007, "step": 23870 }, { "epoch": 25.26984126984127, "grad_norm": 0.10436101257801056, "learning_rate": 1.0956090221061655e-05, "loss": 0.0068, "step": 23880 }, { "epoch": 25.280423280423282, "grad_norm": 0.10664622485637665, "learning_rate": 1.0921684115501208e-05, "loss": 0.0075, "step": 23890 }, { "epoch": 25.29100529100529, "grad_norm": 0.07347920536994934, "learning_rate": 1.0887325493779405e-05, "loss": 0.0066, "step": 23900 }, { "epoch": 25.3015873015873, "grad_norm": 0.15668141841888428, "learning_rate": 1.0853014397645207e-05, "loss": 0.009, "step": 23910 }, { "epoch": 25.312169312169313, "grad_norm": 0.21209704875946045, "learning_rate": 1.0818750868789829e-05, "loss": 0.008, "step": 23920 }, { "epoch": 25.322751322751323, "grad_norm": 0.22931601107120514, "learning_rate": 1.0784534948846703e-05, "loss": 0.0082, "step": 23930 }, { "epoch": 25.333333333333332, "grad_norm": 0.12197685241699219, "learning_rate": 1.0750366679391393e-05, "loss": 0.0075, "step": 23940 }, { "epoch": 25.343915343915345, "grad_norm": 0.1369449943304062, "learning_rate": 1.0716246101941557e-05, "loss": 0.0086, "step": 23950 }, { "epoch": 25.354497354497354, "grad_norm": 0.09662582725286484, "learning_rate": 1.0682173257956934e-05, "loss": 0.0092, "step": 23960 }, { "epoch": 25.365079365079364, "grad_norm": 0.1498357504606247, "learning_rate": 1.064814818883922e-05, "loss": 0.0098, "step": 23970 }, { "epoch": 25.375661375661377, "grad_norm": 0.1009807139635086, "learning_rate": 1.06141709359321e-05, "loss": 0.0063, "step": 23980 }, { "epoch": 25.386243386243386, "grad_norm": 0.0933009535074234, "learning_rate": 1.058024154052114e-05, "loss": 0.0085, "step": 23990 }, { "epoch": 25.396825396825395, "grad_norm": 0.1485666036605835, "learning_rate": 1.0546360043833747e-05, "loss": 0.0069, "step": 24000 }, { "epoch": 25.40740740740741, "grad_norm": 0.10986007750034332, "learning_rate": 1.0512526487039138e-05, "loss": 0.0082, "step": 24010 }, { "epoch": 25.417989417989418, "grad_norm": 0.10271146893501282, "learning_rate": 1.0478740911248258e-05, "loss": 0.0075, "step": 24020 }, { "epoch": 25.428571428571427, "grad_norm": 0.10578035563230515, "learning_rate": 1.0445003357513793e-05, "loss": 0.0067, "step": 24030 }, { "epoch": 25.43915343915344, "grad_norm": 0.31737303733825684, "learning_rate": 1.0411313866830042e-05, "loss": 0.0077, "step": 24040 }, { "epoch": 25.44973544973545, "grad_norm": 0.074001744389534, "learning_rate": 1.0377672480132915e-05, "loss": 0.0081, "step": 24050 }, { "epoch": 25.46031746031746, "grad_norm": 0.11929246038198471, "learning_rate": 1.0344079238299864e-05, "loss": 0.005, "step": 24060 }, { "epoch": 25.47089947089947, "grad_norm": 0.13094021379947662, "learning_rate": 1.0310534182149833e-05, "loss": 0.0087, "step": 24070 }, { "epoch": 25.48148148148148, "grad_norm": 0.09435004740953445, "learning_rate": 1.0277037352443258e-05, "loss": 0.0078, "step": 24080 }, { "epoch": 25.49206349206349, "grad_norm": 0.10119498521089554, "learning_rate": 1.024358878988193e-05, "loss": 0.0067, "step": 24090 }, { "epoch": 25.502645502645503, "grad_norm": 0.1338646560907364, "learning_rate": 1.0210188535108995e-05, "loss": 0.0073, "step": 24100 }, { "epoch": 25.513227513227513, "grad_norm": 0.14772634208202362, "learning_rate": 1.0176836628708935e-05, "loss": 0.0088, "step": 24110 }, { "epoch": 25.523809523809526, "grad_norm": 0.2470131367444992, "learning_rate": 1.0143533111207454e-05, "loss": 0.0092, "step": 24120 }, { "epoch": 25.534391534391535, "grad_norm": 0.13493338227272034, "learning_rate": 1.0110278023071446e-05, "loss": 0.0073, "step": 24130 }, { "epoch": 25.544973544973544, "grad_norm": 0.16616863012313843, "learning_rate": 1.007707140470901e-05, "loss": 0.0078, "step": 24140 }, { "epoch": 25.555555555555557, "grad_norm": 0.09464598447084427, "learning_rate": 1.0043913296469298e-05, "loss": 0.0082, "step": 24150 }, { "epoch": 25.566137566137566, "grad_norm": 0.1316216140985489, "learning_rate": 1.001080373864255e-05, "loss": 0.0104, "step": 24160 }, { "epoch": 25.576719576719576, "grad_norm": 0.1551922708749771, "learning_rate": 9.977742771459991e-06, "loss": 0.0097, "step": 24170 }, { "epoch": 25.58730158730159, "grad_norm": 0.16884398460388184, "learning_rate": 9.944730435093802e-06, "loss": 0.0083, "step": 24180 }, { "epoch": 25.597883597883598, "grad_norm": 0.13197271525859833, "learning_rate": 9.911766769657115e-06, "loss": 0.0057, "step": 24190 }, { "epoch": 25.608465608465607, "grad_norm": 0.14048592746257782, "learning_rate": 9.878851815203882e-06, "loss": 0.0093, "step": 24200 }, { "epoch": 25.61904761904762, "grad_norm": 0.10269637405872345, "learning_rate": 9.845985611728864e-06, "loss": 0.0085, "step": 24210 }, { "epoch": 25.62962962962963, "grad_norm": 0.14330245554447174, "learning_rate": 9.813168199167606e-06, "loss": 0.0074, "step": 24220 }, { "epoch": 25.64021164021164, "grad_norm": 0.11029817909002304, "learning_rate": 9.780399617396341e-06, "loss": 0.0085, "step": 24230 }, { "epoch": 25.650793650793652, "grad_norm": 0.1149207353591919, "learning_rate": 9.747679906232016e-06, "loss": 0.0085, "step": 24240 }, { "epoch": 25.66137566137566, "grad_norm": 0.1127074658870697, "learning_rate": 9.715009105432138e-06, "loss": 0.0092, "step": 24250 }, { "epoch": 25.67195767195767, "grad_norm": 0.15734101831912994, "learning_rate": 9.682387254694835e-06, "loss": 0.0084, "step": 24260 }, { "epoch": 25.682539682539684, "grad_norm": 0.18902800977230072, "learning_rate": 9.649814393658723e-06, "loss": 0.0063, "step": 24270 }, { "epoch": 25.693121693121693, "grad_norm": 0.08239717781543732, "learning_rate": 9.617290561902881e-06, "loss": 0.0087, "step": 24280 }, { "epoch": 25.703703703703702, "grad_norm": 0.19560681283473969, "learning_rate": 9.584815798946861e-06, "loss": 0.007, "step": 24290 }, { "epoch": 25.714285714285715, "grad_norm": 0.08827908337116241, "learning_rate": 9.55239014425055e-06, "loss": 0.0084, "step": 24300 }, { "epoch": 25.724867724867725, "grad_norm": 0.23426872491836548, "learning_rate": 9.520013637214176e-06, "loss": 0.0097, "step": 24310 }, { "epoch": 25.735449735449734, "grad_norm": 0.13183319568634033, "learning_rate": 9.48768631717824e-06, "loss": 0.0075, "step": 24320 }, { "epoch": 25.746031746031747, "grad_norm": 0.12523230910301208, "learning_rate": 9.455408223423496e-06, "loss": 0.0077, "step": 24330 }, { "epoch": 25.756613756613756, "grad_norm": 0.18666675686836243, "learning_rate": 9.423179395170845e-06, "loss": 0.0074, "step": 24340 }, { "epoch": 25.767195767195766, "grad_norm": 0.21558308601379395, "learning_rate": 9.390999871581391e-06, "loss": 0.0078, "step": 24350 }, { "epoch": 25.77777777777778, "grad_norm": 0.08427365869283676, "learning_rate": 9.358869691756273e-06, "loss": 0.0097, "step": 24360 }, { "epoch": 25.788359788359788, "grad_norm": 0.18169133365154266, "learning_rate": 9.326788894736688e-06, "loss": 0.0082, "step": 24370 }, { "epoch": 25.798941798941797, "grad_norm": 0.18481160700321198, "learning_rate": 9.294757519503811e-06, "loss": 0.0074, "step": 24380 }, { "epoch": 25.80952380952381, "grad_norm": 0.05775134265422821, "learning_rate": 9.262775604978819e-06, "loss": 0.008, "step": 24390 }, { "epoch": 25.82010582010582, "grad_norm": 0.053017307072877884, "learning_rate": 9.230843190022725e-06, "loss": 0.0083, "step": 24400 }, { "epoch": 25.83068783068783, "grad_norm": 0.255342036485672, "learning_rate": 9.198960313436445e-06, "loss": 0.0086, "step": 24410 }, { "epoch": 25.841269841269842, "grad_norm": 0.04432034119963646, "learning_rate": 9.16712701396067e-06, "loss": 0.0074, "step": 24420 }, { "epoch": 25.85185185185185, "grad_norm": 0.17343959212303162, "learning_rate": 9.135343330275864e-06, "loss": 0.0084, "step": 24430 }, { "epoch": 25.862433862433864, "grad_norm": 0.0821344330906868, "learning_rate": 9.10360930100218e-06, "loss": 0.0081, "step": 24440 }, { "epoch": 25.873015873015873, "grad_norm": 0.16862758994102478, "learning_rate": 9.07192496469949e-06, "loss": 0.0104, "step": 24450 }, { "epoch": 25.883597883597883, "grad_norm": 0.08400041610002518, "learning_rate": 9.040290359867231e-06, "loss": 0.0072, "step": 24460 }, { "epoch": 25.894179894179896, "grad_norm": 0.07896804064512253, "learning_rate": 9.008705524944438e-06, "loss": 0.0059, "step": 24470 }, { "epoch": 25.904761904761905, "grad_norm": 0.09062916785478592, "learning_rate": 8.977170498309651e-06, "loss": 0.0072, "step": 24480 }, { "epoch": 25.915343915343914, "grad_norm": 0.11748789250850677, "learning_rate": 8.945685318280916e-06, "loss": 0.006, "step": 24490 }, { "epoch": 25.925925925925927, "grad_norm": 0.12092553824186325, "learning_rate": 8.914250023115672e-06, "loss": 0.0083, "step": 24500 }, { "epoch": 25.936507936507937, "grad_norm": 0.12476074695587158, "learning_rate": 8.882864651010798e-06, "loss": 0.0076, "step": 24510 }, { "epoch": 25.947089947089946, "grad_norm": 0.06381344795227051, "learning_rate": 8.851529240102462e-06, "loss": 0.0097, "step": 24520 }, { "epoch": 25.95767195767196, "grad_norm": 0.15390612185001373, "learning_rate": 8.820243828466134e-06, "loss": 0.0075, "step": 24530 }, { "epoch": 25.96825396825397, "grad_norm": 0.060380712151527405, "learning_rate": 8.789008454116566e-06, "loss": 0.0056, "step": 24540 }, { "epoch": 25.978835978835978, "grad_norm": 0.06130677089095116, "learning_rate": 8.757823155007655e-06, "loss": 0.007, "step": 24550 }, { "epoch": 25.98941798941799, "grad_norm": 0.1497763991355896, "learning_rate": 8.72668796903251e-06, "loss": 0.0075, "step": 24560 }, { "epoch": 26.0, "grad_norm": 0.1681707352399826, "learning_rate": 8.695602934023301e-06, "loss": 0.0064, "step": 24570 }, { "epoch": 26.01058201058201, "grad_norm": 0.09299781918525696, "learning_rate": 8.664568087751274e-06, "loss": 0.0093, "step": 24580 }, { "epoch": 26.021164021164022, "grad_norm": 0.10348882526159286, "learning_rate": 8.633583467926697e-06, "loss": 0.0085, "step": 24590 }, { "epoch": 26.03174603174603, "grad_norm": 0.13774070143699646, "learning_rate": 8.602649112198796e-06, "loss": 0.0075, "step": 24600 }, { "epoch": 26.04232804232804, "grad_norm": 0.1336725652217865, "learning_rate": 8.571765058155745e-06, "loss": 0.0083, "step": 24610 }, { "epoch": 26.052910052910054, "grad_norm": 0.08580826222896576, "learning_rate": 8.540931343324583e-06, "loss": 0.0061, "step": 24620 }, { "epoch": 26.063492063492063, "grad_norm": 0.1120038628578186, "learning_rate": 8.51014800517117e-06, "loss": 0.0084, "step": 24630 }, { "epoch": 26.074074074074073, "grad_norm": 0.09131010621786118, "learning_rate": 8.479415081100167e-06, "loss": 0.0088, "step": 24640 }, { "epoch": 26.084656084656086, "grad_norm": 0.07070837169885635, "learning_rate": 8.448732608454967e-06, "loss": 0.0068, "step": 24650 }, { "epoch": 26.095238095238095, "grad_norm": 0.12021137773990631, "learning_rate": 8.418100624517688e-06, "loss": 0.0068, "step": 24660 }, { "epoch": 26.105820105820104, "grad_norm": 0.19730687141418457, "learning_rate": 8.387519166509062e-06, "loss": 0.0092, "step": 24670 }, { "epoch": 26.116402116402117, "grad_norm": 0.1478320062160492, "learning_rate": 8.356988271588445e-06, "loss": 0.0058, "step": 24680 }, { "epoch": 26.126984126984127, "grad_norm": 0.17055165767669678, "learning_rate": 8.326507976853765e-06, "loss": 0.008, "step": 24690 }, { "epoch": 26.137566137566136, "grad_norm": 0.10331761837005615, "learning_rate": 8.296078319341443e-06, "loss": 0.0069, "step": 24700 }, { "epoch": 26.14814814814815, "grad_norm": 0.12712807953357697, "learning_rate": 8.265699336026383e-06, "loss": 0.0088, "step": 24710 }, { "epoch": 26.158730158730158, "grad_norm": 0.1525522917509079, "learning_rate": 8.235371063821923e-06, "loss": 0.0062, "step": 24720 }, { "epoch": 26.16931216931217, "grad_norm": 0.18665069341659546, "learning_rate": 8.205093539579767e-06, "loss": 0.006, "step": 24730 }, { "epoch": 26.17989417989418, "grad_norm": 0.1811317652463913, "learning_rate": 8.174866800089963e-06, "loss": 0.0086, "step": 24740 }, { "epoch": 26.19047619047619, "grad_norm": 0.14154858887195587, "learning_rate": 8.144690882080853e-06, "loss": 0.0093, "step": 24750 }, { "epoch": 26.201058201058203, "grad_norm": 0.09494614601135254, "learning_rate": 8.114565822219006e-06, "loss": 0.0076, "step": 24760 }, { "epoch": 26.211640211640212, "grad_norm": 0.08373267203569412, "learning_rate": 8.084491657109233e-06, "loss": 0.0075, "step": 24770 }, { "epoch": 26.22222222222222, "grad_norm": 0.12180037796497345, "learning_rate": 8.054468423294476e-06, "loss": 0.0082, "step": 24780 }, { "epoch": 26.232804232804234, "grad_norm": 0.2177240550518036, "learning_rate": 8.024496157255785e-06, "loss": 0.0076, "step": 24790 }, { "epoch": 26.243386243386244, "grad_norm": 0.08863195031881332, "learning_rate": 7.994574895412293e-06, "loss": 0.0077, "step": 24800 }, { "epoch": 26.253968253968253, "grad_norm": 0.10401099920272827, "learning_rate": 7.964704674121149e-06, "loss": 0.0077, "step": 24810 }, { "epoch": 26.264550264550266, "grad_norm": 0.1304955929517746, "learning_rate": 7.934885529677505e-06, "loss": 0.0074, "step": 24820 }, { "epoch": 26.275132275132275, "grad_norm": 0.07396501302719116, "learning_rate": 7.905117498314412e-06, "loss": 0.0056, "step": 24830 }, { "epoch": 26.285714285714285, "grad_norm": 0.21921367943286896, "learning_rate": 7.87540061620286e-06, "loss": 0.0086, "step": 24840 }, { "epoch": 26.296296296296298, "grad_norm": 0.07847380638122559, "learning_rate": 7.845734919451647e-06, "loss": 0.0057, "step": 24850 }, { "epoch": 26.306878306878307, "grad_norm": 0.08492527157068253, "learning_rate": 7.816120444107383e-06, "loss": 0.0084, "step": 24860 }, { "epoch": 26.317460317460316, "grad_norm": 0.07644650340080261, "learning_rate": 7.786557226154473e-06, "loss": 0.007, "step": 24870 }, { "epoch": 26.32804232804233, "grad_norm": 0.227617084980011, "learning_rate": 7.757045301514998e-06, "loss": 0.0083, "step": 24880 }, { "epoch": 26.33862433862434, "grad_norm": 0.08148089051246643, "learning_rate": 7.727584706048735e-06, "loss": 0.0077, "step": 24890 }, { "epoch": 26.349206349206348, "grad_norm": 0.11342889070510864, "learning_rate": 7.698175475553077e-06, "loss": 0.0081, "step": 24900 }, { "epoch": 26.35978835978836, "grad_norm": 0.10659900307655334, "learning_rate": 7.668817645763021e-06, "loss": 0.0081, "step": 24910 }, { "epoch": 26.37037037037037, "grad_norm": 0.116643525660038, "learning_rate": 7.639511252351088e-06, "loss": 0.0082, "step": 24920 }, { "epoch": 26.38095238095238, "grad_norm": 0.15195006132125854, "learning_rate": 7.610256330927324e-06, "loss": 0.007, "step": 24930 }, { "epoch": 26.391534391534393, "grad_norm": 0.09907189011573792, "learning_rate": 7.581052917039211e-06, "loss": 0.0074, "step": 24940 }, { "epoch": 26.402116402116402, "grad_norm": 0.12138514965772629, "learning_rate": 7.551901046171645e-06, "loss": 0.0108, "step": 24950 }, { "epoch": 26.41269841269841, "grad_norm": 0.24688255786895752, "learning_rate": 7.522800753746895e-06, "loss": 0.0086, "step": 24960 }, { "epoch": 26.423280423280424, "grad_norm": 0.19129908084869385, "learning_rate": 7.493752075124577e-06, "loss": 0.0061, "step": 24970 }, { "epoch": 26.433862433862434, "grad_norm": 0.12331803143024445, "learning_rate": 7.464755045601557e-06, "loss": 0.0074, "step": 24980 }, { "epoch": 26.444444444444443, "grad_norm": 0.17388053238391876, "learning_rate": 7.435809700411972e-06, "loss": 0.0072, "step": 24990 }, { "epoch": 26.455026455026456, "grad_norm": 0.12578460574150085, "learning_rate": 7.40691607472715e-06, "loss": 0.0084, "step": 25000 }, { "epoch": 26.465608465608465, "grad_norm": 0.06991247087717056, "learning_rate": 7.3780742036555606e-06, "loss": 0.0083, "step": 25010 }, { "epoch": 26.476190476190474, "grad_norm": 0.07223046571016312, "learning_rate": 7.349284122242783e-06, "loss": 0.0061, "step": 25020 }, { "epoch": 26.486772486772487, "grad_norm": 0.16021451354026794, "learning_rate": 7.320545865471512e-06, "loss": 0.008, "step": 25030 }, { "epoch": 26.497354497354497, "grad_norm": 0.1207735612988472, "learning_rate": 7.291859468261425e-06, "loss": 0.0091, "step": 25040 }, { "epoch": 26.507936507936506, "grad_norm": 0.07252541184425354, "learning_rate": 7.263224965469195e-06, "loss": 0.008, "step": 25050 }, { "epoch": 26.51851851851852, "grad_norm": 0.05995796620845795, "learning_rate": 7.234642391888446e-06, "loss": 0.0085, "step": 25060 }, { "epoch": 26.52910052910053, "grad_norm": 0.9467735886573792, "learning_rate": 7.206111782249697e-06, "loss": 0.0083, "step": 25070 }, { "epoch": 26.53968253968254, "grad_norm": 0.11514415591955185, "learning_rate": 7.1776331712203385e-06, "loss": 0.0082, "step": 25080 }, { "epoch": 26.55026455026455, "grad_norm": 0.07256613671779633, "learning_rate": 7.149206593404561e-06, "loss": 0.0084, "step": 25090 }, { "epoch": 26.56084656084656, "grad_norm": 0.08623909950256348, "learning_rate": 7.120832083343337e-06, "loss": 0.0073, "step": 25100 }, { "epoch": 26.571428571428573, "grad_norm": 0.13016292452812195, "learning_rate": 7.092509675514369e-06, "loss": 0.0068, "step": 25110 }, { "epoch": 26.582010582010582, "grad_norm": 0.22980135679244995, "learning_rate": 7.064239404332063e-06, "loss": 0.007, "step": 25120 }, { "epoch": 26.59259259259259, "grad_norm": 0.24668413400650024, "learning_rate": 7.03602130414745e-06, "loss": 0.0055, "step": 25130 }, { "epoch": 26.603174603174605, "grad_norm": 0.08578503131866455, "learning_rate": 7.007855409248198e-06, "loss": 0.008, "step": 25140 }, { "epoch": 26.613756613756614, "grad_norm": 0.16234572231769562, "learning_rate": 6.979741753858521e-06, "loss": 0.0071, "step": 25150 }, { "epoch": 26.624338624338623, "grad_norm": 0.10835228860378265, "learning_rate": 6.951680372139158e-06, "loss": 0.0071, "step": 25160 }, { "epoch": 26.634920634920636, "grad_norm": 0.09848982095718384, "learning_rate": 6.923671298187334e-06, "loss": 0.0075, "step": 25170 }, { "epoch": 26.645502645502646, "grad_norm": 0.12906554341316223, "learning_rate": 6.895714566036704e-06, "loss": 0.0066, "step": 25180 }, { "epoch": 26.656084656084655, "grad_norm": 0.06415915489196777, "learning_rate": 6.86781020965736e-06, "loss": 0.0068, "step": 25190 }, { "epoch": 26.666666666666668, "grad_norm": 0.08083491027355194, "learning_rate": 6.839958262955709e-06, "loss": 0.0069, "step": 25200 }, { "epoch": 26.677248677248677, "grad_norm": 0.1394011229276657, "learning_rate": 6.8121587597744894e-06, "loss": 0.0092, "step": 25210 }, { "epoch": 26.687830687830687, "grad_norm": 0.22256878018379211, "learning_rate": 6.784411733892732e-06, "loss": 0.0076, "step": 25220 }, { "epoch": 26.6984126984127, "grad_norm": 0.07933635264635086, "learning_rate": 6.756717219025666e-06, "loss": 0.0074, "step": 25230 }, { "epoch": 26.70899470899471, "grad_norm": 0.2813841700553894, "learning_rate": 6.729075248824762e-06, "loss": 0.0083, "step": 25240 }, { "epoch": 26.719576719576718, "grad_norm": 0.0715981125831604, "learning_rate": 6.701485856877615e-06, "loss": 0.0077, "step": 25250 }, { "epoch": 26.73015873015873, "grad_norm": 0.11728864163160324, "learning_rate": 6.673949076707925e-06, "loss": 0.0082, "step": 25260 }, { "epoch": 26.74074074074074, "grad_norm": 0.22656911611557007, "learning_rate": 6.646464941775499e-06, "loss": 0.0076, "step": 25270 }, { "epoch": 26.75132275132275, "grad_norm": 0.09773758053779602, "learning_rate": 6.619033485476128e-06, "loss": 0.008, "step": 25280 }, { "epoch": 26.761904761904763, "grad_norm": 0.16704943776130676, "learning_rate": 6.59165474114164e-06, "loss": 0.0057, "step": 25290 }, { "epoch": 26.772486772486772, "grad_norm": 0.16464689373970032, "learning_rate": 6.564328742039783e-06, "loss": 0.0074, "step": 25300 }, { "epoch": 26.78306878306878, "grad_norm": 0.0895526260137558, "learning_rate": 6.537055521374219e-06, "loss": 0.0061, "step": 25310 }, { "epoch": 26.793650793650794, "grad_norm": 0.161675825715065, "learning_rate": 6.509835112284485e-06, "loss": 0.0077, "step": 25320 }, { "epoch": 26.804232804232804, "grad_norm": 0.1768595427274704, "learning_rate": 6.482667547845944e-06, "loss": 0.0072, "step": 25330 }, { "epoch": 26.814814814814813, "grad_norm": 0.17395925521850586, "learning_rate": 6.455552861069736e-06, "loss": 0.0068, "step": 25340 }, { "epoch": 26.825396825396826, "grad_norm": 0.08788488060235977, "learning_rate": 6.428491084902788e-06, "loss": 0.007, "step": 25350 }, { "epoch": 26.835978835978835, "grad_norm": 0.1949339359998703, "learning_rate": 6.401482252227698e-06, "loss": 0.0077, "step": 25360 }, { "epoch": 26.84656084656085, "grad_norm": 0.0842994824051857, "learning_rate": 6.37452639586274e-06, "loss": 0.0069, "step": 25370 }, { "epoch": 26.857142857142858, "grad_norm": 0.07696864753961563, "learning_rate": 6.347623548561826e-06, "loss": 0.0064, "step": 25380 }, { "epoch": 26.867724867724867, "grad_norm": 0.21263457834720612, "learning_rate": 6.320773743014441e-06, "loss": 0.0076, "step": 25390 }, { "epoch": 26.87830687830688, "grad_norm": 0.30173590779304504, "learning_rate": 6.293977011845648e-06, "loss": 0.0064, "step": 25400 }, { "epoch": 26.88888888888889, "grad_norm": 0.1099831610918045, "learning_rate": 6.2672333876159835e-06, "loss": 0.0074, "step": 25410 }, { "epoch": 26.8994708994709, "grad_norm": 0.1258636713027954, "learning_rate": 6.2405429028215e-06, "loss": 0.009, "step": 25420 }, { "epoch": 26.91005291005291, "grad_norm": 0.14041799306869507, "learning_rate": 6.213905589893631e-06, "loss": 0.0092, "step": 25430 }, { "epoch": 26.92063492063492, "grad_norm": 0.11411046236753464, "learning_rate": 6.187321481199221e-06, "loss": 0.0064, "step": 25440 }, { "epoch": 26.93121693121693, "grad_norm": 0.11292960494756699, "learning_rate": 6.1607906090404895e-06, "loss": 0.0088, "step": 25450 }, { "epoch": 26.941798941798943, "grad_norm": 0.18220406770706177, "learning_rate": 6.134313005654929e-06, "loss": 0.0075, "step": 25460 }, { "epoch": 26.952380952380953, "grad_norm": 0.10573024302721024, "learning_rate": 6.107888703215337e-06, "loss": 0.008, "step": 25470 }, { "epoch": 26.962962962962962, "grad_norm": 0.1403624713420868, "learning_rate": 6.081517733829722e-06, "loss": 0.0084, "step": 25480 }, { "epoch": 26.973544973544975, "grad_norm": 0.1835319995880127, "learning_rate": 6.055200129541294e-06, "loss": 0.0082, "step": 25490 }, { "epoch": 26.984126984126984, "grad_norm": 0.1301507204771042, "learning_rate": 6.0289359223284434e-06, "loss": 0.0062, "step": 25500 }, { "epoch": 26.994708994708994, "grad_norm": 0.09520348161458969, "learning_rate": 6.002725144104648e-06, "loss": 0.0062, "step": 25510 }, { "epoch": 27.005291005291006, "grad_norm": 0.09970922023057938, "learning_rate": 5.976567826718477e-06, "loss": 0.0078, "step": 25520 }, { "epoch": 27.015873015873016, "grad_norm": 0.0699843242764473, "learning_rate": 5.9504640019535315e-06, "loss": 0.0081, "step": 25530 }, { "epoch": 27.026455026455025, "grad_norm": 0.10524575412273407, "learning_rate": 5.924413701528414e-06, "loss": 0.0059, "step": 25540 }, { "epoch": 27.037037037037038, "grad_norm": 0.10867656022310257, "learning_rate": 5.898416957096703e-06, "loss": 0.0067, "step": 25550 }, { "epoch": 27.047619047619047, "grad_norm": 0.21412216126918793, "learning_rate": 5.872473800246913e-06, "loss": 0.0081, "step": 25560 }, { "epoch": 27.058201058201057, "grad_norm": 0.14717839658260345, "learning_rate": 5.846584262502402e-06, "loss": 0.008, "step": 25570 }, { "epoch": 27.06878306878307, "grad_norm": 0.07591203600168228, "learning_rate": 5.820748375321411e-06, "loss": 0.0071, "step": 25580 }, { "epoch": 27.07936507936508, "grad_norm": 0.12001851201057434, "learning_rate": 5.794966170096977e-06, "loss": 0.0093, "step": 25590 }, { "epoch": 27.08994708994709, "grad_norm": 0.13798892498016357, "learning_rate": 5.769237678156897e-06, "loss": 0.0083, "step": 25600 }, { "epoch": 27.1005291005291, "grad_norm": 0.12375987321138382, "learning_rate": 5.743562930763735e-06, "loss": 0.0082, "step": 25610 }, { "epoch": 27.11111111111111, "grad_norm": 0.07062825560569763, "learning_rate": 5.717941959114726e-06, "loss": 0.0057, "step": 25620 }, { "epoch": 27.12169312169312, "grad_norm": 0.07096422463655472, "learning_rate": 5.69237479434176e-06, "loss": 0.0074, "step": 25630 }, { "epoch": 27.132275132275133, "grad_norm": 0.08946280181407928, "learning_rate": 5.666861467511353e-06, "loss": 0.0085, "step": 25640 }, { "epoch": 27.142857142857142, "grad_norm": 0.1270720511674881, "learning_rate": 5.641402009624591e-06, "loss": 0.0091, "step": 25650 }, { "epoch": 27.15343915343915, "grad_norm": 0.08128730952739716, "learning_rate": 5.615996451617145e-06, "loss": 0.0076, "step": 25660 }, { "epoch": 27.164021164021165, "grad_norm": 0.09195903688669205, "learning_rate": 5.590644824359148e-06, "loss": 0.0055, "step": 25670 }, { "epoch": 27.174603174603174, "grad_norm": 0.07630094140768051, "learning_rate": 5.56534715865521e-06, "loss": 0.0083, "step": 25680 }, { "epoch": 27.185185185185187, "grad_norm": 0.18981727957725525, "learning_rate": 5.5401034852443965e-06, "loss": 0.0067, "step": 25690 }, { "epoch": 27.195767195767196, "grad_norm": 0.11845400929450989, "learning_rate": 5.514913834800134e-06, "loss": 0.0067, "step": 25700 }, { "epoch": 27.206349206349206, "grad_norm": 0.07213514298200607, "learning_rate": 5.489778237930238e-06, "loss": 0.0062, "step": 25710 }, { "epoch": 27.21693121693122, "grad_norm": 0.39520886540412903, "learning_rate": 5.464696725176827e-06, "loss": 0.0083, "step": 25720 }, { "epoch": 27.227513227513228, "grad_norm": 0.10000025480985641, "learning_rate": 5.439669327016294e-06, "loss": 0.0079, "step": 25730 }, { "epoch": 27.238095238095237, "grad_norm": 0.23467056453227997, "learning_rate": 5.4146960738592985e-06, "loss": 0.0066, "step": 25740 }, { "epoch": 27.24867724867725, "grad_norm": 0.19177348911762238, "learning_rate": 5.389776996050694e-06, "loss": 0.0092, "step": 25750 }, { "epoch": 27.25925925925926, "grad_norm": 0.14530493319034576, "learning_rate": 5.364912123869492e-06, "loss": 0.0072, "step": 25760 }, { "epoch": 27.26984126984127, "grad_norm": 0.25171959400177, "learning_rate": 5.3401014875288864e-06, "loss": 0.0076, "step": 25770 }, { "epoch": 27.280423280423282, "grad_norm": 0.13694019615650177, "learning_rate": 5.315345117176129e-06, "loss": 0.0072, "step": 25780 }, { "epoch": 27.29100529100529, "grad_norm": 0.1492043435573578, "learning_rate": 5.2906430428925405e-06, "loss": 0.0069, "step": 25790 }, { "epoch": 27.3015873015873, "grad_norm": 0.1149340271949768, "learning_rate": 5.265995294693471e-06, "loss": 0.01, "step": 25800 }, { "epoch": 27.312169312169313, "grad_norm": 0.11539540439844131, "learning_rate": 5.241401902528254e-06, "loss": 0.0074, "step": 25810 }, { "epoch": 27.322751322751323, "grad_norm": 0.06066780537366867, "learning_rate": 5.216862896280189e-06, "loss": 0.0058, "step": 25820 }, { "epoch": 27.333333333333332, "grad_norm": 0.12756764888763428, "learning_rate": 5.192378305766471e-06, "loss": 0.0071, "step": 25830 }, { "epoch": 27.343915343915345, "grad_norm": 0.1691998690366745, "learning_rate": 5.1679481607382065e-06, "loss": 0.007, "step": 25840 }, { "epoch": 27.354497354497354, "grad_norm": 0.2794641852378845, "learning_rate": 5.1435724908803105e-06, "loss": 0.008, "step": 25850 }, { "epoch": 27.365079365079364, "grad_norm": 0.3980655372142792, "learning_rate": 5.119251325811514e-06, "loss": 0.0079, "step": 25860 }, { "epoch": 27.375661375661377, "grad_norm": 0.06481165438890457, "learning_rate": 5.094984695084348e-06, "loss": 0.006, "step": 25870 }, { "epoch": 27.386243386243386, "grad_norm": 0.0979704037308693, "learning_rate": 5.070772628185039e-06, "loss": 0.0073, "step": 25880 }, { "epoch": 27.396825396825395, "grad_norm": 0.08046741038560867, "learning_rate": 5.046615154533535e-06, "loss": 0.0081, "step": 25890 }, { "epoch": 27.40740740740741, "grad_norm": 0.09923584759235382, "learning_rate": 5.022512303483451e-06, "loss": 0.0076, "step": 25900 }, { "epoch": 27.417989417989418, "grad_norm": 0.24978315830230713, "learning_rate": 4.998464104322015e-06, "loss": 0.0075, "step": 25910 }, { "epoch": 27.428571428571427, "grad_norm": 0.12836377322673798, "learning_rate": 4.974470586270047e-06, "loss": 0.0081, "step": 25920 }, { "epoch": 27.43915343915344, "grad_norm": 0.15597403049468994, "learning_rate": 4.950531778481965e-06, "loss": 0.0059, "step": 25930 }, { "epoch": 27.44973544973545, "grad_norm": 0.12172301113605499, "learning_rate": 4.926647710045651e-06, "loss": 0.0092, "step": 25940 }, { "epoch": 27.46031746031746, "grad_norm": 0.23006397485733032, "learning_rate": 4.9028184099825125e-06, "loss": 0.0083, "step": 25950 }, { "epoch": 27.47089947089947, "grad_norm": 0.06933967024087906, "learning_rate": 4.879043907247382e-06, "loss": 0.009, "step": 25960 }, { "epoch": 27.48148148148148, "grad_norm": 0.09389746189117432, "learning_rate": 4.8553242307285414e-06, "loss": 0.0065, "step": 25970 }, { "epoch": 27.49206349206349, "grad_norm": 0.09737752377986908, "learning_rate": 4.8316594092476185e-06, "loss": 0.0081, "step": 25980 }, { "epoch": 27.502645502645503, "grad_norm": 0.12768538296222687, "learning_rate": 4.808049471559617e-06, "loss": 0.0092, "step": 25990 }, { "epoch": 27.513227513227513, "grad_norm": 0.13975109159946442, "learning_rate": 4.784494446352833e-06, "loss": 0.007, "step": 26000 }, { "epoch": 27.523809523809526, "grad_norm": 0.08345731347799301, "learning_rate": 4.760994362248833e-06, "loss": 0.0057, "step": 26010 }, { "epoch": 27.534391534391535, "grad_norm": 0.1944485306739807, "learning_rate": 4.737549247802442e-06, "loss": 0.009, "step": 26020 }, { "epoch": 27.544973544973544, "grad_norm": 0.13475744426250458, "learning_rate": 4.714159131501689e-06, "loss": 0.0058, "step": 26030 }, { "epoch": 27.555555555555557, "grad_norm": 0.08833500742912292, "learning_rate": 4.690824041767766e-06, "loss": 0.0081, "step": 26040 }, { "epoch": 27.566137566137566, "grad_norm": 0.14767873287200928, "learning_rate": 4.6675440069550105e-06, "loss": 0.006, "step": 26050 }, { "epoch": 27.576719576719576, "grad_norm": 0.0844602882862091, "learning_rate": 4.64431905535086e-06, "loss": 0.006, "step": 26060 }, { "epoch": 27.58730158730159, "grad_norm": 0.17904053628444672, "learning_rate": 4.621149215175808e-06, "loss": 0.0072, "step": 26070 }, { "epoch": 27.597883597883598, "grad_norm": 0.07360252737998962, "learning_rate": 4.5980345145834154e-06, "loss": 0.0071, "step": 26080 }, { "epoch": 27.608465608465607, "grad_norm": 0.12696576118469238, "learning_rate": 4.574974981660213e-06, "loss": 0.0064, "step": 26090 }, { "epoch": 27.61904761904762, "grad_norm": 0.1338837891817093, "learning_rate": 4.551970644425707e-06, "loss": 0.0047, "step": 26100 }, { "epoch": 27.62962962962963, "grad_norm": 0.1197524294257164, "learning_rate": 4.529021530832328e-06, "loss": 0.0065, "step": 26110 }, { "epoch": 27.64021164021164, "grad_norm": 0.11711903661489487, "learning_rate": 4.5061276687654306e-06, "loss": 0.0071, "step": 26120 }, { "epoch": 27.650793650793652, "grad_norm": 0.08960852026939392, "learning_rate": 4.4832890860431965e-06, "loss": 0.0093, "step": 26130 }, { "epoch": 27.66137566137566, "grad_norm": 0.0934315100312233, "learning_rate": 4.460505810416682e-06, "loss": 0.0077, "step": 26140 }, { "epoch": 27.67195767195767, "grad_norm": 0.22079887986183167, "learning_rate": 4.4377778695696984e-06, "loss": 0.0067, "step": 26150 }, { "epoch": 27.682539682539684, "grad_norm": 0.0849977433681488, "learning_rate": 4.415105291118843e-06, "loss": 0.0072, "step": 26160 }, { "epoch": 27.693121693121693, "grad_norm": 0.11208323389291763, "learning_rate": 4.392488102613435e-06, "loss": 0.0076, "step": 26170 }, { "epoch": 27.703703703703702, "grad_norm": 0.10209309309720993, "learning_rate": 4.369926331535473e-06, "loss": 0.0068, "step": 26180 }, { "epoch": 27.714285714285715, "grad_norm": 0.14496435225009918, "learning_rate": 4.347420005299668e-06, "loss": 0.0075, "step": 26190 }, { "epoch": 27.724867724867725, "grad_norm": 0.13579875230789185, "learning_rate": 4.324969151253317e-06, "loss": 0.0063, "step": 26200 }, { "epoch": 27.735449735449734, "grad_norm": 0.11479581147432327, "learning_rate": 4.302573796676312e-06, "loss": 0.0074, "step": 26210 }, { "epoch": 27.746031746031747, "grad_norm": 0.07273384183645248, "learning_rate": 4.2802339687811385e-06, "loss": 0.0046, "step": 26220 }, { "epoch": 27.756613756613756, "grad_norm": 0.11777371913194656, "learning_rate": 4.257949694712771e-06, "loss": 0.0052, "step": 26230 }, { "epoch": 27.767195767195766, "grad_norm": 0.08152996748685837, "learning_rate": 4.235721001548726e-06, "loss": 0.008, "step": 26240 }, { "epoch": 27.77777777777778, "grad_norm": 0.10567642748355865, "learning_rate": 4.21354791629896e-06, "loss": 0.0078, "step": 26250 }, { "epoch": 27.788359788359788, "grad_norm": 0.19587886333465576, "learning_rate": 4.191430465905843e-06, "loss": 0.0091, "step": 26260 }, { "epoch": 27.798941798941797, "grad_norm": 0.20515446364879608, "learning_rate": 4.169368677244184e-06, "loss": 0.0076, "step": 26270 }, { "epoch": 27.80952380952381, "grad_norm": 0.06653302162885666, "learning_rate": 4.147362577121122e-06, "loss": 0.0062, "step": 26280 }, { "epoch": 27.82010582010582, "grad_norm": 0.1402067393064499, "learning_rate": 4.125412192276157e-06, "loss": 0.0062, "step": 26290 }, { "epoch": 27.83068783068783, "grad_norm": 0.08168792724609375, "learning_rate": 4.10351754938107e-06, "loss": 0.0072, "step": 26300 }, { "epoch": 27.841269841269842, "grad_norm": 0.1703365594148636, "learning_rate": 4.081678675039913e-06, "loss": 0.0085, "step": 26310 }, { "epoch": 27.85185185185185, "grad_norm": 0.08587846904993057, "learning_rate": 4.0598955957889876e-06, "loss": 0.0122, "step": 26320 }, { "epoch": 27.862433862433864, "grad_norm": 0.09259912371635437, "learning_rate": 4.038168338096776e-06, "loss": 0.0072, "step": 26330 }, { "epoch": 27.873015873015873, "grad_norm": 0.11673422902822495, "learning_rate": 4.016496928363944e-06, "loss": 0.0105, "step": 26340 }, { "epoch": 27.883597883597883, "grad_norm": 0.12656958401203156, "learning_rate": 3.994881392923317e-06, "loss": 0.0085, "step": 26350 }, { "epoch": 27.894179894179896, "grad_norm": 0.051860060542821884, "learning_rate": 3.9733217580397945e-06, "loss": 0.0073, "step": 26360 }, { "epoch": 27.904761904761905, "grad_norm": 0.182510644197464, "learning_rate": 3.951818049910366e-06, "loss": 0.0058, "step": 26370 }, { "epoch": 27.915343915343914, "grad_norm": 0.09900026768445969, "learning_rate": 3.930370294664071e-06, "loss": 0.0068, "step": 26380 }, { "epoch": 27.925925925925927, "grad_norm": 0.13125447928905487, "learning_rate": 3.908978518361938e-06, "loss": 0.0073, "step": 26390 }, { "epoch": 27.936507936507937, "grad_norm": 0.2156379371881485, "learning_rate": 3.887642746997017e-06, "loss": 0.0082, "step": 26400 }, { "epoch": 27.947089947089946, "grad_norm": 0.14093849062919617, "learning_rate": 3.866363006494255e-06, "loss": 0.0076, "step": 26410 }, { "epoch": 27.95767195767196, "grad_norm": 0.19358955323696136, "learning_rate": 3.8451393227105725e-06, "loss": 0.0081, "step": 26420 }, { "epoch": 27.96825396825397, "grad_norm": 0.1785634160041809, "learning_rate": 3.823971721434727e-06, "loss": 0.0098, "step": 26430 }, { "epoch": 27.978835978835978, "grad_norm": 0.28616026043891907, "learning_rate": 3.8028602283873505e-06, "loss": 0.0065, "step": 26440 }, { "epoch": 27.98941798941799, "grad_norm": 0.24611668288707733, "learning_rate": 3.781804869220912e-06, "loss": 0.0072, "step": 26450 }, { "epoch": 28.0, "grad_norm": 0.07767170667648315, "learning_rate": 3.760805669519646e-06, "loss": 0.0068, "step": 26460 }, { "epoch": 28.01058201058201, "grad_norm": 0.08419696241617203, "learning_rate": 3.7398626547995587e-06, "loss": 0.0079, "step": 26470 }, { "epoch": 28.021164021164022, "grad_norm": 0.17779769003391266, "learning_rate": 3.7189758505083995e-06, "loss": 0.0066, "step": 26480 }, { "epoch": 28.03174603174603, "grad_norm": 0.06287797540426254, "learning_rate": 3.6981452820255837e-06, "loss": 0.007, "step": 26490 }, { "epoch": 28.04232804232804, "grad_norm": 0.1726861447095871, "learning_rate": 3.6773709746622363e-06, "loss": 0.005, "step": 26500 }, { "epoch": 28.052910052910054, "grad_norm": 0.08843417465686798, "learning_rate": 3.6566529536610995e-06, "loss": 0.0082, "step": 26510 }, { "epoch": 28.063492063492063, "grad_norm": 0.0911419540643692, "learning_rate": 3.635991244196513e-06, "loss": 0.0074, "step": 26520 }, { "epoch": 28.074074074074073, "grad_norm": 0.3347350060939789, "learning_rate": 3.615385871374405e-06, "loss": 0.0069, "step": 26530 }, { "epoch": 28.084656084656086, "grad_norm": 0.22755853831768036, "learning_rate": 3.5948368602322367e-06, "loss": 0.0054, "step": 26540 }, { "epoch": 28.095238095238095, "grad_norm": 0.1333092302083969, "learning_rate": 3.5743442357390187e-06, "loss": 0.0074, "step": 26550 }, { "epoch": 28.105820105820104, "grad_norm": 0.11262001097202301, "learning_rate": 3.5539080227951937e-06, "loss": 0.0057, "step": 26560 }, { "epoch": 28.116402116402117, "grad_norm": 0.23192596435546875, "learning_rate": 3.533528246232709e-06, "loss": 0.0095, "step": 26570 }, { "epoch": 28.126984126984127, "grad_norm": 0.18039904534816742, "learning_rate": 3.5132049308149117e-06, "loss": 0.0072, "step": 26580 }, { "epoch": 28.137566137566136, "grad_norm": 0.16119323670864105, "learning_rate": 3.4929381012365425e-06, "loss": 0.0054, "step": 26590 }, { "epoch": 28.14814814814815, "grad_norm": 0.07561906427145004, "learning_rate": 3.472727782123697e-06, "loss": 0.0081, "step": 26600 }, { "epoch": 28.158730158730158, "grad_norm": 0.17781125009059906, "learning_rate": 3.4525739980338425e-06, "loss": 0.0086, "step": 26610 }, { "epoch": 28.16931216931217, "grad_norm": 0.10528447479009628, "learning_rate": 3.432476773455712e-06, "loss": 0.0072, "step": 26620 }, { "epoch": 28.17989417989418, "grad_norm": 0.11645950376987457, "learning_rate": 3.4124361328093378e-06, "loss": 0.0074, "step": 26630 }, { "epoch": 28.19047619047619, "grad_norm": 0.16198205947875977, "learning_rate": 3.392452100445975e-06, "loss": 0.0087, "step": 26640 }, { "epoch": 28.201058201058203, "grad_norm": 0.10009289532899857, "learning_rate": 3.3725247006481163e-06, "loss": 0.0081, "step": 26650 }, { "epoch": 28.211640211640212, "grad_norm": 0.2910792827606201, "learning_rate": 3.3526539576294313e-06, "loss": 0.008, "step": 26660 }, { "epoch": 28.22222222222222, "grad_norm": 0.12124162912368774, "learning_rate": 3.332839895534745e-06, "loss": 0.0082, "step": 26670 }, { "epoch": 28.232804232804234, "grad_norm": 0.07427642494440079, "learning_rate": 3.3130825384400155e-06, "loss": 0.0069, "step": 26680 }, { "epoch": 28.243386243386244, "grad_norm": 0.10609009861946106, "learning_rate": 3.293381910352278e-06, "loss": 0.0078, "step": 26690 }, { "epoch": 28.253968253968253, "grad_norm": 0.09436683356761932, "learning_rate": 3.273738035209678e-06, "loss": 0.0072, "step": 26700 }, { "epoch": 28.264550264550266, "grad_norm": 0.1763303428888321, "learning_rate": 3.254150936881356e-06, "loss": 0.0062, "step": 26710 }, { "epoch": 28.275132275132275, "grad_norm": 0.08549419790506363, "learning_rate": 3.234620639167496e-06, "loss": 0.0081, "step": 26720 }, { "epoch": 28.285714285714285, "grad_norm": 0.1783219426870346, "learning_rate": 3.2151471657992482e-06, "loss": 0.0079, "step": 26730 }, { "epoch": 28.296296296296298, "grad_norm": 0.15859834849834442, "learning_rate": 3.195730540438718e-06, "loss": 0.0071, "step": 26740 }, { "epoch": 28.306878306878307, "grad_norm": 0.0737132653594017, "learning_rate": 3.1763707866789336e-06, "loss": 0.0066, "step": 26750 }, { "epoch": 28.317460317460316, "grad_norm": 0.1104796975851059, "learning_rate": 3.1570679280438165e-06, "loss": 0.0073, "step": 26760 }, { "epoch": 28.32804232804233, "grad_norm": 0.3413125276565552, "learning_rate": 3.1378219879881767e-06, "loss": 0.0084, "step": 26770 }, { "epoch": 28.33862433862434, "grad_norm": 0.08502422273159027, "learning_rate": 3.11863298989763e-06, "loss": 0.0076, "step": 26780 }, { "epoch": 28.349206349206348, "grad_norm": 0.1977088451385498, "learning_rate": 3.0995009570886303e-06, "loss": 0.0111, "step": 26790 }, { "epoch": 28.35978835978836, "grad_norm": 0.19955949485301971, "learning_rate": 3.0804259128083936e-06, "loss": 0.0076, "step": 26800 }, { "epoch": 28.37037037037037, "grad_norm": 0.13389959931373596, "learning_rate": 3.06140788023489e-06, "loss": 0.0058, "step": 26810 }, { "epoch": 28.38095238095238, "grad_norm": 0.1619662195444107, "learning_rate": 3.042446882476846e-06, "loss": 0.0048, "step": 26820 }, { "epoch": 28.391534391534393, "grad_norm": 0.10449915379285812, "learning_rate": 3.023542942573643e-06, "loss": 0.0058, "step": 26830 }, { "epoch": 28.402116402116402, "grad_norm": 0.08202816545963287, "learning_rate": 3.004696083495351e-06, "loss": 0.0058, "step": 26840 }, { "epoch": 28.41269841269841, "grad_norm": 0.3518655002117157, "learning_rate": 2.98590632814269e-06, "loss": 0.006, "step": 26850 }, { "epoch": 28.423280423280424, "grad_norm": 0.1522909551858902, "learning_rate": 2.9671736993469702e-06, "loss": 0.0077, "step": 26860 }, { "epoch": 28.433862433862434, "grad_norm": 0.09724295139312744, "learning_rate": 2.9484982198701216e-06, "loss": 0.0073, "step": 26870 }, { "epoch": 28.444444444444443, "grad_norm": 0.08344422280788422, "learning_rate": 2.929879912404604e-06, "loss": 0.0071, "step": 26880 }, { "epoch": 28.455026455026456, "grad_norm": 0.14369624853134155, "learning_rate": 2.9113187995734147e-06, "loss": 0.0081, "step": 26890 }, { "epoch": 28.465608465608465, "grad_norm": 0.08649647235870361, "learning_rate": 2.8928149039300524e-06, "loss": 0.006, "step": 26900 }, { "epoch": 28.476190476190474, "grad_norm": 0.487926185131073, "learning_rate": 2.8743682479584975e-06, "loss": 0.0082, "step": 26910 }, { "epoch": 28.486772486772487, "grad_norm": 0.10829430818557739, "learning_rate": 2.8559788540731824e-06, "loss": 0.0073, "step": 26920 }, { "epoch": 28.497354497354497, "grad_norm": 0.13071830570697784, "learning_rate": 2.8376467446189493e-06, "loss": 0.006, "step": 26930 }, { "epoch": 28.507936507936506, "grad_norm": 0.24453799426555634, "learning_rate": 2.8193719418710406e-06, "loss": 0.0081, "step": 26940 }, { "epoch": 28.51851851851852, "grad_norm": 0.18756212294101715, "learning_rate": 2.8011544680350667e-06, "loss": 0.0064, "step": 26950 }, { "epoch": 28.52910052910053, "grad_norm": 0.13723669946193695, "learning_rate": 2.7829943452469753e-06, "loss": 0.008, "step": 26960 }, { "epoch": 28.53968253968254, "grad_norm": 0.17550741136074066, "learning_rate": 2.7648915955730213e-06, "loss": 0.0051, "step": 26970 }, { "epoch": 28.55026455026455, "grad_norm": 0.15751288831233978, "learning_rate": 2.746846241009765e-06, "loss": 0.0063, "step": 26980 }, { "epoch": 28.56084656084656, "grad_norm": 0.06783192604780197, "learning_rate": 2.7288583034839945e-06, "loss": 0.0079, "step": 26990 }, { "epoch": 28.571428571428573, "grad_norm": 0.16807110607624054, "learning_rate": 2.7109278048527752e-06, "loss": 0.0087, "step": 27000 }, { "epoch": 28.582010582010582, "grad_norm": 0.13607145845890045, "learning_rate": 2.6930547669033413e-06, "loss": 0.0081, "step": 27010 }, { "epoch": 28.59259259259259, "grad_norm": 0.10763829201459885, "learning_rate": 2.675239211353109e-06, "loss": 0.0062, "step": 27020 }, { "epoch": 28.603174603174605, "grad_norm": 0.11858341097831726, "learning_rate": 2.6574811598496786e-06, "loss": 0.0079, "step": 27030 }, { "epoch": 28.613756613756614, "grad_norm": 0.11814308911561966, "learning_rate": 2.6397806339707456e-06, "loss": 0.0097, "step": 27040 }, { "epoch": 28.624338624338623, "grad_norm": 0.05410943180322647, "learning_rate": 2.6221376552241217e-06, "loss": 0.0078, "step": 27050 }, { "epoch": 28.634920634920636, "grad_norm": 0.13065962493419647, "learning_rate": 2.604552245047681e-06, "loss": 0.0105, "step": 27060 }, { "epoch": 28.645502645502646, "grad_norm": 0.11066734045743942, "learning_rate": 2.587024424809359e-06, "loss": 0.0074, "step": 27070 }, { "epoch": 28.656084656084655, "grad_norm": 0.18148411810398102, "learning_rate": 2.5695542158071186e-06, "loss": 0.0065, "step": 27080 }, { "epoch": 28.666666666666668, "grad_norm": 0.21252748370170593, "learning_rate": 2.5521416392689068e-06, "loss": 0.0082, "step": 27090 }, { "epoch": 28.677248677248677, "grad_norm": 0.06766245514154434, "learning_rate": 2.5347867163526384e-06, "loss": 0.0064, "step": 27100 }, { "epoch": 28.687830687830687, "grad_norm": 0.11234276741743088, "learning_rate": 2.517489468146189e-06, "loss": 0.0065, "step": 27110 }, { "epoch": 28.6984126984127, "grad_norm": 0.30887570977211, "learning_rate": 2.500249915667341e-06, "loss": 0.006, "step": 27120 }, { "epoch": 28.70899470899471, "grad_norm": 0.1418578177690506, "learning_rate": 2.4830680798637817e-06, "loss": 0.0067, "step": 27130 }, { "epoch": 28.719576719576718, "grad_norm": 0.05227776989340782, "learning_rate": 2.465943981613056e-06, "loss": 0.0061, "step": 27140 }, { "epoch": 28.73015873015873, "grad_norm": 0.3719782531261444, "learning_rate": 2.448877641722569e-06, "loss": 0.0062, "step": 27150 }, { "epoch": 28.74074074074074, "grad_norm": 0.17759184539318085, "learning_rate": 2.431869080929522e-06, "loss": 0.008, "step": 27160 }, { "epoch": 28.75132275132275, "grad_norm": 0.153383269906044, "learning_rate": 2.4149183199009216e-06, "loss": 0.0054, "step": 27170 }, { "epoch": 28.761904761904763, "grad_norm": 0.04717289283871651, "learning_rate": 2.3980253792335427e-06, "loss": 0.0063, "step": 27180 }, { "epoch": 28.772486772486772, "grad_norm": 0.08819758147001266, "learning_rate": 2.381190279453899e-06, "loss": 0.0077, "step": 27190 }, { "epoch": 28.78306878306878, "grad_norm": 0.22150643169879913, "learning_rate": 2.364413041018232e-06, "loss": 0.0066, "step": 27200 }, { "epoch": 28.793650793650794, "grad_norm": 0.10488045960664749, "learning_rate": 2.347693684312463e-06, "loss": 0.0065, "step": 27210 }, { "epoch": 28.804232804232804, "grad_norm": 0.06970519572496414, "learning_rate": 2.331032229652186e-06, "loss": 0.0072, "step": 27220 }, { "epoch": 28.814814814814813, "grad_norm": 0.15832164883613586, "learning_rate": 2.314428697282628e-06, "loss": 0.007, "step": 27230 }, { "epoch": 28.825396825396826, "grad_norm": 0.19421854615211487, "learning_rate": 2.297883107378673e-06, "loss": 0.0068, "step": 27240 }, { "epoch": 28.835978835978835, "grad_norm": 0.1817639023065567, "learning_rate": 2.2813954800447513e-06, "loss": 0.0066, "step": 27250 }, { "epoch": 28.84656084656085, "grad_norm": 0.40689775347709656, "learning_rate": 2.2649658353148974e-06, "loss": 0.0084, "step": 27260 }, { "epoch": 28.857142857142858, "grad_norm": 0.21273918449878693, "learning_rate": 2.2485941931526645e-06, "loss": 0.0063, "step": 27270 }, { "epoch": 28.867724867724867, "grad_norm": 0.074540875852108, "learning_rate": 2.232280573451151e-06, "loss": 0.0072, "step": 27280 }, { "epoch": 28.87830687830688, "grad_norm": 0.15135565400123596, "learning_rate": 2.216024996032945e-06, "loss": 0.0083, "step": 27290 }, { "epoch": 28.88888888888889, "grad_norm": 0.11510252207517624, "learning_rate": 2.1998274806501074e-06, "loss": 0.0071, "step": 27300 }, { "epoch": 28.8994708994709, "grad_norm": 0.1700301468372345, "learning_rate": 2.183688046984139e-06, "loss": 0.005, "step": 27310 }, { "epoch": 28.91005291005291, "grad_norm": 0.09854134172201157, "learning_rate": 2.167606714645981e-06, "loss": 0.0068, "step": 27320 }, { "epoch": 28.92063492063492, "grad_norm": 0.11077456921339035, "learning_rate": 2.151583503175958e-06, "loss": 0.0084, "step": 27330 }, { "epoch": 28.93121693121693, "grad_norm": 0.10567762702703476, "learning_rate": 2.1356184320437957e-06, "loss": 0.0073, "step": 27340 }, { "epoch": 28.941798941798943, "grad_norm": 0.0667891725897789, "learning_rate": 2.119711520648554e-06, "loss": 0.0073, "step": 27350 }, { "epoch": 28.952380952380953, "grad_norm": 0.22067174315452576, "learning_rate": 2.103862788318628e-06, "loss": 0.0051, "step": 27360 }, { "epoch": 28.962962962962962, "grad_norm": 0.11839604377746582, "learning_rate": 2.088072254311729e-06, "loss": 0.0054, "step": 27370 }, { "epoch": 28.973544973544975, "grad_norm": 0.09207990020513535, "learning_rate": 2.0723399378148435e-06, "loss": 0.006, "step": 27380 }, { "epoch": 28.984126984126984, "grad_norm": 0.21961501240730286, "learning_rate": 2.0566658579442065e-06, "loss": 0.0056, "step": 27390 }, { "epoch": 28.994708994708994, "grad_norm": 0.3922137916088104, "learning_rate": 2.0410500337453176e-06, "loss": 0.0056, "step": 27400 }, { "epoch": 29.005291005291006, "grad_norm": 0.19239741563796997, "learning_rate": 2.0254924841928645e-06, "loss": 0.0067, "step": 27410 }, { "epoch": 29.015873015873016, "grad_norm": 0.1454186588525772, "learning_rate": 2.009993228190754e-06, "loss": 0.0069, "step": 27420 }, { "epoch": 29.026455026455025, "grad_norm": 0.0733763799071312, "learning_rate": 1.9945522845720323e-06, "loss": 0.0067, "step": 27430 }, { "epoch": 29.037037037037038, "grad_norm": 0.10015389323234558, "learning_rate": 1.9791696720988963e-06, "loss": 0.0081, "step": 27440 }, { "epoch": 29.047619047619047, "grad_norm": 0.15628018975257874, "learning_rate": 1.9638454094626833e-06, "loss": 0.0071, "step": 27450 }, { "epoch": 29.058201058201057, "grad_norm": 0.5262669920921326, "learning_rate": 1.9485795152838104e-06, "loss": 0.0059, "step": 27460 }, { "epoch": 29.06878306878307, "grad_norm": 0.07777169346809387, "learning_rate": 1.933372008111778e-06, "loss": 0.0058, "step": 27470 }, { "epoch": 29.07936507936508, "grad_norm": 0.09713108837604523, "learning_rate": 1.918222906425143e-06, "loss": 0.008, "step": 27480 }, { "epoch": 29.08994708994709, "grad_norm": 0.16425344347953796, "learning_rate": 1.903132228631488e-06, "loss": 0.007, "step": 27490 }, { "epoch": 29.1005291005291, "grad_norm": 0.24594102799892426, "learning_rate": 1.8880999930674215e-06, "loss": 0.0061, "step": 27500 }, { "epoch": 29.11111111111111, "grad_norm": 0.08636486530303955, "learning_rate": 1.8731262179985166e-06, "loss": 0.0071, "step": 27510 }, { "epoch": 29.12169312169312, "grad_norm": 0.1352970004081726, "learning_rate": 1.8582109216193244e-06, "loss": 0.0065, "step": 27520 }, { "epoch": 29.132275132275133, "grad_norm": 0.14080391824245453, "learning_rate": 1.8433541220533368e-06, "loss": 0.0072, "step": 27530 }, { "epoch": 29.142857142857142, "grad_norm": 0.11617743223905563, "learning_rate": 1.8285558373529577e-06, "loss": 0.0062, "step": 27540 }, { "epoch": 29.15343915343915, "grad_norm": 0.20124514400959015, "learning_rate": 1.8138160854995145e-06, "loss": 0.0079, "step": 27550 }, { "epoch": 29.164021164021165, "grad_norm": 0.11954380571842194, "learning_rate": 1.7991348844031864e-06, "loss": 0.0069, "step": 27560 }, { "epoch": 29.174603174603174, "grad_norm": 0.06125573441386223, "learning_rate": 1.78451225190302e-06, "loss": 0.007, "step": 27570 }, { "epoch": 29.185185185185187, "grad_norm": 0.07403701543807983, "learning_rate": 1.7699482057668914e-06, "loss": 0.0074, "step": 27580 }, { "epoch": 29.195767195767196, "grad_norm": 0.17483501136302948, "learning_rate": 1.7554427636914895e-06, "loss": 0.0054, "step": 27590 }, { "epoch": 29.206349206349206, "grad_norm": 0.13876301050186157, "learning_rate": 1.7409959433022871e-06, "loss": 0.0058, "step": 27600 }, { "epoch": 29.21693121693122, "grad_norm": 0.13250881433486938, "learning_rate": 1.7266077621535481e-06, "loss": 0.0074, "step": 27610 }, { "epoch": 29.227513227513228, "grad_norm": 0.13855762779712677, "learning_rate": 1.7122782377282597e-06, "loss": 0.0073, "step": 27620 }, { "epoch": 29.238095238095237, "grad_norm": 0.08878463506698608, "learning_rate": 1.6980073874381497e-06, "loss": 0.0082, "step": 27630 }, { "epoch": 29.24867724867725, "grad_norm": 0.1427966058254242, "learning_rate": 1.6837952286236413e-06, "loss": 0.0069, "step": 27640 }, { "epoch": 29.25925925925926, "grad_norm": 0.2419036626815796, "learning_rate": 1.6696417785538488e-06, "loss": 0.0071, "step": 27650 }, { "epoch": 29.26984126984127, "grad_norm": 0.10249986499547958, "learning_rate": 1.6555470544265538e-06, "loss": 0.0066, "step": 27660 }, { "epoch": 29.280423280423282, "grad_norm": 0.08227784186601639, "learning_rate": 1.6415110733681738e-06, "loss": 0.007, "step": 27670 }, { "epoch": 29.29100529100529, "grad_norm": 0.16720663011074066, "learning_rate": 1.6275338524337436e-06, "loss": 0.0068, "step": 27680 }, { "epoch": 29.3015873015873, "grad_norm": 0.06390467286109924, "learning_rate": 1.6136154086069056e-06, "loss": 0.0067, "step": 27690 }, { "epoch": 29.312169312169313, "grad_norm": 0.12180763483047485, "learning_rate": 1.5997557587998868e-06, "loss": 0.0072, "step": 27700 }, { "epoch": 29.322751322751323, "grad_norm": 0.15705882012844086, "learning_rate": 1.5859549198534606e-06, "loss": 0.007, "step": 27710 }, { "epoch": 29.333333333333332, "grad_norm": 0.13679039478302002, "learning_rate": 1.572212908536963e-06, "loss": 0.0084, "step": 27720 }, { "epoch": 29.343915343915345, "grad_norm": 0.3364245295524597, "learning_rate": 1.5585297415482202e-06, "loss": 0.0082, "step": 27730 }, { "epoch": 29.354497354497354, "grad_norm": 0.1373998373746872, "learning_rate": 1.5449054355135717e-06, "loss": 0.0058, "step": 27740 }, { "epoch": 29.365079365079364, "grad_norm": 0.24700583517551422, "learning_rate": 1.5313400069878415e-06, "loss": 0.0078, "step": 27750 }, { "epoch": 29.375661375661377, "grad_norm": 0.15415754914283752, "learning_rate": 1.5178334724542887e-06, "loss": 0.0076, "step": 27760 }, { "epoch": 29.386243386243386, "grad_norm": 0.17774516344070435, "learning_rate": 1.504385848324641e-06, "loss": 0.0077, "step": 27770 }, { "epoch": 29.396825396825395, "grad_norm": 0.3300551176071167, "learning_rate": 1.490997150939033e-06, "loss": 0.0073, "step": 27780 }, { "epoch": 29.40740740740741, "grad_norm": 0.15368689596652985, "learning_rate": 1.4776673965659793e-06, "loss": 0.0072, "step": 27790 }, { "epoch": 29.417989417989418, "grad_norm": 0.1520949751138687, "learning_rate": 1.4643966014023957e-06, "loss": 0.0087, "step": 27800 }, { "epoch": 29.428571428571427, "grad_norm": 0.260299950838089, "learning_rate": 1.4511847815735503e-06, "loss": 0.0058, "step": 27810 }, { "epoch": 29.43915343915344, "grad_norm": 0.16604633629322052, "learning_rate": 1.438031953133051e-06, "loss": 0.007, "step": 27820 }, { "epoch": 29.44973544973545, "grad_norm": 0.37235984206199646, "learning_rate": 1.4249381320628197e-06, "loss": 0.0082, "step": 27830 }, { "epoch": 29.46031746031746, "grad_norm": 0.2650263011455536, "learning_rate": 1.4119033342730902e-06, "loss": 0.0073, "step": 27840 }, { "epoch": 29.47089947089947, "grad_norm": 0.1196594163775444, "learning_rate": 1.3989275756023656e-06, "loss": 0.0053, "step": 27850 }, { "epoch": 29.48148148148148, "grad_norm": 0.1310596913099289, "learning_rate": 1.386010871817417e-06, "loss": 0.0064, "step": 27860 }, { "epoch": 29.49206349206349, "grad_norm": 0.1415482461452484, "learning_rate": 1.3731532386132616e-06, "loss": 0.0062, "step": 27870 }, { "epoch": 29.502645502645503, "grad_norm": 0.1555139273405075, "learning_rate": 1.360354691613136e-06, "loss": 0.007, "step": 27880 }, { "epoch": 29.513227513227513, "grad_norm": 0.3027600347995758, "learning_rate": 1.3476152463684776e-06, "loss": 0.0079, "step": 27890 }, { "epoch": 29.523809523809526, "grad_norm": 0.1732601374387741, "learning_rate": 1.3349349183589155e-06, "loss": 0.008, "step": 27900 }, { "epoch": 29.534391534391535, "grad_norm": 0.09355268627405167, "learning_rate": 1.3223137229922356e-06, "loss": 0.0078, "step": 27910 }, { "epoch": 29.544973544973544, "grad_norm": 0.10645715147256851, "learning_rate": 1.3097516756043981e-06, "loss": 0.0083, "step": 27920 }, { "epoch": 29.555555555555557, "grad_norm": 0.2474582940340042, "learning_rate": 1.29724879145946e-06, "loss": 0.0093, "step": 27930 }, { "epoch": 29.566137566137566, "grad_norm": 0.21245472133159637, "learning_rate": 1.284805085749613e-06, "loss": 0.0068, "step": 27940 }, { "epoch": 29.576719576719576, "grad_norm": 0.17744529247283936, "learning_rate": 1.2724205735951288e-06, "loss": 0.007, "step": 27950 }, { "epoch": 29.58730158730159, "grad_norm": 0.05158408358693123, "learning_rate": 1.2600952700443591e-06, "loss": 0.0053, "step": 27960 }, { "epoch": 29.597883597883598, "grad_norm": 0.09823416918516159, "learning_rate": 1.247829190073707e-06, "loss": 0.0069, "step": 27970 }, { "epoch": 29.608465608465607, "grad_norm": 0.12666167318820953, "learning_rate": 1.2356223485876173e-06, "loss": 0.0076, "step": 27980 }, { "epoch": 29.61904761904762, "grad_norm": 0.211853489279747, "learning_rate": 1.2234747604185526e-06, "loss": 0.0076, "step": 27990 }, { "epoch": 29.62962962962963, "grad_norm": 0.05677640065550804, "learning_rate": 1.2113864403269836e-06, "loss": 0.0082, "step": 28000 }, { "epoch": 29.64021164021164, "grad_norm": 0.08339288085699081, "learning_rate": 1.1993574030013555e-06, "loss": 0.0049, "step": 28010 }, { "epoch": 29.650793650793652, "grad_norm": 0.13791494071483612, "learning_rate": 1.1873876630580816e-06, "loss": 0.0058, "step": 28020 }, { "epoch": 29.66137566137566, "grad_norm": 0.10467016696929932, "learning_rate": 1.1754772350415278e-06, "loss": 0.0073, "step": 28030 }, { "epoch": 29.67195767195767, "grad_norm": 0.08289342373609543, "learning_rate": 1.163626133423984e-06, "loss": 0.0076, "step": 28040 }, { "epoch": 29.682539682539684, "grad_norm": 0.10421443730592728, "learning_rate": 1.1518343726056591e-06, "loss": 0.0087, "step": 28050 }, { "epoch": 29.693121693121693, "grad_norm": 0.36156970262527466, "learning_rate": 1.1401019669146474e-06, "loss": 0.008, "step": 28060 }, { "epoch": 29.703703703703702, "grad_norm": 0.08096849173307419, "learning_rate": 1.128428930606934e-06, "loss": 0.0062, "step": 28070 }, { "epoch": 29.714285714285715, "grad_norm": 0.10973234474658966, "learning_rate": 1.1168152778663621e-06, "loss": 0.007, "step": 28080 }, { "epoch": 29.724867724867725, "grad_norm": 0.17875948548316956, "learning_rate": 1.10526102280461e-06, "loss": 0.0061, "step": 28090 }, { "epoch": 29.735449735449734, "grad_norm": 0.14746300876140594, "learning_rate": 1.0937661794611864e-06, "loss": 0.0073, "step": 28100 }, { "epoch": 29.746031746031747, "grad_norm": 0.10701144486665726, "learning_rate": 1.082330761803413e-06, "loss": 0.0063, "step": 28110 }, { "epoch": 29.756613756613756, "grad_norm": 0.12973631918430328, "learning_rate": 1.0709547837263966e-06, "loss": 0.0079, "step": 28120 }, { "epoch": 29.767195767195766, "grad_norm": 0.31304118037223816, "learning_rate": 1.059638259053025e-06, "loss": 0.0061, "step": 28130 }, { "epoch": 29.77777777777778, "grad_norm": 0.15835361182689667, "learning_rate": 1.0483812015339434e-06, "loss": 0.0072, "step": 28140 }, { "epoch": 29.788359788359788, "grad_norm": 0.12289755791425705, "learning_rate": 1.037183624847543e-06, "loss": 0.0069, "step": 28150 }, { "epoch": 29.798941798941797, "grad_norm": 0.07068626582622528, "learning_rate": 1.0260455425999238e-06, "loss": 0.006, "step": 28160 }, { "epoch": 29.80952380952381, "grad_norm": 0.05692445486783981, "learning_rate": 1.0149669683249208e-06, "loss": 0.0065, "step": 28170 }, { "epoch": 29.82010582010582, "grad_norm": 0.1818799376487732, "learning_rate": 1.003947915484027e-06, "loss": 0.0076, "step": 28180 }, { "epoch": 29.83068783068783, "grad_norm": 0.1283659189939499, "learning_rate": 9.92988397466449e-07, "loss": 0.0045, "step": 28190 }, { "epoch": 29.841269841269842, "grad_norm": 0.12541954219341278, "learning_rate": 9.820884275890286e-07, "loss": 0.0061, "step": 28200 }, { "epoch": 29.85185185185185, "grad_norm": 0.44727274775505066, "learning_rate": 9.712480190962546e-07, "loss": 0.0063, "step": 28210 }, { "epoch": 29.862433862433864, "grad_norm": 0.1396133154630661, "learning_rate": 9.604671851602464e-07, "loss": 0.0069, "step": 28220 }, { "epoch": 29.873015873015873, "grad_norm": 0.21243621408939362, "learning_rate": 9.497459388807306e-07, "loss": 0.0051, "step": 28230 }, { "epoch": 29.883597883597883, "grad_norm": 0.16550812125205994, "learning_rate": 9.390842932850364e-07, "loss": 0.0065, "step": 28240 }, { "epoch": 29.894179894179896, "grad_norm": 0.5934991836547852, "learning_rate": 9.28482261328073e-07, "loss": 0.0074, "step": 28250 }, { "epoch": 29.904761904761905, "grad_norm": 0.09381117671728134, "learning_rate": 9.179398558923025e-07, "loss": 0.0074, "step": 28260 }, { "epoch": 29.915343915343914, "grad_norm": 0.1215338185429573, "learning_rate": 9.074570897877388e-07, "loss": 0.0079, "step": 28270 }, { "epoch": 29.925925925925927, "grad_norm": 0.27361923456192017, "learning_rate": 8.970339757519375e-07, "loss": 0.005, "step": 28280 }, { "epoch": 29.936507936507937, "grad_norm": 0.06647294014692307, "learning_rate": 8.866705264499619e-07, "loss": 0.006, "step": 28290 }, { "epoch": 29.947089947089946, "grad_norm": 0.05391930416226387, "learning_rate": 8.763667544743836e-07, "loss": 0.0064, "step": 28300 }, { "epoch": 29.95767195767196, "grad_norm": 0.23300257325172424, "learning_rate": 8.661226723452543e-07, "loss": 0.007, "step": 28310 }, { "epoch": 29.96825396825397, "grad_norm": 0.7719760537147522, "learning_rate": 8.559382925101001e-07, "loss": 0.0079, "step": 28320 }, { "epoch": 29.978835978835978, "grad_norm": 0.10133923590183258, "learning_rate": 8.458136273438943e-07, "loss": 0.0064, "step": 28330 }, { "epoch": 29.98941798941799, "grad_norm": 0.1425493359565735, "learning_rate": 8.357486891490795e-07, "loss": 0.0086, "step": 28340 }, { "epoch": 30.0, "grad_norm": 0.1251571923494339, "learning_rate": 8.257434901554895e-07, "loss": 0.0066, "step": 28350 }, { "epoch": 30.01058201058201, "grad_norm": 0.2770492732524872, "learning_rate": 8.157980425203937e-07, "loss": 0.0082, "step": 28360 }, { "epoch": 30.021164021164022, "grad_norm": 0.11097675561904907, "learning_rate": 8.059123583284367e-07, "loss": 0.0056, "step": 28370 }, { "epoch": 30.03174603174603, "grad_norm": 0.354047954082489, "learning_rate": 7.960864495916654e-07, "loss": 0.0059, "step": 28380 }, { "epoch": 30.04232804232804, "grad_norm": 0.10687369853258133, "learning_rate": 7.863203282494847e-07, "loss": 0.0071, "step": 28390 }, { "epoch": 30.052910052910054, "grad_norm": 0.8124653100967407, "learning_rate": 7.766140061686522e-07, "loss": 0.0078, "step": 28400 }, { "epoch": 30.063492063492063, "grad_norm": 0.08500286936759949, "learning_rate": 7.669674951432615e-07, "loss": 0.0061, "step": 28410 }, { "epoch": 30.074074074074073, "grad_norm": 0.4606158137321472, "learning_rate": 7.573808068947363e-07, "loss": 0.0073, "step": 28420 }, { "epoch": 30.084656084656086, "grad_norm": 0.16987265646457672, "learning_rate": 7.478539530718087e-07, "loss": 0.0056, "step": 28430 }, { "epoch": 30.095238095238095, "grad_norm": 0.17087024450302124, "learning_rate": 7.383869452504965e-07, "loss": 0.0084, "step": 28440 }, { "epoch": 30.105820105820104, "grad_norm": 0.07115305960178375, "learning_rate": 7.289797949341204e-07, "loss": 0.0073, "step": 28450 }, { "epoch": 30.116402116402117, "grad_norm": 0.07818962633609772, "learning_rate": 7.196325135532423e-07, "loss": 0.0084, "step": 28460 }, { "epoch": 30.126984126984127, "grad_norm": 0.20185944437980652, "learning_rate": 7.10345112465699e-07, "loss": 0.0068, "step": 28470 }, { "epoch": 30.137566137566136, "grad_norm": 0.12489299476146698, "learning_rate": 7.011176029565525e-07, "loss": 0.0076, "step": 28480 }, { "epoch": 30.14814814814815, "grad_norm": 0.06899841874837875, "learning_rate": 6.919499962381004e-07, "loss": 0.0064, "step": 28490 }, { "epoch": 30.158730158730158, "grad_norm": 0.13634207844734192, "learning_rate": 6.828423034498488e-07, "loss": 0.0076, "step": 28500 }, { "epoch": 30.16931216931217, "grad_norm": 0.27131202816963196, "learning_rate": 6.737945356585007e-07, "loss": 0.0086, "step": 28510 }, { "epoch": 30.17989417989418, "grad_norm": 0.0916399285197258, "learning_rate": 6.648067038579508e-07, "loss": 0.0059, "step": 28520 }, { "epoch": 30.19047619047619, "grad_norm": 0.0802793800830841, "learning_rate": 6.558788189692578e-07, "loss": 0.007, "step": 28530 }, { "epoch": 30.201058201058203, "grad_norm": 0.10408990830183029, "learning_rate": 6.470108918406492e-07, "loss": 0.0064, "step": 28540 }, { "epoch": 30.211640211640212, "grad_norm": 0.06924672424793243, "learning_rate": 6.382029332474893e-07, "loss": 0.0078, "step": 28550 }, { "epoch": 30.22222222222222, "grad_norm": 0.09604575484991074, "learning_rate": 6.294549538922778e-07, "loss": 0.0076, "step": 28560 }, { "epoch": 30.232804232804234, "grad_norm": 0.09334269911050797, "learning_rate": 6.207669644046344e-07, "loss": 0.007, "step": 28570 }, { "epoch": 30.243386243386244, "grad_norm": 0.0879959762096405, "learning_rate": 6.121389753412865e-07, "loss": 0.0075, "step": 28580 }, { "epoch": 30.253968253968253, "grad_norm": 0.09319759160280228, "learning_rate": 6.035709971860592e-07, "loss": 0.0076, "step": 28590 }, { "epoch": 30.264550264550266, "grad_norm": 0.10726936906576157, "learning_rate": 5.950630403498469e-07, "loss": 0.0059, "step": 28600 }, { "epoch": 30.275132275132275, "grad_norm": 0.3103163242340088, "learning_rate": 5.866151151706189e-07, "loss": 0.0084, "step": 28610 }, { "epoch": 30.285714285714285, "grad_norm": 0.18083372712135315, "learning_rate": 5.782272319134086e-07, "loss": 0.0065, "step": 28620 }, { "epoch": 30.296296296296298, "grad_norm": 0.15742945671081543, "learning_rate": 5.698994007702796e-07, "loss": 0.0081, "step": 28630 }, { "epoch": 30.306878306878307, "grad_norm": 0.0833008885383606, "learning_rate": 5.616316318603321e-07, "loss": 0.0075, "step": 28640 }, { "epoch": 30.317460317460316, "grad_norm": 0.05225185677409172, "learning_rate": 5.534239352296799e-07, "loss": 0.0075, "step": 28650 }, { "epoch": 30.32804232804233, "grad_norm": 0.12440568208694458, "learning_rate": 5.452763208514621e-07, "loss": 0.0069, "step": 28660 }, { "epoch": 30.33862433862434, "grad_norm": 0.0752844586968422, "learning_rate": 5.371887986257873e-07, "loss": 0.0067, "step": 28670 }, { "epoch": 30.349206349206348, "grad_norm": 0.1230759248137474, "learning_rate": 5.291613783797611e-07, "loss": 0.0077, "step": 28680 }, { "epoch": 30.35978835978836, "grad_norm": 0.1595439314842224, "learning_rate": 5.211940698674534e-07, "loss": 0.0073, "step": 28690 }, { "epoch": 30.37037037037037, "grad_norm": 0.09817776083946228, "learning_rate": 5.132868827698978e-07, "loss": 0.0066, "step": 28700 }, { "epoch": 30.38095238095238, "grad_norm": 0.08551474660634995, "learning_rate": 5.054398266950755e-07, "loss": 0.0069, "step": 28710 }, { "epoch": 30.391534391534393, "grad_norm": 0.13036344945430756, "learning_rate": 4.976529111778872e-07, "loss": 0.0065, "step": 28720 }, { "epoch": 30.402116402116402, "grad_norm": 0.1437460035085678, "learning_rate": 4.899261456801862e-07, "loss": 0.0072, "step": 28730 }, { "epoch": 30.41269841269841, "grad_norm": 0.14377503097057343, "learning_rate": 4.822595395907126e-07, "loss": 0.0066, "step": 28740 }, { "epoch": 30.423280423280424, "grad_norm": 0.12284031510353088, "learning_rate": 4.7465310222510886e-07, "loss": 0.0065, "step": 28750 }, { "epoch": 30.433862433862434, "grad_norm": 0.08140567690134048, "learning_rate": 4.6710684282593175e-07, "loss": 0.009, "step": 28760 }, { "epoch": 30.444444444444443, "grad_norm": 0.08295447379350662, "learning_rate": 4.5962077056257993e-07, "loss": 0.007, "step": 28770 }, { "epoch": 30.455026455026456, "grad_norm": 0.5447463989257812, "learning_rate": 4.521948945313492e-07, "loss": 0.0075, "step": 28780 }, { "epoch": 30.465608465608465, "grad_norm": 0.19612935185432434, "learning_rate": 4.4482922375537195e-07, "loss": 0.0064, "step": 28790 }, { "epoch": 30.476190476190474, "grad_norm": 0.0940689817070961, "learning_rate": 4.375237671846333e-07, "loss": 0.0054, "step": 28800 }, { "epoch": 30.486772486772487, "grad_norm": 0.11478488892316818, "learning_rate": 4.302785336959547e-07, "loss": 0.0068, "step": 28810 }, { "epoch": 30.497354497354497, "grad_norm": 0.20745086669921875, "learning_rate": 4.2309353209297744e-07, "loss": 0.0081, "step": 28820 }, { "epoch": 30.507936507936506, "grad_norm": 0.1510360836982727, "learning_rate": 4.159687711061566e-07, "loss": 0.0066, "step": 28830 }, { "epoch": 30.51851851851852, "grad_norm": 0.09426994621753693, "learning_rate": 4.0890425939275055e-07, "loss": 0.0068, "step": 28840 }, { "epoch": 30.52910052910053, "grad_norm": 0.12960904836654663, "learning_rate": 4.0190000553679827e-07, "loss": 0.0081, "step": 28850 }, { "epoch": 30.53968253968254, "grad_norm": 0.1757441908121109, "learning_rate": 3.9495601804913627e-07, "loss": 0.0075, "step": 28860 }, { "epoch": 30.55026455026455, "grad_norm": 0.062065489590168, "learning_rate": 3.880723053673652e-07, "loss": 0.0053, "step": 28870 }, { "epoch": 30.56084656084656, "grad_norm": 0.15079304575920105, "learning_rate": 3.812488758558386e-07, "loss": 0.0095, "step": 28880 }, { "epoch": 30.571428571428573, "grad_norm": 0.1729137897491455, "learning_rate": 3.744857378056743e-07, "loss": 0.0077, "step": 28890 }, { "epoch": 30.582010582010582, "grad_norm": 0.17349578440189362, "learning_rate": 3.677828994347154e-07, "loss": 0.0079, "step": 28900 }, { "epoch": 30.59259259259259, "grad_norm": 0.1361820101737976, "learning_rate": 3.61140368887547e-07, "loss": 0.0077, "step": 28910 }, { "epoch": 30.603174603174605, "grad_norm": 0.26987016201019287, "learning_rate": 3.545581542354681e-07, "loss": 0.0074, "step": 28920 }, { "epoch": 30.613756613756614, "grad_norm": 0.08602481335401535, "learning_rate": 3.480362634764922e-07, "loss": 0.0056, "step": 28930 }, { "epoch": 30.624338624338623, "grad_norm": 0.24035650491714478, "learning_rate": 3.4157470453533014e-07, "loss": 0.0061, "step": 28940 }, { "epoch": 30.634920634920636, "grad_norm": 0.08407130837440491, "learning_rate": 3.3517348526339034e-07, "loss": 0.0065, "step": 28950 }, { "epoch": 30.645502645502646, "grad_norm": 0.0927412137389183, "learning_rate": 3.288326134387454e-07, "loss": 0.0048, "step": 28960 }, { "epoch": 30.656084656084655, "grad_norm": 0.26102694869041443, "learning_rate": 3.225520967661655e-07, "loss": 0.0076, "step": 28970 }, { "epoch": 30.666666666666668, "grad_norm": 0.1602589637041092, "learning_rate": 3.163319428770628e-07, "loss": 0.0067, "step": 28980 }, { "epoch": 30.677248677248677, "grad_norm": 0.1365956813097, "learning_rate": 3.1017215932951374e-07, "loss": 0.0058, "step": 28990 }, { "epoch": 30.687830687830687, "grad_norm": 0.10541553795337677, "learning_rate": 3.040727536082366e-07, "loss": 0.0065, "step": 29000 }, { "epoch": 30.6984126984127, "grad_norm": 0.08033527433872223, "learning_rate": 2.980337331245864e-07, "loss": 0.0061, "step": 29010 }, { "epoch": 30.70899470899471, "grad_norm": 0.24331046640872955, "learning_rate": 2.9205510521653213e-07, "loss": 0.0079, "step": 29020 }, { "epoch": 30.719576719576718, "grad_norm": 0.45762789249420166, "learning_rate": 2.86136877148685e-07, "loss": 0.0082, "step": 29030 }, { "epoch": 30.73015873015873, "grad_norm": 0.12694931030273438, "learning_rate": 2.8027905611223704e-07, "loss": 0.0077, "step": 29040 }, { "epoch": 30.74074074074074, "grad_norm": 0.10647358745336533, "learning_rate": 2.7448164922500573e-07, "loss": 0.0097, "step": 29050 }, { "epoch": 30.75132275132275, "grad_norm": 0.12274744361639023, "learning_rate": 2.687446635313784e-07, "loss": 0.0074, "step": 29060 }, { "epoch": 30.761904761904763, "grad_norm": 0.13186337053775787, "learning_rate": 2.630681060023343e-07, "loss": 0.007, "step": 29070 }, { "epoch": 30.772486772486772, "grad_norm": 0.4734317362308502, "learning_rate": 2.5745198353542833e-07, "loss": 0.0067, "step": 29080 }, { "epoch": 30.78306878306878, "grad_norm": 0.06555823236703873, "learning_rate": 2.518963029547794e-07, "loss": 0.0068, "step": 29090 }, { "epoch": 30.793650793650794, "grad_norm": 0.17213629186153412, "learning_rate": 2.464010710110598e-07, "loss": 0.008, "step": 29100 }, { "epoch": 30.804232804232804, "grad_norm": 0.09087402373552322, "learning_rate": 2.4096629438150054e-07, "loss": 0.0069, "step": 29110 }, { "epoch": 30.814814814814813, "grad_norm": 0.17611996829509735, "learning_rate": 2.3559197966985802e-07, "loss": 0.0065, "step": 29120 }, { "epoch": 30.825396825396826, "grad_norm": 0.33800870180130005, "learning_rate": 2.3027813340644188e-07, "loss": 0.0068, "step": 29130 }, { "epoch": 30.835978835978835, "grad_norm": 0.13566116988658905, "learning_rate": 2.2502476204807055e-07, "loss": 0.0073, "step": 29140 }, { "epoch": 30.84656084656085, "grad_norm": 0.2923640310764313, "learning_rate": 2.1983187197808786e-07, "loss": 0.0095, "step": 29150 }, { "epoch": 30.857142857142858, "grad_norm": 0.12090735882520676, "learning_rate": 2.1469946950634644e-07, "loss": 0.0062, "step": 29160 }, { "epoch": 30.867724867724867, "grad_norm": 0.13309788703918457, "learning_rate": 2.096275608691911e-07, "loss": 0.0076, "step": 29170 }, { "epoch": 30.87830687830688, "grad_norm": 0.06025158241391182, "learning_rate": 2.046161522294754e-07, "loss": 0.0083, "step": 29180 }, { "epoch": 30.88888888888889, "grad_norm": 0.07784081250429153, "learning_rate": 1.9966524967653944e-07, "loss": 0.0068, "step": 29190 }, { "epoch": 30.8994708994709, "grad_norm": 0.0778302252292633, "learning_rate": 1.9477485922618222e-07, "loss": 0.0069, "step": 29200 }, { "epoch": 30.91005291005291, "grad_norm": 0.12230559438467026, "learning_rate": 1.899449868207004e-07, "loss": 0.0064, "step": 29210 }, { "epoch": 30.92063492063492, "grad_norm": 0.16633111238479614, "learning_rate": 1.851756383288439e-07, "loss": 0.0071, "step": 29220 }, { "epoch": 30.93121693121693, "grad_norm": 0.11429347097873688, "learning_rate": 1.8046681954581035e-07, "loss": 0.0072, "step": 29230 }, { "epoch": 30.941798941798943, "grad_norm": 0.17791734635829926, "learning_rate": 1.7581853619327294e-07, "loss": 0.0073, "step": 29240 }, { "epoch": 30.952380952380953, "grad_norm": 0.13837683200836182, "learning_rate": 1.7123079391932472e-07, "loss": 0.0079, "step": 29250 }, { "epoch": 30.962962962962962, "grad_norm": 0.2952558994293213, "learning_rate": 1.6670359829850657e-07, "loss": 0.0085, "step": 29260 }, { "epoch": 30.973544973544975, "grad_norm": 0.09609473496675491, "learning_rate": 1.6223695483179037e-07, "loss": 0.0058, "step": 29270 }, { "epoch": 30.984126984126984, "grad_norm": 0.06499308347702026, "learning_rate": 1.5783086894656795e-07, "loss": 0.007, "step": 29280 }, { "epoch": 30.994708994708994, "grad_norm": 0.139460027217865, "learning_rate": 1.5348534599665121e-07, "loss": 0.0055, "step": 29290 }, { "epoch": 31.005291005291006, "grad_norm": 0.40578633546829224, "learning_rate": 1.4920039126225527e-07, "loss": 0.0071, "step": 29300 }, { "epoch": 31.015873015873016, "grad_norm": 0.14723652601242065, "learning_rate": 1.449760099500097e-07, "loss": 0.0074, "step": 29310 }, { "epoch": 31.026455026455025, "grad_norm": 0.16122955083847046, "learning_rate": 1.4081220719293075e-07, "loss": 0.0084, "step": 29320 }, { "epoch": 31.037037037037038, "grad_norm": 0.12705178558826447, "learning_rate": 1.3670898805043797e-07, "loss": 0.0058, "step": 29330 }, { "epoch": 31.047619047619047, "grad_norm": 0.17664480209350586, "learning_rate": 1.326663575083209e-07, "loss": 0.0063, "step": 29340 }, { "epoch": 31.058201058201057, "grad_norm": 0.15733912587165833, "learning_rate": 1.2868432047876688e-07, "loss": 0.0085, "step": 29350 }, { "epoch": 31.06878306878307, "grad_norm": 0.3306112289428711, "learning_rate": 1.2476288180032213e-07, "loss": 0.0072, "step": 29360 }, { "epoch": 31.07936507936508, "grad_norm": 0.0479251965880394, "learning_rate": 1.209020462379029e-07, "loss": 0.005, "step": 29370 }, { "epoch": 31.08994708994709, "grad_norm": 0.11965290457010269, "learning_rate": 1.1710181848278435e-07, "loss": 0.0044, "step": 29380 }, { "epoch": 31.1005291005291, "grad_norm": 0.14488115906715393, "learning_rate": 1.133622031526116e-07, "loss": 0.0056, "step": 29390 }, { "epoch": 31.11111111111111, "grad_norm": 0.08765584230422974, "learning_rate": 1.0968320479136095e-07, "loss": 0.006, "step": 29400 }, { "epoch": 31.12169312169312, "grad_norm": 0.19013342261314392, "learning_rate": 1.0606482786936767e-07, "loss": 0.0065, "step": 29410 }, { "epoch": 31.132275132275133, "grad_norm": 0.11198524385690689, "learning_rate": 1.0250707678329808e-07, "loss": 0.0057, "step": 29420 }, { "epoch": 31.142857142857142, "grad_norm": 0.16120460629463196, "learning_rate": 9.900995585615525e-08, "loss": 0.0066, "step": 29430 }, { "epoch": 31.15343915343915, "grad_norm": 0.18561072647571564, "learning_rate": 9.55734693372734e-08, "loss": 0.0073, "step": 29440 }, { "epoch": 31.164021164021165, "grad_norm": 0.13351111114025116, "learning_rate": 9.219762140231236e-08, "loss": 0.0066, "step": 29450 }, { "epoch": 31.174603174603174, "grad_norm": 0.07596392184495926, "learning_rate": 8.888241615322978e-08, "loss": 0.0064, "step": 29460 }, { "epoch": 31.185185185185187, "grad_norm": 0.0857609212398529, "learning_rate": 8.562785761833114e-08, "loss": 0.0088, "step": 29470 }, { "epoch": 31.195767195767196, "grad_norm": 0.1178661659359932, "learning_rate": 8.243394975219753e-08, "loss": 0.008, "step": 29480 }, { "epoch": 31.206349206349206, "grad_norm": 0.07858454436063766, "learning_rate": 7.930069643573568e-08, "loss": 0.0051, "step": 29490 }, { "epoch": 31.21693121693122, "grad_norm": 0.2028738111257553, "learning_rate": 7.622810147614456e-08, "loss": 0.0092, "step": 29500 }, { "epoch": 31.227513227513228, "grad_norm": 0.15929238498210907, "learning_rate": 7.321616860690995e-08, "loss": 0.0083, "step": 29510 }, { "epoch": 31.238095238095237, "grad_norm": 0.22736676037311554, "learning_rate": 7.026490148782095e-08, "loss": 0.0074, "step": 29520 }, { "epoch": 31.24867724867725, "grad_norm": 0.1307639628648758, "learning_rate": 6.737430370494236e-08, "loss": 0.007, "step": 29530 }, { "epoch": 31.25925925925926, "grad_norm": 0.0713730975985527, "learning_rate": 6.454437877062569e-08, "loss": 0.0061, "step": 29540 }, { "epoch": 31.26984126984127, "grad_norm": 0.12002429366111755, "learning_rate": 6.177513012349256e-08, "loss": 0.0091, "step": 29550 }, { "epoch": 31.280423280423282, "grad_norm": 0.10917627811431885, "learning_rate": 5.9066561128445775e-08, "loss": 0.0063, "step": 29560 }, { "epoch": 31.29100529100529, "grad_norm": 0.17165903747081757, "learning_rate": 5.6418675076641556e-08, "loss": 0.0061, "step": 29570 }, { "epoch": 31.3015873015873, "grad_norm": 0.08978458493947983, "learning_rate": 5.383147518552845e-08, "loss": 0.0071, "step": 29580 }, { "epoch": 31.312169312169313, "grad_norm": 0.1344803273677826, "learning_rate": 5.1304964598786196e-08, "loss": 0.0066, "step": 29590 }, { "epoch": 31.322751322751323, "grad_norm": 0.07440406829118729, "learning_rate": 4.883914638636467e-08, "loss": 0.0069, "step": 29600 }, { "epoch": 31.333333333333332, "grad_norm": 0.13957320153713226, "learning_rate": 4.643402354446713e-08, "loss": 0.0081, "step": 29610 }, { "epoch": 31.343915343915345, "grad_norm": 0.07065512239933014, "learning_rate": 4.4089598995544766e-08, "loss": 0.0079, "step": 29620 }, { "epoch": 31.354497354497354, "grad_norm": 0.048440031707286835, "learning_rate": 4.180587558829663e-08, "loss": 0.0069, "step": 29630 }, { "epoch": 31.365079365079364, "grad_norm": 0.10456069558858871, "learning_rate": 3.958285609765855e-08, "loss": 0.0066, "step": 29640 }, { "epoch": 31.375661375661377, "grad_norm": 0.15276610851287842, "learning_rate": 3.74205432248087e-08, "loss": 0.0061, "step": 29650 }, { "epoch": 31.386243386243386, "grad_norm": 0.20709136128425598, "learning_rate": 3.531893959716204e-08, "loss": 0.0079, "step": 29660 }, { "epoch": 31.396825396825395, "grad_norm": 0.15138497948646545, "learning_rate": 3.327804776837029e-08, "loss": 0.0067, "step": 29670 }, { "epoch": 31.40740740740741, "grad_norm": 0.16533030569553375, "learning_rate": 3.129787021829977e-08, "loss": 0.0066, "step": 29680 }, { "epoch": 31.417989417989418, "grad_norm": 0.2970920205116272, "learning_rate": 2.9378409353059133e-08, "loss": 0.0072, "step": 29690 }, { "epoch": 31.428571428571427, "grad_norm": 0.12679509818553925, "learning_rate": 2.7519667504971593e-08, "loss": 0.0072, "step": 29700 }, { "epoch": 31.43915343915344, "grad_norm": 0.08025428652763367, "learning_rate": 2.572164693258605e-08, "loss": 0.006, "step": 29710 }, { "epoch": 31.44973544973545, "grad_norm": 0.08054651319980621, "learning_rate": 2.3984349820665996e-08, "loss": 0.0051, "step": 29720 }, { "epoch": 31.46031746031746, "grad_norm": 0.16615915298461914, "learning_rate": 2.2307778280189485e-08, "loss": 0.0063, "step": 29730 }, { "epoch": 31.47089947089947, "grad_norm": 0.10970409214496613, "learning_rate": 2.0691934348354704e-08, "loss": 0.0076, "step": 29740 }, { "epoch": 31.48148148148148, "grad_norm": 0.1142452210187912, "learning_rate": 1.9136819988557763e-08, "loss": 0.0055, "step": 29750 }, { "epoch": 31.49206349206349, "grad_norm": 0.10398027300834656, "learning_rate": 1.7642437090414908e-08, "loss": 0.0063, "step": 29760 }, { "epoch": 31.502645502645503, "grad_norm": 0.1458752155303955, "learning_rate": 1.6208787469734755e-08, "loss": 0.0056, "step": 29770 }, { "epoch": 31.513227513227513, "grad_norm": 0.10636114329099655, "learning_rate": 1.483587286854604e-08, "loss": 0.009, "step": 29780 }, { "epoch": 31.523809523809526, "grad_norm": 0.08823571354150772, "learning_rate": 1.3523694955064336e-08, "loss": 0.0056, "step": 29790 }, { "epoch": 31.534391534391535, "grad_norm": 0.15484894812107086, "learning_rate": 1.2272255323708681e-08, "loss": 0.0077, "step": 29800 }, { "epoch": 31.544973544973544, "grad_norm": 0.16487932205200195, "learning_rate": 1.1081555495096042e-08, "loss": 0.008, "step": 29810 }, { "epoch": 31.555555555555557, "grad_norm": 0.054375674575567245, "learning_rate": 9.951596916041306e-09, "loss": 0.0065, "step": 29820 }, { "epoch": 31.566137566137566, "grad_norm": 0.1022561639547348, "learning_rate": 8.882380959551739e-09, "loss": 0.008, "step": 29830 }, { "epoch": 31.576719576719576, "grad_norm": 0.13162197172641754, "learning_rate": 7.873908924821428e-09, "loss": 0.007, "step": 29840 }, { "epoch": 31.58730158730159, "grad_norm": 0.4315018951892853, "learning_rate": 6.926182037242379e-09, "loss": 0.007, "step": 29850 }, { "epoch": 31.597883597883598, "grad_norm": 0.18685129284858704, "learning_rate": 6.039201448387876e-09, "loss": 0.0063, "step": 29860 }, { "epoch": 31.608465608465607, "grad_norm": 0.0756327286362648, "learning_rate": 5.212968236029125e-09, "loss": 0.0063, "step": 29870 }, { "epoch": 31.61904761904762, "grad_norm": 0.5953189730644226, "learning_rate": 4.447483404118602e-09, "loss": 0.0064, "step": 29880 }, { "epoch": 31.62962962962963, "grad_norm": 0.07357990741729736, "learning_rate": 3.742747882784503e-09, "loss": 0.0065, "step": 29890 }, { "epoch": 31.64021164021164, "grad_norm": 0.07009071111679077, "learning_rate": 3.0987625283473987e-09, "loss": 0.0085, "step": 29900 }, { "epoch": 31.650793650793652, "grad_norm": 0.16222986578941345, "learning_rate": 2.515528123320232e-09, "loss": 0.0058, "step": 29910 }, { "epoch": 31.66137566137566, "grad_norm": 0.13371728360652924, "learning_rate": 1.9930453763750133e-09, "loss": 0.0063, "step": 29920 }, { "epoch": 31.67195767195767, "grad_norm": 0.10468500107526779, "learning_rate": 1.5313149223872281e-09, "loss": 0.0072, "step": 29930 }, { "epoch": 31.682539682539684, "grad_norm": 0.10956578701734543, "learning_rate": 1.1303373224025305e-09, "loss": 0.007, "step": 29940 }, { "epoch": 31.693121693121693, "grad_norm": 0.1075982004404068, "learning_rate": 7.901130636367438e-10, "loss": 0.0073, "step": 29950 }, { "epoch": 31.703703703703702, "grad_norm": 0.06719832122325897, "learning_rate": 5.106425595036158e-10, "loss": 0.0061, "step": 29960 }, { "epoch": 31.714285714285715, "grad_norm": 0.142010897397995, "learning_rate": 2.9192614958706286e-10, "loss": 0.0071, "step": 29970 }, { "epoch": 31.724867724867725, "grad_norm": 0.05848604813218117, "learning_rate": 1.3396409964117064e-10, "loss": 0.006, "step": 29980 }, { "epoch": 31.735449735449734, "grad_norm": 0.07950931042432785, "learning_rate": 3.6756601606846574e-11, "loss": 0.0081, "step": 29990 }, { "epoch": 31.746031746031747, "grad_norm": 0.08285913616418839, "learning_rate": 3.0377360626943076e-13, "loss": 0.0064, "step": 30000 } ], "logging_steps": 10, "max_steps": 30000, "num_input_tokens_seen": 0, "num_train_epochs": 32, "save_steps": 15000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 64, "trial_name": null, "trial_params": null }