{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 283.1858407079646, "eval_steps": 500, "global_step": 32000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.08849557522123894, "grad_norm": 6.029973983764648, "learning_rate": 1.8e-07, "loss": 0.8556, "step": 10 }, { "epoch": 0.17699115044247787, "grad_norm": 5.713200569152832, "learning_rate": 3.8e-07, "loss": 0.8594, "step": 20 }, { "epoch": 0.26548672566371684, "grad_norm": 5.5482683181762695, "learning_rate": 5.8e-07, "loss": 0.8439, "step": 30 }, { "epoch": 0.35398230088495575, "grad_norm": 4.973087310791016, "learning_rate": 7.8e-07, "loss": 0.8031, "step": 40 }, { "epoch": 0.4424778761061947, "grad_norm": 4.316706657409668, "learning_rate": 9.8e-07, "loss": 0.7054, "step": 50 }, { "epoch": 0.5309734513274337, "grad_norm": 2.757664680480957, "learning_rate": 1.18e-06, "loss": 0.6014, "step": 60 }, { "epoch": 0.6194690265486725, "grad_norm": 2.4813029766082764, "learning_rate": 1.3800000000000001e-06, "loss": 0.4621, "step": 70 }, { "epoch": 0.7079646017699115, "grad_norm": 1.1163899898529053, "learning_rate": 1.5800000000000003e-06, "loss": 0.3772, "step": 80 }, { "epoch": 0.7964601769911505, "grad_norm": 0.6677577495574951, "learning_rate": 1.7800000000000001e-06, "loss": 0.2607, "step": 90 }, { "epoch": 0.8849557522123894, "grad_norm": 0.5095245242118835, "learning_rate": 1.98e-06, "loss": 0.2246, "step": 100 }, { "epoch": 0.9734513274336283, "grad_norm": 0.38245514035224915, "learning_rate": 2.1800000000000003e-06, "loss": 0.1869, "step": 110 }, { "epoch": 1.0619469026548674, "grad_norm": 0.3191615343093872, "learning_rate": 2.38e-06, "loss": 0.1633, "step": 120 }, { "epoch": 1.1504424778761062, "grad_norm": 0.2773362398147583, "learning_rate": 2.5800000000000003e-06, "loss": 0.1487, "step": 130 }, { "epoch": 1.238938053097345, "grad_norm": 0.22420527040958405, "learning_rate": 2.78e-06, "loss": 0.1311, "step": 140 }, { "epoch": 1.3274336283185841, "grad_norm": 0.2951527237892151, "learning_rate": 2.9800000000000003e-06, "loss": 0.1256, "step": 150 }, { "epoch": 1.415929203539823, "grad_norm": 0.2016792893409729, "learning_rate": 3.1800000000000005e-06, "loss": 0.1161, "step": 160 }, { "epoch": 1.504424778761062, "grad_norm": 0.23239828646183014, "learning_rate": 3.38e-06, "loss": 0.1121, "step": 170 }, { "epoch": 1.592920353982301, "grad_norm": 0.21232590079307556, "learning_rate": 3.58e-06, "loss": 0.1022, "step": 180 }, { "epoch": 1.6814159292035398, "grad_norm": 0.3535752296447754, "learning_rate": 3.7800000000000002e-06, "loss": 0.099, "step": 190 }, { "epoch": 1.7699115044247788, "grad_norm": 0.16969171166419983, "learning_rate": 3.98e-06, "loss": 0.0942, "step": 200 }, { "epoch": 1.8584070796460177, "grad_norm": 0.18709735572338104, "learning_rate": 4.18e-06, "loss": 0.0852, "step": 210 }, { "epoch": 1.9469026548672566, "grad_norm": 0.18421606719493866, "learning_rate": 4.38e-06, "loss": 0.0827, "step": 220 }, { "epoch": 2.0353982300884956, "grad_norm": 0.18908490240573883, "learning_rate": 4.58e-06, "loss": 0.0785, "step": 230 }, { "epoch": 2.1238938053097347, "grad_norm": 0.18462112545967102, "learning_rate": 4.780000000000001e-06, "loss": 0.0751, "step": 240 }, { "epoch": 2.2123893805309733, "grad_norm": 0.19543930888175964, "learning_rate": 4.98e-06, "loss": 0.0724, "step": 250 }, { "epoch": 2.3008849557522124, "grad_norm": 0.17802758514881134, "learning_rate": 5.18e-06, "loss": 0.0674, "step": 260 }, { "epoch": 2.3893805309734515, "grad_norm": 0.14770545065402985, "learning_rate": 5.38e-06, "loss": 0.0649, "step": 270 }, { "epoch": 2.47787610619469, "grad_norm": 0.17168474197387695, "learning_rate": 5.580000000000001e-06, "loss": 0.0627, "step": 280 }, { "epoch": 2.566371681415929, "grad_norm": 0.16846390068531036, "learning_rate": 5.78e-06, "loss": 0.0592, "step": 290 }, { "epoch": 2.6548672566371683, "grad_norm": 0.17711997032165527, "learning_rate": 5.98e-06, "loss": 0.058, "step": 300 }, { "epoch": 2.7433628318584073, "grad_norm": 0.21773852407932281, "learning_rate": 6.18e-06, "loss": 0.056, "step": 310 }, { "epoch": 2.831858407079646, "grad_norm": 0.16227875649929047, "learning_rate": 6.38e-06, "loss": 0.0552, "step": 320 }, { "epoch": 2.920353982300885, "grad_norm": 0.19825047254562378, "learning_rate": 6.58e-06, "loss": 0.0514, "step": 330 }, { "epoch": 3.0088495575221237, "grad_norm": 0.2886408567428589, "learning_rate": 6.78e-06, "loss": 0.05, "step": 340 }, { "epoch": 3.0973451327433628, "grad_norm": 0.2186298817396164, "learning_rate": 6.98e-06, "loss": 0.047, "step": 350 }, { "epoch": 3.185840707964602, "grad_norm": 0.18055620789527893, "learning_rate": 7.180000000000001e-06, "loss": 0.0481, "step": 360 }, { "epoch": 3.274336283185841, "grad_norm": 0.19108404219150543, "learning_rate": 7.3800000000000005e-06, "loss": 0.0462, "step": 370 }, { "epoch": 3.3628318584070795, "grad_norm": 0.23328030109405518, "learning_rate": 7.580000000000001e-06, "loss": 0.0418, "step": 380 }, { "epoch": 3.4513274336283186, "grad_norm": 0.16056862473487854, "learning_rate": 7.78e-06, "loss": 0.0404, "step": 390 }, { "epoch": 3.5398230088495577, "grad_norm": 0.17877085506916046, "learning_rate": 7.98e-06, "loss": 0.0402, "step": 400 }, { "epoch": 3.6283185840707963, "grad_norm": 0.15461580455303192, "learning_rate": 8.18e-06, "loss": 0.039, "step": 410 }, { "epoch": 3.7168141592920354, "grad_norm": 0.19305895268917084, "learning_rate": 8.380000000000001e-06, "loss": 0.0383, "step": 420 }, { "epoch": 3.8053097345132745, "grad_norm": 0.1814298927783966, "learning_rate": 8.580000000000001e-06, "loss": 0.0357, "step": 430 }, { "epoch": 3.893805309734513, "grad_norm": 0.16759811341762543, "learning_rate": 8.78e-06, "loss": 0.0359, "step": 440 }, { "epoch": 3.982300884955752, "grad_norm": 0.17231112718582153, "learning_rate": 8.98e-06, "loss": 0.0353, "step": 450 }, { "epoch": 4.070796460176991, "grad_norm": 0.1795455813407898, "learning_rate": 9.180000000000002e-06, "loss": 0.0328, "step": 460 }, { "epoch": 4.15929203539823, "grad_norm": 0.19713948667049408, "learning_rate": 9.38e-06, "loss": 0.0327, "step": 470 }, { "epoch": 4.247787610619469, "grad_norm": 0.26041460037231445, "learning_rate": 9.58e-06, "loss": 0.0321, "step": 480 }, { "epoch": 4.336283185840708, "grad_norm": 0.1842440962791443, "learning_rate": 9.78e-06, "loss": 0.0305, "step": 490 }, { "epoch": 4.424778761061947, "grad_norm": 0.14161691069602966, "learning_rate": 9.980000000000001e-06, "loss": 0.032, "step": 500 }, { "epoch": 4.513274336283186, "grad_norm": 0.18040604889392853, "learning_rate": 1.018e-05, "loss": 0.0294, "step": 510 }, { "epoch": 4.601769911504425, "grad_norm": 0.2446688562631607, "learning_rate": 1.038e-05, "loss": 0.0296, "step": 520 }, { "epoch": 4.6902654867256635, "grad_norm": 0.24773721396923065, "learning_rate": 1.058e-05, "loss": 0.0278, "step": 530 }, { "epoch": 4.778761061946903, "grad_norm": 0.18613311648368835, "learning_rate": 1.0780000000000002e-05, "loss": 0.0285, "step": 540 }, { "epoch": 4.867256637168142, "grad_norm": 0.21972256898880005, "learning_rate": 1.098e-05, "loss": 0.0284, "step": 550 }, { "epoch": 4.95575221238938, "grad_norm": 0.1792360544204712, "learning_rate": 1.118e-05, "loss": 0.0275, "step": 560 }, { "epoch": 5.04424778761062, "grad_norm": 0.19835522770881653, "learning_rate": 1.1380000000000001e-05, "loss": 0.0261, "step": 570 }, { "epoch": 5.132743362831858, "grad_norm": 0.19945470988750458, "learning_rate": 1.1580000000000001e-05, "loss": 0.0266, "step": 580 }, { "epoch": 5.221238938053097, "grad_norm": 0.18404513597488403, "learning_rate": 1.178e-05, "loss": 0.0259, "step": 590 }, { "epoch": 5.3097345132743365, "grad_norm": 0.22277390956878662, "learning_rate": 1.198e-05, "loss": 0.0273, "step": 600 }, { "epoch": 5.398230088495575, "grad_norm": 0.2083224356174469, "learning_rate": 1.2180000000000002e-05, "loss": 0.0261, "step": 610 }, { "epoch": 5.486725663716814, "grad_norm": 0.2846670150756836, "learning_rate": 1.238e-05, "loss": 0.0258, "step": 620 }, { "epoch": 5.575221238938053, "grad_norm": 0.2090553641319275, "learning_rate": 1.258e-05, "loss": 0.026, "step": 630 }, { "epoch": 5.663716814159292, "grad_norm": 0.24054387211799622, "learning_rate": 1.278e-05, "loss": 0.024, "step": 640 }, { "epoch": 5.752212389380531, "grad_norm": 0.24390311539173126, "learning_rate": 1.2980000000000001e-05, "loss": 0.0242, "step": 650 }, { "epoch": 5.84070796460177, "grad_norm": 0.19209808111190796, "learning_rate": 1.3180000000000001e-05, "loss": 0.0238, "step": 660 }, { "epoch": 5.929203539823009, "grad_norm": 0.26109975576400757, "learning_rate": 1.338e-05, "loss": 0.0233, "step": 670 }, { "epoch": 6.017699115044247, "grad_norm": 0.22407938539981842, "learning_rate": 1.358e-05, "loss": 0.0236, "step": 680 }, { "epoch": 6.106194690265487, "grad_norm": 0.2744028866291046, "learning_rate": 1.3780000000000002e-05, "loss": 0.0228, "step": 690 }, { "epoch": 6.1946902654867255, "grad_norm": 0.17840787768363953, "learning_rate": 1.3980000000000002e-05, "loss": 0.024, "step": 700 }, { "epoch": 6.283185840707965, "grad_norm": 0.21017034351825714, "learning_rate": 1.4180000000000001e-05, "loss": 0.0236, "step": 710 }, { "epoch": 6.371681415929204, "grad_norm": 0.23242029547691345, "learning_rate": 1.4380000000000001e-05, "loss": 0.0228, "step": 720 }, { "epoch": 6.460176991150442, "grad_norm": 0.18945607542991638, "learning_rate": 1.4580000000000003e-05, "loss": 0.0221, "step": 730 }, { "epoch": 6.548672566371682, "grad_norm": 0.2534150183200836, "learning_rate": 1.4779999999999999e-05, "loss": 0.0215, "step": 740 }, { "epoch": 6.6371681415929205, "grad_norm": 0.2407999187707901, "learning_rate": 1.4979999999999999e-05, "loss": 0.0225, "step": 750 }, { "epoch": 6.725663716814159, "grad_norm": 0.17078283429145813, "learning_rate": 1.518e-05, "loss": 0.0221, "step": 760 }, { "epoch": 6.814159292035399, "grad_norm": 0.20780600607395172, "learning_rate": 1.538e-05, "loss": 0.0223, "step": 770 }, { "epoch": 6.902654867256637, "grad_norm": 0.2091590166091919, "learning_rate": 1.558e-05, "loss": 0.021, "step": 780 }, { "epoch": 6.991150442477876, "grad_norm": 0.1971716284751892, "learning_rate": 1.578e-05, "loss": 0.0204, "step": 790 }, { "epoch": 7.079646017699115, "grad_norm": 0.20692206919193268, "learning_rate": 1.598e-05, "loss": 0.0208, "step": 800 }, { "epoch": 7.168141592920354, "grad_norm": 0.1793445497751236, "learning_rate": 1.618e-05, "loss": 0.0208, "step": 810 }, { "epoch": 7.256637168141593, "grad_norm": 0.2733965218067169, "learning_rate": 1.6380000000000002e-05, "loss": 0.0206, "step": 820 }, { "epoch": 7.345132743362832, "grad_norm": 0.17656664550304413, "learning_rate": 1.658e-05, "loss": 0.0207, "step": 830 }, { "epoch": 7.433628318584071, "grad_norm": 0.2220483422279358, "learning_rate": 1.6780000000000002e-05, "loss": 0.0208, "step": 840 }, { "epoch": 7.522123893805309, "grad_norm": 0.2391643226146698, "learning_rate": 1.698e-05, "loss": 0.0217, "step": 850 }, { "epoch": 7.610619469026549, "grad_norm": 0.18046431243419647, "learning_rate": 1.718e-05, "loss": 0.021, "step": 860 }, { "epoch": 7.699115044247788, "grad_norm": 0.17591039836406708, "learning_rate": 1.7380000000000003e-05, "loss": 0.0201, "step": 870 }, { "epoch": 7.787610619469026, "grad_norm": 0.24627521634101868, "learning_rate": 1.758e-05, "loss": 0.0195, "step": 880 }, { "epoch": 7.876106194690266, "grad_norm": 0.21443326771259308, "learning_rate": 1.7780000000000003e-05, "loss": 0.0194, "step": 890 }, { "epoch": 7.964601769911504, "grad_norm": 0.19286300241947174, "learning_rate": 1.798e-05, "loss": 0.0197, "step": 900 }, { "epoch": 8.053097345132743, "grad_norm": 0.22124329209327698, "learning_rate": 1.818e-05, "loss": 0.0192, "step": 910 }, { "epoch": 8.141592920353983, "grad_norm": 0.22199001908302307, "learning_rate": 1.838e-05, "loss": 0.0198, "step": 920 }, { "epoch": 8.230088495575222, "grad_norm": 0.21194124221801758, "learning_rate": 1.858e-05, "loss": 0.02, "step": 930 }, { "epoch": 8.31858407079646, "grad_norm": 0.24529924988746643, "learning_rate": 1.878e-05, "loss": 0.0187, "step": 940 }, { "epoch": 8.4070796460177, "grad_norm": 0.25683167576789856, "learning_rate": 1.898e-05, "loss": 0.0184, "step": 950 }, { "epoch": 8.495575221238939, "grad_norm": 0.283501535654068, "learning_rate": 1.918e-05, "loss": 0.02, "step": 960 }, { "epoch": 8.584070796460177, "grad_norm": 0.2274293303489685, "learning_rate": 1.938e-05, "loss": 0.0187, "step": 970 }, { "epoch": 8.672566371681416, "grad_norm": 0.18131929636001587, "learning_rate": 1.9580000000000002e-05, "loss": 0.0187, "step": 980 }, { "epoch": 8.761061946902656, "grad_norm": 0.2914557158946991, "learning_rate": 1.978e-05, "loss": 0.019, "step": 990 }, { "epoch": 8.849557522123893, "grad_norm": 0.19533103704452515, "learning_rate": 1.9980000000000002e-05, "loss": 0.0191, "step": 1000 }, { "epoch": 8.938053097345133, "grad_norm": 0.20811727643013, "learning_rate": 2.0180000000000003e-05, "loss": 0.0177, "step": 1010 }, { "epoch": 9.026548672566372, "grad_norm": 0.22807861864566803, "learning_rate": 2.038e-05, "loss": 0.0187, "step": 1020 }, { "epoch": 9.11504424778761, "grad_norm": 0.2197323590517044, "learning_rate": 2.0580000000000003e-05, "loss": 0.0177, "step": 1030 }, { "epoch": 9.20353982300885, "grad_norm": 0.30404001474380493, "learning_rate": 2.078e-05, "loss": 0.0183, "step": 1040 }, { "epoch": 9.29203539823009, "grad_norm": 0.3043852746486664, "learning_rate": 2.098e-05, "loss": 0.0175, "step": 1050 }, { "epoch": 9.380530973451327, "grad_norm": 0.24773834645748138, "learning_rate": 2.118e-05, "loss": 0.0182, "step": 1060 }, { "epoch": 9.469026548672566, "grad_norm": 0.22849597036838531, "learning_rate": 2.138e-05, "loss": 0.018, "step": 1070 }, { "epoch": 9.557522123893806, "grad_norm": 0.25134411454200745, "learning_rate": 2.158e-05, "loss": 0.0187, "step": 1080 }, { "epoch": 9.646017699115044, "grad_norm": 0.21612103283405304, "learning_rate": 2.178e-05, "loss": 0.0178, "step": 1090 }, { "epoch": 9.734513274336283, "grad_norm": 0.2568541169166565, "learning_rate": 2.198e-05, "loss": 0.0178, "step": 1100 }, { "epoch": 9.823008849557523, "grad_norm": 0.2351471185684204, "learning_rate": 2.218e-05, "loss": 0.0176, "step": 1110 }, { "epoch": 9.91150442477876, "grad_norm": 0.17481906712055206, "learning_rate": 2.2380000000000003e-05, "loss": 0.0182, "step": 1120 }, { "epoch": 10.0, "grad_norm": 0.18670059740543365, "learning_rate": 2.258e-05, "loss": 0.0187, "step": 1130 }, { "epoch": 10.08849557522124, "grad_norm": 0.24376435577869415, "learning_rate": 2.2780000000000002e-05, "loss": 0.0169, "step": 1140 }, { "epoch": 10.176991150442477, "grad_norm": 0.2629312574863434, "learning_rate": 2.298e-05, "loss": 0.0171, "step": 1150 }, { "epoch": 10.265486725663717, "grad_norm": 0.2862442433834076, "learning_rate": 2.318e-05, "loss": 0.0176, "step": 1160 }, { "epoch": 10.353982300884956, "grad_norm": 0.21987782418727875, "learning_rate": 2.3380000000000003e-05, "loss": 0.0175, "step": 1170 }, { "epoch": 10.442477876106194, "grad_norm": 0.220542311668396, "learning_rate": 2.358e-05, "loss": 0.0164, "step": 1180 }, { "epoch": 10.530973451327434, "grad_norm": 0.2129259556531906, "learning_rate": 2.3780000000000003e-05, "loss": 0.0165, "step": 1190 }, { "epoch": 10.619469026548673, "grad_norm": 0.23727720975875854, "learning_rate": 2.398e-05, "loss": 0.0165, "step": 1200 }, { "epoch": 10.70796460176991, "grad_norm": 0.20429955422878265, "learning_rate": 2.418e-05, "loss": 0.0161, "step": 1210 }, { "epoch": 10.79646017699115, "grad_norm": 0.1692054122686386, "learning_rate": 2.438e-05, "loss": 0.0168, "step": 1220 }, { "epoch": 10.88495575221239, "grad_norm": 0.22281883656978607, "learning_rate": 2.4580000000000002e-05, "loss": 0.0163, "step": 1230 }, { "epoch": 10.973451327433628, "grad_norm": 0.19232513010501862, "learning_rate": 2.478e-05, "loss": 0.0168, "step": 1240 }, { "epoch": 11.061946902654867, "grad_norm": 0.1965012103319168, "learning_rate": 2.498e-05, "loss": 0.0167, "step": 1250 }, { "epoch": 11.150442477876107, "grad_norm": 0.19214503467082977, "learning_rate": 2.5180000000000003e-05, "loss": 0.0155, "step": 1260 }, { "epoch": 11.238938053097344, "grad_norm": 0.18579241633415222, "learning_rate": 2.5380000000000004e-05, "loss": 0.0165, "step": 1270 }, { "epoch": 11.327433628318584, "grad_norm": 0.21648317575454712, "learning_rate": 2.5580000000000002e-05, "loss": 0.0157, "step": 1280 }, { "epoch": 11.415929203539823, "grad_norm": 0.20621834695339203, "learning_rate": 2.5779999999999997e-05, "loss": 0.016, "step": 1290 }, { "epoch": 11.504424778761061, "grad_norm": 0.26772016286849976, "learning_rate": 2.598e-05, "loss": 0.0153, "step": 1300 }, { "epoch": 11.5929203539823, "grad_norm": 0.23452354967594147, "learning_rate": 2.618e-05, "loss": 0.0154, "step": 1310 }, { "epoch": 11.68141592920354, "grad_norm": 0.23407390713691711, "learning_rate": 2.6379999999999998e-05, "loss": 0.0162, "step": 1320 }, { "epoch": 11.769911504424778, "grad_norm": 0.23372036218643188, "learning_rate": 2.658e-05, "loss": 0.0156, "step": 1330 }, { "epoch": 11.858407079646017, "grad_norm": 0.21054409444332123, "learning_rate": 2.678e-05, "loss": 0.0161, "step": 1340 }, { "epoch": 11.946902654867257, "grad_norm": 0.1630723625421524, "learning_rate": 2.698e-05, "loss": 0.0151, "step": 1350 }, { "epoch": 12.035398230088495, "grad_norm": 0.22612281143665314, "learning_rate": 2.718e-05, "loss": 0.0157, "step": 1360 }, { "epoch": 12.123893805309734, "grad_norm": 0.2516733705997467, "learning_rate": 2.738e-05, "loss": 0.0152, "step": 1370 }, { "epoch": 12.212389380530974, "grad_norm": 0.2558288872241974, "learning_rate": 2.758e-05, "loss": 0.0154, "step": 1380 }, { "epoch": 12.300884955752213, "grad_norm": 0.2218429148197174, "learning_rate": 2.778e-05, "loss": 0.015, "step": 1390 }, { "epoch": 12.389380530973451, "grad_norm": 0.23277941346168518, "learning_rate": 2.798e-05, "loss": 0.0149, "step": 1400 }, { "epoch": 12.47787610619469, "grad_norm": 0.23187731206417084, "learning_rate": 2.818e-05, "loss": 0.0153, "step": 1410 }, { "epoch": 12.56637168141593, "grad_norm": 0.24574844539165497, "learning_rate": 2.8380000000000003e-05, "loss": 0.0143, "step": 1420 }, { "epoch": 12.654867256637168, "grad_norm": 0.1482018530368805, "learning_rate": 2.858e-05, "loss": 0.0151, "step": 1430 }, { "epoch": 12.743362831858407, "grad_norm": 0.19406455755233765, "learning_rate": 2.8780000000000002e-05, "loss": 0.0144, "step": 1440 }, { "epoch": 12.831858407079647, "grad_norm": 0.19292591512203217, "learning_rate": 2.898e-05, "loss": 0.0161, "step": 1450 }, { "epoch": 12.920353982300885, "grad_norm": 0.2019025683403015, "learning_rate": 2.9180000000000002e-05, "loss": 0.016, "step": 1460 }, { "epoch": 13.008849557522124, "grad_norm": 0.26240357756614685, "learning_rate": 2.9380000000000003e-05, "loss": 0.0152, "step": 1470 }, { "epoch": 13.097345132743364, "grad_norm": 0.22358892858028412, "learning_rate": 2.958e-05, "loss": 0.0152, "step": 1480 }, { "epoch": 13.185840707964601, "grad_norm": 0.24458648264408112, "learning_rate": 2.9780000000000003e-05, "loss": 0.0155, "step": 1490 }, { "epoch": 13.274336283185841, "grad_norm": 0.17762477695941925, "learning_rate": 2.998e-05, "loss": 0.0156, "step": 1500 }, { "epoch": 13.36283185840708, "grad_norm": 0.2243274301290512, "learning_rate": 3.0180000000000002e-05, "loss": 0.0152, "step": 1510 }, { "epoch": 13.451327433628318, "grad_norm": 0.19519586861133575, "learning_rate": 3.0380000000000004e-05, "loss": 0.0145, "step": 1520 }, { "epoch": 13.539823008849558, "grad_norm": 0.27256205677986145, "learning_rate": 3.058e-05, "loss": 0.0149, "step": 1530 }, { "epoch": 13.628318584070797, "grad_norm": 0.22428080439567566, "learning_rate": 3.078e-05, "loss": 0.0141, "step": 1540 }, { "epoch": 13.716814159292035, "grad_norm": 0.2076885998249054, "learning_rate": 3.0980000000000005e-05, "loss": 0.0138, "step": 1550 }, { "epoch": 13.805309734513274, "grad_norm": 0.21097923815250397, "learning_rate": 3.118e-05, "loss": 0.014, "step": 1560 }, { "epoch": 13.893805309734514, "grad_norm": 0.214741513133049, "learning_rate": 3.138e-05, "loss": 0.0141, "step": 1570 }, { "epoch": 13.982300884955752, "grad_norm": 0.20153780281543732, "learning_rate": 3.1580000000000006e-05, "loss": 0.0141, "step": 1580 }, { "epoch": 14.070796460176991, "grad_norm": 0.1785677969455719, "learning_rate": 3.1780000000000004e-05, "loss": 0.0142, "step": 1590 }, { "epoch": 14.15929203539823, "grad_norm": 0.20385058224201202, "learning_rate": 3.198e-05, "loss": 0.0147, "step": 1600 }, { "epoch": 14.247787610619469, "grad_norm": 0.23648928105831146, "learning_rate": 3.218e-05, "loss": 0.0143, "step": 1610 }, { "epoch": 14.336283185840708, "grad_norm": 0.28338080644607544, "learning_rate": 3.238e-05, "loss": 0.0141, "step": 1620 }, { "epoch": 14.424778761061948, "grad_norm": 0.20188190042972565, "learning_rate": 3.2579999999999996e-05, "loss": 0.0135, "step": 1630 }, { "epoch": 14.513274336283185, "grad_norm": 0.22046928107738495, "learning_rate": 3.278e-05, "loss": 0.0143, "step": 1640 }, { "epoch": 14.601769911504425, "grad_norm": 0.2269722819328308, "learning_rate": 3.298e-05, "loss": 0.0139, "step": 1650 }, { "epoch": 14.690265486725664, "grad_norm": 0.20017528533935547, "learning_rate": 3.318e-05, "loss": 0.0137, "step": 1660 }, { "epoch": 14.778761061946902, "grad_norm": 0.19861406087875366, "learning_rate": 3.338e-05, "loss": 0.0141, "step": 1670 }, { "epoch": 14.867256637168142, "grad_norm": 0.16092416644096375, "learning_rate": 3.358e-05, "loss": 0.0142, "step": 1680 }, { "epoch": 14.955752212389381, "grad_norm": 0.19187436997890472, "learning_rate": 3.378e-05, "loss": 0.0144, "step": 1690 }, { "epoch": 15.044247787610619, "grad_norm": 0.19412851333618164, "learning_rate": 3.398e-05, "loss": 0.0138, "step": 1700 }, { "epoch": 15.132743362831858, "grad_norm": 0.15217262506484985, "learning_rate": 3.418e-05, "loss": 0.0137, "step": 1710 }, { "epoch": 15.221238938053098, "grad_norm": 0.17059360444545746, "learning_rate": 3.438e-05, "loss": 0.0143, "step": 1720 }, { "epoch": 15.309734513274336, "grad_norm": 0.16499827802181244, "learning_rate": 3.4580000000000004e-05, "loss": 0.0132, "step": 1730 }, { "epoch": 15.398230088495575, "grad_norm": 0.17739605903625488, "learning_rate": 3.478e-05, "loss": 0.0134, "step": 1740 }, { "epoch": 15.486725663716815, "grad_norm": 0.2053276002407074, "learning_rate": 3.498e-05, "loss": 0.0134, "step": 1750 }, { "epoch": 15.575221238938052, "grad_norm": 0.25212031602859497, "learning_rate": 3.518e-05, "loss": 0.0127, "step": 1760 }, { "epoch": 15.663716814159292, "grad_norm": 0.1611550897359848, "learning_rate": 3.5380000000000003e-05, "loss": 0.0127, "step": 1770 }, { "epoch": 15.752212389380531, "grad_norm": 0.18005068600177765, "learning_rate": 3.558e-05, "loss": 0.013, "step": 1780 }, { "epoch": 15.84070796460177, "grad_norm": 0.21542882919311523, "learning_rate": 3.578e-05, "loss": 0.0132, "step": 1790 }, { "epoch": 15.929203539823009, "grad_norm": 0.19802998006343842, "learning_rate": 3.5980000000000004e-05, "loss": 0.0133, "step": 1800 }, { "epoch": 16.01769911504425, "grad_norm": 0.26125791668891907, "learning_rate": 3.618e-05, "loss": 0.0126, "step": 1810 }, { "epoch": 16.106194690265486, "grad_norm": 0.2111397534608841, "learning_rate": 3.638e-05, "loss": 0.0126, "step": 1820 }, { "epoch": 16.194690265486727, "grad_norm": 0.18799275159835815, "learning_rate": 3.6580000000000006e-05, "loss": 0.0126, "step": 1830 }, { "epoch": 16.283185840707965, "grad_norm": 0.19861453771591187, "learning_rate": 3.6780000000000004e-05, "loss": 0.013, "step": 1840 }, { "epoch": 16.371681415929203, "grad_norm": 0.158119797706604, "learning_rate": 3.698e-05, "loss": 0.0124, "step": 1850 }, { "epoch": 16.460176991150444, "grad_norm": 0.22938460111618042, "learning_rate": 3.7180000000000007e-05, "loss": 0.0127, "step": 1860 }, { "epoch": 16.548672566371682, "grad_norm": 0.21520522236824036, "learning_rate": 3.7380000000000005e-05, "loss": 0.0134, "step": 1870 }, { "epoch": 16.63716814159292, "grad_norm": 0.1865280568599701, "learning_rate": 3.758e-05, "loss": 0.0128, "step": 1880 }, { "epoch": 16.72566371681416, "grad_norm": 0.1627875715494156, "learning_rate": 3.778000000000001e-05, "loss": 0.0136, "step": 1890 }, { "epoch": 16.8141592920354, "grad_norm": 0.30490878224372864, "learning_rate": 3.7980000000000006e-05, "loss": 0.0127, "step": 1900 }, { "epoch": 16.902654867256636, "grad_norm": 0.21755605936050415, "learning_rate": 3.818e-05, "loss": 0.0134, "step": 1910 }, { "epoch": 16.991150442477878, "grad_norm": 0.18417157232761383, "learning_rate": 3.838e-05, "loss": 0.0125, "step": 1920 }, { "epoch": 17.079646017699115, "grad_norm": 0.1478920876979828, "learning_rate": 3.858e-05, "loss": 0.0133, "step": 1930 }, { "epoch": 17.168141592920353, "grad_norm": 0.16803410649299622, "learning_rate": 3.878e-05, "loss": 0.0125, "step": 1940 }, { "epoch": 17.256637168141594, "grad_norm": 0.17570087313652039, "learning_rate": 3.898e-05, "loss": 0.0131, "step": 1950 }, { "epoch": 17.345132743362832, "grad_norm": 0.24819160997867584, "learning_rate": 3.918e-05, "loss": 0.0119, "step": 1960 }, { "epoch": 17.43362831858407, "grad_norm": 0.20074355602264404, "learning_rate": 3.938e-05, "loss": 0.012, "step": 1970 }, { "epoch": 17.52212389380531, "grad_norm": 0.19154652953147888, "learning_rate": 3.958e-05, "loss": 0.013, "step": 1980 }, { "epoch": 17.61061946902655, "grad_norm": 0.23709169030189514, "learning_rate": 3.978e-05, "loss": 0.0134, "step": 1990 }, { "epoch": 17.699115044247787, "grad_norm": 0.18035951256752014, "learning_rate": 3.998e-05, "loss": 0.0126, "step": 2000 }, { "epoch": 17.787610619469028, "grad_norm": 0.21263554692268372, "learning_rate": 4.018e-05, "loss": 0.0123, "step": 2010 }, { "epoch": 17.876106194690266, "grad_norm": 0.18528272211551666, "learning_rate": 4.038e-05, "loss": 0.0122, "step": 2020 }, { "epoch": 17.964601769911503, "grad_norm": 0.1647820621728897, "learning_rate": 4.058e-05, "loss": 0.0129, "step": 2030 }, { "epoch": 18.053097345132745, "grad_norm": 0.22848553955554962, "learning_rate": 4.078e-05, "loss": 0.0131, "step": 2040 }, { "epoch": 18.141592920353983, "grad_norm": 0.24731457233428955, "learning_rate": 4.0980000000000004e-05, "loss": 0.0118, "step": 2050 }, { "epoch": 18.23008849557522, "grad_norm": 0.14303353428840637, "learning_rate": 4.118e-05, "loss": 0.0119, "step": 2060 }, { "epoch": 18.31858407079646, "grad_norm": 0.20178769528865814, "learning_rate": 4.138e-05, "loss": 0.0119, "step": 2070 }, { "epoch": 18.4070796460177, "grad_norm": 0.2670722007751465, "learning_rate": 4.1580000000000005e-05, "loss": 0.0126, "step": 2080 }, { "epoch": 18.495575221238937, "grad_norm": 0.18628989160060883, "learning_rate": 4.178e-05, "loss": 0.0133, "step": 2090 }, { "epoch": 18.58407079646018, "grad_norm": 0.23191425204277039, "learning_rate": 4.198e-05, "loss": 0.0117, "step": 2100 }, { "epoch": 18.672566371681416, "grad_norm": 0.20478810369968414, "learning_rate": 4.2180000000000006e-05, "loss": 0.0126, "step": 2110 }, { "epoch": 18.761061946902654, "grad_norm": 0.21395546197891235, "learning_rate": 4.2380000000000004e-05, "loss": 0.0118, "step": 2120 }, { "epoch": 18.849557522123895, "grad_norm": 0.21282713115215302, "learning_rate": 4.258e-05, "loss": 0.0125, "step": 2130 }, { "epoch": 18.938053097345133, "grad_norm": 0.23378048837184906, "learning_rate": 4.278e-05, "loss": 0.0121, "step": 2140 }, { "epoch": 19.02654867256637, "grad_norm": 0.186104416847229, "learning_rate": 4.2980000000000005e-05, "loss": 0.0121, "step": 2150 }, { "epoch": 19.115044247787612, "grad_norm": 0.23305700719356537, "learning_rate": 4.318e-05, "loss": 0.0119, "step": 2160 }, { "epoch": 19.20353982300885, "grad_norm": 0.2158450484275818, "learning_rate": 4.338e-05, "loss": 0.0118, "step": 2170 }, { "epoch": 19.292035398230087, "grad_norm": 0.1941143125295639, "learning_rate": 4.3580000000000006e-05, "loss": 0.0118, "step": 2180 }, { "epoch": 19.38053097345133, "grad_norm": 0.1980721652507782, "learning_rate": 4.3780000000000004e-05, "loss": 0.0116, "step": 2190 }, { "epoch": 19.469026548672566, "grad_norm": 0.22639663517475128, "learning_rate": 4.398e-05, "loss": 0.0127, "step": 2200 }, { "epoch": 19.557522123893804, "grad_norm": 0.1998574286699295, "learning_rate": 4.418000000000001e-05, "loss": 0.0124, "step": 2210 }, { "epoch": 19.646017699115045, "grad_norm": 0.2328738421201706, "learning_rate": 4.438e-05, "loss": 0.0112, "step": 2220 }, { "epoch": 19.734513274336283, "grad_norm": 0.15380622446537018, "learning_rate": 4.458e-05, "loss": 0.0116, "step": 2230 }, { "epoch": 19.82300884955752, "grad_norm": 0.187288299202919, "learning_rate": 4.478e-05, "loss": 0.0114, "step": 2240 }, { "epoch": 19.911504424778762, "grad_norm": 0.22886765003204346, "learning_rate": 4.498e-05, "loss": 0.0114, "step": 2250 }, { "epoch": 20.0, "grad_norm": 0.2300090491771698, "learning_rate": 4.518e-05, "loss": 0.0113, "step": 2260 }, { "epoch": 20.088495575221238, "grad_norm": 0.2091229110956192, "learning_rate": 4.538e-05, "loss": 0.0116, "step": 2270 }, { "epoch": 20.17699115044248, "grad_norm": 0.21108782291412354, "learning_rate": 4.558e-05, "loss": 0.0116, "step": 2280 }, { "epoch": 20.265486725663717, "grad_norm": 0.19910787045955658, "learning_rate": 4.578e-05, "loss": 0.0112, "step": 2290 }, { "epoch": 20.353982300884955, "grad_norm": 0.22218987345695496, "learning_rate": 4.5980000000000004e-05, "loss": 0.0123, "step": 2300 }, { "epoch": 20.442477876106196, "grad_norm": 0.18388235569000244, "learning_rate": 4.618e-05, "loss": 0.0121, "step": 2310 }, { "epoch": 20.530973451327434, "grad_norm": 0.2284725159406662, "learning_rate": 4.638e-05, "loss": 0.0116, "step": 2320 }, { "epoch": 20.61946902654867, "grad_norm": 0.18700768053531647, "learning_rate": 4.6580000000000005e-05, "loss": 0.0114, "step": 2330 }, { "epoch": 20.707964601769913, "grad_norm": 0.1835946887731552, "learning_rate": 4.678e-05, "loss": 0.0108, "step": 2340 }, { "epoch": 20.79646017699115, "grad_norm": 0.16614285111427307, "learning_rate": 4.698e-05, "loss": 0.0113, "step": 2350 }, { "epoch": 20.884955752212388, "grad_norm": 0.2018408626317978, "learning_rate": 4.718e-05, "loss": 0.0117, "step": 2360 }, { "epoch": 20.97345132743363, "grad_norm": 0.15835976600646973, "learning_rate": 4.7380000000000004e-05, "loss": 0.0109, "step": 2370 }, { "epoch": 21.061946902654867, "grad_norm": 0.22443874180316925, "learning_rate": 4.758e-05, "loss": 0.0113, "step": 2380 }, { "epoch": 21.150442477876105, "grad_norm": 0.17791017889976501, "learning_rate": 4.778e-05, "loss": 0.0125, "step": 2390 }, { "epoch": 21.238938053097346, "grad_norm": 0.16210758686065674, "learning_rate": 4.7980000000000005e-05, "loss": 0.0119, "step": 2400 }, { "epoch": 21.327433628318584, "grad_norm": 0.24134404957294464, "learning_rate": 4.818e-05, "loss": 0.0121, "step": 2410 }, { "epoch": 21.41592920353982, "grad_norm": 0.16999419033527374, "learning_rate": 4.838e-05, "loss": 0.0114, "step": 2420 }, { "epoch": 21.504424778761063, "grad_norm": 0.16048188507556915, "learning_rate": 4.8580000000000006e-05, "loss": 0.0119, "step": 2430 }, { "epoch": 21.5929203539823, "grad_norm": 0.19450658559799194, "learning_rate": 4.8780000000000004e-05, "loss": 0.0106, "step": 2440 }, { "epoch": 21.68141592920354, "grad_norm": 0.16528622806072235, "learning_rate": 4.898e-05, "loss": 0.0113, "step": 2450 }, { "epoch": 21.76991150442478, "grad_norm": 0.20057730376720428, "learning_rate": 4.918000000000001e-05, "loss": 0.0111, "step": 2460 }, { "epoch": 21.858407079646017, "grad_norm": 0.2268998920917511, "learning_rate": 4.9380000000000005e-05, "loss": 0.0105, "step": 2470 }, { "epoch": 21.946902654867255, "grad_norm": 0.21985943615436554, "learning_rate": 4.958e-05, "loss": 0.0116, "step": 2480 }, { "epoch": 22.035398230088497, "grad_norm": 0.18145669996738434, "learning_rate": 4.978e-05, "loss": 0.011, "step": 2490 }, { "epoch": 22.123893805309734, "grad_norm": 0.21360571682453156, "learning_rate": 4.9980000000000006e-05, "loss": 0.0115, "step": 2500 }, { "epoch": 22.212389380530972, "grad_norm": 0.16881686449050903, "learning_rate": 5.0180000000000004e-05, "loss": 0.0105, "step": 2510 }, { "epoch": 22.300884955752213, "grad_norm": 0.18654261529445648, "learning_rate": 5.038e-05, "loss": 0.0106, "step": 2520 }, { "epoch": 22.38938053097345, "grad_norm": 0.1934080868959427, "learning_rate": 5.058000000000001e-05, "loss": 0.0114, "step": 2530 }, { "epoch": 22.47787610619469, "grad_norm": 0.2071368396282196, "learning_rate": 5.0780000000000005e-05, "loss": 0.0112, "step": 2540 }, { "epoch": 22.56637168141593, "grad_norm": 0.19993095099925995, "learning_rate": 5.098e-05, "loss": 0.011, "step": 2550 }, { "epoch": 22.654867256637168, "grad_norm": 0.16395840048789978, "learning_rate": 5.118000000000001e-05, "loss": 0.0104, "step": 2560 }, { "epoch": 22.743362831858406, "grad_norm": 0.18830649554729462, "learning_rate": 5.1380000000000006e-05, "loss": 0.0104, "step": 2570 }, { "epoch": 22.831858407079647, "grad_norm": 0.18220649659633636, "learning_rate": 5.1580000000000004e-05, "loss": 0.0104, "step": 2580 }, { "epoch": 22.920353982300885, "grad_norm": 0.20872625708580017, "learning_rate": 5.178000000000001e-05, "loss": 0.0106, "step": 2590 }, { "epoch": 23.008849557522122, "grad_norm": 0.1854247897863388, "learning_rate": 5.198000000000001e-05, "loss": 0.0106, "step": 2600 }, { "epoch": 23.097345132743364, "grad_norm": 0.1710224449634552, "learning_rate": 5.2180000000000005e-05, "loss": 0.0102, "step": 2610 }, { "epoch": 23.1858407079646, "grad_norm": 0.18648681044578552, "learning_rate": 5.238000000000001e-05, "loss": 0.0108, "step": 2620 }, { "epoch": 23.27433628318584, "grad_norm": 0.18982253968715668, "learning_rate": 5.258000000000001e-05, "loss": 0.0106, "step": 2630 }, { "epoch": 23.36283185840708, "grad_norm": 0.2088329941034317, "learning_rate": 5.2780000000000006e-05, "loss": 0.0102, "step": 2640 }, { "epoch": 23.451327433628318, "grad_norm": 0.2174072116613388, "learning_rate": 5.2980000000000004e-05, "loss": 0.0105, "step": 2650 }, { "epoch": 23.539823008849556, "grad_norm": 0.21603617072105408, "learning_rate": 5.318000000000001e-05, "loss": 0.0104, "step": 2660 }, { "epoch": 23.628318584070797, "grad_norm": 0.21778415143489838, "learning_rate": 5.338000000000001e-05, "loss": 0.0104, "step": 2670 }, { "epoch": 23.716814159292035, "grad_norm": 0.19035662710666656, "learning_rate": 5.3580000000000005e-05, "loss": 0.0107, "step": 2680 }, { "epoch": 23.805309734513273, "grad_norm": 0.20847085118293762, "learning_rate": 5.378e-05, "loss": 0.0105, "step": 2690 }, { "epoch": 23.893805309734514, "grad_norm": 0.21918515861034393, "learning_rate": 5.3979999999999995e-05, "loss": 0.0106, "step": 2700 }, { "epoch": 23.98230088495575, "grad_norm": 0.18985942006111145, "learning_rate": 5.418e-05, "loss": 0.01, "step": 2710 }, { "epoch": 24.07079646017699, "grad_norm": 0.1921069622039795, "learning_rate": 5.438e-05, "loss": 0.0102, "step": 2720 }, { "epoch": 24.15929203539823, "grad_norm": 0.18723104894161224, "learning_rate": 5.4579999999999996e-05, "loss": 0.0102, "step": 2730 }, { "epoch": 24.24778761061947, "grad_norm": 0.1974979043006897, "learning_rate": 5.478e-05, "loss": 0.0102, "step": 2740 }, { "epoch": 24.336283185840706, "grad_norm": 0.17769354581832886, "learning_rate": 5.498e-05, "loss": 0.0103, "step": 2750 }, { "epoch": 24.424778761061948, "grad_norm": 0.17133192718029022, "learning_rate": 5.518e-05, "loss": 0.0097, "step": 2760 }, { "epoch": 24.513274336283185, "grad_norm": 0.12975075840950012, "learning_rate": 5.538e-05, "loss": 0.0101, "step": 2770 }, { "epoch": 24.601769911504427, "grad_norm": 0.1950993686914444, "learning_rate": 5.558e-05, "loss": 0.0106, "step": 2780 }, { "epoch": 24.690265486725664, "grad_norm": 0.17881418764591217, "learning_rate": 5.578e-05, "loss": 0.0097, "step": 2790 }, { "epoch": 24.778761061946902, "grad_norm": 0.15642932057380676, "learning_rate": 5.5979999999999996e-05, "loss": 0.01, "step": 2800 }, { "epoch": 24.86725663716814, "grad_norm": 0.15516893565654755, "learning_rate": 5.618e-05, "loss": 0.0112, "step": 2810 }, { "epoch": 24.95575221238938, "grad_norm": 0.17176637053489685, "learning_rate": 5.638e-05, "loss": 0.0098, "step": 2820 }, { "epoch": 25.04424778761062, "grad_norm": 0.1794217824935913, "learning_rate": 5.658e-05, "loss": 0.0104, "step": 2830 }, { "epoch": 25.13274336283186, "grad_norm": 0.1998187154531479, "learning_rate": 5.678e-05, "loss": 0.0108, "step": 2840 }, { "epoch": 25.221238938053098, "grad_norm": 0.17469826340675354, "learning_rate": 5.698e-05, "loss": 0.0105, "step": 2850 }, { "epoch": 25.309734513274336, "grad_norm": 0.22570648789405823, "learning_rate": 5.718e-05, "loss": 0.0095, "step": 2860 }, { "epoch": 25.398230088495577, "grad_norm": 0.1735209822654724, "learning_rate": 5.738e-05, "loss": 0.011, "step": 2870 }, { "epoch": 25.486725663716815, "grad_norm": 0.2552860379219055, "learning_rate": 5.758e-05, "loss": 0.0098, "step": 2880 }, { "epoch": 25.575221238938052, "grad_norm": 0.1956593245267868, "learning_rate": 5.778e-05, "loss": 0.0102, "step": 2890 }, { "epoch": 25.663716814159294, "grad_norm": 0.17074844241142273, "learning_rate": 5.7980000000000004e-05, "loss": 0.0107, "step": 2900 }, { "epoch": 25.75221238938053, "grad_norm": 0.20826052129268646, "learning_rate": 5.818e-05, "loss": 0.0095, "step": 2910 }, { "epoch": 25.84070796460177, "grad_norm": 0.20487037301063538, "learning_rate": 5.838e-05, "loss": 0.0101, "step": 2920 }, { "epoch": 25.92920353982301, "grad_norm": 0.17378631234169006, "learning_rate": 5.858e-05, "loss": 0.0099, "step": 2930 }, { "epoch": 26.01769911504425, "grad_norm": 0.2046569287776947, "learning_rate": 5.878e-05, "loss": 0.0103, "step": 2940 }, { "epoch": 26.106194690265486, "grad_norm": 0.17090077698230743, "learning_rate": 5.898e-05, "loss": 0.0107, "step": 2950 }, { "epoch": 26.194690265486727, "grad_norm": 0.20197150111198425, "learning_rate": 5.918e-05, "loss": 0.0099, "step": 2960 }, { "epoch": 26.283185840707965, "grad_norm": 0.190754234790802, "learning_rate": 5.9380000000000004e-05, "loss": 0.01, "step": 2970 }, { "epoch": 26.371681415929203, "grad_norm": 0.19185955822467804, "learning_rate": 5.958e-05, "loss": 0.0106, "step": 2980 }, { "epoch": 26.460176991150444, "grad_norm": 0.2416859269142151, "learning_rate": 5.978e-05, "loss": 0.0102, "step": 2990 }, { "epoch": 26.548672566371682, "grad_norm": 0.2122340202331543, "learning_rate": 5.9980000000000005e-05, "loss": 0.0101, "step": 3000 }, { "epoch": 26.63716814159292, "grad_norm": 0.18418727815151215, "learning_rate": 6.018e-05, "loss": 0.0095, "step": 3010 }, { "epoch": 26.72566371681416, "grad_norm": 0.15898945927619934, "learning_rate": 6.038e-05, "loss": 0.0099, "step": 3020 }, { "epoch": 26.8141592920354, "grad_norm": 0.21353374421596527, "learning_rate": 6.0580000000000006e-05, "loss": 0.0099, "step": 3030 }, { "epoch": 26.902654867256636, "grad_norm": 0.25403711199760437, "learning_rate": 6.0780000000000004e-05, "loss": 0.0098, "step": 3040 }, { "epoch": 26.991150442477878, "grad_norm": 0.2205997109413147, "learning_rate": 6.098e-05, "loss": 0.0103, "step": 3050 }, { "epoch": 27.079646017699115, "grad_norm": 0.24625472724437714, "learning_rate": 6.118000000000001e-05, "loss": 0.0108, "step": 3060 }, { "epoch": 27.168141592920353, "grad_norm": 0.22611097991466522, "learning_rate": 6.138e-05, "loss": 0.0096, "step": 3070 }, { "epoch": 27.256637168141594, "grad_norm": 0.20855627954006195, "learning_rate": 6.158e-05, "loss": 0.0096, "step": 3080 }, { "epoch": 27.345132743362832, "grad_norm": 0.18382352590560913, "learning_rate": 6.178000000000001e-05, "loss": 0.0094, "step": 3090 }, { "epoch": 27.43362831858407, "grad_norm": 0.21632954478263855, "learning_rate": 6.198e-05, "loss": 0.0097, "step": 3100 }, { "epoch": 27.52212389380531, "grad_norm": 0.1993696689605713, "learning_rate": 6.218e-05, "loss": 0.0102, "step": 3110 }, { "epoch": 27.61061946902655, "grad_norm": 0.15271496772766113, "learning_rate": 6.238000000000001e-05, "loss": 0.01, "step": 3120 }, { "epoch": 27.699115044247787, "grad_norm": 0.1937088966369629, "learning_rate": 6.258e-05, "loss": 0.0102, "step": 3130 }, { "epoch": 27.787610619469028, "grad_norm": 0.1129690408706665, "learning_rate": 6.278e-05, "loss": 0.01, "step": 3140 }, { "epoch": 27.876106194690266, "grad_norm": 0.16758713126182556, "learning_rate": 6.298000000000001e-05, "loss": 0.0098, "step": 3150 }, { "epoch": 27.964601769911503, "grad_norm": 0.1712367832660675, "learning_rate": 6.318e-05, "loss": 0.0099, "step": 3160 }, { "epoch": 28.053097345132745, "grad_norm": 0.18849895894527435, "learning_rate": 6.338e-05, "loss": 0.0091, "step": 3170 }, { "epoch": 28.141592920353983, "grad_norm": 0.1575203835964203, "learning_rate": 6.358000000000001e-05, "loss": 0.009, "step": 3180 }, { "epoch": 28.23008849557522, "grad_norm": 0.16038940846920013, "learning_rate": 6.378e-05, "loss": 0.0096, "step": 3190 }, { "epoch": 28.31858407079646, "grad_norm": 0.15592043101787567, "learning_rate": 6.398000000000001e-05, "loss": 0.0098, "step": 3200 }, { "epoch": 28.4070796460177, "grad_norm": 0.15957768261432648, "learning_rate": 6.418000000000001e-05, "loss": 0.0094, "step": 3210 }, { "epoch": 28.495575221238937, "grad_norm": 0.23940764367580414, "learning_rate": 6.438e-05, "loss": 0.0095, "step": 3220 }, { "epoch": 28.58407079646018, "grad_norm": 0.22940567135810852, "learning_rate": 6.458000000000001e-05, "loss": 0.0099, "step": 3230 }, { "epoch": 28.672566371681416, "grad_norm": 0.22047995030879974, "learning_rate": 6.478000000000001e-05, "loss": 0.0096, "step": 3240 }, { "epoch": 28.761061946902654, "grad_norm": 0.1963990032672882, "learning_rate": 6.498e-05, "loss": 0.0094, "step": 3250 }, { "epoch": 28.849557522123895, "grad_norm": 0.22907644510269165, "learning_rate": 6.518000000000001e-05, "loss": 0.0098, "step": 3260 }, { "epoch": 28.938053097345133, "grad_norm": 0.17745962738990784, "learning_rate": 6.538000000000001e-05, "loss": 0.0093, "step": 3270 }, { "epoch": 29.02654867256637, "grad_norm": 0.15264318883419037, "learning_rate": 6.558e-05, "loss": 0.0088, "step": 3280 }, { "epoch": 29.115044247787612, "grad_norm": 0.17992503941059113, "learning_rate": 6.578000000000001e-05, "loss": 0.0095, "step": 3290 }, { "epoch": 29.20353982300885, "grad_norm": 0.20190118253231049, "learning_rate": 6.598e-05, "loss": 0.0098, "step": 3300 }, { "epoch": 29.292035398230087, "grad_norm": 0.1557914912700653, "learning_rate": 6.618e-05, "loss": 0.0098, "step": 3310 }, { "epoch": 29.38053097345133, "grad_norm": 0.14843136072158813, "learning_rate": 6.638e-05, "loss": 0.01, "step": 3320 }, { "epoch": 29.469026548672566, "grad_norm": 0.20950397849082947, "learning_rate": 6.658e-05, "loss": 0.0102, "step": 3330 }, { "epoch": 29.557522123893804, "grad_norm": 0.16429200768470764, "learning_rate": 6.678e-05, "loss": 0.0095, "step": 3340 }, { "epoch": 29.646017699115045, "grad_norm": 0.1504441201686859, "learning_rate": 6.698e-05, "loss": 0.0098, "step": 3350 }, { "epoch": 29.734513274336283, "grad_norm": 0.19190770387649536, "learning_rate": 6.718e-05, "loss": 0.0096, "step": 3360 }, { "epoch": 29.82300884955752, "grad_norm": 0.16728892922401428, "learning_rate": 6.738e-05, "loss": 0.0089, "step": 3370 }, { "epoch": 29.911504424778762, "grad_norm": 0.181132972240448, "learning_rate": 6.758e-05, "loss": 0.0088, "step": 3380 }, { "epoch": 30.0, "grad_norm": 0.17006106674671173, "learning_rate": 6.778e-05, "loss": 0.0098, "step": 3390 }, { "epoch": 30.088495575221238, "grad_norm": 0.21980248391628265, "learning_rate": 6.798e-05, "loss": 0.0083, "step": 3400 }, { "epoch": 30.17699115044248, "grad_norm": 0.14729014039039612, "learning_rate": 6.818e-05, "loss": 0.0091, "step": 3410 }, { "epoch": 30.265486725663717, "grad_norm": 0.20811328291893005, "learning_rate": 6.838e-05, "loss": 0.009, "step": 3420 }, { "epoch": 30.353982300884955, "grad_norm": 0.17850981652736664, "learning_rate": 6.858e-05, "loss": 0.0085, "step": 3430 }, { "epoch": 30.442477876106196, "grad_norm": 0.14746429026126862, "learning_rate": 6.878e-05, "loss": 0.0097, "step": 3440 }, { "epoch": 30.530973451327434, "grad_norm": 0.15437854826450348, "learning_rate": 6.898e-05, "loss": 0.009, "step": 3450 }, { "epoch": 30.61946902654867, "grad_norm": 0.14086470007896423, "learning_rate": 6.918e-05, "loss": 0.0091, "step": 3460 }, { "epoch": 30.707964601769913, "grad_norm": 0.12399916350841522, "learning_rate": 6.938e-05, "loss": 0.0098, "step": 3470 }, { "epoch": 30.79646017699115, "grad_norm": 0.19156357645988464, "learning_rate": 6.958e-05, "loss": 0.0092, "step": 3480 }, { "epoch": 30.884955752212388, "grad_norm": 0.1824861615896225, "learning_rate": 6.978e-05, "loss": 0.0092, "step": 3490 }, { "epoch": 30.97345132743363, "grad_norm": 0.17979440093040466, "learning_rate": 6.998e-05, "loss": 0.0096, "step": 3500 }, { "epoch": 31.061946902654867, "grad_norm": 0.1666015386581421, "learning_rate": 7.018e-05, "loss": 0.0097, "step": 3510 }, { "epoch": 31.150442477876105, "grad_norm": 0.20595945417881012, "learning_rate": 7.038e-05, "loss": 0.0098, "step": 3520 }, { "epoch": 31.238938053097346, "grad_norm": 0.22851444780826569, "learning_rate": 7.058e-05, "loss": 0.0094, "step": 3530 }, { "epoch": 31.327433628318584, "grad_norm": 0.17134319245815277, "learning_rate": 7.078e-05, "loss": 0.0095, "step": 3540 }, { "epoch": 31.41592920353982, "grad_norm": 0.16253021359443665, "learning_rate": 7.098e-05, "loss": 0.0088, "step": 3550 }, { "epoch": 31.504424778761063, "grad_norm": 0.19967076182365417, "learning_rate": 7.118e-05, "loss": 0.0089, "step": 3560 }, { "epoch": 31.5929203539823, "grad_norm": 0.14656779170036316, "learning_rate": 7.138e-05, "loss": 0.009, "step": 3570 }, { "epoch": 31.68141592920354, "grad_norm": 0.2233688235282898, "learning_rate": 7.158e-05, "loss": 0.0094, "step": 3580 }, { "epoch": 31.76991150442478, "grad_norm": 0.1803266853094101, "learning_rate": 7.178000000000001e-05, "loss": 0.0093, "step": 3590 }, { "epoch": 31.858407079646017, "grad_norm": 0.19906945526599884, "learning_rate": 7.198e-05, "loss": 0.0094, "step": 3600 }, { "epoch": 31.946902654867255, "grad_norm": 0.1841084361076355, "learning_rate": 7.218e-05, "loss": 0.0096, "step": 3610 }, { "epoch": 32.0353982300885, "grad_norm": 0.14032480120658875, "learning_rate": 7.238000000000001e-05, "loss": 0.0096, "step": 3620 }, { "epoch": 32.123893805309734, "grad_norm": 0.16564306616783142, "learning_rate": 7.258e-05, "loss": 0.0093, "step": 3630 }, { "epoch": 32.21238938053097, "grad_norm": 0.13823221623897552, "learning_rate": 7.278e-05, "loss": 0.0094, "step": 3640 }, { "epoch": 32.30088495575221, "grad_norm": 0.14694763720035553, "learning_rate": 7.298000000000001e-05, "loss": 0.0097, "step": 3650 }, { "epoch": 32.389380530973455, "grad_norm": 0.18370236456394196, "learning_rate": 7.318e-05, "loss": 0.0091, "step": 3660 }, { "epoch": 32.47787610619469, "grad_norm": 0.15265528857707977, "learning_rate": 7.338e-05, "loss": 0.0094, "step": 3670 }, { "epoch": 32.56637168141593, "grad_norm": 0.15920133888721466, "learning_rate": 7.358000000000001e-05, "loss": 0.0088, "step": 3680 }, { "epoch": 32.65486725663717, "grad_norm": 0.17650489509105682, "learning_rate": 7.378e-05, "loss": 0.0093, "step": 3690 }, { "epoch": 32.743362831858406, "grad_norm": 0.145453080534935, "learning_rate": 7.398e-05, "loss": 0.0089, "step": 3700 }, { "epoch": 32.83185840707964, "grad_norm": 0.12905701994895935, "learning_rate": 7.418000000000001e-05, "loss": 0.0092, "step": 3710 }, { "epoch": 32.92035398230089, "grad_norm": 0.13699135184288025, "learning_rate": 7.438e-05, "loss": 0.0085, "step": 3720 }, { "epoch": 33.008849557522126, "grad_norm": 0.18020181357860565, "learning_rate": 7.458000000000001e-05, "loss": 0.0093, "step": 3730 }, { "epoch": 33.097345132743364, "grad_norm": 0.14986072480678558, "learning_rate": 7.478e-05, "loss": 0.009, "step": 3740 }, { "epoch": 33.1858407079646, "grad_norm": 0.14074194431304932, "learning_rate": 7.498e-05, "loss": 0.0093, "step": 3750 }, { "epoch": 33.27433628318584, "grad_norm": 0.1257593184709549, "learning_rate": 7.518000000000001e-05, "loss": 0.0082, "step": 3760 }, { "epoch": 33.36283185840708, "grad_norm": 0.21100081503391266, "learning_rate": 7.538e-05, "loss": 0.0094, "step": 3770 }, { "epoch": 33.45132743362832, "grad_norm": 0.14728806912899017, "learning_rate": 7.558e-05, "loss": 0.0086, "step": 3780 }, { "epoch": 33.53982300884956, "grad_norm": 0.15006926655769348, "learning_rate": 7.578000000000001e-05, "loss": 0.009, "step": 3790 }, { "epoch": 33.6283185840708, "grad_norm": 0.15550263226032257, "learning_rate": 7.598e-05, "loss": 0.0091, "step": 3800 }, { "epoch": 33.716814159292035, "grad_norm": 0.1371162086725235, "learning_rate": 7.618e-05, "loss": 0.0091, "step": 3810 }, { "epoch": 33.80530973451327, "grad_norm": 0.15556631982326508, "learning_rate": 7.638000000000001e-05, "loss": 0.0078, "step": 3820 }, { "epoch": 33.89380530973451, "grad_norm": 0.139182910323143, "learning_rate": 7.658e-05, "loss": 0.009, "step": 3830 }, { "epoch": 33.982300884955755, "grad_norm": 0.12569105625152588, "learning_rate": 7.678000000000001e-05, "loss": 0.0083, "step": 3840 }, { "epoch": 34.07079646017699, "grad_norm": 0.13380029797554016, "learning_rate": 7.698000000000001e-05, "loss": 0.0091, "step": 3850 }, { "epoch": 34.15929203539823, "grad_norm": 0.17689161002635956, "learning_rate": 7.718e-05, "loss": 0.0081, "step": 3860 }, { "epoch": 34.24778761061947, "grad_norm": 0.13733205199241638, "learning_rate": 7.738000000000001e-05, "loss": 0.0086, "step": 3870 }, { "epoch": 34.336283185840706, "grad_norm": 0.16643129289150238, "learning_rate": 7.758000000000001e-05, "loss": 0.0087, "step": 3880 }, { "epoch": 34.424778761061944, "grad_norm": 0.11663112789392471, "learning_rate": 7.778e-05, "loss": 0.0087, "step": 3890 }, { "epoch": 34.51327433628319, "grad_norm": 0.15019580721855164, "learning_rate": 7.798000000000001e-05, "loss": 0.0088, "step": 3900 }, { "epoch": 34.60176991150443, "grad_norm": 0.1732584536075592, "learning_rate": 7.818000000000001e-05, "loss": 0.0088, "step": 3910 }, { "epoch": 34.690265486725664, "grad_norm": 0.14080511033535004, "learning_rate": 7.838e-05, "loss": 0.0085, "step": 3920 }, { "epoch": 34.7787610619469, "grad_norm": 0.182724729180336, "learning_rate": 7.858000000000001e-05, "loss": 0.0087, "step": 3930 }, { "epoch": 34.86725663716814, "grad_norm": 0.15996137261390686, "learning_rate": 7.878e-05, "loss": 0.0085, "step": 3940 }, { "epoch": 34.95575221238938, "grad_norm": 0.16426995396614075, "learning_rate": 7.897999999999999e-05, "loss": 0.0082, "step": 3950 }, { "epoch": 35.04424778761062, "grad_norm": 0.15797816216945648, "learning_rate": 7.918e-05, "loss": 0.0084, "step": 3960 }, { "epoch": 35.13274336283186, "grad_norm": 0.20105281472206116, "learning_rate": 7.938e-05, "loss": 0.0081, "step": 3970 }, { "epoch": 35.2212389380531, "grad_norm": 0.14041103422641754, "learning_rate": 7.958e-05, "loss": 0.0088, "step": 3980 }, { "epoch": 35.309734513274336, "grad_norm": 0.16384132206439972, "learning_rate": 7.978e-05, "loss": 0.0092, "step": 3990 }, { "epoch": 35.39823008849557, "grad_norm": 0.17888638377189636, "learning_rate": 7.998e-05, "loss": 0.0089, "step": 4000 }, { "epoch": 35.48672566371681, "grad_norm": 0.1731533706188202, "learning_rate": 8.018e-05, "loss": 0.0089, "step": 4010 }, { "epoch": 35.575221238938056, "grad_norm": 0.148861825466156, "learning_rate": 8.038e-05, "loss": 0.0086, "step": 4020 }, { "epoch": 35.663716814159294, "grad_norm": 0.17036963999271393, "learning_rate": 8.058e-05, "loss": 0.0086, "step": 4030 }, { "epoch": 35.75221238938053, "grad_norm": 0.1750449538230896, "learning_rate": 8.078e-05, "loss": 0.0086, "step": 4040 }, { "epoch": 35.84070796460177, "grad_norm": 0.16272523999214172, "learning_rate": 8.098e-05, "loss": 0.0091, "step": 4050 }, { "epoch": 35.92920353982301, "grad_norm": 0.14315882325172424, "learning_rate": 8.118e-05, "loss": 0.0095, "step": 4060 }, { "epoch": 36.017699115044245, "grad_norm": 0.1691032499074936, "learning_rate": 8.138e-05, "loss": 0.0092, "step": 4070 }, { "epoch": 36.10619469026549, "grad_norm": 0.15037715435028076, "learning_rate": 8.158e-05, "loss": 0.0091, "step": 4080 }, { "epoch": 36.19469026548673, "grad_norm": 0.1389438509941101, "learning_rate": 8.178e-05, "loss": 0.0083, "step": 4090 }, { "epoch": 36.283185840707965, "grad_norm": 0.16533799469470978, "learning_rate": 8.198e-05, "loss": 0.0087, "step": 4100 }, { "epoch": 36.3716814159292, "grad_norm": 0.13812726736068726, "learning_rate": 8.218e-05, "loss": 0.0082, "step": 4110 }, { "epoch": 36.46017699115044, "grad_norm": 0.15419334173202515, "learning_rate": 8.238000000000001e-05, "loss": 0.0083, "step": 4120 }, { "epoch": 36.54867256637168, "grad_norm": 0.15441758930683136, "learning_rate": 8.258e-05, "loss": 0.0083, "step": 4130 }, { "epoch": 36.63716814159292, "grad_norm": 0.1351751983165741, "learning_rate": 8.278e-05, "loss": 0.0081, "step": 4140 }, { "epoch": 36.72566371681416, "grad_norm": 0.15623344480991364, "learning_rate": 8.298000000000001e-05, "loss": 0.0078, "step": 4150 }, { "epoch": 36.8141592920354, "grad_norm": 0.15462660789489746, "learning_rate": 8.318e-05, "loss": 0.0092, "step": 4160 }, { "epoch": 36.902654867256636, "grad_norm": 0.1266651153564453, "learning_rate": 8.338e-05, "loss": 0.0086, "step": 4170 }, { "epoch": 36.991150442477874, "grad_norm": 0.14301811158657074, "learning_rate": 8.358e-05, "loss": 0.0083, "step": 4180 }, { "epoch": 37.07964601769911, "grad_norm": 0.167973130941391, "learning_rate": 8.378e-05, "loss": 0.0091, "step": 4190 }, { "epoch": 37.16814159292036, "grad_norm": 0.20928312838077545, "learning_rate": 8.398e-05, "loss": 0.009, "step": 4200 }, { "epoch": 37.256637168141594, "grad_norm": 0.17960430681705475, "learning_rate": 8.418e-05, "loss": 0.0087, "step": 4210 }, { "epoch": 37.34513274336283, "grad_norm": 0.12802328169345856, "learning_rate": 8.438e-05, "loss": 0.0086, "step": 4220 }, { "epoch": 37.43362831858407, "grad_norm": 0.17191757261753082, "learning_rate": 8.458e-05, "loss": 0.0089, "step": 4230 }, { "epoch": 37.52212389380531, "grad_norm": 0.16063416004180908, "learning_rate": 8.478e-05, "loss": 0.0087, "step": 4240 }, { "epoch": 37.610619469026545, "grad_norm": 0.15133914351463318, "learning_rate": 8.498e-05, "loss": 0.0088, "step": 4250 }, { "epoch": 37.69911504424779, "grad_norm": 0.13532212376594543, "learning_rate": 8.518000000000001e-05, "loss": 0.0086, "step": 4260 }, { "epoch": 37.78761061946903, "grad_norm": 0.15464913845062256, "learning_rate": 8.538e-05, "loss": 0.0086, "step": 4270 }, { "epoch": 37.876106194690266, "grad_norm": 0.16056020557880402, "learning_rate": 8.558e-05, "loss": 0.0092, "step": 4280 }, { "epoch": 37.9646017699115, "grad_norm": 0.1944892257452011, "learning_rate": 8.578000000000001e-05, "loss": 0.0081, "step": 4290 }, { "epoch": 38.05309734513274, "grad_norm": 0.18636475503444672, "learning_rate": 8.598e-05, "loss": 0.0098, "step": 4300 }, { "epoch": 38.14159292035398, "grad_norm": 0.1302233338356018, "learning_rate": 8.618e-05, "loss": 0.0087, "step": 4310 }, { "epoch": 38.230088495575224, "grad_norm": 0.15690068900585175, "learning_rate": 8.638000000000001e-05, "loss": 0.0083, "step": 4320 }, { "epoch": 38.31858407079646, "grad_norm": 0.17732751369476318, "learning_rate": 8.658e-05, "loss": 0.0098, "step": 4330 }, { "epoch": 38.4070796460177, "grad_norm": 0.16375420987606049, "learning_rate": 8.678e-05, "loss": 0.0078, "step": 4340 }, { "epoch": 38.49557522123894, "grad_norm": 0.15327563881874084, "learning_rate": 8.698000000000001e-05, "loss": 0.0083, "step": 4350 }, { "epoch": 38.584070796460175, "grad_norm": 0.17066507041454315, "learning_rate": 8.718e-05, "loss": 0.0085, "step": 4360 }, { "epoch": 38.67256637168141, "grad_norm": 0.13817591965198517, "learning_rate": 8.738000000000001e-05, "loss": 0.0083, "step": 4370 }, { "epoch": 38.76106194690266, "grad_norm": 0.13917043805122375, "learning_rate": 8.758000000000001e-05, "loss": 0.0082, "step": 4380 }, { "epoch": 38.849557522123895, "grad_norm": 0.1588297188282013, "learning_rate": 8.778e-05, "loss": 0.0085, "step": 4390 }, { "epoch": 38.93805309734513, "grad_norm": 0.1482231169939041, "learning_rate": 8.798000000000001e-05, "loss": 0.0077, "step": 4400 }, { "epoch": 39.02654867256637, "grad_norm": 0.19748488068580627, "learning_rate": 8.818000000000001e-05, "loss": 0.008, "step": 4410 }, { "epoch": 39.11504424778761, "grad_norm": 0.15280547738075256, "learning_rate": 8.838e-05, "loss": 0.008, "step": 4420 }, { "epoch": 39.203539823008846, "grad_norm": 0.1811227649450302, "learning_rate": 8.858000000000001e-05, "loss": 0.0088, "step": 4430 }, { "epoch": 39.29203539823009, "grad_norm": 0.21726331114768982, "learning_rate": 8.878000000000001e-05, "loss": 0.0086, "step": 4440 }, { "epoch": 39.38053097345133, "grad_norm": 0.2437189519405365, "learning_rate": 8.898e-05, "loss": 0.0086, "step": 4450 }, { "epoch": 39.469026548672566, "grad_norm": 0.14687137305736542, "learning_rate": 8.918000000000001e-05, "loss": 0.0082, "step": 4460 }, { "epoch": 39.557522123893804, "grad_norm": 0.1408999264240265, "learning_rate": 8.938e-05, "loss": 0.0083, "step": 4470 }, { "epoch": 39.64601769911504, "grad_norm": 0.12296683341264725, "learning_rate": 8.958e-05, "loss": 0.0076, "step": 4480 }, { "epoch": 39.73451327433628, "grad_norm": 0.12216699123382568, "learning_rate": 8.978000000000001e-05, "loss": 0.0078, "step": 4490 }, { "epoch": 39.823008849557525, "grad_norm": 0.12778812646865845, "learning_rate": 8.998e-05, "loss": 0.0083, "step": 4500 }, { "epoch": 39.91150442477876, "grad_norm": 0.13679400086402893, "learning_rate": 9.018000000000001e-05, "loss": 0.0084, "step": 4510 }, { "epoch": 40.0, "grad_norm": 0.16394869983196259, "learning_rate": 9.038000000000001e-05, "loss": 0.0081, "step": 4520 }, { "epoch": 40.08849557522124, "grad_norm": 0.16781149804592133, "learning_rate": 9.058e-05, "loss": 0.008, "step": 4530 }, { "epoch": 40.176991150442475, "grad_norm": 0.15661022067070007, "learning_rate": 9.078000000000001e-05, "loss": 0.0088, "step": 4540 }, { "epoch": 40.26548672566372, "grad_norm": 0.12360215187072754, "learning_rate": 9.098000000000001e-05, "loss": 0.0076, "step": 4550 }, { "epoch": 40.35398230088496, "grad_norm": 0.12684562802314758, "learning_rate": 9.118e-05, "loss": 0.009, "step": 4560 }, { "epoch": 40.442477876106196, "grad_norm": 0.10499118268489838, "learning_rate": 9.138e-05, "loss": 0.0082, "step": 4570 }, { "epoch": 40.530973451327434, "grad_norm": 0.12011920660734177, "learning_rate": 9.158e-05, "loss": 0.0081, "step": 4580 }, { "epoch": 40.61946902654867, "grad_norm": 0.22020132839679718, "learning_rate": 9.178e-05, "loss": 0.0086, "step": 4590 }, { "epoch": 40.70796460176991, "grad_norm": 0.14828668534755707, "learning_rate": 9.198e-05, "loss": 0.0079, "step": 4600 }, { "epoch": 40.796460176991154, "grad_norm": 0.17221546173095703, "learning_rate": 9.218e-05, "loss": 0.0083, "step": 4610 }, { "epoch": 40.88495575221239, "grad_norm": 0.15954197943210602, "learning_rate": 9.238e-05, "loss": 0.0085, "step": 4620 }, { "epoch": 40.97345132743363, "grad_norm": 0.15018020570278168, "learning_rate": 9.258e-05, "loss": 0.0081, "step": 4630 }, { "epoch": 41.06194690265487, "grad_norm": 0.13175387680530548, "learning_rate": 9.278e-05, "loss": 0.0089, "step": 4640 }, { "epoch": 41.150442477876105, "grad_norm": 0.15653151273727417, "learning_rate": 9.298e-05, "loss": 0.0079, "step": 4650 }, { "epoch": 41.23893805309734, "grad_norm": 0.17663529515266418, "learning_rate": 9.318e-05, "loss": 0.0085, "step": 4660 }, { "epoch": 41.32743362831859, "grad_norm": 0.18750809133052826, "learning_rate": 9.338e-05, "loss": 0.0088, "step": 4670 }, { "epoch": 41.415929203539825, "grad_norm": 0.1502334028482437, "learning_rate": 9.358e-05, "loss": 0.0079, "step": 4680 }, { "epoch": 41.50442477876106, "grad_norm": 0.14593996107578278, "learning_rate": 9.378e-05, "loss": 0.0082, "step": 4690 }, { "epoch": 41.5929203539823, "grad_norm": 0.1764802187681198, "learning_rate": 9.398e-05, "loss": 0.0093, "step": 4700 }, { "epoch": 41.68141592920354, "grad_norm": 0.18359750509262085, "learning_rate": 9.418e-05, "loss": 0.0087, "step": 4710 }, { "epoch": 41.769911504424776, "grad_norm": 0.12710478901863098, "learning_rate": 9.438e-05, "loss": 0.0079, "step": 4720 }, { "epoch": 41.85840707964602, "grad_norm": 0.12599807977676392, "learning_rate": 9.458e-05, "loss": 0.0085, "step": 4730 }, { "epoch": 41.94690265486726, "grad_norm": 0.13770687580108643, "learning_rate": 9.478e-05, "loss": 0.0086, "step": 4740 }, { "epoch": 42.0353982300885, "grad_norm": 0.11630912870168686, "learning_rate": 9.498e-05, "loss": 0.0084, "step": 4750 }, { "epoch": 42.123893805309734, "grad_norm": 0.13092055916786194, "learning_rate": 9.518000000000001e-05, "loss": 0.0076, "step": 4760 }, { "epoch": 42.21238938053097, "grad_norm": 0.17325466871261597, "learning_rate": 9.538e-05, "loss": 0.0083, "step": 4770 }, { "epoch": 42.30088495575221, "grad_norm": 0.15955308079719543, "learning_rate": 9.558e-05, "loss": 0.0086, "step": 4780 }, { "epoch": 42.389380530973455, "grad_norm": 0.12556494772434235, "learning_rate": 9.578000000000001e-05, "loss": 0.0076, "step": 4790 }, { "epoch": 42.47787610619469, "grad_norm": 0.12370289117097855, "learning_rate": 9.598e-05, "loss": 0.008, "step": 4800 }, { "epoch": 42.56637168141593, "grad_norm": 0.15087510645389557, "learning_rate": 9.618e-05, "loss": 0.0076, "step": 4810 }, { "epoch": 42.65486725663717, "grad_norm": 0.1375018209218979, "learning_rate": 9.638000000000001e-05, "loss": 0.0082, "step": 4820 }, { "epoch": 42.743362831858406, "grad_norm": 0.16676568984985352, "learning_rate": 9.658e-05, "loss": 0.0084, "step": 4830 }, { "epoch": 42.83185840707964, "grad_norm": 0.1747582107782364, "learning_rate": 9.678e-05, "loss": 0.0079, "step": 4840 }, { "epoch": 42.92035398230089, "grad_norm": 0.17830878496170044, "learning_rate": 9.698000000000001e-05, "loss": 0.0081, "step": 4850 }, { "epoch": 43.008849557522126, "grad_norm": 0.17418724298477173, "learning_rate": 9.718e-05, "loss": 0.0082, "step": 4860 }, { "epoch": 43.097345132743364, "grad_norm": 0.15233151614665985, "learning_rate": 9.738e-05, "loss": 0.0081, "step": 4870 }, { "epoch": 43.1858407079646, "grad_norm": 0.11910518258810043, "learning_rate": 9.758000000000001e-05, "loss": 0.0077, "step": 4880 }, { "epoch": 43.27433628318584, "grad_norm": 0.12808698415756226, "learning_rate": 9.778e-05, "loss": 0.0075, "step": 4890 }, { "epoch": 43.36283185840708, "grad_norm": 0.12413737922906876, "learning_rate": 9.798000000000001e-05, "loss": 0.0073, "step": 4900 }, { "epoch": 43.45132743362832, "grad_norm": 0.1637522429227829, "learning_rate": 9.818000000000001e-05, "loss": 0.0081, "step": 4910 }, { "epoch": 43.53982300884956, "grad_norm": 0.1581503301858902, "learning_rate": 9.838e-05, "loss": 0.0082, "step": 4920 }, { "epoch": 43.6283185840708, "grad_norm": 0.13720406591892242, "learning_rate": 9.858000000000001e-05, "loss": 0.0091, "step": 4930 }, { "epoch": 43.716814159292035, "grad_norm": 0.1338191032409668, "learning_rate": 9.878e-05, "loss": 0.008, "step": 4940 }, { "epoch": 43.80530973451327, "grad_norm": 0.14985616505146027, "learning_rate": 9.898e-05, "loss": 0.0078, "step": 4950 }, { "epoch": 43.89380530973451, "grad_norm": 0.12583546340465546, "learning_rate": 9.918000000000001e-05, "loss": 0.0084, "step": 4960 }, { "epoch": 43.982300884955755, "grad_norm": 0.12459380179643631, "learning_rate": 9.938e-05, "loss": 0.0083, "step": 4970 }, { "epoch": 44.07079646017699, "grad_norm": 0.17305994033813477, "learning_rate": 9.958e-05, "loss": 0.008, "step": 4980 }, { "epoch": 44.15929203539823, "grad_norm": 0.1261778473854065, "learning_rate": 9.978000000000001e-05, "loss": 0.0084, "step": 4990 }, { "epoch": 44.24778761061947, "grad_norm": 0.12196177244186401, "learning_rate": 9.998e-05, "loss": 0.0077, "step": 5000 }, { "epoch": 44.336283185840706, "grad_norm": 0.1706259697675705, "learning_rate": 9.999999778549045e-05, "loss": 0.0088, "step": 5010 }, { "epoch": 44.424778761061944, "grad_norm": 0.16474901139736176, "learning_rate": 9.999999013039593e-05, "loss": 0.0085, "step": 5020 }, { "epoch": 44.51327433628319, "grad_norm": 0.14632219076156616, "learning_rate": 9.999997700737766e-05, "loss": 0.0082, "step": 5030 }, { "epoch": 44.60176991150443, "grad_norm": 0.13608916103839874, "learning_rate": 9.999995841643709e-05, "loss": 0.0082, "step": 5040 }, { "epoch": 44.690265486725664, "grad_norm": 0.12573711574077606, "learning_rate": 9.999993435757623e-05, "loss": 0.0085, "step": 5050 }, { "epoch": 44.7787610619469, "grad_norm": 0.12018602341413498, "learning_rate": 9.999990483079773e-05, "loss": 0.0083, "step": 5060 }, { "epoch": 44.86725663716814, "grad_norm": 0.13197782635688782, "learning_rate": 9.999986983610481e-05, "loss": 0.0075, "step": 5070 }, { "epoch": 44.95575221238938, "grad_norm": 0.1478177160024643, "learning_rate": 9.99998293735013e-05, "loss": 0.0075, "step": 5080 }, { "epoch": 45.04424778761062, "grad_norm": 0.1418713927268982, "learning_rate": 9.999978344299161e-05, "loss": 0.0079, "step": 5090 }, { "epoch": 45.13274336283186, "grad_norm": 0.18489570915699005, "learning_rate": 9.99997320445808e-05, "loss": 0.0083, "step": 5100 }, { "epoch": 45.2212389380531, "grad_norm": 0.17373768985271454, "learning_rate": 9.999967517827444e-05, "loss": 0.008, "step": 5110 }, { "epoch": 45.309734513274336, "grad_norm": 0.12457728385925293, "learning_rate": 9.999961284407879e-05, "loss": 0.0079, "step": 5120 }, { "epoch": 45.39823008849557, "grad_norm": 0.16301515698432922, "learning_rate": 9.999954504200067e-05, "loss": 0.0085, "step": 5130 }, { "epoch": 45.48672566371681, "grad_norm": 0.15682262182235718, "learning_rate": 9.999947177204744e-05, "loss": 0.0079, "step": 5140 }, { "epoch": 45.575221238938056, "grad_norm": 0.08579237014055252, "learning_rate": 9.999939303422718e-05, "loss": 0.0078, "step": 5150 }, { "epoch": 45.663716814159294, "grad_norm": 0.1308777928352356, "learning_rate": 9.999930882854847e-05, "loss": 0.0077, "step": 5160 }, { "epoch": 45.75221238938053, "grad_norm": 0.1625610589981079, "learning_rate": 9.999921915502051e-05, "loss": 0.0078, "step": 5170 }, { "epoch": 45.84070796460177, "grad_norm": 0.16500815749168396, "learning_rate": 9.99991240136531e-05, "loss": 0.0084, "step": 5180 }, { "epoch": 45.92920353982301, "grad_norm": 0.1339222937822342, "learning_rate": 9.999902340445668e-05, "loss": 0.0078, "step": 5190 }, { "epoch": 46.017699115044245, "grad_norm": 0.136415034532547, "learning_rate": 9.999891732744224e-05, "loss": 0.0076, "step": 5200 }, { "epoch": 46.10619469026549, "grad_norm": 0.14090347290039062, "learning_rate": 9.999880578262135e-05, "loss": 0.0077, "step": 5210 }, { "epoch": 46.19469026548673, "grad_norm": 0.1553945541381836, "learning_rate": 9.999868877000624e-05, "loss": 0.0076, "step": 5220 }, { "epoch": 46.283185840707965, "grad_norm": 0.17248113453388214, "learning_rate": 9.99985662896097e-05, "loss": 0.0079, "step": 5230 }, { "epoch": 46.3716814159292, "grad_norm": 0.1252346783876419, "learning_rate": 9.999843834144513e-05, "loss": 0.0078, "step": 5240 }, { "epoch": 46.46017699115044, "grad_norm": 0.21699823439121246, "learning_rate": 9.99983049255265e-05, "loss": 0.0079, "step": 5250 }, { "epoch": 46.54867256637168, "grad_norm": 0.16120117902755737, "learning_rate": 9.999816604186843e-05, "loss": 0.0076, "step": 5260 }, { "epoch": 46.63716814159292, "grad_norm": 0.1501779556274414, "learning_rate": 9.999802169048609e-05, "loss": 0.0085, "step": 5270 }, { "epoch": 46.72566371681416, "grad_norm": 0.18957437574863434, "learning_rate": 9.999787187139527e-05, "loss": 0.0079, "step": 5280 }, { "epoch": 46.8141592920354, "grad_norm": 0.17378948628902435, "learning_rate": 9.999771658461234e-05, "loss": 0.0079, "step": 5290 }, { "epoch": 46.902654867256636, "grad_norm": 0.18934838473796844, "learning_rate": 9.999755583015431e-05, "loss": 0.008, "step": 5300 }, { "epoch": 46.991150442477874, "grad_norm": 0.16995805501937866, "learning_rate": 9.999738960803874e-05, "loss": 0.0087, "step": 5310 }, { "epoch": 47.07964601769911, "grad_norm": 0.16987305879592896, "learning_rate": 9.99972179182838e-05, "loss": 0.0079, "step": 5320 }, { "epoch": 47.16814159292036, "grad_norm": 0.15868712961673737, "learning_rate": 9.99970407609083e-05, "loss": 0.0082, "step": 5330 }, { "epoch": 47.256637168141594, "grad_norm": 0.11736416816711426, "learning_rate": 9.999685813593159e-05, "loss": 0.0079, "step": 5340 }, { "epoch": 47.34513274336283, "grad_norm": 0.1414930820465088, "learning_rate": 9.999667004337362e-05, "loss": 0.0075, "step": 5350 }, { "epoch": 47.43362831858407, "grad_norm": 0.1305140256881714, "learning_rate": 9.9996476483255e-05, "loss": 0.0071, "step": 5360 }, { "epoch": 47.52212389380531, "grad_norm": 0.11197655647993088, "learning_rate": 9.999627745559688e-05, "loss": 0.0078, "step": 5370 }, { "epoch": 47.610619469026545, "grad_norm": 0.1341860443353653, "learning_rate": 9.999607296042101e-05, "loss": 0.0073, "step": 5380 }, { "epoch": 47.69911504424779, "grad_norm": 0.13850046694278717, "learning_rate": 9.99958629977498e-05, "loss": 0.0081, "step": 5390 }, { "epoch": 47.78761061946903, "grad_norm": 0.1535767912864685, "learning_rate": 9.999564756760615e-05, "loss": 0.0078, "step": 5400 }, { "epoch": 47.876106194690266, "grad_norm": 0.14185595512390137, "learning_rate": 9.999542667001366e-05, "loss": 0.0085, "step": 5410 }, { "epoch": 47.9646017699115, "grad_norm": 0.13457666337490082, "learning_rate": 9.999520030499647e-05, "loss": 0.0079, "step": 5420 }, { "epoch": 48.05309734513274, "grad_norm": 0.14653809368610382, "learning_rate": 9.999496847257936e-05, "loss": 0.0078, "step": 5430 }, { "epoch": 48.14159292035398, "grad_norm": 0.13704976439476013, "learning_rate": 9.999473117278764e-05, "loss": 0.007, "step": 5440 }, { "epoch": 48.230088495575224, "grad_norm": 0.16365082561969757, "learning_rate": 9.999448840564731e-05, "loss": 0.0078, "step": 5450 }, { "epoch": 48.31858407079646, "grad_norm": 0.13804985582828522, "learning_rate": 9.999424017118488e-05, "loss": 0.0082, "step": 5460 }, { "epoch": 48.4070796460177, "grad_norm": 0.14704294502735138, "learning_rate": 9.999398646942751e-05, "loss": 0.0075, "step": 5470 }, { "epoch": 48.49557522123894, "grad_norm": 0.12413077801465988, "learning_rate": 9.999372730040296e-05, "loss": 0.0076, "step": 5480 }, { "epoch": 48.584070796460175, "grad_norm": 0.16119691729545593, "learning_rate": 9.999346266413953e-05, "loss": 0.0081, "step": 5490 }, { "epoch": 48.67256637168141, "grad_norm": 0.17906896770000458, "learning_rate": 9.99931925606662e-05, "loss": 0.008, "step": 5500 }, { "epoch": 48.76106194690266, "grad_norm": 0.1275365799665451, "learning_rate": 9.99929169900125e-05, "loss": 0.0072, "step": 5510 }, { "epoch": 48.849557522123895, "grad_norm": 0.12519137561321259, "learning_rate": 9.999263595220855e-05, "loss": 0.008, "step": 5520 }, { "epoch": 48.93805309734513, "grad_norm": 0.15194182097911835, "learning_rate": 9.99923494472851e-05, "loss": 0.0081, "step": 5530 }, { "epoch": 49.02654867256637, "grad_norm": 0.12054912000894547, "learning_rate": 9.999205747527348e-05, "loss": 0.0081, "step": 5540 }, { "epoch": 49.11504424778761, "grad_norm": 0.17035503685474396, "learning_rate": 9.999176003620561e-05, "loss": 0.0074, "step": 5550 }, { "epoch": 49.203539823008846, "grad_norm": 0.16036124527454376, "learning_rate": 9.999145713011405e-05, "loss": 0.0071, "step": 5560 }, { "epoch": 49.29203539823009, "grad_norm": 0.16443262994289398, "learning_rate": 9.999114875703186e-05, "loss": 0.0073, "step": 5570 }, { "epoch": 49.38053097345133, "grad_norm": 0.1346704214811325, "learning_rate": 9.999083491699281e-05, "loss": 0.0076, "step": 5580 }, { "epoch": 49.469026548672566, "grad_norm": 0.1312287151813507, "learning_rate": 9.999051561003123e-05, "loss": 0.0069, "step": 5590 }, { "epoch": 49.557522123893804, "grad_norm": 0.1326473206281662, "learning_rate": 9.999019083618202e-05, "loss": 0.008, "step": 5600 }, { "epoch": 49.64601769911504, "grad_norm": 0.13939642906188965, "learning_rate": 9.99898605954807e-05, "loss": 0.0071, "step": 5610 }, { "epoch": 49.73451327433628, "grad_norm": 0.09773382544517517, "learning_rate": 9.998952488796338e-05, "loss": 0.0072, "step": 5620 }, { "epoch": 49.823008849557525, "grad_norm": 0.09923142939805984, "learning_rate": 9.998918371366676e-05, "loss": 0.007, "step": 5630 }, { "epoch": 49.91150442477876, "grad_norm": 0.13490958511829376, "learning_rate": 9.99888370726282e-05, "loss": 0.0081, "step": 5640 }, { "epoch": 50.0, "grad_norm": 0.10558434575796127, "learning_rate": 9.998848496488556e-05, "loss": 0.0074, "step": 5650 }, { "epoch": 50.08849557522124, "grad_norm": 0.11885306239128113, "learning_rate": 9.998812739047736e-05, "loss": 0.007, "step": 5660 }, { "epoch": 50.176991150442475, "grad_norm": 0.12306637316942215, "learning_rate": 9.99877643494427e-05, "loss": 0.0081, "step": 5670 }, { "epoch": 50.26548672566372, "grad_norm": 0.13808202743530273, "learning_rate": 9.998739584182128e-05, "loss": 0.0067, "step": 5680 }, { "epoch": 50.35398230088496, "grad_norm": 0.11992965638637543, "learning_rate": 9.998702186765342e-05, "loss": 0.0073, "step": 5690 }, { "epoch": 50.442477876106196, "grad_norm": 0.14634229242801666, "learning_rate": 9.998664242698e-05, "loss": 0.0078, "step": 5700 }, { "epoch": 50.530973451327434, "grad_norm": 0.1344883143901825, "learning_rate": 9.998625751984251e-05, "loss": 0.0072, "step": 5710 }, { "epoch": 50.61946902654867, "grad_norm": 0.12016019970178604, "learning_rate": 9.998586714628307e-05, "loss": 0.0073, "step": 5720 }, { "epoch": 50.70796460176991, "grad_norm": 0.10786448419094086, "learning_rate": 9.998547130634432e-05, "loss": 0.0068, "step": 5730 }, { "epoch": 50.796460176991154, "grad_norm": 0.11596783250570297, "learning_rate": 9.99850700000696e-05, "loss": 0.0066, "step": 5740 }, { "epoch": 50.88495575221239, "grad_norm": 0.10513979941606522, "learning_rate": 9.998466322750278e-05, "loss": 0.0071, "step": 5750 }, { "epoch": 50.97345132743363, "grad_norm": 0.11052510142326355, "learning_rate": 9.998425098868834e-05, "loss": 0.0074, "step": 5760 }, { "epoch": 51.06194690265487, "grad_norm": 0.10767388343811035, "learning_rate": 9.998383328367136e-05, "loss": 0.0072, "step": 5770 }, { "epoch": 51.150442477876105, "grad_norm": 0.11172880232334137, "learning_rate": 9.99834101124975e-05, "loss": 0.0071, "step": 5780 }, { "epoch": 51.23893805309734, "grad_norm": 0.11890020221471786, "learning_rate": 9.998298147521309e-05, "loss": 0.0068, "step": 5790 }, { "epoch": 51.32743362831859, "grad_norm": 0.12728281319141388, "learning_rate": 9.998254737186496e-05, "loss": 0.0069, "step": 5800 }, { "epoch": 51.415929203539825, "grad_norm": 0.14971046149730682, "learning_rate": 9.99821078025006e-05, "loss": 0.0075, "step": 5810 }, { "epoch": 51.50442477876106, "grad_norm": 0.13278022408485413, "learning_rate": 9.998166276716807e-05, "loss": 0.0076, "step": 5820 }, { "epoch": 51.5929203539823, "grad_norm": 0.13928380608558655, "learning_rate": 9.998121226591606e-05, "loss": 0.0069, "step": 5830 }, { "epoch": 51.68141592920354, "grad_norm": 0.13920563459396362, "learning_rate": 9.998075629879382e-05, "loss": 0.0072, "step": 5840 }, { "epoch": 51.769911504424776, "grad_norm": 0.13636727631092072, "learning_rate": 9.99802948658512e-05, "loss": 0.0074, "step": 5850 }, { "epoch": 51.85840707964602, "grad_norm": 0.12887609004974365, "learning_rate": 9.99798279671387e-05, "loss": 0.0076, "step": 5860 }, { "epoch": 51.94690265486726, "grad_norm": 0.13336510956287384, "learning_rate": 9.997935560270734e-05, "loss": 0.0074, "step": 5870 }, { "epoch": 52.0353982300885, "grad_norm": 0.1491890251636505, "learning_rate": 9.997887777260879e-05, "loss": 0.0075, "step": 5880 }, { "epoch": 52.123893805309734, "grad_norm": 0.11994275450706482, "learning_rate": 9.997839447689532e-05, "loss": 0.0068, "step": 5890 }, { "epoch": 52.21238938053097, "grad_norm": 0.14876966178417206, "learning_rate": 9.997790571561978e-05, "loss": 0.0073, "step": 5900 }, { "epoch": 52.30088495575221, "grad_norm": 0.20354466140270233, "learning_rate": 9.99774114888356e-05, "loss": 0.0079, "step": 5910 }, { "epoch": 52.389380530973455, "grad_norm": 0.1251549869775772, "learning_rate": 9.997691179659684e-05, "loss": 0.0069, "step": 5920 }, { "epoch": 52.47787610619469, "grad_norm": 0.13673116266727448, "learning_rate": 9.997640663895815e-05, "loss": 0.0071, "step": 5930 }, { "epoch": 52.56637168141593, "grad_norm": 0.11641037464141846, "learning_rate": 9.997589601597477e-05, "loss": 0.0071, "step": 5940 }, { "epoch": 52.65486725663717, "grad_norm": 0.15281692147254944, "learning_rate": 9.997537992770252e-05, "loss": 0.0076, "step": 5950 }, { "epoch": 52.743362831858406, "grad_norm": 0.15915822982788086, "learning_rate": 9.997485837419788e-05, "loss": 0.0079, "step": 5960 }, { "epoch": 52.83185840707964, "grad_norm": 0.1358008086681366, "learning_rate": 9.997433135551786e-05, "loss": 0.007, "step": 5970 }, { "epoch": 52.92035398230089, "grad_norm": 0.18115267157554626, "learning_rate": 9.997379887172009e-05, "loss": 0.0074, "step": 5980 }, { "epoch": 53.008849557522126, "grad_norm": 0.14527107775211334, "learning_rate": 9.997326092286281e-05, "loss": 0.0073, "step": 5990 }, { "epoch": 53.097345132743364, "grad_norm": 0.1449854075908661, "learning_rate": 9.997271750900486e-05, "loss": 0.0075, "step": 6000 }, { "epoch": 53.1858407079646, "grad_norm": 0.12055668979883194, "learning_rate": 9.997216863020565e-05, "loss": 0.007, "step": 6010 }, { "epoch": 53.27433628318584, "grad_norm": 0.14306789636611938, "learning_rate": 9.99716142865252e-05, "loss": 0.0072, "step": 6020 }, { "epoch": 53.36283185840708, "grad_norm": 0.12405459582805634, "learning_rate": 9.997105447802415e-05, "loss": 0.007, "step": 6030 }, { "epoch": 53.45132743362832, "grad_norm": 0.1439598649740219, "learning_rate": 9.997048920476373e-05, "loss": 0.007, "step": 6040 }, { "epoch": 53.53982300884956, "grad_norm": 0.1553235799074173, "learning_rate": 9.996991846680572e-05, "loss": 0.007, "step": 6050 }, { "epoch": 53.6283185840708, "grad_norm": 0.16419124603271484, "learning_rate": 9.996934226421257e-05, "loss": 0.0066, "step": 6060 }, { "epoch": 53.716814159292035, "grad_norm": 0.1346033364534378, "learning_rate": 9.996876059704726e-05, "loss": 0.0073, "step": 6070 }, { "epoch": 53.80530973451327, "grad_norm": 0.13124015927314758, "learning_rate": 9.996817346537343e-05, "loss": 0.0076, "step": 6080 }, { "epoch": 53.89380530973451, "grad_norm": 0.11873197555541992, "learning_rate": 9.996758086925526e-05, "loss": 0.0077, "step": 6090 }, { "epoch": 53.982300884955755, "grad_norm": 0.12637655436992645, "learning_rate": 9.996698280875759e-05, "loss": 0.0069, "step": 6100 }, { "epoch": 54.07079646017699, "grad_norm": 0.13177549839019775, "learning_rate": 9.99663792839458e-05, "loss": 0.0082, "step": 6110 }, { "epoch": 54.15929203539823, "grad_norm": 0.09472975879907608, "learning_rate": 9.99657702948859e-05, "loss": 0.0066, "step": 6120 }, { "epoch": 54.24778761061947, "grad_norm": 0.1014070063829422, "learning_rate": 9.996515584164448e-05, "loss": 0.0068, "step": 6130 }, { "epoch": 54.336283185840706, "grad_norm": 0.130171537399292, "learning_rate": 9.996453592428873e-05, "loss": 0.0078, "step": 6140 }, { "epoch": 54.424778761061944, "grad_norm": 0.15171340107917786, "learning_rate": 9.996391054288646e-05, "loss": 0.007, "step": 6150 }, { "epoch": 54.51327433628319, "grad_norm": 0.13298299908638, "learning_rate": 9.996327969750605e-05, "loss": 0.0078, "step": 6160 }, { "epoch": 54.60176991150443, "grad_norm": 0.1474461853504181, "learning_rate": 9.996264338821649e-05, "loss": 0.0073, "step": 6170 }, { "epoch": 54.690265486725664, "grad_norm": 0.11146021634340286, "learning_rate": 9.996200161508735e-05, "loss": 0.0069, "step": 6180 }, { "epoch": 54.7787610619469, "grad_norm": 0.10253606736660004, "learning_rate": 9.996135437818885e-05, "loss": 0.007, "step": 6190 }, { "epoch": 54.86725663716814, "grad_norm": 0.12463326752185822, "learning_rate": 9.996070167759175e-05, "loss": 0.007, "step": 6200 }, { "epoch": 54.95575221238938, "grad_norm": 0.16125406324863434, "learning_rate": 9.996004351336743e-05, "loss": 0.0076, "step": 6210 }, { "epoch": 55.04424778761062, "grad_norm": 0.15930014848709106, "learning_rate": 9.995937988558785e-05, "loss": 0.0074, "step": 6220 }, { "epoch": 55.13274336283186, "grad_norm": 0.13451719284057617, "learning_rate": 9.995871079432561e-05, "loss": 0.0072, "step": 6230 }, { "epoch": 55.2212389380531, "grad_norm": 0.13616570830345154, "learning_rate": 9.995803623965389e-05, "loss": 0.008, "step": 6240 }, { "epoch": 55.309734513274336, "grad_norm": 0.1808609962463379, "learning_rate": 9.995735622164641e-05, "loss": 0.0071, "step": 6250 }, { "epoch": 55.39823008849557, "grad_norm": 0.14456044137477875, "learning_rate": 9.995667074037758e-05, "loss": 0.0067, "step": 6260 }, { "epoch": 55.48672566371681, "grad_norm": 0.14145857095718384, "learning_rate": 9.995597979592232e-05, "loss": 0.0077, "step": 6270 }, { "epoch": 55.575221238938056, "grad_norm": 0.10300381481647491, "learning_rate": 9.995528338835625e-05, "loss": 0.0071, "step": 6280 }, { "epoch": 55.663716814159294, "grad_norm": 0.0987062156200409, "learning_rate": 9.995458151775547e-05, "loss": 0.007, "step": 6290 }, { "epoch": 55.75221238938053, "grad_norm": 0.10733798146247864, "learning_rate": 9.995387418419677e-05, "loss": 0.0066, "step": 6300 }, { "epoch": 55.84070796460177, "grad_norm": 0.11436575651168823, "learning_rate": 9.99531613877575e-05, "loss": 0.0069, "step": 6310 }, { "epoch": 55.92920353982301, "grad_norm": 0.11065594851970673, "learning_rate": 9.995244312851559e-05, "loss": 0.007, "step": 6320 }, { "epoch": 56.017699115044245, "grad_norm": 0.10757574439048767, "learning_rate": 9.995171940654961e-05, "loss": 0.0069, "step": 6330 }, { "epoch": 56.10619469026549, "grad_norm": 0.1453833431005478, "learning_rate": 9.995099022193871e-05, "loss": 0.0074, "step": 6340 }, { "epoch": 56.19469026548673, "grad_norm": 0.09507931768894196, "learning_rate": 9.995025557476261e-05, "loss": 0.0065, "step": 6350 }, { "epoch": 56.283185840707965, "grad_norm": 0.098049595952034, "learning_rate": 9.994951546510165e-05, "loss": 0.0077, "step": 6360 }, { "epoch": 56.3716814159292, "grad_norm": 0.10645747929811478, "learning_rate": 9.994876989303679e-05, "loss": 0.0066, "step": 6370 }, { "epoch": 56.46017699115044, "grad_norm": 0.11773432791233063, "learning_rate": 9.994801885864955e-05, "loss": 0.007, "step": 6380 }, { "epoch": 56.54867256637168, "grad_norm": 0.10352558642625809, "learning_rate": 9.994726236202205e-05, "loss": 0.0071, "step": 6390 }, { "epoch": 56.63716814159292, "grad_norm": 0.12138586491346359, "learning_rate": 9.994650040323704e-05, "loss": 0.0063, "step": 6400 }, { "epoch": 56.72566371681416, "grad_norm": 0.1313965916633606, "learning_rate": 9.994573298237784e-05, "loss": 0.0072, "step": 6410 }, { "epoch": 56.8141592920354, "grad_norm": 0.10712376236915588, "learning_rate": 9.994496009952837e-05, "loss": 0.0068, "step": 6420 }, { "epoch": 56.902654867256636, "grad_norm": 0.14054933190345764, "learning_rate": 9.994418175477316e-05, "loss": 0.0067, "step": 6430 }, { "epoch": 56.991150442477874, "grad_norm": 0.1412844955921173, "learning_rate": 9.994339794819733e-05, "loss": 0.0066, "step": 6440 }, { "epoch": 57.07964601769911, "grad_norm": 0.11112283170223236, "learning_rate": 9.994260867988658e-05, "loss": 0.0068, "step": 6450 }, { "epoch": 57.16814159292036, "grad_norm": 0.13258463144302368, "learning_rate": 9.994181394992723e-05, "loss": 0.0071, "step": 6460 }, { "epoch": 57.256637168141594, "grad_norm": 0.12830524146556854, "learning_rate": 9.994101375840618e-05, "loss": 0.007, "step": 6470 }, { "epoch": 57.34513274336283, "grad_norm": 0.12997086346149445, "learning_rate": 9.994020810541098e-05, "loss": 0.0069, "step": 6480 }, { "epoch": 57.43362831858407, "grad_norm": 0.11829020828008652, "learning_rate": 9.99393969910297e-05, "loss": 0.0062, "step": 6490 }, { "epoch": 57.52212389380531, "grad_norm": 0.134749174118042, "learning_rate": 9.993858041535104e-05, "loss": 0.0066, "step": 6500 }, { "epoch": 57.610619469026545, "grad_norm": 0.12809590995311737, "learning_rate": 9.99377583784643e-05, "loss": 0.0071, "step": 6510 }, { "epoch": 57.69911504424779, "grad_norm": 0.10998570173978806, "learning_rate": 9.993693088045939e-05, "loss": 0.0067, "step": 6520 }, { "epoch": 57.78761061946903, "grad_norm": 0.10405268520116806, "learning_rate": 9.99360979214268e-05, "loss": 0.006, "step": 6530 }, { "epoch": 57.876106194690266, "grad_norm": 0.0972352996468544, "learning_rate": 9.99352595014576e-05, "loss": 0.0068, "step": 6540 }, { "epoch": 57.9646017699115, "grad_norm": 0.10262265801429749, "learning_rate": 9.993441562064354e-05, "loss": 0.0064, "step": 6550 }, { "epoch": 58.05309734513274, "grad_norm": 0.11251109838485718, "learning_rate": 9.993356627907685e-05, "loss": 0.0071, "step": 6560 }, { "epoch": 58.14159292035398, "grad_norm": 0.12802760303020477, "learning_rate": 9.99327114768504e-05, "loss": 0.0068, "step": 6570 }, { "epoch": 58.230088495575224, "grad_norm": 0.1458670049905777, "learning_rate": 9.99318512140577e-05, "loss": 0.0067, "step": 6580 }, { "epoch": 58.31858407079646, "grad_norm": 0.1487755924463272, "learning_rate": 9.993098549079284e-05, "loss": 0.0073, "step": 6590 }, { "epoch": 58.4070796460177, "grad_norm": 0.15512724220752716, "learning_rate": 9.993011430715047e-05, "loss": 0.0073, "step": 6600 }, { "epoch": 58.49557522123894, "grad_norm": 0.12231585383415222, "learning_rate": 9.992923766322586e-05, "loss": 0.0068, "step": 6610 }, { "epoch": 58.584070796460175, "grad_norm": 0.13225847482681274, "learning_rate": 9.99283555591149e-05, "loss": 0.0065, "step": 6620 }, { "epoch": 58.67256637168141, "grad_norm": 0.12105485796928406, "learning_rate": 9.992746799491404e-05, "loss": 0.0065, "step": 6630 }, { "epoch": 58.76106194690266, "grad_norm": 0.12060397118330002, "learning_rate": 9.992657497072033e-05, "loss": 0.0068, "step": 6640 }, { "epoch": 58.849557522123895, "grad_norm": 0.11555910110473633, "learning_rate": 9.992567648663147e-05, "loss": 0.0069, "step": 6650 }, { "epoch": 58.93805309734513, "grad_norm": 0.137675940990448, "learning_rate": 9.992477254274568e-05, "loss": 0.0061, "step": 6660 }, { "epoch": 59.02654867256637, "grad_norm": 0.14340656995773315, "learning_rate": 9.992386313916183e-05, "loss": 0.0068, "step": 6670 }, { "epoch": 59.11504424778761, "grad_norm": 0.12909331917762756, "learning_rate": 9.992294827597934e-05, "loss": 0.0072, "step": 6680 }, { "epoch": 59.203539823008846, "grad_norm": 0.11933132261037827, "learning_rate": 9.992202795329831e-05, "loss": 0.0063, "step": 6690 }, { "epoch": 59.29203539823009, "grad_norm": 0.12800657749176025, "learning_rate": 9.992110217121936e-05, "loss": 0.0065, "step": 6700 }, { "epoch": 59.38053097345133, "grad_norm": 0.1375381499528885, "learning_rate": 9.992017092984372e-05, "loss": 0.0068, "step": 6710 }, { "epoch": 59.469026548672566, "grad_norm": 0.1585198938846588, "learning_rate": 9.991923422927326e-05, "loss": 0.0071, "step": 6720 }, { "epoch": 59.557522123893804, "grad_norm": 0.1125902459025383, "learning_rate": 9.991829206961037e-05, "loss": 0.0069, "step": 6730 }, { "epoch": 59.64601769911504, "grad_norm": 0.1029604971408844, "learning_rate": 9.991734445095813e-05, "loss": 0.0066, "step": 6740 }, { "epoch": 59.73451327433628, "grad_norm": 0.08860950171947479, "learning_rate": 9.991639137342015e-05, "loss": 0.007, "step": 6750 }, { "epoch": 59.823008849557525, "grad_norm": 0.13004404306411743, "learning_rate": 9.991543283710064e-05, "loss": 0.007, "step": 6760 }, { "epoch": 59.91150442477876, "grad_norm": 0.1068708598613739, "learning_rate": 9.991446884210445e-05, "loss": 0.007, "step": 6770 }, { "epoch": 60.0, "grad_norm": 0.11480708420276642, "learning_rate": 9.9913499388537e-05, "loss": 0.0067, "step": 6780 }, { "epoch": 60.08849557522124, "grad_norm": 0.14268812537193298, "learning_rate": 9.99125244765043e-05, "loss": 0.0068, "step": 6790 }, { "epoch": 60.176991150442475, "grad_norm": 0.11171834915876389, "learning_rate": 9.991154410611296e-05, "loss": 0.0068, "step": 6800 }, { "epoch": 60.26548672566372, "grad_norm": 0.11970166862010956, "learning_rate": 9.99105582774702e-05, "loss": 0.0069, "step": 6810 }, { "epoch": 60.35398230088496, "grad_norm": 0.11299673467874527, "learning_rate": 9.990956699068384e-05, "loss": 0.0067, "step": 6820 }, { "epoch": 60.442477876106196, "grad_norm": 0.08252069354057312, "learning_rate": 9.990857024586224e-05, "loss": 0.0062, "step": 6830 }, { "epoch": 60.530973451327434, "grad_norm": 0.09128489345312119, "learning_rate": 9.990756804311446e-05, "loss": 0.0068, "step": 6840 }, { "epoch": 60.61946902654867, "grad_norm": 0.1355484277009964, "learning_rate": 9.990656038255006e-05, "loss": 0.0066, "step": 6850 }, { "epoch": 60.70796460176991, "grad_norm": 0.16328756511211395, "learning_rate": 9.990554726427926e-05, "loss": 0.0066, "step": 6860 }, { "epoch": 60.796460176991154, "grad_norm": 0.10482525080442429, "learning_rate": 9.990452868841284e-05, "loss": 0.0069, "step": 6870 }, { "epoch": 60.88495575221239, "grad_norm": 0.10071635991334915, "learning_rate": 9.99035046550622e-05, "loss": 0.0066, "step": 6880 }, { "epoch": 60.97345132743363, "grad_norm": 0.11753587424755096, "learning_rate": 9.99024751643393e-05, "loss": 0.0065, "step": 6890 }, { "epoch": 61.06194690265487, "grad_norm": 0.14748500287532806, "learning_rate": 9.990144021635677e-05, "loss": 0.0067, "step": 6900 }, { "epoch": 61.150442477876105, "grad_norm": 0.09329412877559662, "learning_rate": 9.990039981122775e-05, "loss": 0.0067, "step": 6910 }, { "epoch": 61.23893805309734, "grad_norm": 0.1277143657207489, "learning_rate": 9.989935394906602e-05, "loss": 0.0068, "step": 6920 }, { "epoch": 61.32743362831859, "grad_norm": 0.1615787297487259, "learning_rate": 9.989830262998598e-05, "loss": 0.0067, "step": 6930 }, { "epoch": 61.415929203539825, "grad_norm": 0.14806464314460754, "learning_rate": 9.989724585410259e-05, "loss": 0.0067, "step": 6940 }, { "epoch": 61.50442477876106, "grad_norm": 0.1545518934726715, "learning_rate": 9.989618362153139e-05, "loss": 0.0067, "step": 6950 }, { "epoch": 61.5929203539823, "grad_norm": 0.13788571953773499, "learning_rate": 9.989511593238859e-05, "loss": 0.0072, "step": 6960 }, { "epoch": 61.68141592920354, "grad_norm": 0.11123031377792358, "learning_rate": 9.98940427867909e-05, "loss": 0.0067, "step": 6970 }, { "epoch": 61.769911504424776, "grad_norm": 0.11635732650756836, "learning_rate": 9.989296418485573e-05, "loss": 0.0067, "step": 6980 }, { "epoch": 61.85840707964602, "grad_norm": 0.10851877927780151, "learning_rate": 9.989188012670101e-05, "loss": 0.0063, "step": 6990 }, { "epoch": 61.94690265486726, "grad_norm": 0.15028974413871765, "learning_rate": 9.989079061244528e-05, "loss": 0.0064, "step": 7000 }, { "epoch": 62.0353982300885, "grad_norm": 0.13226105272769928, "learning_rate": 9.988969564220769e-05, "loss": 0.0069, "step": 7010 }, { "epoch": 62.123893805309734, "grad_norm": 0.10781238973140717, "learning_rate": 9.988859521610801e-05, "loss": 0.0074, "step": 7020 }, { "epoch": 62.21238938053097, "grad_norm": 0.12379418313503265, "learning_rate": 9.988748933426656e-05, "loss": 0.0069, "step": 7030 }, { "epoch": 62.30088495575221, "grad_norm": 0.16668306291103363, "learning_rate": 9.988637799680428e-05, "loss": 0.0069, "step": 7040 }, { "epoch": 62.389380530973455, "grad_norm": 0.15208269655704498, "learning_rate": 9.98852612038427e-05, "loss": 0.0075, "step": 7050 }, { "epoch": 62.47787610619469, "grad_norm": 0.13797883689403534, "learning_rate": 9.988413895550397e-05, "loss": 0.0075, "step": 7060 }, { "epoch": 62.56637168141593, "grad_norm": 0.10546509921550751, "learning_rate": 9.98830112519108e-05, "loss": 0.0067, "step": 7070 }, { "epoch": 62.65486725663717, "grad_norm": 0.09016519784927368, "learning_rate": 9.98818780931865e-05, "loss": 0.0066, "step": 7080 }, { "epoch": 62.743362831858406, "grad_norm": 0.12119634449481964, "learning_rate": 9.988073947945502e-05, "loss": 0.0069, "step": 7090 }, { "epoch": 62.83185840707964, "grad_norm": 0.139201819896698, "learning_rate": 9.987959541084087e-05, "loss": 0.0066, "step": 7100 }, { "epoch": 62.92035398230089, "grad_norm": 0.12145806849002838, "learning_rate": 9.987844588746915e-05, "loss": 0.0066, "step": 7110 }, { "epoch": 63.008849557522126, "grad_norm": 0.15809254348278046, "learning_rate": 9.987729090946558e-05, "loss": 0.0071, "step": 7120 }, { "epoch": 63.097345132743364, "grad_norm": 0.15733498334884644, "learning_rate": 9.987613047695647e-05, "loss": 0.0067, "step": 7130 }, { "epoch": 63.1858407079646, "grad_norm": 0.15272538363933563, "learning_rate": 9.987496459006871e-05, "loss": 0.0064, "step": 7140 }, { "epoch": 63.27433628318584, "grad_norm": 0.13281561434268951, "learning_rate": 9.987379324892982e-05, "loss": 0.0065, "step": 7150 }, { "epoch": 63.36283185840708, "grad_norm": 0.14352570474147797, "learning_rate": 9.987261645366788e-05, "loss": 0.0067, "step": 7160 }, { "epoch": 63.45132743362832, "grad_norm": 0.12036270648241043, "learning_rate": 9.987143420441158e-05, "loss": 0.0067, "step": 7170 }, { "epoch": 63.53982300884956, "grad_norm": 0.10853111743927002, "learning_rate": 9.987024650129022e-05, "loss": 0.0067, "step": 7180 }, { "epoch": 63.6283185840708, "grad_norm": 0.11569415777921677, "learning_rate": 9.986905334443368e-05, "loss": 0.0069, "step": 7190 }, { "epoch": 63.716814159292035, "grad_norm": 0.1018272191286087, "learning_rate": 9.986785473397245e-05, "loss": 0.0062, "step": 7200 }, { "epoch": 63.80530973451327, "grad_norm": 0.11825605481863022, "learning_rate": 9.98666506700376e-05, "loss": 0.0061, "step": 7210 }, { "epoch": 63.89380530973451, "grad_norm": 0.09540525823831558, "learning_rate": 9.986544115276081e-05, "loss": 0.0062, "step": 7220 }, { "epoch": 63.982300884955755, "grad_norm": 0.09156974405050278, "learning_rate": 9.986422618227433e-05, "loss": 0.0059, "step": 7230 }, { "epoch": 64.070796460177, "grad_norm": 0.11161953210830688, "learning_rate": 9.986300575871106e-05, "loss": 0.0062, "step": 7240 }, { "epoch": 64.15929203539822, "grad_norm": 0.11510982364416122, "learning_rate": 9.986177988220444e-05, "loss": 0.0058, "step": 7250 }, { "epoch": 64.24778761061947, "grad_norm": 0.1127602681517601, "learning_rate": 9.986054855288856e-05, "loss": 0.0058, "step": 7260 }, { "epoch": 64.33628318584071, "grad_norm": 0.09882752597332001, "learning_rate": 9.985931177089802e-05, "loss": 0.0064, "step": 7270 }, { "epoch": 64.42477876106194, "grad_norm": 0.13131412863731384, "learning_rate": 9.985806953636814e-05, "loss": 0.0062, "step": 7280 }, { "epoch": 64.51327433628319, "grad_norm": 0.12704169750213623, "learning_rate": 9.985682184943471e-05, "loss": 0.0071, "step": 7290 }, { "epoch": 64.60176991150442, "grad_norm": 0.12674878537654877, "learning_rate": 9.98555687102342e-05, "loss": 0.006, "step": 7300 }, { "epoch": 64.69026548672566, "grad_norm": 0.1061272844672203, "learning_rate": 9.985431011890367e-05, "loss": 0.0065, "step": 7310 }, { "epoch": 64.77876106194691, "grad_norm": 0.12144777923822403, "learning_rate": 9.985304607558075e-05, "loss": 0.0063, "step": 7320 }, { "epoch": 64.86725663716814, "grad_norm": 0.12685328722000122, "learning_rate": 9.985177658040364e-05, "loss": 0.0068, "step": 7330 }, { "epoch": 64.95575221238938, "grad_norm": 0.11787043511867523, "learning_rate": 9.985050163351119e-05, "loss": 0.0061, "step": 7340 }, { "epoch": 65.04424778761062, "grad_norm": 0.12329311668872833, "learning_rate": 9.984922123504286e-05, "loss": 0.0067, "step": 7350 }, { "epoch": 65.13274336283186, "grad_norm": 0.10366512835025787, "learning_rate": 9.984793538513862e-05, "loss": 0.0072, "step": 7360 }, { "epoch": 65.22123893805309, "grad_norm": 0.13781490921974182, "learning_rate": 9.984664408393912e-05, "loss": 0.0063, "step": 7370 }, { "epoch": 65.30973451327434, "grad_norm": 0.10985226184129715, "learning_rate": 9.984534733158556e-05, "loss": 0.0069, "step": 7380 }, { "epoch": 65.39823008849558, "grad_norm": 0.10374374687671661, "learning_rate": 9.984404512821977e-05, "loss": 0.0065, "step": 7390 }, { "epoch": 65.48672566371681, "grad_norm": 0.1328171342611313, "learning_rate": 9.984273747398411e-05, "loss": 0.006, "step": 7400 }, { "epoch": 65.57522123893806, "grad_norm": 0.11941342800855637, "learning_rate": 9.984142436902165e-05, "loss": 0.0072, "step": 7410 }, { "epoch": 65.66371681415929, "grad_norm": 0.11243950575590134, "learning_rate": 9.984010581347596e-05, "loss": 0.006, "step": 7420 }, { "epoch": 65.75221238938053, "grad_norm": 0.1336316019296646, "learning_rate": 9.983878180749121e-05, "loss": 0.0063, "step": 7430 }, { "epoch": 65.84070796460178, "grad_norm": 0.10539887100458145, "learning_rate": 9.983745235121222e-05, "loss": 0.0061, "step": 7440 }, { "epoch": 65.929203539823, "grad_norm": 0.10091380774974823, "learning_rate": 9.983611744478438e-05, "loss": 0.0069, "step": 7450 }, { "epoch": 66.01769911504425, "grad_norm": 0.11845533549785614, "learning_rate": 9.983477708835365e-05, "loss": 0.0066, "step": 7460 }, { "epoch": 66.10619469026548, "grad_norm": 0.1373676359653473, "learning_rate": 9.983343128206664e-05, "loss": 0.006, "step": 7470 }, { "epoch": 66.19469026548673, "grad_norm": 0.1283421516418457, "learning_rate": 9.983208002607049e-05, "loss": 0.0062, "step": 7480 }, { "epoch": 66.28318584070796, "grad_norm": 0.14404000341892242, "learning_rate": 9.9830723320513e-05, "loss": 0.0064, "step": 7490 }, { "epoch": 66.3716814159292, "grad_norm": 0.11007724702358246, "learning_rate": 9.982936116554254e-05, "loss": 0.0065, "step": 7500 }, { "epoch": 66.46017699115045, "grad_norm": 0.09327288717031479, "learning_rate": 9.982799356130803e-05, "loss": 0.0067, "step": 7510 }, { "epoch": 66.54867256637168, "grad_norm": 0.15277142822742462, "learning_rate": 9.982662050795908e-05, "loss": 0.0059, "step": 7520 }, { "epoch": 66.63716814159292, "grad_norm": 0.10053522139787674, "learning_rate": 9.982524200564583e-05, "loss": 0.0063, "step": 7530 }, { "epoch": 66.72566371681415, "grad_norm": 0.16435828804969788, "learning_rate": 9.982385805451901e-05, "loss": 0.0065, "step": 7540 }, { "epoch": 66.8141592920354, "grad_norm": 0.12222518026828766, "learning_rate": 9.982246865472998e-05, "loss": 0.006, "step": 7550 }, { "epoch": 66.90265486725664, "grad_norm": 0.1254880130290985, "learning_rate": 9.982107380643069e-05, "loss": 0.0066, "step": 7560 }, { "epoch": 66.99115044247787, "grad_norm": 0.10239735245704651, "learning_rate": 9.981967350977368e-05, "loss": 0.0067, "step": 7570 }, { "epoch": 67.07964601769912, "grad_norm": 0.13554519414901733, "learning_rate": 9.981826776491208e-05, "loss": 0.0062, "step": 7580 }, { "epoch": 67.16814159292035, "grad_norm": 0.1134534478187561, "learning_rate": 9.98168565719996e-05, "loss": 0.0066, "step": 7590 }, { "epoch": 67.2566371681416, "grad_norm": 0.10414533317089081, "learning_rate": 9.98154399311906e-05, "loss": 0.0061, "step": 7600 }, { "epoch": 67.34513274336283, "grad_norm": 0.09506098926067352, "learning_rate": 9.981401784263997e-05, "loss": 0.0067, "step": 7610 }, { "epoch": 67.43362831858407, "grad_norm": 0.10589203983545303, "learning_rate": 9.981259030650326e-05, "loss": 0.0063, "step": 7620 }, { "epoch": 67.52212389380531, "grad_norm": 0.11228547990322113, "learning_rate": 9.981115732293655e-05, "loss": 0.006, "step": 7630 }, { "epoch": 67.61061946902655, "grad_norm": 0.10999303311109543, "learning_rate": 9.980971889209659e-05, "loss": 0.0063, "step": 7640 }, { "epoch": 67.69911504424779, "grad_norm": 0.12506988644599915, "learning_rate": 9.980827501414064e-05, "loss": 0.0059, "step": 7650 }, { "epoch": 67.78761061946902, "grad_norm": 0.10974645614624023, "learning_rate": 9.980682568922663e-05, "loss": 0.0066, "step": 7660 }, { "epoch": 67.87610619469027, "grad_norm": 0.13396316766738892, "learning_rate": 9.980537091751304e-05, "loss": 0.0066, "step": 7670 }, { "epoch": 67.96460176991151, "grad_norm": 0.15663084387779236, "learning_rate": 9.980391069915897e-05, "loss": 0.0064, "step": 7680 }, { "epoch": 68.05309734513274, "grad_norm": 0.14433486759662628, "learning_rate": 9.98024450343241e-05, "loss": 0.0065, "step": 7690 }, { "epoch": 68.14159292035399, "grad_norm": 0.13201414048671722, "learning_rate": 9.980097392316872e-05, "loss": 0.0069, "step": 7700 }, { "epoch": 68.23008849557522, "grad_norm": 0.14836303889751434, "learning_rate": 9.97994973658537e-05, "loss": 0.0059, "step": 7710 }, { "epoch": 68.31858407079646, "grad_norm": 0.1281977891921997, "learning_rate": 9.979801536254054e-05, "loss": 0.0062, "step": 7720 }, { "epoch": 68.40707964601769, "grad_norm": 0.1255827695131302, "learning_rate": 9.979652791339127e-05, "loss": 0.0064, "step": 7730 }, { "epoch": 68.49557522123894, "grad_norm": 0.13699287176132202, "learning_rate": 9.97950350185686e-05, "loss": 0.007, "step": 7740 }, { "epoch": 68.58407079646018, "grad_norm": 0.09662818163633347, "learning_rate": 9.979353667823574e-05, "loss": 0.0059, "step": 7750 }, { "epoch": 68.67256637168141, "grad_norm": 0.14224949479103088, "learning_rate": 9.979203289255658e-05, "loss": 0.0057, "step": 7760 }, { "epoch": 68.76106194690266, "grad_norm": 0.1116119846701622, "learning_rate": 9.979052366169557e-05, "loss": 0.0061, "step": 7770 }, { "epoch": 68.84955752212389, "grad_norm": 0.11664599180221558, "learning_rate": 9.978900898581775e-05, "loss": 0.0055, "step": 7780 }, { "epoch": 68.93805309734513, "grad_norm": 0.12078734487295151, "learning_rate": 9.978748886508875e-05, "loss": 0.0063, "step": 7790 }, { "epoch": 69.02654867256638, "grad_norm": 0.13062354922294617, "learning_rate": 9.978596329967484e-05, "loss": 0.0063, "step": 7800 }, { "epoch": 69.11504424778761, "grad_norm": 0.10523324459791183, "learning_rate": 9.978443228974284e-05, "loss": 0.0066, "step": 7810 }, { "epoch": 69.20353982300885, "grad_norm": 0.12760022282600403, "learning_rate": 9.978289583546015e-05, "loss": 0.007, "step": 7820 }, { "epoch": 69.29203539823008, "grad_norm": 0.08580560982227325, "learning_rate": 9.978135393699484e-05, "loss": 0.0064, "step": 7830 }, { "epoch": 69.38053097345133, "grad_norm": 0.0944862887263298, "learning_rate": 9.977980659451548e-05, "loss": 0.0066, "step": 7840 }, { "epoch": 69.46902654867256, "grad_norm": 0.1618262231349945, "learning_rate": 9.977825380819135e-05, "loss": 0.007, "step": 7850 }, { "epoch": 69.5575221238938, "grad_norm": 0.0991116613149643, "learning_rate": 9.97766955781922e-05, "loss": 0.007, "step": 7860 }, { "epoch": 69.64601769911505, "grad_norm": 0.09711875021457672, "learning_rate": 9.977513190468848e-05, "loss": 0.0059, "step": 7870 }, { "epoch": 69.73451327433628, "grad_norm": 0.14101259410381317, "learning_rate": 9.977356278785116e-05, "loss": 0.0061, "step": 7880 }, { "epoch": 69.82300884955752, "grad_norm": 0.13027630746364594, "learning_rate": 9.977198822785184e-05, "loss": 0.0073, "step": 7890 }, { "epoch": 69.91150442477876, "grad_norm": 0.1128426194190979, "learning_rate": 9.977040822486273e-05, "loss": 0.006, "step": 7900 }, { "epoch": 70.0, "grad_norm": 0.1288958489894867, "learning_rate": 9.97688227790566e-05, "loss": 0.0066, "step": 7910 }, { "epoch": 70.08849557522124, "grad_norm": 0.10770505666732788, "learning_rate": 9.976723189060684e-05, "loss": 0.0065, "step": 7920 }, { "epoch": 70.17699115044248, "grad_norm": 0.10949558764696121, "learning_rate": 9.976563555968742e-05, "loss": 0.0063, "step": 7930 }, { "epoch": 70.26548672566372, "grad_norm": 0.11015713959932327, "learning_rate": 9.976403378647292e-05, "loss": 0.0063, "step": 7940 }, { "epoch": 70.35398230088495, "grad_norm": 0.0960339903831482, "learning_rate": 9.97624265711385e-05, "loss": 0.0057, "step": 7950 }, { "epoch": 70.4424778761062, "grad_norm": 0.10930323600769043, "learning_rate": 9.976081391385993e-05, "loss": 0.0064, "step": 7960 }, { "epoch": 70.53097345132744, "grad_norm": 0.09943617135286331, "learning_rate": 9.975919581481356e-05, "loss": 0.0065, "step": 7970 }, { "epoch": 70.61946902654867, "grad_norm": 0.12512046098709106, "learning_rate": 9.975757227417634e-05, "loss": 0.0066, "step": 7980 }, { "epoch": 70.70796460176992, "grad_norm": 0.12245571613311768, "learning_rate": 9.975594329212586e-05, "loss": 0.0063, "step": 7990 }, { "epoch": 70.79646017699115, "grad_norm": 0.09935732930898666, "learning_rate": 9.97543088688402e-05, "loss": 0.0062, "step": 8000 }, { "epoch": 70.88495575221239, "grad_norm": 0.1041388288140297, "learning_rate": 9.975266900449814e-05, "loss": 0.006, "step": 8010 }, { "epoch": 70.97345132743362, "grad_norm": 0.13429324328899384, "learning_rate": 9.975102369927898e-05, "loss": 0.0066, "step": 8020 }, { "epoch": 71.06194690265487, "grad_norm": 0.12102712690830231, "learning_rate": 9.974937295336269e-05, "loss": 0.0066, "step": 8030 }, { "epoch": 71.15044247787611, "grad_norm": 0.09954701364040375, "learning_rate": 9.974771676692975e-05, "loss": 0.0061, "step": 8040 }, { "epoch": 71.23893805309734, "grad_norm": 0.12255644053220749, "learning_rate": 9.974605514016131e-05, "loss": 0.0061, "step": 8050 }, { "epoch": 71.32743362831859, "grad_norm": 0.1163419783115387, "learning_rate": 9.974438807323907e-05, "loss": 0.0064, "step": 8060 }, { "epoch": 71.41592920353982, "grad_norm": 0.0953826829791069, "learning_rate": 9.974271556634535e-05, "loss": 0.0061, "step": 8070 }, { "epoch": 71.50442477876106, "grad_norm": 0.08752670139074326, "learning_rate": 9.974103761966302e-05, "loss": 0.0057, "step": 8080 }, { "epoch": 71.59292035398231, "grad_norm": 0.10170669108629227, "learning_rate": 9.973935423337563e-05, "loss": 0.0058, "step": 8090 }, { "epoch": 71.68141592920354, "grad_norm": 0.11900454014539719, "learning_rate": 9.973766540766722e-05, "loss": 0.0061, "step": 8100 }, { "epoch": 71.76991150442478, "grad_norm": 0.10666965693235397, "learning_rate": 9.97359711427225e-05, "loss": 0.0058, "step": 8110 }, { "epoch": 71.85840707964601, "grad_norm": 0.09945285320281982, "learning_rate": 9.973427143872677e-05, "loss": 0.0064, "step": 8120 }, { "epoch": 71.94690265486726, "grad_norm": 0.11118035018444061, "learning_rate": 9.973256629586589e-05, "loss": 0.0052, "step": 8130 }, { "epoch": 72.03539823008849, "grad_norm": 0.12074369192123413, "learning_rate": 9.973085571432632e-05, "loss": 0.0065, "step": 8140 }, { "epoch": 72.12389380530973, "grad_norm": 0.12977924942970276, "learning_rate": 9.972913969429513e-05, "loss": 0.0057, "step": 8150 }, { "epoch": 72.21238938053098, "grad_norm": 0.11370422691106796, "learning_rate": 9.972741823596e-05, "loss": 0.006, "step": 8160 }, { "epoch": 72.30088495575221, "grad_norm": 0.1010235995054245, "learning_rate": 9.972569133950917e-05, "loss": 0.0063, "step": 8170 }, { "epoch": 72.38938053097345, "grad_norm": 0.07953961193561554, "learning_rate": 9.972395900513151e-05, "loss": 0.0063, "step": 8180 }, { "epoch": 72.47787610619469, "grad_norm": 0.08898427337408066, "learning_rate": 9.972222123301645e-05, "loss": 0.0062, "step": 8190 }, { "epoch": 72.56637168141593, "grad_norm": 0.09729045629501343, "learning_rate": 9.972047802335403e-05, "loss": 0.0057, "step": 8200 }, { "epoch": 72.65486725663717, "grad_norm": 0.10975847393274307, "learning_rate": 9.971872937633488e-05, "loss": 0.0054, "step": 8210 }, { "epoch": 72.7433628318584, "grad_norm": 0.07432346045970917, "learning_rate": 9.971697529215024e-05, "loss": 0.0059, "step": 8220 }, { "epoch": 72.83185840707965, "grad_norm": 0.08630833029747009, "learning_rate": 9.971521577099192e-05, "loss": 0.0063, "step": 8230 }, { "epoch": 72.92035398230088, "grad_norm": 0.09515399485826492, "learning_rate": 9.971345081305236e-05, "loss": 0.0061, "step": 8240 }, { "epoch": 73.00884955752213, "grad_norm": 0.1094130128622055, "learning_rate": 9.971168041852456e-05, "loss": 0.0061, "step": 8250 }, { "epoch": 73.09734513274336, "grad_norm": 0.11208491772413254, "learning_rate": 9.970990458760215e-05, "loss": 0.006, "step": 8260 }, { "epoch": 73.1858407079646, "grad_norm": 0.12518024444580078, "learning_rate": 9.970812332047929e-05, "loss": 0.0058, "step": 8270 }, { "epoch": 73.27433628318585, "grad_norm": 0.14046713709831238, "learning_rate": 9.97063366173508e-05, "loss": 0.0062, "step": 8280 }, { "epoch": 73.36283185840708, "grad_norm": 0.13894730806350708, "learning_rate": 9.970454447841207e-05, "loss": 0.006, "step": 8290 }, { "epoch": 73.45132743362832, "grad_norm": 0.14054197072982788, "learning_rate": 9.970274690385909e-05, "loss": 0.0064, "step": 8300 }, { "epoch": 73.53982300884955, "grad_norm": 0.12739135324954987, "learning_rate": 9.970094389388844e-05, "loss": 0.0061, "step": 8310 }, { "epoch": 73.6283185840708, "grad_norm": 0.11821544170379639, "learning_rate": 9.969913544869728e-05, "loss": 0.0058, "step": 8320 }, { "epoch": 73.71681415929204, "grad_norm": 0.13413073122501373, "learning_rate": 9.96973215684834e-05, "loss": 0.0063, "step": 8330 }, { "epoch": 73.80530973451327, "grad_norm": 0.14590054750442505, "learning_rate": 9.969550225344513e-05, "loss": 0.0062, "step": 8340 }, { "epoch": 73.89380530973452, "grad_norm": 0.10180841386318207, "learning_rate": 9.969367750378147e-05, "loss": 0.0059, "step": 8350 }, { "epoch": 73.98230088495575, "grad_norm": 0.09979245811700821, "learning_rate": 9.969184731969194e-05, "loss": 0.0058, "step": 8360 }, { "epoch": 74.070796460177, "grad_norm": 0.09676596522331238, "learning_rate": 9.96900117013767e-05, "loss": 0.0058, "step": 8370 }, { "epoch": 74.15929203539822, "grad_norm": 0.10069845616817474, "learning_rate": 9.96881706490365e-05, "loss": 0.006, "step": 8380 }, { "epoch": 74.24778761061947, "grad_norm": 0.11834061145782471, "learning_rate": 9.968632416287265e-05, "loss": 0.0058, "step": 8390 }, { "epoch": 74.33628318584071, "grad_norm": 0.1127222403883934, "learning_rate": 9.96844722430871e-05, "loss": 0.0057, "step": 8400 }, { "epoch": 74.42477876106194, "grad_norm": 0.09758228063583374, "learning_rate": 9.968261488988235e-05, "loss": 0.0064, "step": 8410 }, { "epoch": 74.51327433628319, "grad_norm": 0.09847403317689896, "learning_rate": 9.968075210346155e-05, "loss": 0.0057, "step": 8420 }, { "epoch": 74.60176991150442, "grad_norm": 0.10120010375976562, "learning_rate": 9.967888388402839e-05, "loss": 0.0061, "step": 8430 }, { "epoch": 74.69026548672566, "grad_norm": 0.11648057401180267, "learning_rate": 9.967701023178717e-05, "loss": 0.0061, "step": 8440 }, { "epoch": 74.77876106194691, "grad_norm": 0.11761264503002167, "learning_rate": 9.967513114694282e-05, "loss": 0.0059, "step": 8450 }, { "epoch": 74.86725663716814, "grad_norm": 0.14571155607700348, "learning_rate": 9.967324662970079e-05, "loss": 0.0059, "step": 8460 }, { "epoch": 74.95575221238938, "grad_norm": 0.13059626519680023, "learning_rate": 9.96713566802672e-05, "loss": 0.0059, "step": 8470 }, { "epoch": 75.04424778761062, "grad_norm": 0.08886721730232239, "learning_rate": 9.966946129884873e-05, "loss": 0.0061, "step": 8480 }, { "epoch": 75.13274336283186, "grad_norm": 0.10290578752756119, "learning_rate": 9.966756048565265e-05, "loss": 0.006, "step": 8490 }, { "epoch": 75.22123893805309, "grad_norm": 0.10942812263965607, "learning_rate": 9.966565424088681e-05, "loss": 0.0062, "step": 8500 }, { "epoch": 75.30973451327434, "grad_norm": 0.10852043330669403, "learning_rate": 9.96637425647597e-05, "loss": 0.0065, "step": 8510 }, { "epoch": 75.39823008849558, "grad_norm": 0.09682053327560425, "learning_rate": 9.966182545748038e-05, "loss": 0.0058, "step": 8520 }, { "epoch": 75.48672566371681, "grad_norm": 0.1184111088514328, "learning_rate": 9.96599029192585e-05, "loss": 0.0066, "step": 8530 }, { "epoch": 75.57522123893806, "grad_norm": 0.12149938941001892, "learning_rate": 9.965797495030428e-05, "loss": 0.0064, "step": 8540 }, { "epoch": 75.66371681415929, "grad_norm": 0.08611765503883362, "learning_rate": 9.96560415508286e-05, "loss": 0.0062, "step": 8550 }, { "epoch": 75.75221238938053, "grad_norm": 0.09456919878721237, "learning_rate": 9.965410272104286e-05, "loss": 0.0059, "step": 8560 }, { "epoch": 75.84070796460178, "grad_norm": 0.08759572356939316, "learning_rate": 9.96521584611591e-05, "loss": 0.0062, "step": 8570 }, { "epoch": 75.929203539823, "grad_norm": 0.11711636930704117, "learning_rate": 9.965020877138994e-05, "loss": 0.0058, "step": 8580 }, { "epoch": 76.01769911504425, "grad_norm": 0.1080760657787323, "learning_rate": 9.964825365194861e-05, "loss": 0.0066, "step": 8590 }, { "epoch": 76.10619469026548, "grad_norm": 0.14110954105854034, "learning_rate": 9.96462931030489e-05, "loss": 0.0061, "step": 8600 }, { "epoch": 76.19469026548673, "grad_norm": 0.08578823506832123, "learning_rate": 9.96443271249052e-05, "loss": 0.0062, "step": 8610 }, { "epoch": 76.28318584070796, "grad_norm": 0.09732786566019058, "learning_rate": 9.964235571773255e-05, "loss": 0.0065, "step": 8620 }, { "epoch": 76.3716814159292, "grad_norm": 0.08971934765577316, "learning_rate": 9.96403788817465e-05, "loss": 0.0057, "step": 8630 }, { "epoch": 76.46017699115045, "grad_norm": 0.0840202271938324, "learning_rate": 9.963839661716325e-05, "loss": 0.0062, "step": 8640 }, { "epoch": 76.54867256637168, "grad_norm": 0.10182677954435349, "learning_rate": 9.963640892419958e-05, "loss": 0.0062, "step": 8650 }, { "epoch": 76.63716814159292, "grad_norm": 0.10846767574548721, "learning_rate": 9.963441580307286e-05, "loss": 0.0064, "step": 8660 }, { "epoch": 76.72566371681415, "grad_norm": 0.107025645673275, "learning_rate": 9.963241725400104e-05, "loss": 0.0063, "step": 8670 }, { "epoch": 76.8141592920354, "grad_norm": 0.12509597837924957, "learning_rate": 9.963041327720271e-05, "loss": 0.0059, "step": 8680 }, { "epoch": 76.90265486725664, "grad_norm": 0.11883611977100372, "learning_rate": 9.962840387289697e-05, "loss": 0.0063, "step": 8690 }, { "epoch": 76.99115044247787, "grad_norm": 0.11385417729616165, "learning_rate": 9.962638904130363e-05, "loss": 0.0064, "step": 8700 }, { "epoch": 77.07964601769912, "grad_norm": 0.11951186507940292, "learning_rate": 9.962436878264298e-05, "loss": 0.0062, "step": 8710 }, { "epoch": 77.16814159292035, "grad_norm": 0.10156366229057312, "learning_rate": 9.962234309713598e-05, "loss": 0.0061, "step": 8720 }, { "epoch": 77.2566371681416, "grad_norm": 0.1006225198507309, "learning_rate": 9.962031198500414e-05, "loss": 0.0062, "step": 8730 }, { "epoch": 77.34513274336283, "grad_norm": 0.0937986820936203, "learning_rate": 9.961827544646958e-05, "loss": 0.0061, "step": 8740 }, { "epoch": 77.43362831858407, "grad_norm": 0.09878598153591156, "learning_rate": 9.961623348175501e-05, "loss": 0.0062, "step": 8750 }, { "epoch": 77.52212389380531, "grad_norm": 0.09665126353502274, "learning_rate": 9.961418609108377e-05, "loss": 0.0059, "step": 8760 }, { "epoch": 77.61061946902655, "grad_norm": 0.10199528187513351, "learning_rate": 9.961213327467971e-05, "loss": 0.0055, "step": 8770 }, { "epoch": 77.69911504424779, "grad_norm": 0.1394171565771103, "learning_rate": 9.961007503276736e-05, "loss": 0.006, "step": 8780 }, { "epoch": 77.78761061946902, "grad_norm": 0.1384783834218979, "learning_rate": 9.960801136557179e-05, "loss": 0.0064, "step": 8790 }, { "epoch": 77.87610619469027, "grad_norm": 0.12698636949062347, "learning_rate": 9.960594227331866e-05, "loss": 0.0059, "step": 8800 }, { "epoch": 77.96460176991151, "grad_norm": 0.10136426985263824, "learning_rate": 9.960386775623429e-05, "loss": 0.006, "step": 8810 }, { "epoch": 78.05309734513274, "grad_norm": 0.12089662998914719, "learning_rate": 9.96017878145455e-05, "loss": 0.0067, "step": 8820 }, { "epoch": 78.14159292035399, "grad_norm": 0.1091749519109726, "learning_rate": 9.959970244847977e-05, "loss": 0.0057, "step": 8830 }, { "epoch": 78.23008849557522, "grad_norm": 0.12326713651418686, "learning_rate": 9.959761165826518e-05, "loss": 0.0061, "step": 8840 }, { "epoch": 78.31858407079646, "grad_norm": 0.12532973289489746, "learning_rate": 9.959551544413033e-05, "loss": 0.0063, "step": 8850 }, { "epoch": 78.40707964601769, "grad_norm": 0.13419322669506073, "learning_rate": 9.959341380630448e-05, "loss": 0.0058, "step": 8860 }, { "epoch": 78.49557522123894, "grad_norm": 0.123810775578022, "learning_rate": 9.959130674501746e-05, "loss": 0.0058, "step": 8870 }, { "epoch": 78.58407079646018, "grad_norm": 0.07981380820274353, "learning_rate": 9.958919426049968e-05, "loss": 0.0058, "step": 8880 }, { "epoch": 78.67256637168141, "grad_norm": 0.11350315064191818, "learning_rate": 9.958707635298219e-05, "loss": 0.0056, "step": 8890 }, { "epoch": 78.76106194690266, "grad_norm": 0.12435969710350037, "learning_rate": 9.958495302269657e-05, "loss": 0.0061, "step": 8900 }, { "epoch": 78.84955752212389, "grad_norm": 0.10741688311100006, "learning_rate": 9.958282426987503e-05, "loss": 0.0062, "step": 8910 }, { "epoch": 78.93805309734513, "grad_norm": 0.11767306923866272, "learning_rate": 9.95806900947504e-05, "loss": 0.0058, "step": 8920 }, { "epoch": 79.02654867256638, "grad_norm": 0.09329789876937866, "learning_rate": 9.957855049755604e-05, "loss": 0.0056, "step": 8930 }, { "epoch": 79.11504424778761, "grad_norm": 0.07824363559484482, "learning_rate": 9.957640547852593e-05, "loss": 0.0061, "step": 8940 }, { "epoch": 79.20353982300885, "grad_norm": 0.11748840659856796, "learning_rate": 9.957425503789466e-05, "loss": 0.006, "step": 8950 }, { "epoch": 79.29203539823008, "grad_norm": 0.12106886506080627, "learning_rate": 9.957209917589738e-05, "loss": 0.0054, "step": 8960 }, { "epoch": 79.38053097345133, "grad_norm": 0.10347803682088852, "learning_rate": 9.956993789276987e-05, "loss": 0.0053, "step": 8970 }, { "epoch": 79.46902654867256, "grad_norm": 0.09693463146686554, "learning_rate": 9.956777118874847e-05, "loss": 0.0056, "step": 8980 }, { "epoch": 79.5575221238938, "grad_norm": 0.08462657034397125, "learning_rate": 9.956559906407016e-05, "loss": 0.0057, "step": 8990 }, { "epoch": 79.64601769911505, "grad_norm": 0.10918508470058441, "learning_rate": 9.956342151897245e-05, "loss": 0.0054, "step": 9000 }, { "epoch": 79.73451327433628, "grad_norm": 0.08692421019077301, "learning_rate": 9.956123855369346e-05, "loss": 0.0062, "step": 9010 }, { "epoch": 79.82300884955752, "grad_norm": 0.10115757584571838, "learning_rate": 9.955905016847196e-05, "loss": 0.006, "step": 9020 }, { "epoch": 79.91150442477876, "grad_norm": 0.07491198182106018, "learning_rate": 9.955685636354723e-05, "loss": 0.0056, "step": 9030 }, { "epoch": 80.0, "grad_norm": 0.09081827104091644, "learning_rate": 9.95546571391592e-05, "loss": 0.0057, "step": 9040 }, { "epoch": 80.08849557522124, "grad_norm": 0.11967772245407104, "learning_rate": 9.955245249554837e-05, "loss": 0.0061, "step": 9050 }, { "epoch": 80.17699115044248, "grad_norm": 0.08820559084415436, "learning_rate": 9.955024243295582e-05, "loss": 0.0057, "step": 9060 }, { "epoch": 80.26548672566372, "grad_norm": 0.0963023453950882, "learning_rate": 9.954802695162328e-05, "loss": 0.0062, "step": 9070 }, { "epoch": 80.35398230088495, "grad_norm": 0.12457380443811417, "learning_rate": 9.954580605179302e-05, "loss": 0.0058, "step": 9080 }, { "epoch": 80.4424778761062, "grad_norm": 0.09814894944429398, "learning_rate": 9.954357973370788e-05, "loss": 0.006, "step": 9090 }, { "epoch": 80.53097345132744, "grad_norm": 0.10084599256515503, "learning_rate": 9.954134799761135e-05, "loss": 0.0061, "step": 9100 }, { "epoch": 80.61946902654867, "grad_norm": 0.10183996707201004, "learning_rate": 9.953911084374748e-05, "loss": 0.0064, "step": 9110 }, { "epoch": 80.70796460176992, "grad_norm": 0.11158488690853119, "learning_rate": 9.953686827236093e-05, "loss": 0.0055, "step": 9120 }, { "epoch": 80.79646017699115, "grad_norm": 0.15042245388031006, "learning_rate": 9.953462028369695e-05, "loss": 0.0061, "step": 9130 }, { "epoch": 80.88495575221239, "grad_norm": 0.11170530319213867, "learning_rate": 9.953236687800136e-05, "loss": 0.0062, "step": 9140 }, { "epoch": 80.97345132743362, "grad_norm": 0.08855316042900085, "learning_rate": 9.95301080555206e-05, "loss": 0.0066, "step": 9150 }, { "epoch": 81.06194690265487, "grad_norm": 0.08837416768074036, "learning_rate": 9.952784381650171e-05, "loss": 0.006, "step": 9160 }, { "epoch": 81.15044247787611, "grad_norm": 0.12177439779043198, "learning_rate": 9.952557416119226e-05, "loss": 0.0061, "step": 9170 }, { "epoch": 81.23893805309734, "grad_norm": 0.08791584521532059, "learning_rate": 9.95232990898405e-05, "loss": 0.0053, "step": 9180 }, { "epoch": 81.32743362831859, "grad_norm": 0.11918691545724869, "learning_rate": 9.95210186026952e-05, "loss": 0.0061, "step": 9190 }, { "epoch": 81.41592920353982, "grad_norm": 0.11663874238729477, "learning_rate": 9.951873270000576e-05, "loss": 0.006, "step": 9200 }, { "epoch": 81.50442477876106, "grad_norm": 0.15835988521575928, "learning_rate": 9.951644138202216e-05, "loss": 0.0059, "step": 9210 }, { "epoch": 81.59292035398231, "grad_norm": 0.10509777814149857, "learning_rate": 9.951414464899498e-05, "loss": 0.0065, "step": 9220 }, { "epoch": 81.68141592920354, "grad_norm": 0.10519393533468246, "learning_rate": 9.951184250117538e-05, "loss": 0.0056, "step": 9230 }, { "epoch": 81.76991150442478, "grad_norm": 0.11766769737005234, "learning_rate": 9.950953493881513e-05, "loss": 0.006, "step": 9240 }, { "epoch": 81.85840707964601, "grad_norm": 0.11554163694381714, "learning_rate": 9.950722196216658e-05, "loss": 0.0053, "step": 9250 }, { "epoch": 81.94690265486726, "grad_norm": 0.10996510088443756, "learning_rate": 9.950490357148265e-05, "loss": 0.0058, "step": 9260 }, { "epoch": 82.03539823008849, "grad_norm": 0.1464570015668869, "learning_rate": 9.950257976701692e-05, "loss": 0.0061, "step": 9270 }, { "epoch": 82.12389380530973, "grad_norm": 0.07254662364721298, "learning_rate": 9.950025054902348e-05, "loss": 0.0056, "step": 9280 }, { "epoch": 82.21238938053098, "grad_norm": 0.11178135126829147, "learning_rate": 9.949791591775706e-05, "loss": 0.0056, "step": 9290 }, { "epoch": 82.30088495575221, "grad_norm": 0.07963807880878448, "learning_rate": 9.949557587347298e-05, "loss": 0.0052, "step": 9300 }, { "epoch": 82.38938053097345, "grad_norm": 0.06853527575731277, "learning_rate": 9.949323041642713e-05, "loss": 0.0057, "step": 9310 }, { "epoch": 82.47787610619469, "grad_norm": 0.1029864028096199, "learning_rate": 9.949087954687602e-05, "loss": 0.0055, "step": 9320 }, { "epoch": 82.56637168141593, "grad_norm": 0.10096902400255203, "learning_rate": 9.948852326507672e-05, "loss": 0.0059, "step": 9330 }, { "epoch": 82.65486725663717, "grad_norm": 0.12288044393062592, "learning_rate": 9.948616157128694e-05, "loss": 0.006, "step": 9340 }, { "epoch": 82.7433628318584, "grad_norm": 0.09727030992507935, "learning_rate": 9.948379446576493e-05, "loss": 0.0065, "step": 9350 }, { "epoch": 82.83185840707965, "grad_norm": 0.11985905468463898, "learning_rate": 9.948142194876952e-05, "loss": 0.0057, "step": 9360 }, { "epoch": 82.92035398230088, "grad_norm": 0.10850052535533905, "learning_rate": 9.947904402056024e-05, "loss": 0.0053, "step": 9370 }, { "epoch": 83.00884955752213, "grad_norm": 0.1047726571559906, "learning_rate": 9.947666068139708e-05, "loss": 0.0059, "step": 9380 }, { "epoch": 83.09734513274336, "grad_norm": 0.09545407444238663, "learning_rate": 9.947427193154071e-05, "loss": 0.0056, "step": 9390 }, { "epoch": 83.1858407079646, "grad_norm": 0.09111058712005615, "learning_rate": 9.947187777125233e-05, "loss": 0.0059, "step": 9400 }, { "epoch": 83.27433628318585, "grad_norm": 0.0800376608967781, "learning_rate": 9.946947820079377e-05, "loss": 0.0053, "step": 9410 }, { "epoch": 83.36283185840708, "grad_norm": 0.09253442287445068, "learning_rate": 9.946707322042747e-05, "loss": 0.0051, "step": 9420 }, { "epoch": 83.45132743362832, "grad_norm": 0.08025803416967392, "learning_rate": 9.94646628304164e-05, "loss": 0.0052, "step": 9430 }, { "epoch": 83.53982300884955, "grad_norm": 0.10385815054178238, "learning_rate": 9.946224703102418e-05, "loss": 0.0056, "step": 9440 }, { "epoch": 83.6283185840708, "grad_norm": 0.10036700963973999, "learning_rate": 9.945982582251498e-05, "loss": 0.0059, "step": 9450 }, { "epoch": 83.71681415929204, "grad_norm": 0.11064866185188293, "learning_rate": 9.94573992051536e-05, "loss": 0.0056, "step": 9460 }, { "epoch": 83.80530973451327, "grad_norm": 0.10301430523395538, "learning_rate": 9.94549671792054e-05, "loss": 0.0064, "step": 9470 }, { "epoch": 83.89380530973452, "grad_norm": 0.12252218276262283, "learning_rate": 9.945252974493635e-05, "loss": 0.0056, "step": 9480 }, { "epoch": 83.98230088495575, "grad_norm": 0.12609660625457764, "learning_rate": 9.9450086902613e-05, "loss": 0.006, "step": 9490 }, { "epoch": 84.070796460177, "grad_norm": 0.10264957696199417, "learning_rate": 9.944763865250248e-05, "loss": 0.0054, "step": 9500 }, { "epoch": 84.15929203539822, "grad_norm": 0.10181338340044022, "learning_rate": 9.944518499487254e-05, "loss": 0.0065, "step": 9510 }, { "epoch": 84.24778761061947, "grad_norm": 0.10405100136995316, "learning_rate": 9.944272592999151e-05, "loss": 0.006, "step": 9520 }, { "epoch": 84.33628318584071, "grad_norm": 0.10661379992961884, "learning_rate": 9.94402614581283e-05, "loss": 0.0068, "step": 9530 }, { "epoch": 84.42477876106194, "grad_norm": 0.09384993463754654, "learning_rate": 9.943779157955244e-05, "loss": 0.0055, "step": 9540 }, { "epoch": 84.51327433628319, "grad_norm": 0.08623424917459488, "learning_rate": 9.943531629453403e-05, "loss": 0.0054, "step": 9550 }, { "epoch": 84.60176991150442, "grad_norm": 0.09496702253818512, "learning_rate": 9.943283560334375e-05, "loss": 0.0056, "step": 9560 }, { "epoch": 84.69026548672566, "grad_norm": 0.1038193479180336, "learning_rate": 9.943034950625288e-05, "loss": 0.0051, "step": 9570 }, { "epoch": 84.77876106194691, "grad_norm": 0.09917920082807541, "learning_rate": 9.942785800353332e-05, "loss": 0.0063, "step": 9580 }, { "epoch": 84.86725663716814, "grad_norm": 0.1258620172739029, "learning_rate": 9.942536109545751e-05, "loss": 0.0055, "step": 9590 }, { "epoch": 84.95575221238938, "grad_norm": 0.12064798921346664, "learning_rate": 9.942285878229853e-05, "loss": 0.0053, "step": 9600 }, { "epoch": 85.04424778761062, "grad_norm": 0.08963096141815186, "learning_rate": 9.942035106433001e-05, "loss": 0.0054, "step": 9610 }, { "epoch": 85.13274336283186, "grad_norm": 0.1056583970785141, "learning_rate": 9.94178379418262e-05, "loss": 0.0053, "step": 9620 }, { "epoch": 85.22123893805309, "grad_norm": 0.07988014817237854, "learning_rate": 9.941531941506194e-05, "loss": 0.0056, "step": 9630 }, { "epoch": 85.30973451327434, "grad_norm": 0.12046244740486145, "learning_rate": 9.941279548431263e-05, "loss": 0.0061, "step": 9640 }, { "epoch": 85.39823008849558, "grad_norm": 0.08098091930150986, "learning_rate": 9.941026614985431e-05, "loss": 0.0055, "step": 9650 }, { "epoch": 85.48672566371681, "grad_norm": 0.09164562821388245, "learning_rate": 9.940773141196357e-05, "loss": 0.0055, "step": 9660 }, { "epoch": 85.57522123893806, "grad_norm": 0.0929412692785263, "learning_rate": 9.94051912709176e-05, "loss": 0.0059, "step": 9670 }, { "epoch": 85.66371681415929, "grad_norm": 0.12282797694206238, "learning_rate": 9.940264572699421e-05, "loss": 0.0055, "step": 9680 }, { "epoch": 85.75221238938053, "grad_norm": 0.09457139670848846, "learning_rate": 9.940009478047174e-05, "loss": 0.0058, "step": 9690 }, { "epoch": 85.84070796460178, "grad_norm": 0.09769842028617859, "learning_rate": 9.939753843162918e-05, "loss": 0.0052, "step": 9700 }, { "epoch": 85.929203539823, "grad_norm": 0.10988950729370117, "learning_rate": 9.939497668074609e-05, "loss": 0.0062, "step": 9710 }, { "epoch": 86.01769911504425, "grad_norm": 0.10237492620944977, "learning_rate": 9.93924095281026e-05, "loss": 0.006, "step": 9720 }, { "epoch": 86.10619469026548, "grad_norm": 0.12422050535678864, "learning_rate": 9.938983697397948e-05, "loss": 0.0053, "step": 9730 }, { "epoch": 86.19469026548673, "grad_norm": 0.1295003443956375, "learning_rate": 9.938725901865805e-05, "loss": 0.0057, "step": 9740 }, { "epoch": 86.28318584070796, "grad_norm": 0.12800505757331848, "learning_rate": 9.93846756624202e-05, "loss": 0.0053, "step": 9750 }, { "epoch": 86.3716814159292, "grad_norm": 0.12015777081251144, "learning_rate": 9.938208690554849e-05, "loss": 0.0062, "step": 9760 }, { "epoch": 86.46017699115045, "grad_norm": 0.11969970166683197, "learning_rate": 9.9379492748326e-05, "loss": 0.006, "step": 9770 }, { "epoch": 86.54867256637168, "grad_norm": 0.1453525424003601, "learning_rate": 9.937689319103641e-05, "loss": 0.0061, "step": 9780 }, { "epoch": 86.63716814159292, "grad_norm": 0.13354043662548065, "learning_rate": 9.937428823396404e-05, "loss": 0.0061, "step": 9790 }, { "epoch": 86.72566371681415, "grad_norm": 0.09945225715637207, "learning_rate": 9.937167787739372e-05, "loss": 0.006, "step": 9800 }, { "epoch": 86.8141592920354, "grad_norm": 0.13246048986911774, "learning_rate": 9.936906212161095e-05, "loss": 0.0054, "step": 9810 }, { "epoch": 86.90265486725664, "grad_norm": 0.1504821926355362, "learning_rate": 9.936644096690176e-05, "loss": 0.0054, "step": 9820 }, { "epoch": 86.99115044247787, "grad_norm": 0.12351587414741516, "learning_rate": 9.936381441355282e-05, "loss": 0.0057, "step": 9830 }, { "epoch": 87.07964601769912, "grad_norm": 0.10668089985847473, "learning_rate": 9.936118246185136e-05, "loss": 0.0057, "step": 9840 }, { "epoch": 87.16814159292035, "grad_norm": 0.11420907080173492, "learning_rate": 9.935854511208518e-05, "loss": 0.0055, "step": 9850 }, { "epoch": 87.2566371681416, "grad_norm": 0.11495466530323029, "learning_rate": 9.935590236454272e-05, "loss": 0.0056, "step": 9860 }, { "epoch": 87.34513274336283, "grad_norm": 0.1109667494893074, "learning_rate": 9.935325421951298e-05, "loss": 0.0058, "step": 9870 }, { "epoch": 87.43362831858407, "grad_norm": 0.1166212260723114, "learning_rate": 9.935060067728557e-05, "loss": 0.0052, "step": 9880 }, { "epoch": 87.52212389380531, "grad_norm": 0.10015451163053513, "learning_rate": 9.934794173815067e-05, "loss": 0.0056, "step": 9890 }, { "epoch": 87.61061946902655, "grad_norm": 0.12183625251054764, "learning_rate": 9.934527740239906e-05, "loss": 0.0053, "step": 9900 }, { "epoch": 87.69911504424779, "grad_norm": 0.1307184249162674, "learning_rate": 9.934260767032209e-05, "loss": 0.0054, "step": 9910 }, { "epoch": 87.78761061946902, "grad_norm": 0.09311528503894806, "learning_rate": 9.933993254221172e-05, "loss": 0.005, "step": 9920 }, { "epoch": 87.87610619469027, "grad_norm": 0.11214279383420944, "learning_rate": 9.933725201836053e-05, "loss": 0.0061, "step": 9930 }, { "epoch": 87.96460176991151, "grad_norm": 0.125685915350914, "learning_rate": 9.933456609906162e-05, "loss": 0.0058, "step": 9940 }, { "epoch": 88.05309734513274, "grad_norm": 0.09989435225725174, "learning_rate": 9.933187478460875e-05, "loss": 0.0055, "step": 9950 }, { "epoch": 88.14159292035399, "grad_norm": 0.13455693423748016, "learning_rate": 9.93291780752962e-05, "loss": 0.0063, "step": 9960 }, { "epoch": 88.23008849557522, "grad_norm": 0.125112846493721, "learning_rate": 9.932647597141893e-05, "loss": 0.0061, "step": 9970 }, { "epoch": 88.31858407079646, "grad_norm": 0.09922801703214645, "learning_rate": 9.932376847327239e-05, "loss": 0.0058, "step": 9980 }, { "epoch": 88.40707964601769, "grad_norm": 0.09822017699480057, "learning_rate": 9.932105558115268e-05, "loss": 0.0056, "step": 9990 }, { "epoch": 88.49557522123894, "grad_norm": 0.07514345645904541, "learning_rate": 9.931833729535651e-05, "loss": 0.0065, "step": 10000 }, { "epoch": 88.58407079646018, "grad_norm": 0.09033774584531784, "learning_rate": 9.931561361618111e-05, "loss": 0.006, "step": 10010 }, { "epoch": 88.67256637168141, "grad_norm": 0.08522652089595795, "learning_rate": 9.931288454392435e-05, "loss": 0.0054, "step": 10020 }, { "epoch": 88.76106194690266, "grad_norm": 0.09163916110992432, "learning_rate": 9.931015007888467e-05, "loss": 0.0053, "step": 10030 }, { "epoch": 88.84955752212389, "grad_norm": 0.0859520360827446, "learning_rate": 9.930741022136112e-05, "loss": 0.005, "step": 10040 }, { "epoch": 88.93805309734513, "grad_norm": 0.10434073954820633, "learning_rate": 9.930466497165333e-05, "loss": 0.0052, "step": 10050 }, { "epoch": 89.02654867256638, "grad_norm": 0.08592240512371063, "learning_rate": 9.93019143300615e-05, "loss": 0.0054, "step": 10060 }, { "epoch": 89.11504424778761, "grad_norm": 0.0753011703491211, "learning_rate": 9.929915829688644e-05, "loss": 0.0058, "step": 10070 }, { "epoch": 89.20353982300885, "grad_norm": 0.09075010567903519, "learning_rate": 9.929639687242955e-05, "loss": 0.0052, "step": 10080 }, { "epoch": 89.29203539823008, "grad_norm": 0.10076374560594559, "learning_rate": 9.929363005699281e-05, "loss": 0.0054, "step": 10090 }, { "epoch": 89.38053097345133, "grad_norm": 0.11675793677568436, "learning_rate": 9.92908578508788e-05, "loss": 0.0053, "step": 10100 }, { "epoch": 89.46902654867256, "grad_norm": 0.10183381289243698, "learning_rate": 9.928808025439069e-05, "loss": 0.0052, "step": 10110 }, { "epoch": 89.5575221238938, "grad_norm": 0.09888213127851486, "learning_rate": 9.928529726783223e-05, "loss": 0.0061, "step": 10120 }, { "epoch": 89.64601769911505, "grad_norm": 0.08462341129779816, "learning_rate": 9.928250889150774e-05, "loss": 0.0056, "step": 10130 }, { "epoch": 89.73451327433628, "grad_norm": 0.09935176372528076, "learning_rate": 9.92797151257222e-05, "loss": 0.0052, "step": 10140 }, { "epoch": 89.82300884955752, "grad_norm": 0.08787396550178528, "learning_rate": 9.927691597078108e-05, "loss": 0.0058, "step": 10150 }, { "epoch": 89.91150442477876, "grad_norm": 0.08054458349943161, "learning_rate": 9.927411142699053e-05, "loss": 0.0056, "step": 10160 }, { "epoch": 90.0, "grad_norm": 0.09531901031732559, "learning_rate": 9.927130149465725e-05, "loss": 0.0056, "step": 10170 }, { "epoch": 90.08849557522124, "grad_norm": 0.12852004170417786, "learning_rate": 9.92684861740885e-05, "loss": 0.005, "step": 10180 }, { "epoch": 90.17699115044248, "grad_norm": 0.11645930260419846, "learning_rate": 9.926566546559217e-05, "loss": 0.0053, "step": 10190 }, { "epoch": 90.26548672566372, "grad_norm": 0.09866490960121155, "learning_rate": 9.926283936947673e-05, "loss": 0.0059, "step": 10200 }, { "epoch": 90.35398230088495, "grad_norm": 0.1029471755027771, "learning_rate": 9.926000788605126e-05, "loss": 0.0053, "step": 10210 }, { "epoch": 90.4424778761062, "grad_norm": 0.09673412889242172, "learning_rate": 9.92571710156254e-05, "loss": 0.0053, "step": 10220 }, { "epoch": 90.53097345132744, "grad_norm": 0.12121190875768661, "learning_rate": 9.925432875850936e-05, "loss": 0.0054, "step": 10230 }, { "epoch": 90.61946902654867, "grad_norm": 0.09911256283521652, "learning_rate": 9.925148111501396e-05, "loss": 0.006, "step": 10240 }, { "epoch": 90.70796460176992, "grad_norm": 0.09052865952253342, "learning_rate": 9.924862808545066e-05, "loss": 0.0051, "step": 10250 }, { "epoch": 90.79646017699115, "grad_norm": 0.08117509633302689, "learning_rate": 9.924576967013141e-05, "loss": 0.0058, "step": 10260 }, { "epoch": 90.88495575221239, "grad_norm": 0.08937954902648926, "learning_rate": 9.924290586936887e-05, "loss": 0.006, "step": 10270 }, { "epoch": 90.97345132743362, "grad_norm": 0.13125444948673248, "learning_rate": 9.924003668347614e-05, "loss": 0.0058, "step": 10280 }, { "epoch": 91.06194690265487, "grad_norm": 0.07319777458906174, "learning_rate": 9.923716211276704e-05, "loss": 0.0053, "step": 10290 }, { "epoch": 91.15044247787611, "grad_norm": 0.11353085935115814, "learning_rate": 9.923428215755594e-05, "loss": 0.0057, "step": 10300 }, { "epoch": 91.23893805309734, "grad_norm": 0.12361222505569458, "learning_rate": 9.923139681815775e-05, "loss": 0.0061, "step": 10310 }, { "epoch": 91.32743362831859, "grad_norm": 0.10526737570762634, "learning_rate": 9.922850609488801e-05, "loss": 0.0054, "step": 10320 }, { "epoch": 91.41592920353982, "grad_norm": 0.08422639220952988, "learning_rate": 9.922560998806287e-05, "loss": 0.0057, "step": 10330 }, { "epoch": 91.50442477876106, "grad_norm": 0.07958678156137466, "learning_rate": 9.922270849799905e-05, "loss": 0.0055, "step": 10340 }, { "epoch": 91.59292035398231, "grad_norm": 0.11012106388807297, "learning_rate": 9.92198016250138e-05, "loss": 0.0058, "step": 10350 }, { "epoch": 91.68141592920354, "grad_norm": 0.11095649003982544, "learning_rate": 9.921688936942506e-05, "loss": 0.0051, "step": 10360 }, { "epoch": 91.76991150442478, "grad_norm": 0.11602571606636047, "learning_rate": 9.921397173155129e-05, "loss": 0.006, "step": 10370 }, { "epoch": 91.85840707964601, "grad_norm": 0.11211151629686356, "learning_rate": 9.921104871171157e-05, "loss": 0.006, "step": 10380 }, { "epoch": 91.94690265486726, "grad_norm": 0.11110955476760864, "learning_rate": 9.920812031022554e-05, "loss": 0.0054, "step": 10390 }, { "epoch": 92.03539823008849, "grad_norm": 0.11620385199785233, "learning_rate": 9.920518652741348e-05, "loss": 0.0052, "step": 10400 }, { "epoch": 92.12389380530973, "grad_norm": 0.10567715764045715, "learning_rate": 9.920224736359618e-05, "loss": 0.0051, "step": 10410 }, { "epoch": 92.21238938053098, "grad_norm": 0.12158272415399551, "learning_rate": 9.91993028190951e-05, "loss": 0.0051, "step": 10420 }, { "epoch": 92.30088495575221, "grad_norm": 0.08298440277576447, "learning_rate": 9.919635289423222e-05, "loss": 0.0062, "step": 10430 }, { "epoch": 92.38938053097345, "grad_norm": 0.11056473106145859, "learning_rate": 9.919339758933015e-05, "loss": 0.0054, "step": 10440 }, { "epoch": 92.47787610619469, "grad_norm": 0.08665698766708374, "learning_rate": 9.919043690471209e-05, "loss": 0.0053, "step": 10450 }, { "epoch": 92.56637168141593, "grad_norm": 0.12419696152210236, "learning_rate": 9.91874708407018e-05, "loss": 0.0052, "step": 10460 }, { "epoch": 92.65486725663717, "grad_norm": 0.14422670006752014, "learning_rate": 9.918449939762367e-05, "loss": 0.0053, "step": 10470 }, { "epoch": 92.7433628318584, "grad_norm": 0.1112806499004364, "learning_rate": 9.91815225758026e-05, "loss": 0.0055, "step": 10480 }, { "epoch": 92.83185840707965, "grad_norm": 0.10156608372926712, "learning_rate": 9.917854037556419e-05, "loss": 0.0052, "step": 10490 }, { "epoch": 92.92035398230088, "grad_norm": 0.09362796694040298, "learning_rate": 9.917555279723454e-05, "loss": 0.0054, "step": 10500 }, { "epoch": 93.00884955752213, "grad_norm": 0.10930593311786652, "learning_rate": 9.917255984114036e-05, "loss": 0.0054, "step": 10510 }, { "epoch": 93.09734513274336, "grad_norm": 0.08442426472902298, "learning_rate": 9.916956150760896e-05, "loss": 0.0056, "step": 10520 }, { "epoch": 93.1858407079646, "grad_norm": 0.08293269574642181, "learning_rate": 9.916655779696826e-05, "loss": 0.0051, "step": 10530 }, { "epoch": 93.27433628318585, "grad_norm": 0.10199480503797531, "learning_rate": 9.916354870954671e-05, "loss": 0.0058, "step": 10540 }, { "epoch": 93.36283185840708, "grad_norm": 0.10303441435098648, "learning_rate": 9.91605342456734e-05, "loss": 0.0049, "step": 10550 }, { "epoch": 93.45132743362832, "grad_norm": 0.09612749516963959, "learning_rate": 9.915751440567795e-05, "loss": 0.0051, "step": 10560 }, { "epoch": 93.53982300884955, "grad_norm": 0.09685444831848145, "learning_rate": 9.915448918989066e-05, "loss": 0.0047, "step": 10570 }, { "epoch": 93.6283185840708, "grad_norm": 0.12692221999168396, "learning_rate": 9.915145859864232e-05, "loss": 0.0052, "step": 10580 }, { "epoch": 93.71681415929204, "grad_norm": 0.08728212118148804, "learning_rate": 9.914842263226437e-05, "loss": 0.005, "step": 10590 }, { "epoch": 93.80530973451327, "grad_norm": 0.1048746407032013, "learning_rate": 9.914538129108882e-05, "loss": 0.0058, "step": 10600 }, { "epoch": 93.89380530973452, "grad_norm": 0.12184499204158783, "learning_rate": 9.914233457544825e-05, "loss": 0.0052, "step": 10610 }, { "epoch": 93.98230088495575, "grad_norm": 0.12471144646406174, "learning_rate": 9.913928248567586e-05, "loss": 0.0054, "step": 10620 }, { "epoch": 94.070796460177, "grad_norm": 0.13259096443653107, "learning_rate": 9.913622502210542e-05, "loss": 0.0047, "step": 10630 }, { "epoch": 94.15929203539822, "grad_norm": 0.09159719944000244, "learning_rate": 9.913316218507128e-05, "loss": 0.0057, "step": 10640 }, { "epoch": 94.24778761061947, "grad_norm": 0.09752307087182999, "learning_rate": 9.91300939749084e-05, "loss": 0.0054, "step": 10650 }, { "epoch": 94.33628318584071, "grad_norm": 0.09806368499994278, "learning_rate": 9.91270203919523e-05, "loss": 0.0052, "step": 10660 }, { "epoch": 94.42477876106194, "grad_norm": 0.08339997380971909, "learning_rate": 9.912394143653912e-05, "loss": 0.0054, "step": 10670 }, { "epoch": 94.51327433628319, "grad_norm": 0.09601221233606339, "learning_rate": 9.912085710900555e-05, "loss": 0.0055, "step": 10680 }, { "epoch": 94.60176991150442, "grad_norm": 0.10113587230443954, "learning_rate": 9.911776740968892e-05, "loss": 0.0053, "step": 10690 }, { "epoch": 94.69026548672566, "grad_norm": 0.08769881725311279, "learning_rate": 9.911467233892709e-05, "loss": 0.0055, "step": 10700 }, { "epoch": 94.77876106194691, "grad_norm": 0.08661729842424393, "learning_rate": 9.911157189705853e-05, "loss": 0.0055, "step": 10710 }, { "epoch": 94.86725663716814, "grad_norm": 0.08152610063552856, "learning_rate": 9.910846608442229e-05, "loss": 0.0054, "step": 10720 }, { "epoch": 94.95575221238938, "grad_norm": 0.11567069590091705, "learning_rate": 9.910535490135805e-05, "loss": 0.0049, "step": 10730 }, { "epoch": 95.04424778761062, "grad_norm": 0.12678343057632446, "learning_rate": 9.910223834820603e-05, "loss": 0.0057, "step": 10740 }, { "epoch": 95.13274336283186, "grad_norm": 0.12780961394309998, "learning_rate": 9.909911642530703e-05, "loss": 0.0048, "step": 10750 }, { "epoch": 95.22123893805309, "grad_norm": 0.08420619368553162, "learning_rate": 9.909598913300249e-05, "loss": 0.0052, "step": 10760 }, { "epoch": 95.30973451327434, "grad_norm": 0.1518271267414093, "learning_rate": 9.909285647163438e-05, "loss": 0.0053, "step": 10770 }, { "epoch": 95.39823008849558, "grad_norm": 0.12385930120944977, "learning_rate": 9.908971844154531e-05, "loss": 0.0061, "step": 10780 }, { "epoch": 95.48672566371681, "grad_norm": 0.12428860366344452, "learning_rate": 9.908657504307843e-05, "loss": 0.005, "step": 10790 }, { "epoch": 95.57522123893806, "grad_norm": 0.09221476316452026, "learning_rate": 9.908342627657751e-05, "loss": 0.0052, "step": 10800 }, { "epoch": 95.66371681415929, "grad_norm": 0.11753755807876587, "learning_rate": 9.908027214238689e-05, "loss": 0.0057, "step": 10810 }, { "epoch": 95.75221238938053, "grad_norm": 0.11843214929103851, "learning_rate": 9.90771126408515e-05, "loss": 0.0053, "step": 10820 }, { "epoch": 95.84070796460178, "grad_norm": 0.109309621155262, "learning_rate": 9.907394777231685e-05, "loss": 0.0062, "step": 10830 }, { "epoch": 95.929203539823, "grad_norm": 0.10851550102233887, "learning_rate": 9.907077753712905e-05, "loss": 0.0055, "step": 10840 }, { "epoch": 96.01769911504425, "grad_norm": 0.07705362886190414, "learning_rate": 9.906760193563482e-05, "loss": 0.0054, "step": 10850 }, { "epoch": 96.10619469026548, "grad_norm": 0.08829394727945328, "learning_rate": 9.906442096818139e-05, "loss": 0.0054, "step": 10860 }, { "epoch": 96.19469026548673, "grad_norm": 0.09787177294492722, "learning_rate": 9.906123463511665e-05, "loss": 0.0058, "step": 10870 }, { "epoch": 96.28318584070796, "grad_norm": 0.091305211186409, "learning_rate": 9.905804293678907e-05, "loss": 0.005, "step": 10880 }, { "epoch": 96.3716814159292, "grad_norm": 0.08605948835611343, "learning_rate": 9.905484587354766e-05, "loss": 0.0051, "step": 10890 }, { "epoch": 96.46017699115045, "grad_norm": 0.08073023706674576, "learning_rate": 9.905164344574205e-05, "loss": 0.0053, "step": 10900 }, { "epoch": 96.54867256637168, "grad_norm": 0.10548191517591476, "learning_rate": 9.904843565372248e-05, "loss": 0.0051, "step": 10910 }, { "epoch": 96.63716814159292, "grad_norm": 0.11649339646100998, "learning_rate": 9.904522249783972e-05, "loss": 0.0055, "step": 10920 }, { "epoch": 96.72566371681415, "grad_norm": 0.12158945947885513, "learning_rate": 9.904200397844517e-05, "loss": 0.0057, "step": 10930 }, { "epoch": 96.8141592920354, "grad_norm": 0.13256609439849854, "learning_rate": 9.903878009589078e-05, "loss": 0.0062, "step": 10940 }, { "epoch": 96.90265486725664, "grad_norm": 0.07447865605354309, "learning_rate": 9.903555085052915e-05, "loss": 0.0057, "step": 10950 }, { "epoch": 96.99115044247787, "grad_norm": 0.10352752357721329, "learning_rate": 9.903231624271338e-05, "loss": 0.0061, "step": 10960 }, { "epoch": 97.07964601769912, "grad_norm": 0.09371993690729141, "learning_rate": 9.902907627279724e-05, "loss": 0.0061, "step": 10970 }, { "epoch": 97.16814159292035, "grad_norm": 0.12164576351642609, "learning_rate": 9.902583094113504e-05, "loss": 0.0049, "step": 10980 }, { "epoch": 97.2566371681416, "grad_norm": 0.09476112574338913, "learning_rate": 9.902258024808168e-05, "loss": 0.0052, "step": 10990 }, { "epoch": 97.34513274336283, "grad_norm": 0.06818000972270966, "learning_rate": 9.901932419399264e-05, "loss": 0.0048, "step": 11000 }, { "epoch": 97.43362831858407, "grad_norm": 0.08018700033426285, "learning_rate": 9.9016062779224e-05, "loss": 0.0048, "step": 11010 }, { "epoch": 97.52212389380531, "grad_norm": 0.07797984033823013, "learning_rate": 9.901279600413242e-05, "loss": 0.005, "step": 11020 }, { "epoch": 97.61061946902655, "grad_norm": 0.0751151368021965, "learning_rate": 9.900952386907518e-05, "loss": 0.0057, "step": 11030 }, { "epoch": 97.69911504424779, "grad_norm": 0.0867956131696701, "learning_rate": 9.90062463744101e-05, "loss": 0.0049, "step": 11040 }, { "epoch": 97.78761061946902, "grad_norm": 0.07608005404472351, "learning_rate": 9.900296352049558e-05, "loss": 0.0054, "step": 11050 }, { "epoch": 97.87610619469027, "grad_norm": 0.11258876323699951, "learning_rate": 9.899967530769065e-05, "loss": 0.0052, "step": 11060 }, { "epoch": 97.96460176991151, "grad_norm": 0.0988629162311554, "learning_rate": 9.899638173635489e-05, "loss": 0.0051, "step": 11070 }, { "epoch": 98.05309734513274, "grad_norm": 0.1009017750620842, "learning_rate": 9.899308280684849e-05, "loss": 0.0055, "step": 11080 }, { "epoch": 98.14159292035399, "grad_norm": 0.13422739505767822, "learning_rate": 9.898977851953222e-05, "loss": 0.0055, "step": 11090 }, { "epoch": 98.23008849557522, "grad_norm": 0.10034146904945374, "learning_rate": 9.898646887476741e-05, "loss": 0.0051, "step": 11100 }, { "epoch": 98.31858407079646, "grad_norm": 0.08099722862243652, "learning_rate": 9.898315387291603e-05, "loss": 0.0056, "step": 11110 }, { "epoch": 98.40707964601769, "grad_norm": 0.1098143681883812, "learning_rate": 9.89798335143406e-05, "loss": 0.0055, "step": 11120 }, { "epoch": 98.49557522123894, "grad_norm": 0.12781833112239838, "learning_rate": 9.897650779940419e-05, "loss": 0.006, "step": 11130 }, { "epoch": 98.58407079646018, "grad_norm": 0.12174452096223831, "learning_rate": 9.897317672847054e-05, "loss": 0.0055, "step": 11140 }, { "epoch": 98.67256637168141, "grad_norm": 0.10445994138717651, "learning_rate": 9.89698403019039e-05, "loss": 0.006, "step": 11150 }, { "epoch": 98.76106194690266, "grad_norm": 0.11261439323425293, "learning_rate": 9.896649852006917e-05, "loss": 0.0053, "step": 11160 }, { "epoch": 98.84955752212389, "grad_norm": 0.10478182137012482, "learning_rate": 9.896315138333177e-05, "loss": 0.0057, "step": 11170 }, { "epoch": 98.93805309734513, "grad_norm": 0.08884542435407639, "learning_rate": 9.895979889205774e-05, "loss": 0.005, "step": 11180 }, { "epoch": 99.02654867256638, "grad_norm": 0.12726913392543793, "learning_rate": 9.895644104661372e-05, "loss": 0.0051, "step": 11190 }, { "epoch": 99.11504424778761, "grad_norm": 0.06636649370193481, "learning_rate": 9.895307784736691e-05, "loss": 0.005, "step": 11200 }, { "epoch": 99.20353982300885, "grad_norm": 0.0950116440653801, "learning_rate": 9.894970929468512e-05, "loss": 0.005, "step": 11210 }, { "epoch": 99.29203539823008, "grad_norm": 0.09199380874633789, "learning_rate": 9.89463353889367e-05, "loss": 0.0048, "step": 11220 }, { "epoch": 99.38053097345133, "grad_norm": 0.09105435013771057, "learning_rate": 9.894295613049065e-05, "loss": 0.0047, "step": 11230 }, { "epoch": 99.46902654867256, "grad_norm": 0.11524466425180435, "learning_rate": 9.893957151971649e-05, "loss": 0.005, "step": 11240 }, { "epoch": 99.5575221238938, "grad_norm": 0.09950827062129974, "learning_rate": 9.893618155698436e-05, "loss": 0.0046, "step": 11250 }, { "epoch": 99.64601769911505, "grad_norm": 0.08869663625955582, "learning_rate": 9.8932786242665e-05, "loss": 0.0049, "step": 11260 }, { "epoch": 99.73451327433628, "grad_norm": 0.14323453605175018, "learning_rate": 9.89293855771297e-05, "loss": 0.0049, "step": 11270 }, { "epoch": 99.82300884955752, "grad_norm": 0.09014832228422165, "learning_rate": 9.892597956075036e-05, "loss": 0.0052, "step": 11280 }, { "epoch": 99.91150442477876, "grad_norm": 0.08269821852445602, "learning_rate": 9.892256819389947e-05, "loss": 0.005, "step": 11290 }, { "epoch": 100.0, "grad_norm": 0.08946595340967178, "learning_rate": 9.891915147695006e-05, "loss": 0.0057, "step": 11300 }, { "epoch": 100.08849557522124, "grad_norm": 0.0897507593035698, "learning_rate": 9.891572941027577e-05, "loss": 0.0056, "step": 11310 }, { "epoch": 100.17699115044248, "grad_norm": 0.11029284447431564, "learning_rate": 9.89123019942509e-05, "loss": 0.0054, "step": 11320 }, { "epoch": 100.26548672566372, "grad_norm": 0.10562395304441452, "learning_rate": 9.89088692292502e-05, "loss": 0.0051, "step": 11330 }, { "epoch": 100.35398230088495, "grad_norm": 0.09461378306150436, "learning_rate": 9.89054311156491e-05, "loss": 0.0056, "step": 11340 }, { "epoch": 100.4424778761062, "grad_norm": 0.12478769570589066, "learning_rate": 9.890198765382357e-05, "loss": 0.0055, "step": 11350 }, { "epoch": 100.53097345132744, "grad_norm": 0.07728267461061478, "learning_rate": 9.889853884415021e-05, "loss": 0.0049, "step": 11360 }, { "epoch": 100.61946902654867, "grad_norm": 0.11941128969192505, "learning_rate": 9.889508468700614e-05, "loss": 0.005, "step": 11370 }, { "epoch": 100.70796460176992, "grad_norm": 0.09899173676967621, "learning_rate": 9.889162518276915e-05, "loss": 0.0051, "step": 11380 }, { "epoch": 100.79646017699115, "grad_norm": 0.12666434049606323, "learning_rate": 9.888816033181752e-05, "loss": 0.0057, "step": 11390 }, { "epoch": 100.88495575221239, "grad_norm": 0.10586833208799362, "learning_rate": 9.888469013453018e-05, "loss": 0.006, "step": 11400 }, { "epoch": 100.97345132743362, "grad_norm": 0.11104032397270203, "learning_rate": 9.888121459128663e-05, "loss": 0.005, "step": 11410 }, { "epoch": 101.06194690265487, "grad_norm": 0.13559183478355408, "learning_rate": 9.887773370246693e-05, "loss": 0.0048, "step": 11420 }, { "epoch": 101.15044247787611, "grad_norm": 0.13161902129650116, "learning_rate": 9.887424746845177e-05, "loss": 0.0056, "step": 11430 }, { "epoch": 101.23893805309734, "grad_norm": 0.12117984890937805, "learning_rate": 9.887075588962239e-05, "loss": 0.0056, "step": 11440 }, { "epoch": 101.32743362831859, "grad_norm": 0.1159241572022438, "learning_rate": 9.88672589663606e-05, "loss": 0.0056, "step": 11450 }, { "epoch": 101.41592920353982, "grad_norm": 0.10363749414682388, "learning_rate": 9.886375669904886e-05, "loss": 0.0052, "step": 11460 }, { "epoch": 101.50442477876106, "grad_norm": 0.1449795812368393, "learning_rate": 9.886024908807014e-05, "loss": 0.0053, "step": 11470 }, { "epoch": 101.59292035398231, "grad_norm": 0.10984648764133453, "learning_rate": 9.885673613380806e-05, "loss": 0.0055, "step": 11480 }, { "epoch": 101.68141592920354, "grad_norm": 0.10604984313249588, "learning_rate": 9.885321783664676e-05, "loss": 0.0054, "step": 11490 }, { "epoch": 101.76991150442478, "grad_norm": 0.08602876216173172, "learning_rate": 9.884969419697101e-05, "loss": 0.0048, "step": 11500 }, { "epoch": 101.85840707964601, "grad_norm": 0.09405429661273956, "learning_rate": 9.884616521516614e-05, "loss": 0.0053, "step": 11510 }, { "epoch": 101.94690265486726, "grad_norm": 0.0969252809882164, "learning_rate": 9.88426308916181e-05, "loss": 0.006, "step": 11520 }, { "epoch": 102.03539823008849, "grad_norm": 0.15012577176094055, "learning_rate": 9.883909122671335e-05, "loss": 0.0051, "step": 11530 }, { "epoch": 102.12389380530973, "grad_norm": 0.09544498473405838, "learning_rate": 9.883554622083904e-05, "loss": 0.0062, "step": 11540 }, { "epoch": 102.21238938053098, "grad_norm": 0.13681212067604065, "learning_rate": 9.88319958743828e-05, "loss": 0.0055, "step": 11550 }, { "epoch": 102.30088495575221, "grad_norm": 0.09109208732843399, "learning_rate": 9.882844018773291e-05, "loss": 0.0055, "step": 11560 }, { "epoch": 102.38938053097345, "grad_norm": 0.11888206750154495, "learning_rate": 9.882487916127823e-05, "loss": 0.0054, "step": 11570 }, { "epoch": 102.47787610619469, "grad_norm": 0.10582287609577179, "learning_rate": 9.882131279540815e-05, "loss": 0.005, "step": 11580 }, { "epoch": 102.56637168141593, "grad_norm": 0.113967165350914, "learning_rate": 9.881774109051271e-05, "loss": 0.0052, "step": 11590 }, { "epoch": 102.65486725663717, "grad_norm": 0.10604744404554367, "learning_rate": 9.881416404698252e-05, "loss": 0.0053, "step": 11600 }, { "epoch": 102.7433628318584, "grad_norm": 0.08425791561603546, "learning_rate": 9.881058166520873e-05, "loss": 0.0055, "step": 11610 }, { "epoch": 102.83185840707965, "grad_norm": 0.0939178317785263, "learning_rate": 9.880699394558311e-05, "loss": 0.0054, "step": 11620 }, { "epoch": 102.92035398230088, "grad_norm": 0.08087138831615448, "learning_rate": 9.880340088849801e-05, "loss": 0.0054, "step": 11630 }, { "epoch": 103.00884955752213, "grad_norm": 0.08242262154817581, "learning_rate": 9.879980249434637e-05, "loss": 0.0054, "step": 11640 }, { "epoch": 103.09734513274336, "grad_norm": 0.10663232952356339, "learning_rate": 9.879619876352168e-05, "loss": 0.0049, "step": 11650 }, { "epoch": 103.1858407079646, "grad_norm": 0.0819629356265068, "learning_rate": 9.879258969641809e-05, "loss": 0.0051, "step": 11660 }, { "epoch": 103.27433628318585, "grad_norm": 0.07466502487659454, "learning_rate": 9.878897529343023e-05, "loss": 0.0046, "step": 11670 }, { "epoch": 103.36283185840708, "grad_norm": 0.07112641632556915, "learning_rate": 9.878535555495338e-05, "loss": 0.0057, "step": 11680 }, { "epoch": 103.45132743362832, "grad_norm": 0.09865057468414307, "learning_rate": 9.87817304813834e-05, "loss": 0.0049, "step": 11690 }, { "epoch": 103.53982300884955, "grad_norm": 0.0825251042842865, "learning_rate": 9.877810007311671e-05, "loss": 0.0055, "step": 11700 }, { "epoch": 103.6283185840708, "grad_norm": 0.09424883127212524, "learning_rate": 9.877446433055035e-05, "loss": 0.005, "step": 11710 }, { "epoch": 103.71681415929204, "grad_norm": 0.09876972436904907, "learning_rate": 9.877082325408191e-05, "loss": 0.005, "step": 11720 }, { "epoch": 103.80530973451327, "grad_norm": 0.10787871479988098, "learning_rate": 9.876717684410954e-05, "loss": 0.0054, "step": 11730 }, { "epoch": 103.89380530973452, "grad_norm": 0.08507556468248367, "learning_rate": 9.876352510103204e-05, "loss": 0.0047, "step": 11740 }, { "epoch": 103.98230088495575, "grad_norm": 0.09287779778242111, "learning_rate": 9.875986802524875e-05, "loss": 0.0051, "step": 11750 }, { "epoch": 104.070796460177, "grad_norm": 0.09439730644226074, "learning_rate": 9.87562056171596e-05, "loss": 0.0056, "step": 11760 }, { "epoch": 104.15929203539822, "grad_norm": 0.07966262102127075, "learning_rate": 9.875253787716511e-05, "loss": 0.0053, "step": 11770 }, { "epoch": 104.24778761061947, "grad_norm": 0.10832482576370239, "learning_rate": 9.874886480566637e-05, "loss": 0.0049, "step": 11780 }, { "epoch": 104.33628318584071, "grad_norm": 0.08569982647895813, "learning_rate": 9.874518640306507e-05, "loss": 0.005, "step": 11790 }, { "epoch": 104.42477876106194, "grad_norm": 0.07777139544487, "learning_rate": 9.874150266976347e-05, "loss": 0.0052, "step": 11800 }, { "epoch": 104.51327433628319, "grad_norm": 0.09986227005720139, "learning_rate": 9.873781360616443e-05, "loss": 0.0056, "step": 11810 }, { "epoch": 104.60176991150442, "grad_norm": 0.08663097023963928, "learning_rate": 9.873411921267137e-05, "loss": 0.0052, "step": 11820 }, { "epoch": 104.69026548672566, "grad_norm": 0.10029438883066177, "learning_rate": 9.873041948968829e-05, "loss": 0.0057, "step": 11830 }, { "epoch": 104.77876106194691, "grad_norm": 0.09535370767116547, "learning_rate": 9.872671443761981e-05, "loss": 0.0051, "step": 11840 }, { "epoch": 104.86725663716814, "grad_norm": 0.09979565441608429, "learning_rate": 9.872300405687109e-05, "loss": 0.0054, "step": 11850 }, { "epoch": 104.95575221238938, "grad_norm": 0.0894877091050148, "learning_rate": 9.871928834784792e-05, "loss": 0.0049, "step": 11860 }, { "epoch": 105.04424778761062, "grad_norm": 0.11560477316379547, "learning_rate": 9.871556731095661e-05, "loss": 0.0047, "step": 11870 }, { "epoch": 105.13274336283186, "grad_norm": 0.1062682718038559, "learning_rate": 9.871184094660411e-05, "loss": 0.0054, "step": 11880 }, { "epoch": 105.22123893805309, "grad_norm": 0.1089135929942131, "learning_rate": 9.870810925519791e-05, "loss": 0.0047, "step": 11890 }, { "epoch": 105.30973451327434, "grad_norm": 0.10968752950429916, "learning_rate": 9.870437223714612e-05, "loss": 0.005, "step": 11900 }, { "epoch": 105.39823008849558, "grad_norm": 0.12240723520517349, "learning_rate": 9.87006298928574e-05, "loss": 0.0052, "step": 11910 }, { "epoch": 105.48672566371681, "grad_norm": 0.0977071225643158, "learning_rate": 9.869688222274103e-05, "loss": 0.0056, "step": 11920 }, { "epoch": 105.57522123893806, "grad_norm": 0.10155139118432999, "learning_rate": 9.869312922720681e-05, "loss": 0.0055, "step": 11930 }, { "epoch": 105.66371681415929, "grad_norm": 0.08629415184259415, "learning_rate": 9.868937090666521e-05, "loss": 0.0051, "step": 11940 }, { "epoch": 105.75221238938053, "grad_norm": 0.11148345470428467, "learning_rate": 9.86856072615272e-05, "loss": 0.0049, "step": 11950 }, { "epoch": 105.84070796460178, "grad_norm": 0.10339879989624023, "learning_rate": 9.868183829220438e-05, "loss": 0.0055, "step": 11960 }, { "epoch": 105.929203539823, "grad_norm": 0.1088627427816391, "learning_rate": 9.867806399910893e-05, "loss": 0.0055, "step": 11970 }, { "epoch": 106.01769911504425, "grad_norm": 0.09286243468523026, "learning_rate": 9.867428438265356e-05, "loss": 0.0053, "step": 11980 }, { "epoch": 106.10619469026548, "grad_norm": 0.10349633544683456, "learning_rate": 9.867049944325165e-05, "loss": 0.0049, "step": 11990 }, { "epoch": 106.19469026548673, "grad_norm": 0.10118602961301804, "learning_rate": 9.86667091813171e-05, "loss": 0.0055, "step": 12000 }, { "epoch": 106.28318584070796, "grad_norm": 0.0856422632932663, "learning_rate": 9.866291359726438e-05, "loss": 0.0054, "step": 12010 }, { "epoch": 106.3716814159292, "grad_norm": 0.08684667944908142, "learning_rate": 9.865911269150861e-05, "loss": 0.0047, "step": 12020 }, { "epoch": 106.46017699115045, "grad_norm": 0.07805564254522324, "learning_rate": 9.865530646446544e-05, "loss": 0.0043, "step": 12030 }, { "epoch": 106.54867256637168, "grad_norm": 0.09507829695940018, "learning_rate": 9.86514949165511e-05, "loss": 0.0046, "step": 12040 }, { "epoch": 106.63716814159292, "grad_norm": 0.11667557060718536, "learning_rate": 9.864767804818243e-05, "loss": 0.0047, "step": 12050 }, { "epoch": 106.72566371681415, "grad_norm": 0.09120805561542511, "learning_rate": 9.86438558597768e-05, "loss": 0.0046, "step": 12060 }, { "epoch": 106.8141592920354, "grad_norm": 0.09021265059709549, "learning_rate": 9.864002835175225e-05, "loss": 0.0052, "step": 12070 }, { "epoch": 106.90265486725664, "grad_norm": 0.10057032853364944, "learning_rate": 9.863619552452734e-05, "loss": 0.0049, "step": 12080 }, { "epoch": 106.99115044247787, "grad_norm": 0.07605649530887604, "learning_rate": 9.863235737852119e-05, "loss": 0.0053, "step": 12090 }, { "epoch": 107.07964601769912, "grad_norm": 0.08129508793354034, "learning_rate": 9.862851391415356e-05, "loss": 0.0051, "step": 12100 }, { "epoch": 107.16814159292035, "grad_norm": 0.09423434734344482, "learning_rate": 9.862466513184477e-05, "loss": 0.0052, "step": 12110 }, { "epoch": 107.2566371681416, "grad_norm": 0.10215440392494202, "learning_rate": 9.86208110320157e-05, "loss": 0.0061, "step": 12120 }, { "epoch": 107.34513274336283, "grad_norm": 0.10903831571340561, "learning_rate": 9.861695161508784e-05, "loss": 0.0054, "step": 12130 }, { "epoch": 107.43362831858407, "grad_norm": 0.10834818333387375, "learning_rate": 9.861308688148324e-05, "loss": 0.0057, "step": 12140 }, { "epoch": 107.52212389380531, "grad_norm": 0.07090001553297043, "learning_rate": 9.860921683162455e-05, "loss": 0.0053, "step": 12150 }, { "epoch": 107.61061946902655, "grad_norm": 0.13069690763950348, "learning_rate": 9.860534146593499e-05, "loss": 0.0052, "step": 12160 }, { "epoch": 107.69911504424779, "grad_norm": 0.10894755274057388, "learning_rate": 9.860146078483836e-05, "loss": 0.0048, "step": 12170 }, { "epoch": 107.78761061946902, "grad_norm": 0.1547512412071228, "learning_rate": 9.859757478875905e-05, "loss": 0.0058, "step": 12180 }, { "epoch": 107.87610619469027, "grad_norm": 0.0878150537610054, "learning_rate": 9.859368347812204e-05, "loss": 0.0052, "step": 12190 }, { "epoch": 107.96460176991151, "grad_norm": 0.08880646526813507, "learning_rate": 9.858978685335285e-05, "loss": 0.0046, "step": 12200 }, { "epoch": 108.05309734513274, "grad_norm": 0.09298834204673767, "learning_rate": 9.858588491487763e-05, "loss": 0.0051, "step": 12210 }, { "epoch": 108.14159292035399, "grad_norm": 0.07077731192111969, "learning_rate": 9.858197766312308e-05, "loss": 0.0048, "step": 12220 }, { "epoch": 108.23008849557522, "grad_norm": 0.07037153095006943, "learning_rate": 9.857806509851649e-05, "loss": 0.0048, "step": 12230 }, { "epoch": 108.31858407079646, "grad_norm": 0.07296419143676758, "learning_rate": 9.857414722148574e-05, "loss": 0.0055, "step": 12240 }, { "epoch": 108.40707964601769, "grad_norm": 0.10267648845911026, "learning_rate": 9.857022403245928e-05, "loss": 0.0049, "step": 12250 }, { "epoch": 108.49557522123894, "grad_norm": 0.111343614757061, "learning_rate": 9.856629553186615e-05, "loss": 0.0049, "step": 12260 }, { "epoch": 108.58407079646018, "grad_norm": 0.11200346052646637, "learning_rate": 9.856236172013595e-05, "loss": 0.0052, "step": 12270 }, { "epoch": 108.67256637168141, "grad_norm": 0.15367485582828522, "learning_rate": 9.85584225976989e-05, "loss": 0.0045, "step": 12280 }, { "epoch": 108.76106194690266, "grad_norm": 0.07386550307273865, "learning_rate": 9.855447816498575e-05, "loss": 0.0047, "step": 12290 }, { "epoch": 108.84955752212389, "grad_norm": 0.11371751874685287, "learning_rate": 9.855052842242787e-05, "loss": 0.0045, "step": 12300 }, { "epoch": 108.93805309734513, "grad_norm": 0.08958017826080322, "learning_rate": 9.85465733704572e-05, "loss": 0.005, "step": 12310 }, { "epoch": 109.02654867256638, "grad_norm": 0.08157729357481003, "learning_rate": 9.854261300950624e-05, "loss": 0.0047, "step": 12320 }, { "epoch": 109.11504424778761, "grad_norm": 0.07012051343917847, "learning_rate": 9.853864734000813e-05, "loss": 0.0051, "step": 12330 }, { "epoch": 109.20353982300885, "grad_norm": 0.09253772348165512, "learning_rate": 9.85346763623965e-05, "loss": 0.005, "step": 12340 }, { "epoch": 109.29203539823008, "grad_norm": 0.08400961011648178, "learning_rate": 9.853070007710564e-05, "loss": 0.0051, "step": 12350 }, { "epoch": 109.38053097345133, "grad_norm": 0.1163819208741188, "learning_rate": 9.85267184845704e-05, "loss": 0.0048, "step": 12360 }, { "epoch": 109.46902654867256, "grad_norm": 0.0943228155374527, "learning_rate": 9.852273158522616e-05, "loss": 0.0049, "step": 12370 }, { "epoch": 109.5575221238938, "grad_norm": 0.10741299390792847, "learning_rate": 9.851873937950896e-05, "loss": 0.0052, "step": 12380 }, { "epoch": 109.64601769911505, "grad_norm": 0.0818244218826294, "learning_rate": 9.851474186785537e-05, "loss": 0.0052, "step": 12390 }, { "epoch": 109.73451327433628, "grad_norm": 0.08463114500045776, "learning_rate": 9.851073905070254e-05, "loss": 0.0045, "step": 12400 }, { "epoch": 109.82300884955752, "grad_norm": 0.11294638365507126, "learning_rate": 9.850673092848824e-05, "loss": 0.0053, "step": 12410 }, { "epoch": 109.91150442477876, "grad_norm": 0.10436401516199112, "learning_rate": 9.850271750165077e-05, "loss": 0.0053, "step": 12420 }, { "epoch": 110.0, "grad_norm": 0.09833112359046936, "learning_rate": 9.849869877062902e-05, "loss": 0.0049, "step": 12430 }, { "epoch": 110.08849557522124, "grad_norm": 0.13551482558250427, "learning_rate": 9.849467473586252e-05, "loss": 0.0051, "step": 12440 }, { "epoch": 110.17699115044248, "grad_norm": 0.12051155418157578, "learning_rate": 9.849064539779127e-05, "loss": 0.0055, "step": 12450 }, { "epoch": 110.26548672566372, "grad_norm": 0.10150273889303207, "learning_rate": 9.848661075685594e-05, "loss": 0.0054, "step": 12460 }, { "epoch": 110.35398230088495, "grad_norm": 0.10090801864862442, "learning_rate": 9.848257081349778e-05, "loss": 0.0056, "step": 12470 }, { "epoch": 110.4424778761062, "grad_norm": 0.08186084032058716, "learning_rate": 9.847852556815856e-05, "loss": 0.0049, "step": 12480 }, { "epoch": 110.53097345132744, "grad_norm": 0.08123543113470078, "learning_rate": 9.847447502128067e-05, "loss": 0.0048, "step": 12490 }, { "epoch": 110.61946902654867, "grad_norm": 0.07736730575561523, "learning_rate": 9.847041917330708e-05, "loss": 0.0045, "step": 12500 }, { "epoch": 110.70796460176992, "grad_norm": 0.08575046807527542, "learning_rate": 9.846635802468132e-05, "loss": 0.0046, "step": 12510 }, { "epoch": 110.79646017699115, "grad_norm": 0.07963137328624725, "learning_rate": 9.84622915758475e-05, "loss": 0.0046, "step": 12520 }, { "epoch": 110.88495575221239, "grad_norm": 0.09060648083686829, "learning_rate": 9.845821982725034e-05, "loss": 0.0045, "step": 12530 }, { "epoch": 110.97345132743362, "grad_norm": 0.10538797825574875, "learning_rate": 9.845414277933514e-05, "loss": 0.0054, "step": 12540 }, { "epoch": 111.06194690265487, "grad_norm": 0.13742206990718842, "learning_rate": 9.845006043254771e-05, "loss": 0.0049, "step": 12550 }, { "epoch": 111.15044247787611, "grad_norm": 0.1161801740527153, "learning_rate": 9.844597278733451e-05, "loss": 0.0045, "step": 12560 }, { "epoch": 111.23893805309734, "grad_norm": 0.10717505216598511, "learning_rate": 9.844187984414259e-05, "loss": 0.0054, "step": 12570 }, { "epoch": 111.32743362831859, "grad_norm": 0.14164257049560547, "learning_rate": 9.84377816034195e-05, "loss": 0.0052, "step": 12580 }, { "epoch": 111.41592920353982, "grad_norm": 0.07494541257619858, "learning_rate": 9.843367806561345e-05, "loss": 0.0056, "step": 12590 }, { "epoch": 111.50442477876106, "grad_norm": 0.11627686768770218, "learning_rate": 9.842956923117317e-05, "loss": 0.0046, "step": 12600 }, { "epoch": 111.59292035398231, "grad_norm": 0.09290684014558792, "learning_rate": 9.842545510054802e-05, "loss": 0.005, "step": 12610 }, { "epoch": 111.68141592920354, "grad_norm": 0.08084304630756378, "learning_rate": 9.842133567418792e-05, "loss": 0.0048, "step": 12620 }, { "epoch": 111.76991150442478, "grad_norm": 0.10311734676361084, "learning_rate": 9.841721095254333e-05, "loss": 0.0049, "step": 12630 }, { "epoch": 111.85840707964601, "grad_norm": 0.11525673419237137, "learning_rate": 9.841308093606537e-05, "loss": 0.0052, "step": 12640 }, { "epoch": 111.94690265486726, "grad_norm": 0.09082310646772385, "learning_rate": 9.840894562520565e-05, "loss": 0.0047, "step": 12650 }, { "epoch": 112.03539823008849, "grad_norm": 0.08389148116111755, "learning_rate": 9.840480502041642e-05, "loss": 0.0046, "step": 12660 }, { "epoch": 112.12389380530973, "grad_norm": 0.08969353139400482, "learning_rate": 9.840065912215049e-05, "loss": 0.0048, "step": 12670 }, { "epoch": 112.21238938053098, "grad_norm": 0.10022334009408951, "learning_rate": 9.839650793086124e-05, "loss": 0.0051, "step": 12680 }, { "epoch": 112.30088495575221, "grad_norm": 0.080583356320858, "learning_rate": 9.839235144700265e-05, "loss": 0.0043, "step": 12690 }, { "epoch": 112.38938053097345, "grad_norm": 0.11581997573375702, "learning_rate": 9.838818967102926e-05, "loss": 0.0049, "step": 12700 }, { "epoch": 112.47787610619469, "grad_norm": 0.12856446206569672, "learning_rate": 9.83840226033962e-05, "loss": 0.0046, "step": 12710 }, { "epoch": 112.56637168141593, "grad_norm": 0.09050169587135315, "learning_rate": 9.837985024455918e-05, "loss": 0.0054, "step": 12720 }, { "epoch": 112.65486725663717, "grad_norm": 0.11309484392404556, "learning_rate": 9.837567259497447e-05, "loss": 0.0053, "step": 12730 }, { "epoch": 112.7433628318584, "grad_norm": 0.07470899075269699, "learning_rate": 9.837148965509894e-05, "loss": 0.0054, "step": 12740 }, { "epoch": 112.83185840707965, "grad_norm": 0.10359036177396774, "learning_rate": 9.836730142539001e-05, "loss": 0.0052, "step": 12750 }, { "epoch": 112.92035398230088, "grad_norm": 0.07076502591371536, "learning_rate": 9.836310790630574e-05, "loss": 0.0049, "step": 12760 }, { "epoch": 113.00884955752213, "grad_norm": 0.06969419866800308, "learning_rate": 9.83589090983047e-05, "loss": 0.006, "step": 12770 }, { "epoch": 113.09734513274336, "grad_norm": 0.08438348770141602, "learning_rate": 9.835470500184605e-05, "loss": 0.0054, "step": 12780 }, { "epoch": 113.1858407079646, "grad_norm": 0.0791611298918724, "learning_rate": 9.835049561738957e-05, "loss": 0.0046, "step": 12790 }, { "epoch": 113.27433628318585, "grad_norm": 0.08097337186336517, "learning_rate": 9.834628094539558e-05, "loss": 0.0045, "step": 12800 }, { "epoch": 113.36283185840708, "grad_norm": 0.09380268305540085, "learning_rate": 9.834206098632499e-05, "loss": 0.005, "step": 12810 }, { "epoch": 113.45132743362832, "grad_norm": 0.09866824001073837, "learning_rate": 9.833783574063931e-05, "loss": 0.0052, "step": 12820 }, { "epoch": 113.53982300884955, "grad_norm": 0.09071388840675354, "learning_rate": 9.833360520880058e-05, "loss": 0.0056, "step": 12830 }, { "epoch": 113.6283185840708, "grad_norm": 0.09898029267787933, "learning_rate": 9.832936939127144e-05, "loss": 0.0053, "step": 12840 }, { "epoch": 113.71681415929204, "grad_norm": 0.11208298057317734, "learning_rate": 9.832512828851515e-05, "loss": 0.0052, "step": 12850 }, { "epoch": 113.80530973451327, "grad_norm": 0.10317390412092209, "learning_rate": 9.832088190099546e-05, "loss": 0.0054, "step": 12860 }, { "epoch": 113.89380530973452, "grad_norm": 0.12791317701339722, "learning_rate": 9.831663022917679e-05, "loss": 0.0047, "step": 12870 }, { "epoch": 113.98230088495575, "grad_norm": 0.09146826714277267, "learning_rate": 9.831237327352407e-05, "loss": 0.0047, "step": 12880 }, { "epoch": 114.070796460177, "grad_norm": 0.07679908722639084, "learning_rate": 9.830811103450286e-05, "loss": 0.0049, "step": 12890 }, { "epoch": 114.15929203539822, "grad_norm": 0.10419469326734543, "learning_rate": 9.830384351257924e-05, "loss": 0.0055, "step": 12900 }, { "epoch": 114.24778761061947, "grad_norm": 0.09728535264730453, "learning_rate": 9.829957070821993e-05, "loss": 0.0049, "step": 12910 }, { "epoch": 114.33628318584071, "grad_norm": 0.1001863181591034, "learning_rate": 9.829529262189218e-05, "loss": 0.005, "step": 12920 }, { "epoch": 114.42477876106194, "grad_norm": 0.11207375675439835, "learning_rate": 9.829100925406385e-05, "loss": 0.0051, "step": 12930 }, { "epoch": 114.51327433628319, "grad_norm": 0.10799598693847656, "learning_rate": 9.828672060520333e-05, "loss": 0.0051, "step": 12940 }, { "epoch": 114.60176991150442, "grad_norm": 0.09029443562030792, "learning_rate": 9.828242667577966e-05, "loss": 0.0051, "step": 12950 }, { "epoch": 114.69026548672566, "grad_norm": 0.09480349719524384, "learning_rate": 9.82781274662624e-05, "loss": 0.0051, "step": 12960 }, { "epoch": 114.77876106194691, "grad_norm": 0.08139754086732864, "learning_rate": 9.82738229771217e-05, "loss": 0.0049, "step": 12970 }, { "epoch": 114.86725663716814, "grad_norm": 0.10518685728311539, "learning_rate": 9.826951320882829e-05, "loss": 0.005, "step": 12980 }, { "epoch": 114.95575221238938, "grad_norm": 0.07408828288316727, "learning_rate": 9.826519816185351e-05, "loss": 0.0051, "step": 12990 }, { "epoch": 115.04424778761062, "grad_norm": 0.06902866065502167, "learning_rate": 9.826087783666921e-05, "loss": 0.0047, "step": 13000 }, { "epoch": 115.13274336283186, "grad_norm": 0.09459097683429718, "learning_rate": 9.825655223374787e-05, "loss": 0.0048, "step": 13010 }, { "epoch": 115.22123893805309, "grad_norm": 0.08853932470083237, "learning_rate": 9.825222135356253e-05, "loss": 0.0045, "step": 13020 }, { "epoch": 115.30973451327434, "grad_norm": 0.06905514001846313, "learning_rate": 9.82478851965868e-05, "loss": 0.0045, "step": 13030 }, { "epoch": 115.39823008849558, "grad_norm": 0.06915145367383957, "learning_rate": 9.82435437632949e-05, "loss": 0.0049, "step": 13040 }, { "epoch": 115.48672566371681, "grad_norm": 0.10482807457447052, "learning_rate": 9.823919705416158e-05, "loss": 0.0047, "step": 13050 }, { "epoch": 115.57522123893806, "grad_norm": 0.09655066579580307, "learning_rate": 9.82348450696622e-05, "loss": 0.0047, "step": 13060 }, { "epoch": 115.66371681415929, "grad_norm": 0.12596549093723297, "learning_rate": 9.823048781027268e-05, "loss": 0.005, "step": 13070 }, { "epoch": 115.75221238938053, "grad_norm": 0.10428296029567719, "learning_rate": 9.822612527646953e-05, "loss": 0.0046, "step": 13080 }, { "epoch": 115.84070796460178, "grad_norm": 0.0876854658126831, "learning_rate": 9.822175746872984e-05, "loss": 0.0048, "step": 13090 }, { "epoch": 115.929203539823, "grad_norm": 0.1134437546133995, "learning_rate": 9.821738438753123e-05, "loss": 0.0045, "step": 13100 }, { "epoch": 116.01769911504425, "grad_norm": 0.11000191420316696, "learning_rate": 9.821300603335196e-05, "loss": 0.0054, "step": 13110 }, { "epoch": 116.10619469026548, "grad_norm": 0.09956730902194977, "learning_rate": 9.820862240667085e-05, "loss": 0.0047, "step": 13120 }, { "epoch": 116.19469026548673, "grad_norm": 0.10938471555709839, "learning_rate": 9.820423350796726e-05, "loss": 0.0053, "step": 13130 }, { "epoch": 116.28318584070796, "grad_norm": 0.15229512751102448, "learning_rate": 9.819983933772118e-05, "loss": 0.0054, "step": 13140 }, { "epoch": 116.3716814159292, "grad_norm": 0.14460870623588562, "learning_rate": 9.819543989641314e-05, "loss": 0.0045, "step": 13150 }, { "epoch": 116.46017699115045, "grad_norm": 0.09999319165945053, "learning_rate": 9.819103518452423e-05, "loss": 0.0048, "step": 13160 }, { "epoch": 116.54867256637168, "grad_norm": 0.09677735716104507, "learning_rate": 9.818662520253618e-05, "loss": 0.0051, "step": 13170 }, { "epoch": 116.63716814159292, "grad_norm": 0.09551084786653519, "learning_rate": 9.818220995093126e-05, "loss": 0.0048, "step": 13180 }, { "epoch": 116.72566371681415, "grad_norm": 0.1081274151802063, "learning_rate": 9.817778943019228e-05, "loss": 0.0052, "step": 13190 }, { "epoch": 116.8141592920354, "grad_norm": 0.09280046075582504, "learning_rate": 9.81733636408027e-05, "loss": 0.0043, "step": 13200 }, { "epoch": 116.90265486725664, "grad_norm": 0.09336479008197784, "learning_rate": 9.816893258324649e-05, "loss": 0.0048, "step": 13210 }, { "epoch": 116.99115044247787, "grad_norm": 0.08059985935688019, "learning_rate": 9.816449625800823e-05, "loss": 0.0049, "step": 13220 }, { "epoch": 117.07964601769912, "grad_norm": 0.08991356194019318, "learning_rate": 9.816005466557308e-05, "loss": 0.0047, "step": 13230 }, { "epoch": 117.16814159292035, "grad_norm": 0.09440635144710541, "learning_rate": 9.815560780642674e-05, "loss": 0.0047, "step": 13240 }, { "epoch": 117.2566371681416, "grad_norm": 0.11046204715967178, "learning_rate": 9.815115568105555e-05, "loss": 0.0049, "step": 13250 }, { "epoch": 117.34513274336283, "grad_norm": 0.13764484226703644, "learning_rate": 9.814669828994638e-05, "loss": 0.0048, "step": 13260 }, { "epoch": 117.43362831858407, "grad_norm": 0.10590597242116928, "learning_rate": 9.814223563358665e-05, "loss": 0.005, "step": 13270 }, { "epoch": 117.52212389380531, "grad_norm": 0.10310439020395279, "learning_rate": 9.813776771246443e-05, "loss": 0.0055, "step": 13280 }, { "epoch": 117.61061946902655, "grad_norm": 0.09680131077766418, "learning_rate": 9.813329452706829e-05, "loss": 0.005, "step": 13290 }, { "epoch": 117.69911504424779, "grad_norm": 0.096352718770504, "learning_rate": 9.812881607788744e-05, "loss": 0.0044, "step": 13300 }, { "epoch": 117.78761061946902, "grad_norm": 0.10829818248748779, "learning_rate": 9.812433236541163e-05, "loss": 0.0045, "step": 13310 }, { "epoch": 117.87610619469027, "grad_norm": 0.08044736087322235, "learning_rate": 9.811984339013116e-05, "loss": 0.005, "step": 13320 }, { "epoch": 117.96460176991151, "grad_norm": 0.09623504430055618, "learning_rate": 9.811534915253698e-05, "loss": 0.0054, "step": 13330 }, { "epoch": 118.05309734513274, "grad_norm": 0.08268078416585922, "learning_rate": 9.811084965312056e-05, "loss": 0.0045, "step": 13340 }, { "epoch": 118.14159292035399, "grad_norm": 0.09187091886997223, "learning_rate": 9.810634489237396e-05, "loss": 0.005, "step": 13350 }, { "epoch": 118.23008849557522, "grad_norm": 0.07927794754505157, "learning_rate": 9.81018348707898e-05, "loss": 0.0051, "step": 13360 }, { "epoch": 118.31858407079646, "grad_norm": 0.09707525372505188, "learning_rate": 9.809731958886131e-05, "loss": 0.0048, "step": 13370 }, { "epoch": 118.40707964601769, "grad_norm": 0.08279309421777725, "learning_rate": 9.809279904708224e-05, "loss": 0.0053, "step": 13380 }, { "epoch": 118.49557522123894, "grad_norm": 0.07647504657506943, "learning_rate": 9.808827324594699e-05, "loss": 0.0047, "step": 13390 }, { "epoch": 118.58407079646018, "grad_norm": 0.10878463089466095, "learning_rate": 9.808374218595046e-05, "loss": 0.0053, "step": 13400 }, { "epoch": 118.67256637168141, "grad_norm": 0.0870056077837944, "learning_rate": 9.80792058675882e-05, "loss": 0.0046, "step": 13410 }, { "epoch": 118.76106194690266, "grad_norm": 0.10078975558280945, "learning_rate": 9.807466429135627e-05, "loss": 0.0045, "step": 13420 }, { "epoch": 118.84955752212389, "grad_norm": 0.10372629761695862, "learning_rate": 9.807011745775132e-05, "loss": 0.0048, "step": 13430 }, { "epoch": 118.93805309734513, "grad_norm": 0.08950068801641464, "learning_rate": 9.806556536727061e-05, "loss": 0.0048, "step": 13440 }, { "epoch": 119.02654867256638, "grad_norm": 0.0969741940498352, "learning_rate": 9.806100802041193e-05, "loss": 0.0049, "step": 13450 }, { "epoch": 119.11504424778761, "grad_norm": 0.10920450091362, "learning_rate": 9.805644541767368e-05, "loss": 0.0042, "step": 13460 }, { "epoch": 119.20353982300885, "grad_norm": 0.08094905316829681, "learning_rate": 9.805187755955478e-05, "loss": 0.005, "step": 13470 }, { "epoch": 119.29203539823008, "grad_norm": 0.11694866418838501, "learning_rate": 9.804730444655483e-05, "loss": 0.0048, "step": 13480 }, { "epoch": 119.38053097345133, "grad_norm": 0.09283019602298737, "learning_rate": 9.804272607917388e-05, "loss": 0.0045, "step": 13490 }, { "epoch": 119.46902654867256, "grad_norm": 0.09850625693798065, "learning_rate": 9.803814245791265e-05, "loss": 0.0048, "step": 13500 }, { "epoch": 119.5575221238938, "grad_norm": 0.08881653100252151, "learning_rate": 9.803355358327239e-05, "loss": 0.0048, "step": 13510 }, { "epoch": 119.64601769911505, "grad_norm": 0.08178632706403732, "learning_rate": 9.802895945575492e-05, "loss": 0.0049, "step": 13520 }, { "epoch": 119.73451327433628, "grad_norm": 0.09740108996629715, "learning_rate": 9.802436007586266e-05, "loss": 0.0051, "step": 13530 }, { "epoch": 119.82300884955752, "grad_norm": 0.073703832924366, "learning_rate": 9.801975544409858e-05, "loss": 0.0046, "step": 13540 }, { "epoch": 119.91150442477876, "grad_norm": 0.09771501272916794, "learning_rate": 9.801514556096625e-05, "loss": 0.0046, "step": 13550 }, { "epoch": 120.0, "grad_norm": 0.0936935693025589, "learning_rate": 9.801053042696977e-05, "loss": 0.0046, "step": 13560 }, { "epoch": 120.08849557522124, "grad_norm": 0.06842462718486786, "learning_rate": 9.800591004261388e-05, "loss": 0.0046, "step": 13570 }, { "epoch": 120.17699115044248, "grad_norm": 0.093881756067276, "learning_rate": 9.800128440840385e-05, "loss": 0.005, "step": 13580 }, { "epoch": 120.26548672566372, "grad_norm": 0.09438110888004303, "learning_rate": 9.799665352484552e-05, "loss": 0.0052, "step": 13590 }, { "epoch": 120.35398230088495, "grad_norm": 0.09731823205947876, "learning_rate": 9.799201739244532e-05, "loss": 0.0044, "step": 13600 }, { "epoch": 120.4424778761062, "grad_norm": 0.11149504780769348, "learning_rate": 9.798737601171025e-05, "loss": 0.0047, "step": 13610 }, { "epoch": 120.53097345132744, "grad_norm": 0.10170624405145645, "learning_rate": 9.79827293831479e-05, "loss": 0.0048, "step": 13620 }, { "epoch": 120.61946902654867, "grad_norm": 0.1379348337650299, "learning_rate": 9.797807750726638e-05, "loss": 0.0048, "step": 13630 }, { "epoch": 120.70796460176992, "grad_norm": 0.14172613620758057, "learning_rate": 9.797342038457446e-05, "loss": 0.0048, "step": 13640 }, { "epoch": 120.79646017699115, "grad_norm": 0.10578557848930359, "learning_rate": 9.796875801558141e-05, "loss": 0.0051, "step": 13650 }, { "epoch": 120.88495575221239, "grad_norm": 0.10718676447868347, "learning_rate": 9.79640904007971e-05, "loss": 0.0049, "step": 13660 }, { "epoch": 120.97345132743362, "grad_norm": 0.09708257764577866, "learning_rate": 9.795941754073199e-05, "loss": 0.0051, "step": 13670 }, { "epoch": 121.06194690265487, "grad_norm": 0.07529598474502563, "learning_rate": 9.795473943589705e-05, "loss": 0.0044, "step": 13680 }, { "epoch": 121.15044247787611, "grad_norm": 0.08835124224424362, "learning_rate": 9.795005608680394e-05, "loss": 0.005, "step": 13690 }, { "epoch": 121.23893805309734, "grad_norm": 0.08687232434749603, "learning_rate": 9.794536749396477e-05, "loss": 0.0043, "step": 13700 }, { "epoch": 121.32743362831859, "grad_norm": 0.08317497372627258, "learning_rate": 9.79406736578923e-05, "loss": 0.0045, "step": 13710 }, { "epoch": 121.41592920353982, "grad_norm": 0.09224767982959747, "learning_rate": 9.793597457909984e-05, "loss": 0.0046, "step": 13720 }, { "epoch": 121.50442477876106, "grad_norm": 0.10657454282045364, "learning_rate": 9.793127025810127e-05, "loss": 0.0049, "step": 13730 }, { "epoch": 121.59292035398231, "grad_norm": 0.08375581353902817, "learning_rate": 9.792656069541104e-05, "loss": 0.0043, "step": 13740 }, { "epoch": 121.68141592920354, "grad_norm": 0.10253141820430756, "learning_rate": 9.79218458915442e-05, "loss": 0.0046, "step": 13750 }, { "epoch": 121.76991150442478, "grad_norm": 0.12858906388282776, "learning_rate": 9.791712584701634e-05, "loss": 0.0048, "step": 13760 }, { "epoch": 121.85840707964601, "grad_norm": 0.0959785133600235, "learning_rate": 9.791240056234364e-05, "loss": 0.0048, "step": 13770 }, { "epoch": 121.94690265486726, "grad_norm": 0.08991490304470062, "learning_rate": 9.790767003804283e-05, "loss": 0.0044, "step": 13780 }, { "epoch": 122.03539823008849, "grad_norm": 0.0762917622923851, "learning_rate": 9.790293427463126e-05, "loss": 0.0048, "step": 13790 }, { "epoch": 122.12389380530973, "grad_norm": 0.08979620039463043, "learning_rate": 9.789819327262684e-05, "loss": 0.0045, "step": 13800 }, { "epoch": 122.21238938053098, "grad_norm": 0.0948258563876152, "learning_rate": 9.7893447032548e-05, "loss": 0.0043, "step": 13810 }, { "epoch": 122.30088495575221, "grad_norm": 0.09578508883714676, "learning_rate": 9.78886955549138e-05, "loss": 0.0047, "step": 13820 }, { "epoch": 122.38938053097345, "grad_norm": 0.10254698991775513, "learning_rate": 9.788393884024387e-05, "loss": 0.0047, "step": 13830 }, { "epoch": 122.47787610619469, "grad_norm": 0.10174353420734406, "learning_rate": 9.787917688905836e-05, "loss": 0.0051, "step": 13840 }, { "epoch": 122.56637168141593, "grad_norm": 0.07291367650032043, "learning_rate": 9.787440970187807e-05, "loss": 0.0046, "step": 13850 }, { "epoch": 122.65486725663717, "grad_norm": 0.08008397370576859, "learning_rate": 9.786963727922429e-05, "loss": 0.0048, "step": 13860 }, { "epoch": 122.7433628318584, "grad_norm": 0.08041434735059738, "learning_rate": 9.786485962161897e-05, "loss": 0.0054, "step": 13870 }, { "epoch": 122.83185840707965, "grad_norm": 0.05671537294983864, "learning_rate": 9.786007672958455e-05, "loss": 0.0049, "step": 13880 }, { "epoch": 122.92035398230088, "grad_norm": 0.07369757443666458, "learning_rate": 9.78552886036441e-05, "loss": 0.0047, "step": 13890 }, { "epoch": 123.00884955752213, "grad_norm": 0.09922321140766144, "learning_rate": 9.785049524432124e-05, "loss": 0.0052, "step": 13900 }, { "epoch": 123.09734513274336, "grad_norm": 0.09758707135915756, "learning_rate": 9.784569665214016e-05, "loss": 0.0049, "step": 13910 }, { "epoch": 123.1858407079646, "grad_norm": 0.08706746995449066, "learning_rate": 9.784089282762563e-05, "loss": 0.0044, "step": 13920 }, { "epoch": 123.27433628318585, "grad_norm": 0.1051543727517128, "learning_rate": 9.7836083771303e-05, "loss": 0.005, "step": 13930 }, { "epoch": 123.36283185840708, "grad_norm": 0.11566181480884552, "learning_rate": 9.783126948369817e-05, "loss": 0.0043, "step": 13940 }, { "epoch": 123.45132743362832, "grad_norm": 0.09359104931354523, "learning_rate": 9.78264499653376e-05, "loss": 0.0049, "step": 13950 }, { "epoch": 123.53982300884955, "grad_norm": 0.07956608384847641, "learning_rate": 9.782162521674838e-05, "loss": 0.0053, "step": 13960 }, { "epoch": 123.6283185840708, "grad_norm": 0.06604462116956711, "learning_rate": 9.781679523845812e-05, "loss": 0.0046, "step": 13970 }, { "epoch": 123.71681415929204, "grad_norm": 0.10218661278486252, "learning_rate": 9.781196003099502e-05, "loss": 0.0043, "step": 13980 }, { "epoch": 123.80530973451327, "grad_norm": 0.09171789884567261, "learning_rate": 9.780711959488786e-05, "loss": 0.0044, "step": 13990 }, { "epoch": 123.89380530973452, "grad_norm": 0.10901418328285217, "learning_rate": 9.780227393066599e-05, "loss": 0.0043, "step": 14000 }, { "epoch": 123.98230088495575, "grad_norm": 0.09203574061393738, "learning_rate": 9.77974230388593e-05, "loss": 0.0047, "step": 14010 }, { "epoch": 124.070796460177, "grad_norm": 0.09468352794647217, "learning_rate": 9.779256691999829e-05, "loss": 0.0049, "step": 14020 }, { "epoch": 124.15929203539822, "grad_norm": 0.0860840380191803, "learning_rate": 9.778770557461403e-05, "loss": 0.0046, "step": 14030 }, { "epoch": 124.24778761061947, "grad_norm": 0.0906984806060791, "learning_rate": 9.778283900323812e-05, "loss": 0.0047, "step": 14040 }, { "epoch": 124.33628318584071, "grad_norm": 0.07782988995313644, "learning_rate": 9.777796720640277e-05, "loss": 0.0044, "step": 14050 }, { "epoch": 124.42477876106194, "grad_norm": 0.0823393315076828, "learning_rate": 9.777309018464078e-05, "loss": 0.0044, "step": 14060 }, { "epoch": 124.51327433628319, "grad_norm": 0.08426350355148315, "learning_rate": 9.776820793848547e-05, "loss": 0.0042, "step": 14070 }, { "epoch": 124.60176991150442, "grad_norm": 0.07872885465621948, "learning_rate": 9.776332046847075e-05, "loss": 0.0042, "step": 14080 }, { "epoch": 124.69026548672566, "grad_norm": 0.08927364647388458, "learning_rate": 9.775842777513111e-05, "loss": 0.0046, "step": 14090 }, { "epoch": 124.77876106194691, "grad_norm": 0.10542257875204086, "learning_rate": 9.775352985900163e-05, "loss": 0.0042, "step": 14100 }, { "epoch": 124.86725663716814, "grad_norm": 0.09474708139896393, "learning_rate": 9.774862672061791e-05, "loss": 0.0048, "step": 14110 }, { "epoch": 124.95575221238938, "grad_norm": 0.1077922135591507, "learning_rate": 9.774371836051616e-05, "loss": 0.0041, "step": 14120 }, { "epoch": 125.04424778761062, "grad_norm": 0.08847453445196152, "learning_rate": 9.773880477923315e-05, "loss": 0.0045, "step": 14130 }, { "epoch": 125.13274336283186, "grad_norm": 0.09930987656116486, "learning_rate": 9.773388597730623e-05, "loss": 0.0045, "step": 14140 }, { "epoch": 125.22123893805309, "grad_norm": 0.09038558602333069, "learning_rate": 9.77289619552733e-05, "loss": 0.0043, "step": 14150 }, { "epoch": 125.30973451327434, "grad_norm": 0.07441874593496323, "learning_rate": 9.772403271367285e-05, "loss": 0.0048, "step": 14160 }, { "epoch": 125.39823008849558, "grad_norm": 0.07121234387159348, "learning_rate": 9.771909825304396e-05, "loss": 0.0048, "step": 14170 }, { "epoch": 125.48672566371681, "grad_norm": 0.08724091202020645, "learning_rate": 9.771415857392619e-05, "loss": 0.0043, "step": 14180 }, { "epoch": 125.57522123893806, "grad_norm": 0.08686840534210205, "learning_rate": 9.770921367685978e-05, "loss": 0.0047, "step": 14190 }, { "epoch": 125.66371681415929, "grad_norm": 0.1247321218252182, "learning_rate": 9.770426356238551e-05, "loss": 0.005, "step": 14200 }, { "epoch": 125.75221238938053, "grad_norm": 0.09085719287395477, "learning_rate": 9.769930823104469e-05, "loss": 0.0051, "step": 14210 }, { "epoch": 125.84070796460178, "grad_norm": 0.09506575763225555, "learning_rate": 9.769434768337926e-05, "loss": 0.0047, "step": 14220 }, { "epoch": 125.929203539823, "grad_norm": 0.12057707458734512, "learning_rate": 9.768938191993164e-05, "loss": 0.0049, "step": 14230 }, { "epoch": 126.01769911504425, "grad_norm": 0.08686672151088715, "learning_rate": 9.768441094124494e-05, "loss": 0.0047, "step": 14240 }, { "epoch": 126.10619469026548, "grad_norm": 0.09228711575269699, "learning_rate": 9.767943474786275e-05, "loss": 0.0041, "step": 14250 }, { "epoch": 126.19469026548673, "grad_norm": 0.08965133875608444, "learning_rate": 9.767445334032923e-05, "loss": 0.005, "step": 14260 }, { "epoch": 126.28318584070796, "grad_norm": 0.11491426080465317, "learning_rate": 9.766946671918919e-05, "loss": 0.0048, "step": 14270 }, { "epoch": 126.3716814159292, "grad_norm": 0.10687656700611115, "learning_rate": 9.766447488498796e-05, "loss": 0.005, "step": 14280 }, { "epoch": 126.46017699115045, "grad_norm": 0.11001214385032654, "learning_rate": 9.765947783827139e-05, "loss": 0.0046, "step": 14290 }, { "epoch": 126.54867256637168, "grad_norm": 0.08782150596380234, "learning_rate": 9.765447557958599e-05, "loss": 0.0049, "step": 14300 }, { "epoch": 126.63716814159292, "grad_norm": 0.11031058430671692, "learning_rate": 9.764946810947879e-05, "loss": 0.0045, "step": 14310 }, { "epoch": 126.72566371681415, "grad_norm": 0.09057732671499252, "learning_rate": 9.764445542849738e-05, "loss": 0.0051, "step": 14320 }, { "epoch": 126.8141592920354, "grad_norm": 0.0869070515036583, "learning_rate": 9.763943753718998e-05, "loss": 0.0051, "step": 14330 }, { "epoch": 126.90265486725664, "grad_norm": 0.07889094948768616, "learning_rate": 9.76344144361053e-05, "loss": 0.0047, "step": 14340 }, { "epoch": 126.99115044247787, "grad_norm": 0.09695584326982498, "learning_rate": 9.762938612579269e-05, "loss": 0.0047, "step": 14350 }, { "epoch": 127.07964601769912, "grad_norm": 0.10670273005962372, "learning_rate": 9.762435260680202e-05, "loss": 0.0048, "step": 14360 }, { "epoch": 127.16814159292035, "grad_norm": 0.07563699781894684, "learning_rate": 9.761931387968373e-05, "loss": 0.0049, "step": 14370 }, { "epoch": 127.2566371681416, "grad_norm": 0.09555374085903168, "learning_rate": 9.76142699449889e-05, "loss": 0.0047, "step": 14380 }, { "epoch": 127.34513274336283, "grad_norm": 0.09421759098768234, "learning_rate": 9.760922080326908e-05, "loss": 0.0045, "step": 14390 }, { "epoch": 127.43362831858407, "grad_norm": 0.06916548311710358, "learning_rate": 9.760416645507644e-05, "loss": 0.0046, "step": 14400 }, { "epoch": 127.52212389380531, "grad_norm": 0.0753602534532547, "learning_rate": 9.759910690096375e-05, "loss": 0.0049, "step": 14410 }, { "epoch": 127.61061946902655, "grad_norm": 0.08479373157024384, "learning_rate": 9.759404214148429e-05, "loss": 0.0044, "step": 14420 }, { "epoch": 127.69911504424779, "grad_norm": 0.058024417608976364, "learning_rate": 9.758897217719191e-05, "loss": 0.0046, "step": 14430 }, { "epoch": 127.78761061946902, "grad_norm": 0.06608692556619644, "learning_rate": 9.758389700864113e-05, "loss": 0.005, "step": 14440 }, { "epoch": 127.87610619469027, "grad_norm": 0.06077607348561287, "learning_rate": 9.757881663638688e-05, "loss": 0.005, "step": 14450 }, { "epoch": 127.96460176991151, "grad_norm": 0.0939125120639801, "learning_rate": 9.757373106098478e-05, "loss": 0.0044, "step": 14460 }, { "epoch": 128.05309734513276, "grad_norm": 0.06060545891523361, "learning_rate": 9.756864028299097e-05, "loss": 0.0047, "step": 14470 }, { "epoch": 128.141592920354, "grad_norm": 0.09752415865659714, "learning_rate": 9.75635443029622e-05, "loss": 0.0046, "step": 14480 }, { "epoch": 128.23008849557522, "grad_norm": 0.09237126260995865, "learning_rate": 9.755844312145572e-05, "loss": 0.005, "step": 14490 }, { "epoch": 128.31858407079645, "grad_norm": 0.09960038959980011, "learning_rate": 9.755333673902941e-05, "loss": 0.005, "step": 14500 }, { "epoch": 128.4070796460177, "grad_norm": 0.07071255147457123, "learning_rate": 9.75482251562417e-05, "loss": 0.0047, "step": 14510 }, { "epoch": 128.49557522123894, "grad_norm": 0.09240058064460754, "learning_rate": 9.754310837365155e-05, "loss": 0.0051, "step": 14520 }, { "epoch": 128.58407079646017, "grad_norm": 0.1154487133026123, "learning_rate": 9.753798639181856e-05, "loss": 0.0049, "step": 14530 }, { "epoch": 128.67256637168143, "grad_norm": 0.11557339131832123, "learning_rate": 9.753285921130286e-05, "loss": 0.005, "step": 14540 }, { "epoch": 128.76106194690266, "grad_norm": 0.09677843004465103, "learning_rate": 9.752772683266512e-05, "loss": 0.0043, "step": 14550 }, { "epoch": 128.8495575221239, "grad_norm": 0.07759091258049011, "learning_rate": 9.752258925646665e-05, "loss": 0.0041, "step": 14560 }, { "epoch": 128.93805309734512, "grad_norm": 0.0955912321805954, "learning_rate": 9.751744648326926e-05, "loss": 0.0045, "step": 14570 }, { "epoch": 129.02654867256638, "grad_norm": 0.09074988961219788, "learning_rate": 9.751229851363536e-05, "loss": 0.0045, "step": 14580 }, { "epoch": 129.1150442477876, "grad_norm": 0.1201673224568367, "learning_rate": 9.750714534812793e-05, "loss": 0.0049, "step": 14590 }, { "epoch": 129.20353982300884, "grad_norm": 0.09088577330112457, "learning_rate": 9.750198698731053e-05, "loss": 0.0043, "step": 14600 }, { "epoch": 129.2920353982301, "grad_norm": 0.09523794054985046, "learning_rate": 9.749682343174722e-05, "loss": 0.0047, "step": 14610 }, { "epoch": 129.38053097345133, "grad_norm": 0.07919316738843918, "learning_rate": 9.749165468200272e-05, "loss": 0.0043, "step": 14620 }, { "epoch": 129.46902654867256, "grad_norm": 0.09244667738676071, "learning_rate": 9.748648073864229e-05, "loss": 0.0045, "step": 14630 }, { "epoch": 129.55752212389382, "grad_norm": 0.10094402730464935, "learning_rate": 9.748130160223168e-05, "loss": 0.0044, "step": 14640 }, { "epoch": 129.64601769911505, "grad_norm": 0.08799636363983154, "learning_rate": 9.747611727333734e-05, "loss": 0.0047, "step": 14650 }, { "epoch": 129.73451327433628, "grad_norm": 0.09072507917881012, "learning_rate": 9.74709277525262e-05, "loss": 0.0047, "step": 14660 }, { "epoch": 129.8230088495575, "grad_norm": 0.08687207847833633, "learning_rate": 9.746573304036576e-05, "loss": 0.0046, "step": 14670 }, { "epoch": 129.91150442477877, "grad_norm": 0.08784420788288116, "learning_rate": 9.746053313742412e-05, "loss": 0.0049, "step": 14680 }, { "epoch": 130.0, "grad_norm": 0.07573256641626358, "learning_rate": 9.745532804426994e-05, "loss": 0.0047, "step": 14690 }, { "epoch": 130.08849557522123, "grad_norm": 0.08202871680259705, "learning_rate": 9.745011776147242e-05, "loss": 0.0045, "step": 14700 }, { "epoch": 130.1769911504425, "grad_norm": 0.07722724974155426, "learning_rate": 9.744490228960138e-05, "loss": 0.0045, "step": 14710 }, { "epoch": 130.26548672566372, "grad_norm": 0.08964098244905472, "learning_rate": 9.743968162922713e-05, "loss": 0.0048, "step": 14720 }, { "epoch": 130.35398230088495, "grad_norm": 0.08461961895227432, "learning_rate": 9.743445578092064e-05, "loss": 0.0045, "step": 14730 }, { "epoch": 130.44247787610618, "grad_norm": 0.08182445913553238, "learning_rate": 9.742922474525338e-05, "loss": 0.0046, "step": 14740 }, { "epoch": 130.53097345132744, "grad_norm": 0.09286143630743027, "learning_rate": 9.742398852279741e-05, "loss": 0.0042, "step": 14750 }, { "epoch": 130.61946902654867, "grad_norm": 0.06511912494897842, "learning_rate": 9.741874711412535e-05, "loss": 0.0044, "step": 14760 }, { "epoch": 130.7079646017699, "grad_norm": 0.11930762976408005, "learning_rate": 9.741350051981042e-05, "loss": 0.0047, "step": 14770 }, { "epoch": 130.79646017699116, "grad_norm": 0.09444644302129745, "learning_rate": 9.740824874042633e-05, "loss": 0.0045, "step": 14780 }, { "epoch": 130.8849557522124, "grad_norm": 0.07934024184942245, "learning_rate": 9.740299177654746e-05, "loss": 0.0045, "step": 14790 }, { "epoch": 130.97345132743362, "grad_norm": 0.06642714142799377, "learning_rate": 9.739772962874867e-05, "loss": 0.005, "step": 14800 }, { "epoch": 131.06194690265488, "grad_norm": 0.08939877897500992, "learning_rate": 9.739246229760541e-05, "loss": 0.0044, "step": 14810 }, { "epoch": 131.1504424778761, "grad_norm": 0.06999985873699188, "learning_rate": 9.738718978369376e-05, "loss": 0.0047, "step": 14820 }, { "epoch": 131.23893805309734, "grad_norm": 0.0725906565785408, "learning_rate": 9.738191208759025e-05, "loss": 0.0051, "step": 14830 }, { "epoch": 131.32743362831857, "grad_norm": 0.060734961181879044, "learning_rate": 9.73766292098721e-05, "loss": 0.0048, "step": 14840 }, { "epoch": 131.41592920353983, "grad_norm": 0.07550102472305298, "learning_rate": 9.737134115111699e-05, "loss": 0.0045, "step": 14850 }, { "epoch": 131.50442477876106, "grad_norm": 0.07997506111860275, "learning_rate": 9.736604791190323e-05, "loss": 0.0045, "step": 14860 }, { "epoch": 131.5929203539823, "grad_norm": 0.08172719180583954, "learning_rate": 9.73607494928097e-05, "loss": 0.0045, "step": 14870 }, { "epoch": 131.68141592920355, "grad_norm": 0.09249954670667648, "learning_rate": 9.735544589441581e-05, "loss": 0.0046, "step": 14880 }, { "epoch": 131.76991150442478, "grad_norm": 0.0984351858496666, "learning_rate": 9.735013711730154e-05, "loss": 0.0047, "step": 14890 }, { "epoch": 131.858407079646, "grad_norm": 0.0903792679309845, "learning_rate": 9.734482316204747e-05, "loss": 0.0043, "step": 14900 }, { "epoch": 131.94690265486724, "grad_norm": 0.09786337614059448, "learning_rate": 9.733950402923473e-05, "loss": 0.0051, "step": 14910 }, { "epoch": 132.0353982300885, "grad_norm": 0.06881490349769592, "learning_rate": 9.7334179719445e-05, "loss": 0.0044, "step": 14920 }, { "epoch": 132.12389380530973, "grad_norm": 0.08163861930370331, "learning_rate": 9.732885023326053e-05, "loss": 0.0047, "step": 14930 }, { "epoch": 132.21238938053096, "grad_norm": 0.09966982901096344, "learning_rate": 9.732351557126418e-05, "loss": 0.0046, "step": 14940 }, { "epoch": 132.30088495575222, "grad_norm": 0.10399967432022095, "learning_rate": 9.731817573403929e-05, "loss": 0.005, "step": 14950 }, { "epoch": 132.38938053097345, "grad_norm": 0.09278465807437897, "learning_rate": 9.731283072216985e-05, "loss": 0.0047, "step": 14960 }, { "epoch": 132.47787610619469, "grad_norm": 0.07237947732210159, "learning_rate": 9.730748053624039e-05, "loss": 0.0046, "step": 14970 }, { "epoch": 132.56637168141592, "grad_norm": 0.08664923906326294, "learning_rate": 9.730212517683598e-05, "loss": 0.0048, "step": 14980 }, { "epoch": 132.65486725663717, "grad_norm": 0.09754429012537003, "learning_rate": 9.729676464454228e-05, "loss": 0.0047, "step": 14990 }, { "epoch": 132.7433628318584, "grad_norm": 0.09304405748844147, "learning_rate": 9.72913989399455e-05, "loss": 0.0047, "step": 15000 }, { "epoch": 132.83185840707964, "grad_norm": 0.08593449741601944, "learning_rate": 9.728602806363242e-05, "loss": 0.0045, "step": 15010 }, { "epoch": 132.9203539823009, "grad_norm": 0.08178793638944626, "learning_rate": 9.728065201619043e-05, "loss": 0.0044, "step": 15020 }, { "epoch": 133.00884955752213, "grad_norm": 0.10769564658403397, "learning_rate": 9.727527079820742e-05, "loss": 0.0045, "step": 15030 }, { "epoch": 133.09734513274336, "grad_norm": 0.07890182733535767, "learning_rate": 9.726988441027186e-05, "loss": 0.0048, "step": 15040 }, { "epoch": 133.18584070796462, "grad_norm": 0.09230794757604599, "learning_rate": 9.726449285297281e-05, "loss": 0.0042, "step": 15050 }, { "epoch": 133.27433628318585, "grad_norm": 0.08531497418880463, "learning_rate": 9.72590961268999e-05, "loss": 0.005, "step": 15060 }, { "epoch": 133.36283185840708, "grad_norm": 0.08413286507129669, "learning_rate": 9.725369423264328e-05, "loss": 0.0043, "step": 15070 }, { "epoch": 133.4513274336283, "grad_norm": 0.08414915204048157, "learning_rate": 9.72482871707937e-05, "loss": 0.0042, "step": 15080 }, { "epoch": 133.53982300884957, "grad_norm": 0.08280562609434128, "learning_rate": 9.724287494194247e-05, "loss": 0.0043, "step": 15090 }, { "epoch": 133.6283185840708, "grad_norm": 0.0924462154507637, "learning_rate": 9.723745754668147e-05, "loss": 0.0045, "step": 15100 }, { "epoch": 133.71681415929203, "grad_norm": 0.10409913957118988, "learning_rate": 9.723203498560313e-05, "loss": 0.0044, "step": 15110 }, { "epoch": 133.8053097345133, "grad_norm": 0.07903169095516205, "learning_rate": 9.722660725930046e-05, "loss": 0.0047, "step": 15120 }, { "epoch": 133.89380530973452, "grad_norm": 0.07303314656019211, "learning_rate": 9.722117436836702e-05, "loss": 0.0049, "step": 15130 }, { "epoch": 133.98230088495575, "grad_norm": 0.09858247637748718, "learning_rate": 9.721573631339696e-05, "loss": 0.0043, "step": 15140 }, { "epoch": 134.07079646017698, "grad_norm": 0.08907880634069443, "learning_rate": 9.721029309498494e-05, "loss": 0.0043, "step": 15150 }, { "epoch": 134.15929203539824, "grad_norm": 0.10118086636066437, "learning_rate": 9.720484471372627e-05, "loss": 0.0045, "step": 15160 }, { "epoch": 134.24778761061947, "grad_norm": 0.06842686235904694, "learning_rate": 9.719939117021673e-05, "loss": 0.0046, "step": 15170 }, { "epoch": 134.3362831858407, "grad_norm": 0.10126128792762756, "learning_rate": 9.719393246505275e-05, "loss": 0.0045, "step": 15180 }, { "epoch": 134.42477876106196, "grad_norm": 0.08537789434194565, "learning_rate": 9.718846859883128e-05, "loss": 0.0047, "step": 15190 }, { "epoch": 134.5132743362832, "grad_norm": 0.08163796365261078, "learning_rate": 9.718299957214982e-05, "loss": 0.0051, "step": 15200 }, { "epoch": 134.60176991150442, "grad_norm": 0.10206631571054459, "learning_rate": 9.717752538560646e-05, "loss": 0.0045, "step": 15210 }, { "epoch": 134.69026548672565, "grad_norm": 0.07892101258039474, "learning_rate": 9.717204603979986e-05, "loss": 0.0052, "step": 15220 }, { "epoch": 134.7787610619469, "grad_norm": 0.09631048887968063, "learning_rate": 9.716656153532922e-05, "loss": 0.0044, "step": 15230 }, { "epoch": 134.86725663716814, "grad_norm": 0.0852091908454895, "learning_rate": 9.716107187279434e-05, "loss": 0.0042, "step": 15240 }, { "epoch": 134.95575221238937, "grad_norm": 0.09941623359918594, "learning_rate": 9.715557705279555e-05, "loss": 0.0047, "step": 15250 }, { "epoch": 135.04424778761063, "grad_norm": 0.10643564909696579, "learning_rate": 9.715007707593372e-05, "loss": 0.0049, "step": 15260 }, { "epoch": 135.13274336283186, "grad_norm": 0.07489769905805588, "learning_rate": 9.714457194281036e-05, "loss": 0.0046, "step": 15270 }, { "epoch": 135.2212389380531, "grad_norm": 0.07565651834011078, "learning_rate": 9.713906165402751e-05, "loss": 0.0044, "step": 15280 }, { "epoch": 135.30973451327435, "grad_norm": 0.07323747128248215, "learning_rate": 9.713354621018774e-05, "loss": 0.0044, "step": 15290 }, { "epoch": 135.39823008849558, "grad_norm": 0.08144234120845795, "learning_rate": 9.712802561189422e-05, "loss": 0.0049, "step": 15300 }, { "epoch": 135.4867256637168, "grad_norm": 0.0706813856959343, "learning_rate": 9.712249985975069e-05, "loss": 0.004, "step": 15310 }, { "epoch": 135.57522123893804, "grad_norm": 0.09424728155136108, "learning_rate": 9.71169689543614e-05, "loss": 0.0048, "step": 15320 }, { "epoch": 135.6637168141593, "grad_norm": 0.07848697900772095, "learning_rate": 9.711143289633123e-05, "loss": 0.0044, "step": 15330 }, { "epoch": 135.75221238938053, "grad_norm": 0.0953902080655098, "learning_rate": 9.710589168626561e-05, "loss": 0.0041, "step": 15340 }, { "epoch": 135.84070796460176, "grad_norm": 0.08296086639165878, "learning_rate": 9.710034532477048e-05, "loss": 0.0049, "step": 15350 }, { "epoch": 135.92920353982302, "grad_norm": 0.08497466892004013, "learning_rate": 9.709479381245239e-05, "loss": 0.0047, "step": 15360 }, { "epoch": 136.01769911504425, "grad_norm": 0.09814866632223129, "learning_rate": 9.708923714991847e-05, "loss": 0.0046, "step": 15370 }, { "epoch": 136.10619469026548, "grad_norm": 0.10947033762931824, "learning_rate": 9.708367533777638e-05, "loss": 0.0041, "step": 15380 }, { "epoch": 136.1946902654867, "grad_norm": 0.09069793671369553, "learning_rate": 9.707810837663431e-05, "loss": 0.0052, "step": 15390 }, { "epoch": 136.28318584070797, "grad_norm": 0.06988800317049026, "learning_rate": 9.707253626710113e-05, "loss": 0.0047, "step": 15400 }, { "epoch": 136.3716814159292, "grad_norm": 0.07086635380983353, "learning_rate": 9.706695900978613e-05, "loss": 0.0049, "step": 15410 }, { "epoch": 136.46017699115043, "grad_norm": 0.0840168222784996, "learning_rate": 9.706137660529926e-05, "loss": 0.0042, "step": 15420 }, { "epoch": 136.5486725663717, "grad_norm": 0.09476053714752197, "learning_rate": 9.705578905425101e-05, "loss": 0.0042, "step": 15430 }, { "epoch": 136.63716814159292, "grad_norm": 0.07365026324987411, "learning_rate": 9.705019635725241e-05, "loss": 0.0049, "step": 15440 }, { "epoch": 136.72566371681415, "grad_norm": 0.08042581379413605, "learning_rate": 9.704459851491508e-05, "loss": 0.0043, "step": 15450 }, { "epoch": 136.81415929203538, "grad_norm": 0.0946301817893982, "learning_rate": 9.703899552785118e-05, "loss": 0.0043, "step": 15460 }, { "epoch": 136.90265486725664, "grad_norm": 0.0751660093665123, "learning_rate": 9.703338739667346e-05, "loss": 0.0047, "step": 15470 }, { "epoch": 136.99115044247787, "grad_norm": 0.09506610780954361, "learning_rate": 9.70277741219952e-05, "loss": 0.0047, "step": 15480 }, { "epoch": 137.0796460176991, "grad_norm": 0.09341573715209961, "learning_rate": 9.702215570443027e-05, "loss": 0.0047, "step": 15490 }, { "epoch": 137.16814159292036, "grad_norm": 0.08903171867132187, "learning_rate": 9.701653214459309e-05, "loss": 0.0051, "step": 15500 }, { "epoch": 137.2566371681416, "grad_norm": 0.08235716819763184, "learning_rate": 9.701090344309865e-05, "loss": 0.0049, "step": 15510 }, { "epoch": 137.34513274336283, "grad_norm": 0.11730387061834335, "learning_rate": 9.700526960056247e-05, "loss": 0.0049, "step": 15520 }, { "epoch": 137.43362831858408, "grad_norm": 0.07509349286556244, "learning_rate": 9.699963061760068e-05, "loss": 0.005, "step": 15530 }, { "epoch": 137.52212389380531, "grad_norm": 0.09111277014017105, "learning_rate": 9.699398649482997e-05, "loss": 0.0049, "step": 15540 }, { "epoch": 137.61061946902655, "grad_norm": 0.11541454493999481, "learning_rate": 9.698833723286753e-05, "loss": 0.005, "step": 15550 }, { "epoch": 137.69911504424778, "grad_norm": 0.08394335955381393, "learning_rate": 9.698268283233118e-05, "loss": 0.0046, "step": 15560 }, { "epoch": 137.78761061946904, "grad_norm": 0.0837567076086998, "learning_rate": 9.697702329383929e-05, "loss": 0.0043, "step": 15570 }, { "epoch": 137.87610619469027, "grad_norm": 0.08950358629226685, "learning_rate": 9.697135861801074e-05, "loss": 0.0048, "step": 15580 }, { "epoch": 137.9646017699115, "grad_norm": 0.09696181118488312, "learning_rate": 9.696568880546505e-05, "loss": 0.0045, "step": 15590 }, { "epoch": 138.05309734513276, "grad_norm": 0.08598753064870834, "learning_rate": 9.696001385682223e-05, "loss": 0.0046, "step": 15600 }, { "epoch": 138.141592920354, "grad_norm": 0.09887449443340302, "learning_rate": 9.695433377270291e-05, "loss": 0.0043, "step": 15610 }, { "epoch": 138.23008849557522, "grad_norm": 0.09556949138641357, "learning_rate": 9.694864855372824e-05, "loss": 0.0044, "step": 15620 }, { "epoch": 138.31858407079645, "grad_norm": 0.11367657035589218, "learning_rate": 9.694295820051995e-05, "loss": 0.0045, "step": 15630 }, { "epoch": 138.4070796460177, "grad_norm": 0.0860295370221138, "learning_rate": 9.693726271370032e-05, "loss": 0.0044, "step": 15640 }, { "epoch": 138.49557522123894, "grad_norm": 0.11800909042358398, "learning_rate": 9.693156209389221e-05, "loss": 0.0041, "step": 15650 }, { "epoch": 138.58407079646017, "grad_norm": 0.1515018343925476, "learning_rate": 9.692585634171905e-05, "loss": 0.0044, "step": 15660 }, { "epoch": 138.67256637168143, "grad_norm": 0.08284424245357513, "learning_rate": 9.692014545780476e-05, "loss": 0.0044, "step": 15670 }, { "epoch": 138.76106194690266, "grad_norm": 0.08957313001155853, "learning_rate": 9.691442944277393e-05, "loss": 0.004, "step": 15680 }, { "epoch": 138.8495575221239, "grad_norm": 0.07543960213661194, "learning_rate": 9.690870829725162e-05, "loss": 0.0043, "step": 15690 }, { "epoch": 138.93805309734512, "grad_norm": 0.09393542259931564, "learning_rate": 9.69029820218635e-05, "loss": 0.0046, "step": 15700 }, { "epoch": 139.02654867256638, "grad_norm": 0.09585735946893692, "learning_rate": 9.689725061723579e-05, "loss": 0.0045, "step": 15710 }, { "epoch": 139.1150442477876, "grad_norm": 0.12023745477199554, "learning_rate": 9.689151408399527e-05, "loss": 0.0047, "step": 15720 }, { "epoch": 139.20353982300884, "grad_norm": 0.08876563608646393, "learning_rate": 9.688577242276924e-05, "loss": 0.0043, "step": 15730 }, { "epoch": 139.2920353982301, "grad_norm": 0.09717845171689987, "learning_rate": 9.688002563418566e-05, "loss": 0.0044, "step": 15740 }, { "epoch": 139.38053097345133, "grad_norm": 0.09812776744365692, "learning_rate": 9.687427371887293e-05, "loss": 0.0046, "step": 15750 }, { "epoch": 139.46902654867256, "grad_norm": 0.08406495302915573, "learning_rate": 9.686851667746012e-05, "loss": 0.0043, "step": 15760 }, { "epoch": 139.55752212389382, "grad_norm": 0.1130337119102478, "learning_rate": 9.686275451057677e-05, "loss": 0.0046, "step": 15770 }, { "epoch": 139.64601769911505, "grad_norm": 0.11284122616052628, "learning_rate": 9.685698721885308e-05, "loss": 0.0044, "step": 15780 }, { "epoch": 139.73451327433628, "grad_norm": 0.07276623696088791, "learning_rate": 9.68512148029197e-05, "loss": 0.004, "step": 15790 }, { "epoch": 139.8230088495575, "grad_norm": 0.07430467009544373, "learning_rate": 9.684543726340791e-05, "loss": 0.0044, "step": 15800 }, { "epoch": 139.91150442477877, "grad_norm": 0.07501018047332764, "learning_rate": 9.683965460094952e-05, "loss": 0.0046, "step": 15810 }, { "epoch": 140.0, "grad_norm": 0.0634877011179924, "learning_rate": 9.683386681617694e-05, "loss": 0.0045, "step": 15820 }, { "epoch": 140.08849557522123, "grad_norm": 0.07791174203157425, "learning_rate": 9.68280739097231e-05, "loss": 0.0049, "step": 15830 }, { "epoch": 140.1769911504425, "grad_norm": 0.08390826731920242, "learning_rate": 9.682227588222148e-05, "loss": 0.0043, "step": 15840 }, { "epoch": 140.26548672566372, "grad_norm": 0.09050701558589935, "learning_rate": 9.681647273430618e-05, "loss": 0.0048, "step": 15850 }, { "epoch": 140.35398230088495, "grad_norm": 0.09205297380685806, "learning_rate": 9.681066446661182e-05, "loss": 0.0045, "step": 15860 }, { "epoch": 140.44247787610618, "grad_norm": 0.08391117304563522, "learning_rate": 9.680485107977357e-05, "loss": 0.0047, "step": 15870 }, { "epoch": 140.53097345132744, "grad_norm": 0.10555341094732285, "learning_rate": 9.679903257442716e-05, "loss": 0.0049, "step": 15880 }, { "epoch": 140.61946902654867, "grad_norm": 0.08682692795991898, "learning_rate": 9.679320895120891e-05, "loss": 0.0042, "step": 15890 }, { "epoch": 140.7079646017699, "grad_norm": 0.07572079449892044, "learning_rate": 9.67873802107557e-05, "loss": 0.0048, "step": 15900 }, { "epoch": 140.79646017699116, "grad_norm": 0.07349273562431335, "learning_rate": 9.67815463537049e-05, "loss": 0.0044, "step": 15910 }, { "epoch": 140.8849557522124, "grad_norm": 0.05762498080730438, "learning_rate": 9.677570738069457e-05, "loss": 0.0039, "step": 15920 }, { "epoch": 140.97345132743362, "grad_norm": 0.08590847998857498, "learning_rate": 9.676986329236318e-05, "loss": 0.0044, "step": 15930 }, { "epoch": 141.06194690265488, "grad_norm": 0.08742246776819229, "learning_rate": 9.676401408934987e-05, "loss": 0.0047, "step": 15940 }, { "epoch": 141.1504424778761, "grad_norm": 0.07438519597053528, "learning_rate": 9.675815977229428e-05, "loss": 0.0049, "step": 15950 }, { "epoch": 141.23893805309734, "grad_norm": 0.08556005358695984, "learning_rate": 9.675230034183664e-05, "loss": 0.0043, "step": 15960 }, { "epoch": 141.32743362831857, "grad_norm": 0.08960195630788803, "learning_rate": 9.674643579861773e-05, "loss": 0.0046, "step": 15970 }, { "epoch": 141.41592920353983, "grad_norm": 0.07764564454555511, "learning_rate": 9.674056614327886e-05, "loss": 0.0048, "step": 15980 }, { "epoch": 141.50442477876106, "grad_norm": 0.0662534311413765, "learning_rate": 9.673469137646198e-05, "loss": 0.0045, "step": 15990 }, { "epoch": 141.5929203539823, "grad_norm": 0.06831704080104828, "learning_rate": 9.67288114988095e-05, "loss": 0.0044, "step": 16000 }, { "epoch": 141.68141592920355, "grad_norm": 0.09450788050889969, "learning_rate": 9.672292651096447e-05, "loss": 0.0048, "step": 16010 }, { "epoch": 141.76991150442478, "grad_norm": 0.09383562952280045, "learning_rate": 9.671703641357042e-05, "loss": 0.0043, "step": 16020 }, { "epoch": 141.858407079646, "grad_norm": 0.09770745784044266, "learning_rate": 9.67111412072715e-05, "loss": 0.0046, "step": 16030 }, { "epoch": 141.94690265486724, "grad_norm": 0.08345290273427963, "learning_rate": 9.670524089271242e-05, "loss": 0.0041, "step": 16040 }, { "epoch": 142.0353982300885, "grad_norm": 0.08651567995548248, "learning_rate": 9.669933547053842e-05, "loss": 0.0048, "step": 16050 }, { "epoch": 142.12389380530973, "grad_norm": 0.10266868025064468, "learning_rate": 9.669342494139531e-05, "loss": 0.0045, "step": 16060 }, { "epoch": 142.21238938053096, "grad_norm": 0.10197079181671143, "learning_rate": 9.668750930592943e-05, "loss": 0.0042, "step": 16070 }, { "epoch": 142.30088495575222, "grad_norm": 0.06374242901802063, "learning_rate": 9.668158856478775e-05, "loss": 0.0039, "step": 16080 }, { "epoch": 142.38938053097345, "grad_norm": 0.1120043396949768, "learning_rate": 9.66756627186177e-05, "loss": 0.0048, "step": 16090 }, { "epoch": 142.47787610619469, "grad_norm": 0.0955357551574707, "learning_rate": 9.666973176806737e-05, "loss": 0.0046, "step": 16100 }, { "epoch": 142.56637168141592, "grad_norm": 0.09819258749485016, "learning_rate": 9.666379571378534e-05, "loss": 0.0044, "step": 16110 }, { "epoch": 142.65486725663717, "grad_norm": 0.07241105288267136, "learning_rate": 9.665785455642076e-05, "loss": 0.0045, "step": 16120 }, { "epoch": 142.7433628318584, "grad_norm": 0.07491251826286316, "learning_rate": 9.665190829662337e-05, "loss": 0.0041, "step": 16130 }, { "epoch": 142.83185840707964, "grad_norm": 0.081999272108078, "learning_rate": 9.664595693504342e-05, "loss": 0.0041, "step": 16140 }, { "epoch": 142.9203539823009, "grad_norm": 0.09949837625026703, "learning_rate": 9.664000047233175e-05, "loss": 0.0044, "step": 16150 }, { "epoch": 143.00884955752213, "grad_norm": 0.09236070513725281, "learning_rate": 9.663403890913976e-05, "loss": 0.0042, "step": 16160 }, { "epoch": 143.09734513274336, "grad_norm": 0.07917601615190506, "learning_rate": 9.662807224611938e-05, "loss": 0.0046, "step": 16170 }, { "epoch": 143.18584070796462, "grad_norm": 0.0917157530784607, "learning_rate": 9.662210048392311e-05, "loss": 0.0049, "step": 16180 }, { "epoch": 143.27433628318585, "grad_norm": 0.08320609480142593, "learning_rate": 9.661612362320405e-05, "loss": 0.0047, "step": 16190 }, { "epoch": 143.36283185840708, "grad_norm": 0.10119134187698364, "learning_rate": 9.661014166461579e-05, "loss": 0.0045, "step": 16200 }, { "epoch": 143.4513274336283, "grad_norm": 0.07993099838495255, "learning_rate": 9.66041546088125e-05, "loss": 0.0045, "step": 16210 }, { "epoch": 143.53982300884957, "grad_norm": 0.1103522777557373, "learning_rate": 9.659816245644895e-05, "loss": 0.0048, "step": 16220 }, { "epoch": 143.6283185840708, "grad_norm": 0.0887879952788353, "learning_rate": 9.65921652081804e-05, "loss": 0.0047, "step": 16230 }, { "epoch": 143.71681415929203, "grad_norm": 0.0893658995628357, "learning_rate": 9.658616286466271e-05, "loss": 0.0043, "step": 16240 }, { "epoch": 143.8053097345133, "grad_norm": 0.09164734929800034, "learning_rate": 9.65801554265523e-05, "loss": 0.0049, "step": 16250 }, { "epoch": 143.89380530973452, "grad_norm": 0.0867948830127716, "learning_rate": 9.657414289450612e-05, "loss": 0.0041, "step": 16260 }, { "epoch": 143.98230088495575, "grad_norm": 0.07291430234909058, "learning_rate": 9.656812526918171e-05, "loss": 0.0047, "step": 16270 }, { "epoch": 144.07079646017698, "grad_norm": 0.10674595087766647, "learning_rate": 9.656210255123712e-05, "loss": 0.0048, "step": 16280 }, { "epoch": 144.15929203539824, "grad_norm": 0.10656282305717468, "learning_rate": 9.6556074741331e-05, "loss": 0.0051, "step": 16290 }, { "epoch": 144.24778761061947, "grad_norm": 0.08283402025699615, "learning_rate": 9.655004184012256e-05, "loss": 0.0046, "step": 16300 }, { "epoch": 144.3362831858407, "grad_norm": 0.08003602176904678, "learning_rate": 9.654400384827152e-05, "loss": 0.0044, "step": 16310 }, { "epoch": 144.42477876106196, "grad_norm": 0.07463746517896652, "learning_rate": 9.653796076643818e-05, "loss": 0.0045, "step": 16320 }, { "epoch": 144.5132743362832, "grad_norm": 0.09373288601636887, "learning_rate": 9.653191259528344e-05, "loss": 0.0042, "step": 16330 }, { "epoch": 144.60176991150442, "grad_norm": 0.09234106540679932, "learning_rate": 9.65258593354687e-05, "loss": 0.0041, "step": 16340 }, { "epoch": 144.69026548672565, "grad_norm": 0.07017724961042404, "learning_rate": 9.651980098765591e-05, "loss": 0.0047, "step": 16350 }, { "epoch": 144.7787610619469, "grad_norm": 0.09159772843122482, "learning_rate": 9.651373755250765e-05, "loss": 0.0046, "step": 16360 }, { "epoch": 144.86725663716814, "grad_norm": 0.06195638328790665, "learning_rate": 9.650766903068697e-05, "loss": 0.0046, "step": 16370 }, { "epoch": 144.95575221238937, "grad_norm": 0.08870291709899902, "learning_rate": 9.650159542285753e-05, "loss": 0.0039, "step": 16380 }, { "epoch": 145.04424778761063, "grad_norm": 0.09314273297786713, "learning_rate": 9.649551672968353e-05, "loss": 0.0042, "step": 16390 }, { "epoch": 145.13274336283186, "grad_norm": 0.07327783107757568, "learning_rate": 9.648943295182973e-05, "loss": 0.0047, "step": 16400 }, { "epoch": 145.2212389380531, "grad_norm": 0.11421719938516617, "learning_rate": 9.648334408996144e-05, "loss": 0.0042, "step": 16410 }, { "epoch": 145.30973451327435, "grad_norm": 0.12038262188434601, "learning_rate": 9.647725014474452e-05, "loss": 0.004, "step": 16420 }, { "epoch": 145.39823008849558, "grad_norm": 0.09349304437637329, "learning_rate": 9.64711511168454e-05, "loss": 0.0038, "step": 16430 }, { "epoch": 145.4867256637168, "grad_norm": 0.09387480467557907, "learning_rate": 9.646504700693108e-05, "loss": 0.0038, "step": 16440 }, { "epoch": 145.57522123893804, "grad_norm": 0.10325340926647186, "learning_rate": 9.645893781566907e-05, "loss": 0.0048, "step": 16450 }, { "epoch": 145.6637168141593, "grad_norm": 0.10835103690624237, "learning_rate": 9.645282354372744e-05, "loss": 0.0046, "step": 16460 }, { "epoch": 145.75221238938053, "grad_norm": 0.0900927186012268, "learning_rate": 9.644670419177491e-05, "loss": 0.0045, "step": 16470 }, { "epoch": 145.84070796460176, "grad_norm": 0.09508642554283142, "learning_rate": 9.644057976048062e-05, "loss": 0.0048, "step": 16480 }, { "epoch": 145.92920353982302, "grad_norm": 0.08206213265657425, "learning_rate": 9.643445025051435e-05, "loss": 0.0047, "step": 16490 }, { "epoch": 146.01769911504425, "grad_norm": 0.11639118194580078, "learning_rate": 9.642831566254641e-05, "loss": 0.0044, "step": 16500 }, { "epoch": 146.10619469026548, "grad_norm": 0.09717216342687607, "learning_rate": 9.642217599724769e-05, "loss": 0.0051, "step": 16510 }, { "epoch": 146.1946902654867, "grad_norm": 0.08572854101657867, "learning_rate": 9.64160312552896e-05, "loss": 0.0046, "step": 16520 }, { "epoch": 146.28318584070797, "grad_norm": 0.09049621969461441, "learning_rate": 9.64098814373441e-05, "loss": 0.0044, "step": 16530 }, { "epoch": 146.3716814159292, "grad_norm": 0.07521453499794006, "learning_rate": 9.640372654408374e-05, "loss": 0.0042, "step": 16540 }, { "epoch": 146.46017699115043, "grad_norm": 0.0795065313577652, "learning_rate": 9.639756657618162e-05, "loss": 0.0042, "step": 16550 }, { "epoch": 146.5486725663717, "grad_norm": 0.06960976868867874, "learning_rate": 9.639140153431138e-05, "loss": 0.0043, "step": 16560 }, { "epoch": 146.63716814159292, "grad_norm": 0.07714482396841049, "learning_rate": 9.638523141914721e-05, "loss": 0.0042, "step": 16570 }, { "epoch": 146.72566371681415, "grad_norm": 0.08392421156167984, "learning_rate": 9.637905623136388e-05, "loss": 0.0042, "step": 16580 }, { "epoch": 146.81415929203538, "grad_norm": 0.08617852628231049, "learning_rate": 9.637287597163669e-05, "loss": 0.0043, "step": 16590 }, { "epoch": 146.90265486725664, "grad_norm": 0.09800716489553452, "learning_rate": 9.63666906406415e-05, "loss": 0.0045, "step": 16600 }, { "epoch": 146.99115044247787, "grad_norm": 0.08995994180440903, "learning_rate": 9.636050023905473e-05, "loss": 0.0043, "step": 16610 }, { "epoch": 147.0796460176991, "grad_norm": 0.07552612572908401, "learning_rate": 9.635430476755336e-05, "loss": 0.0043, "step": 16620 }, { "epoch": 147.16814159292036, "grad_norm": 0.08786757290363312, "learning_rate": 9.63481042268149e-05, "loss": 0.0042, "step": 16630 }, { "epoch": 147.2566371681416, "grad_norm": 0.09066076576709747, "learning_rate": 9.634189861751745e-05, "loss": 0.005, "step": 16640 }, { "epoch": 147.34513274336283, "grad_norm": 0.08089663088321686, "learning_rate": 9.633568794033967e-05, "loss": 0.0041, "step": 16650 }, { "epoch": 147.43362831858408, "grad_norm": 0.0810329020023346, "learning_rate": 9.63294721959607e-05, "loss": 0.004, "step": 16660 }, { "epoch": 147.52212389380531, "grad_norm": 0.09006273001432419, "learning_rate": 9.63232513850603e-05, "loss": 0.0044, "step": 16670 }, { "epoch": 147.61061946902655, "grad_norm": 0.10602041333913803, "learning_rate": 9.631702550831878e-05, "loss": 0.0044, "step": 16680 }, { "epoch": 147.69911504424778, "grad_norm": 0.08955232053995132, "learning_rate": 9.631079456641698e-05, "loss": 0.005, "step": 16690 }, { "epoch": 147.78761061946904, "grad_norm": 0.08623801171779633, "learning_rate": 9.630455856003632e-05, "loss": 0.0042, "step": 16700 }, { "epoch": 147.87610619469027, "grad_norm": 0.08050607144832611, "learning_rate": 9.629831748985876e-05, "loss": 0.0045, "step": 16710 }, { "epoch": 147.9646017699115, "grad_norm": 0.09921468049287796, "learning_rate": 9.629207135656679e-05, "loss": 0.0044, "step": 16720 }, { "epoch": 148.05309734513276, "grad_norm": 0.09762436151504517, "learning_rate": 9.628582016084353e-05, "loss": 0.0044, "step": 16730 }, { "epoch": 148.141592920354, "grad_norm": 0.07619350403547287, "learning_rate": 9.627956390337254e-05, "loss": 0.0046, "step": 16740 }, { "epoch": 148.23008849557522, "grad_norm": 0.0831795260310173, "learning_rate": 9.627330258483802e-05, "loss": 0.0039, "step": 16750 }, { "epoch": 148.31858407079645, "grad_norm": 0.07813547551631927, "learning_rate": 9.62670362059247e-05, "loss": 0.0048, "step": 16760 }, { "epoch": 148.4070796460177, "grad_norm": 0.08414070308208466, "learning_rate": 9.626076476731786e-05, "loss": 0.0041, "step": 16770 }, { "epoch": 148.49557522123894, "grad_norm": 0.1004515290260315, "learning_rate": 9.625448826970336e-05, "loss": 0.0045, "step": 16780 }, { "epoch": 148.58407079646017, "grad_norm": 0.09463264048099518, "learning_rate": 9.624820671376755e-05, "loss": 0.0045, "step": 16790 }, { "epoch": 148.67256637168143, "grad_norm": 0.07164070010185242, "learning_rate": 9.62419201001974e-05, "loss": 0.0048, "step": 16800 }, { "epoch": 148.76106194690266, "grad_norm": 0.10227761417627335, "learning_rate": 9.623562842968037e-05, "loss": 0.0042, "step": 16810 }, { "epoch": 148.8495575221239, "grad_norm": 0.09342348575592041, "learning_rate": 9.622933170290454e-05, "loss": 0.0041, "step": 16820 }, { "epoch": 148.93805309734512, "grad_norm": 0.14254695177078247, "learning_rate": 9.622302992055849e-05, "loss": 0.0043, "step": 16830 }, { "epoch": 149.02654867256638, "grad_norm": 0.08960986882448196, "learning_rate": 9.62167230833314e-05, "loss": 0.0045, "step": 16840 }, { "epoch": 149.1150442477876, "grad_norm": 0.1110040619969368, "learning_rate": 9.621041119191295e-05, "loss": 0.0041, "step": 16850 }, { "epoch": 149.20353982300884, "grad_norm": 0.10956564545631409, "learning_rate": 9.620409424699342e-05, "loss": 0.0043, "step": 16860 }, { "epoch": 149.2920353982301, "grad_norm": 0.09771065413951874, "learning_rate": 9.619777224926359e-05, "loss": 0.0044, "step": 16870 }, { "epoch": 149.38053097345133, "grad_norm": 0.08850602060556412, "learning_rate": 9.619144519941485e-05, "loss": 0.0048, "step": 16880 }, { "epoch": 149.46902654867256, "grad_norm": 0.0835912749171257, "learning_rate": 9.618511309813912e-05, "loss": 0.0042, "step": 16890 }, { "epoch": 149.55752212389382, "grad_norm": 0.08654868602752686, "learning_rate": 9.617877594612886e-05, "loss": 0.0038, "step": 16900 }, { "epoch": 149.64601769911505, "grad_norm": 0.09056159853935242, "learning_rate": 9.617243374407707e-05, "loss": 0.0042, "step": 16910 }, { "epoch": 149.73451327433628, "grad_norm": 0.07458975166082382, "learning_rate": 9.616608649267736e-05, "loss": 0.0042, "step": 16920 }, { "epoch": 149.8230088495575, "grad_norm": 0.09526009112596512, "learning_rate": 9.615973419262385e-05, "loss": 0.0045, "step": 16930 }, { "epoch": 149.91150442477877, "grad_norm": 0.07747995108366013, "learning_rate": 9.615337684461119e-05, "loss": 0.0042, "step": 16940 }, { "epoch": 150.0, "grad_norm": 0.07099507749080658, "learning_rate": 9.614701444933465e-05, "loss": 0.0043, "step": 16950 }, { "epoch": 150.08849557522123, "grad_norm": 0.06323894113302231, "learning_rate": 9.614064700748997e-05, "loss": 0.0041, "step": 16960 }, { "epoch": 150.1769911504425, "grad_norm": 0.06224800646305084, "learning_rate": 9.613427451977352e-05, "loss": 0.0041, "step": 16970 }, { "epoch": 150.26548672566372, "grad_norm": 0.06368186324834824, "learning_rate": 9.612789698688216e-05, "loss": 0.0044, "step": 16980 }, { "epoch": 150.35398230088495, "grad_norm": 0.07854883372783661, "learning_rate": 9.612151440951334e-05, "loss": 0.004, "step": 16990 }, { "epoch": 150.44247787610618, "grad_norm": 0.08761570602655411, "learning_rate": 9.611512678836506e-05, "loss": 0.0044, "step": 17000 }, { "epoch": 150.53097345132744, "grad_norm": 0.09755021333694458, "learning_rate": 9.610873412413584e-05, "loss": 0.0041, "step": 17010 }, { "epoch": 150.61946902654867, "grad_norm": 0.0868295282125473, "learning_rate": 9.610233641752476e-05, "loss": 0.0046, "step": 17020 }, { "epoch": 150.7079646017699, "grad_norm": 0.10647405683994293, "learning_rate": 9.609593366923151e-05, "loss": 0.0042, "step": 17030 }, { "epoch": 150.79646017699116, "grad_norm": 0.08765871822834015, "learning_rate": 9.608952587995625e-05, "loss": 0.0044, "step": 17040 }, { "epoch": 150.8849557522124, "grad_norm": 0.10043369233608246, "learning_rate": 9.608311305039972e-05, "loss": 0.0042, "step": 17050 }, { "epoch": 150.97345132743362, "grad_norm": 0.11272019892930984, "learning_rate": 9.607669518126326e-05, "loss": 0.0041, "step": 17060 }, { "epoch": 151.06194690265488, "grad_norm": 0.10871003568172455, "learning_rate": 9.607027227324866e-05, "loss": 0.0049, "step": 17070 }, { "epoch": 151.1504424778761, "grad_norm": 0.09508123248815536, "learning_rate": 9.606384432705837e-05, "loss": 0.004, "step": 17080 }, { "epoch": 151.23893805309734, "grad_norm": 0.07134079933166504, "learning_rate": 9.60574113433953e-05, "loss": 0.0044, "step": 17090 }, { "epoch": 151.32743362831857, "grad_norm": 0.09845523536205292, "learning_rate": 9.6050973322963e-05, "loss": 0.0046, "step": 17100 }, { "epoch": 151.41592920353983, "grad_norm": 0.1103099063038826, "learning_rate": 9.604453026646547e-05, "loss": 0.0044, "step": 17110 }, { "epoch": 151.50442477876106, "grad_norm": 0.07855617254972458, "learning_rate": 9.603808217460735e-05, "loss": 0.0044, "step": 17120 }, { "epoch": 151.5929203539823, "grad_norm": 0.09625886380672455, "learning_rate": 9.603162904809377e-05, "loss": 0.004, "step": 17130 }, { "epoch": 151.68141592920355, "grad_norm": 0.07785475254058838, "learning_rate": 9.602517088763045e-05, "loss": 0.0044, "step": 17140 }, { "epoch": 151.76991150442478, "grad_norm": 0.10237831622362137, "learning_rate": 9.601870769392365e-05, "loss": 0.004, "step": 17150 }, { "epoch": 151.858407079646, "grad_norm": 0.08839760720729828, "learning_rate": 9.601223946768017e-05, "loss": 0.004, "step": 17160 }, { "epoch": 151.94690265486724, "grad_norm": 0.0734468474984169, "learning_rate": 9.600576620960734e-05, "loss": 0.0043, "step": 17170 }, { "epoch": 152.0353982300885, "grad_norm": 0.10750427842140198, "learning_rate": 9.599928792041308e-05, "loss": 0.0041, "step": 17180 }, { "epoch": 152.12389380530973, "grad_norm": 0.09627014398574829, "learning_rate": 9.599280460080587e-05, "loss": 0.0044, "step": 17190 }, { "epoch": 152.21238938053096, "grad_norm": 0.1052168682217598, "learning_rate": 9.59863162514947e-05, "loss": 0.0046, "step": 17200 }, { "epoch": 152.30088495575222, "grad_norm": 0.09223965555429459, "learning_rate": 9.597982287318911e-05, "loss": 0.0044, "step": 17210 }, { "epoch": 152.38938053097345, "grad_norm": 0.10456546396017075, "learning_rate": 9.597332446659923e-05, "loss": 0.0041, "step": 17220 }, { "epoch": 152.47787610619469, "grad_norm": 0.06416191160678864, "learning_rate": 9.59668210324357e-05, "loss": 0.0045, "step": 17230 }, { "epoch": 152.56637168141592, "grad_norm": 0.07435528934001923, "learning_rate": 9.596031257140974e-05, "loss": 0.0036, "step": 17240 }, { "epoch": 152.65486725663717, "grad_norm": 0.09575612097978592, "learning_rate": 9.59537990842331e-05, "loss": 0.0044, "step": 17250 }, { "epoch": 152.7433628318584, "grad_norm": 0.07021874189376831, "learning_rate": 9.594728057161806e-05, "loss": 0.0041, "step": 17260 }, { "epoch": 152.83185840707964, "grad_norm": 0.08878923207521439, "learning_rate": 9.594075703427752e-05, "loss": 0.0044, "step": 17270 }, { "epoch": 152.9203539823009, "grad_norm": 0.10422157496213913, "learning_rate": 9.593422847292486e-05, "loss": 0.0043, "step": 17280 }, { "epoch": 153.00884955752213, "grad_norm": 0.08962541818618774, "learning_rate": 9.592769488827402e-05, "loss": 0.0038, "step": 17290 }, { "epoch": 153.09734513274336, "grad_norm": 0.05347001180052757, "learning_rate": 9.592115628103952e-05, "loss": 0.0046, "step": 17300 }, { "epoch": 153.18584070796462, "grad_norm": 0.07783836126327515, "learning_rate": 9.591461265193643e-05, "loss": 0.0041, "step": 17310 }, { "epoch": 153.27433628318585, "grad_norm": 0.07316505163908005, "learning_rate": 9.590806400168032e-05, "loss": 0.0043, "step": 17320 }, { "epoch": 153.36283185840708, "grad_norm": 0.07716911286115646, "learning_rate": 9.590151033098735e-05, "loss": 0.0042, "step": 17330 }, { "epoch": 153.4513274336283, "grad_norm": 0.0913330614566803, "learning_rate": 9.589495164057423e-05, "loss": 0.0044, "step": 17340 }, { "epoch": 153.53982300884957, "grad_norm": 0.10033827275037766, "learning_rate": 9.58883879311582e-05, "loss": 0.0038, "step": 17350 }, { "epoch": 153.6283185840708, "grad_norm": 0.08237603306770325, "learning_rate": 9.588181920345705e-05, "loss": 0.0045, "step": 17360 }, { "epoch": 153.71681415929203, "grad_norm": 0.10472193360328674, "learning_rate": 9.587524545818913e-05, "loss": 0.0044, "step": 17370 }, { "epoch": 153.8053097345133, "grad_norm": 0.07839012145996094, "learning_rate": 9.586866669607335e-05, "loss": 0.004, "step": 17380 }, { "epoch": 153.89380530973452, "grad_norm": 0.09537231177091599, "learning_rate": 9.586208291782915e-05, "loss": 0.0049, "step": 17390 }, { "epoch": 153.98230088495575, "grad_norm": 0.09925303608179092, "learning_rate": 9.58554941241765e-05, "loss": 0.0042, "step": 17400 }, { "epoch": 154.07079646017698, "grad_norm": 0.08477620035409927, "learning_rate": 9.584890031583596e-05, "loss": 0.0044, "step": 17410 }, { "epoch": 154.15929203539824, "grad_norm": 0.08299144357442856, "learning_rate": 9.584230149352861e-05, "loss": 0.0043, "step": 17420 }, { "epoch": 154.24778761061947, "grad_norm": 0.09542787820100784, "learning_rate": 9.58356976579761e-05, "loss": 0.0043, "step": 17430 }, { "epoch": 154.3362831858407, "grad_norm": 0.07333514839410782, "learning_rate": 9.58290888099006e-05, "loss": 0.0043, "step": 17440 }, { "epoch": 154.42477876106196, "grad_norm": 0.08177866041660309, "learning_rate": 9.582247495002486e-05, "loss": 0.0046, "step": 17450 }, { "epoch": 154.5132743362832, "grad_norm": 0.07352180778980255, "learning_rate": 9.581585607907214e-05, "loss": 0.0042, "step": 17460 }, { "epoch": 154.60176991150442, "grad_norm": 0.08054899424314499, "learning_rate": 9.580923219776628e-05, "loss": 0.0035, "step": 17470 }, { "epoch": 154.69026548672565, "grad_norm": 0.09385788440704346, "learning_rate": 9.580260330683167e-05, "loss": 0.0039, "step": 17480 }, { "epoch": 154.7787610619469, "grad_norm": 0.09144049137830734, "learning_rate": 9.579596940699322e-05, "loss": 0.0042, "step": 17490 }, { "epoch": 154.86725663716814, "grad_norm": 0.10268810391426086, "learning_rate": 9.578933049897643e-05, "loss": 0.0038, "step": 17500 }, { "epoch": 154.95575221238937, "grad_norm": 0.09858044236898422, "learning_rate": 9.578268658350728e-05, "loss": 0.0044, "step": 17510 }, { "epoch": 155.04424778761063, "grad_norm": 0.07640513777732849, "learning_rate": 9.577603766131235e-05, "loss": 0.0041, "step": 17520 }, { "epoch": 155.13274336283186, "grad_norm": 0.09446601569652557, "learning_rate": 9.576938373311878e-05, "loss": 0.0045, "step": 17530 }, { "epoch": 155.2212389380531, "grad_norm": 0.10332854837179184, "learning_rate": 9.576272479965421e-05, "loss": 0.0044, "step": 17540 }, { "epoch": 155.30973451327435, "grad_norm": 0.10205870121717453, "learning_rate": 9.575606086164687e-05, "loss": 0.0045, "step": 17550 }, { "epoch": 155.39823008849558, "grad_norm": 0.10591495782136917, "learning_rate": 9.57493919198255e-05, "loss": 0.0042, "step": 17560 }, { "epoch": 155.4867256637168, "grad_norm": 0.07039690017700195, "learning_rate": 9.57427179749194e-05, "loss": 0.0044, "step": 17570 }, { "epoch": 155.57522123893804, "grad_norm": 0.06022496521472931, "learning_rate": 9.573603902765846e-05, "loss": 0.004, "step": 17580 }, { "epoch": 155.6637168141593, "grad_norm": 0.07746409624814987, "learning_rate": 9.572935507877304e-05, "loss": 0.0038, "step": 17590 }, { "epoch": 155.75221238938053, "grad_norm": 0.0890486016869545, "learning_rate": 9.57226661289941e-05, "loss": 0.0044, "step": 17600 }, { "epoch": 155.84070796460176, "grad_norm": 0.06809765845537186, "learning_rate": 9.571597217905315e-05, "loss": 0.0045, "step": 17610 }, { "epoch": 155.92920353982302, "grad_norm": 0.07350140810012817, "learning_rate": 9.57092732296822e-05, "loss": 0.0047, "step": 17620 }, { "epoch": 156.01769911504425, "grad_norm": 0.11181480437517166, "learning_rate": 9.570256928161385e-05, "loss": 0.004, "step": 17630 }, { "epoch": 156.10619469026548, "grad_norm": 0.10229756683111191, "learning_rate": 9.569586033558126e-05, "loss": 0.0041, "step": 17640 }, { "epoch": 156.1946902654867, "grad_norm": 0.09117792546749115, "learning_rate": 9.568914639231807e-05, "loss": 0.0042, "step": 17650 }, { "epoch": 156.28318584070797, "grad_norm": 0.10809333622455597, "learning_rate": 9.568242745255852e-05, "loss": 0.004, "step": 17660 }, { "epoch": 156.3716814159292, "grad_norm": 0.0819663554430008, "learning_rate": 9.567570351703739e-05, "loss": 0.0041, "step": 17670 }, { "epoch": 156.46017699115043, "grad_norm": 0.07318873703479767, "learning_rate": 9.566897458649001e-05, "loss": 0.0041, "step": 17680 }, { "epoch": 156.5486725663717, "grad_norm": 0.09827159345149994, "learning_rate": 9.566224066165221e-05, "loss": 0.0036, "step": 17690 }, { "epoch": 156.63716814159292, "grad_norm": 0.08058913797140121, "learning_rate": 9.565550174326043e-05, "loss": 0.0042, "step": 17700 }, { "epoch": 156.72566371681415, "grad_norm": 0.07254812121391296, "learning_rate": 9.564875783205162e-05, "loss": 0.0042, "step": 17710 }, { "epoch": 156.81415929203538, "grad_norm": 0.08068995177745819, "learning_rate": 9.564200892876328e-05, "loss": 0.0045, "step": 17720 }, { "epoch": 156.90265486725664, "grad_norm": 0.08793877065181732, "learning_rate": 9.563525503413348e-05, "loss": 0.0041, "step": 17730 }, { "epoch": 156.99115044247787, "grad_norm": 0.07917606830596924, "learning_rate": 9.562849614890079e-05, "loss": 0.0049, "step": 17740 }, { "epoch": 157.0796460176991, "grad_norm": 0.0970199927687645, "learning_rate": 9.562173227380436e-05, "loss": 0.0046, "step": 17750 }, { "epoch": 157.16814159292036, "grad_norm": 0.104884572327137, "learning_rate": 9.561496340958389e-05, "loss": 0.0045, "step": 17760 }, { "epoch": 157.2566371681416, "grad_norm": 0.08023256063461304, "learning_rate": 9.560818955697959e-05, "loss": 0.0046, "step": 17770 }, { "epoch": 157.34513274336283, "grad_norm": 0.07891476899385452, "learning_rate": 9.560141071673228e-05, "loss": 0.0043, "step": 17780 }, { "epoch": 157.43362831858408, "grad_norm": 0.08542822301387787, "learning_rate": 9.559462688958323e-05, "loss": 0.0046, "step": 17790 }, { "epoch": 157.52212389380531, "grad_norm": 0.08797013759613037, "learning_rate": 9.558783807627434e-05, "loss": 0.0045, "step": 17800 }, { "epoch": 157.61061946902655, "grad_norm": 0.09226534515619278, "learning_rate": 9.558104427754801e-05, "loss": 0.004, "step": 17810 }, { "epoch": 157.69911504424778, "grad_norm": 0.08366729319095612, "learning_rate": 9.557424549414722e-05, "loss": 0.004, "step": 17820 }, { "epoch": 157.78761061946904, "grad_norm": 0.0687665194272995, "learning_rate": 9.556744172681546e-05, "loss": 0.0037, "step": 17830 }, { "epoch": 157.87610619469027, "grad_norm": 0.09219760447740555, "learning_rate": 9.556063297629677e-05, "loss": 0.0044, "step": 17840 }, { "epoch": 157.9646017699115, "grad_norm": 0.0749521553516388, "learning_rate": 9.555381924333578e-05, "loss": 0.004, "step": 17850 }, { "epoch": 158.05309734513276, "grad_norm": 0.09271930158138275, "learning_rate": 9.554700052867758e-05, "loss": 0.0043, "step": 17860 }, { "epoch": 158.141592920354, "grad_norm": 0.08980640023946762, "learning_rate": 9.554017683306789e-05, "loss": 0.0045, "step": 17870 }, { "epoch": 158.23008849557522, "grad_norm": 0.11064504086971283, "learning_rate": 9.553334815725294e-05, "loss": 0.0047, "step": 17880 }, { "epoch": 158.31858407079645, "grad_norm": 0.07996485382318497, "learning_rate": 9.552651450197949e-05, "loss": 0.0043, "step": 17890 }, { "epoch": 158.4070796460177, "grad_norm": 0.09444921463727951, "learning_rate": 9.551967586799486e-05, "loss": 0.0039, "step": 17900 }, { "epoch": 158.49557522123894, "grad_norm": 0.0995616763830185, "learning_rate": 9.551283225604692e-05, "loss": 0.0042, "step": 17910 }, { "epoch": 158.58407079646017, "grad_norm": 0.10688190162181854, "learning_rate": 9.550598366688406e-05, "loss": 0.0041, "step": 17920 }, { "epoch": 158.67256637168143, "grad_norm": 0.12246780842542648, "learning_rate": 9.549913010125526e-05, "loss": 0.0043, "step": 17930 }, { "epoch": 158.76106194690266, "grad_norm": 0.08677249401807785, "learning_rate": 9.549227155990999e-05, "loss": 0.0051, "step": 17940 }, { "epoch": 158.8495575221239, "grad_norm": 0.10012728720903397, "learning_rate": 9.548540804359828e-05, "loss": 0.0044, "step": 17950 }, { "epoch": 158.93805309734512, "grad_norm": 0.10381681472063065, "learning_rate": 9.547853955307077e-05, "loss": 0.0039, "step": 17960 }, { "epoch": 159.02654867256638, "grad_norm": 0.0857657715678215, "learning_rate": 9.547166608907853e-05, "loss": 0.0041, "step": 17970 }, { "epoch": 159.1150442477876, "grad_norm": 0.0828576385974884, "learning_rate": 9.546478765237326e-05, "loss": 0.0042, "step": 17980 }, { "epoch": 159.20353982300884, "grad_norm": 0.08854452520608902, "learning_rate": 9.545790424370715e-05, "loss": 0.0045, "step": 17990 }, { "epoch": 159.2920353982301, "grad_norm": 0.09095236659049988, "learning_rate": 9.5451015863833e-05, "loss": 0.0045, "step": 18000 }, { "epoch": 159.38053097345133, "grad_norm": 0.09899206459522247, "learning_rate": 9.544412251350408e-05, "loss": 0.0042, "step": 18010 }, { "epoch": 159.46902654867256, "grad_norm": 0.09673959016799927, "learning_rate": 9.543722419347422e-05, "loss": 0.0039, "step": 18020 }, { "epoch": 159.55752212389382, "grad_norm": 0.10539836436510086, "learning_rate": 9.543032090449788e-05, "loss": 0.0044, "step": 18030 }, { "epoch": 159.64601769911505, "grad_norm": 0.09337548911571503, "learning_rate": 9.542341264732992e-05, "loss": 0.0046, "step": 18040 }, { "epoch": 159.73451327433628, "grad_norm": 0.09492888301610947, "learning_rate": 9.541649942272585e-05, "loss": 0.0041, "step": 18050 }, { "epoch": 159.8230088495575, "grad_norm": 0.11635717749595642, "learning_rate": 9.54095812314417e-05, "loss": 0.0042, "step": 18060 }, { "epoch": 159.91150442477877, "grad_norm": 0.08584687113761902, "learning_rate": 9.540265807423401e-05, "loss": 0.0039, "step": 18070 }, { "epoch": 160.0, "grad_norm": 0.056951262056827545, "learning_rate": 9.53957299518599e-05, "loss": 0.0041, "step": 18080 }, { "epoch": 160.08849557522123, "grad_norm": 0.09164301306009293, "learning_rate": 9.5388796865077e-05, "loss": 0.0042, "step": 18090 }, { "epoch": 160.1769911504425, "grad_norm": 0.09685871750116348, "learning_rate": 9.538185881464353e-05, "loss": 0.0042, "step": 18100 }, { "epoch": 160.26548672566372, "grad_norm": 0.07035042345523834, "learning_rate": 9.537491580131821e-05, "loss": 0.0041, "step": 18110 }, { "epoch": 160.35398230088495, "grad_norm": 0.08510271459817886, "learning_rate": 9.53679678258603e-05, "loss": 0.004, "step": 18120 }, { "epoch": 160.44247787610618, "grad_norm": 0.07103925198316574, "learning_rate": 9.536101488902966e-05, "loss": 0.0041, "step": 18130 }, { "epoch": 160.53097345132744, "grad_norm": 0.08885382860898972, "learning_rate": 9.535405699158663e-05, "loss": 0.0047, "step": 18140 }, { "epoch": 160.61946902654867, "grad_norm": 0.09845031052827835, "learning_rate": 9.53470941342921e-05, "loss": 0.0042, "step": 18150 }, { "epoch": 160.7079646017699, "grad_norm": 0.09310560673475266, "learning_rate": 9.534012631790756e-05, "loss": 0.0039, "step": 18160 }, { "epoch": 160.79646017699116, "grad_norm": 0.07304170727729797, "learning_rate": 9.533315354319494e-05, "loss": 0.0044, "step": 18170 }, { "epoch": 160.8849557522124, "grad_norm": 0.094046950340271, "learning_rate": 9.532617581091682e-05, "loss": 0.0043, "step": 18180 }, { "epoch": 160.97345132743362, "grad_norm": 0.09146838635206223, "learning_rate": 9.531919312183629e-05, "loss": 0.0043, "step": 18190 }, { "epoch": 161.06194690265488, "grad_norm": 0.09784809499979019, "learning_rate": 9.531220547671688e-05, "loss": 0.0037, "step": 18200 }, { "epoch": 161.1504424778761, "grad_norm": 0.08099901676177979, "learning_rate": 9.530521287632285e-05, "loss": 0.0041, "step": 18210 }, { "epoch": 161.23893805309734, "grad_norm": 0.08826098591089249, "learning_rate": 9.529821532141884e-05, "loss": 0.0038, "step": 18220 }, { "epoch": 161.32743362831857, "grad_norm": 0.07766509056091309, "learning_rate": 9.52912128127701e-05, "loss": 0.0045, "step": 18230 }, { "epoch": 161.41592920353983, "grad_norm": 0.0693865418434143, "learning_rate": 9.528420535114244e-05, "loss": 0.0039, "step": 18240 }, { "epoch": 161.50442477876106, "grad_norm": 0.0829935371875763, "learning_rate": 9.527719293730215e-05, "loss": 0.0038, "step": 18250 }, { "epoch": 161.5929203539823, "grad_norm": 0.09733983874320984, "learning_rate": 9.527017557201611e-05, "loss": 0.004, "step": 18260 }, { "epoch": 161.68141592920355, "grad_norm": 0.08739615231752396, "learning_rate": 9.526315325605176e-05, "loss": 0.0045, "step": 18270 }, { "epoch": 161.76991150442478, "grad_norm": 0.07029226422309875, "learning_rate": 9.525612599017699e-05, "loss": 0.0038, "step": 18280 }, { "epoch": 161.858407079646, "grad_norm": 0.07799669355154037, "learning_rate": 9.524909377516033e-05, "loss": 0.0039, "step": 18290 }, { "epoch": 161.94690265486724, "grad_norm": 0.07842861115932465, "learning_rate": 9.524205661177081e-05, "loss": 0.0046, "step": 18300 }, { "epoch": 162.0353982300885, "grad_norm": 0.07585214823484421, "learning_rate": 9.523501450077801e-05, "loss": 0.0041, "step": 18310 }, { "epoch": 162.12389380530973, "grad_norm": 0.08480408787727356, "learning_rate": 9.522796744295202e-05, "loss": 0.0041, "step": 18320 }, { "epoch": 162.21238938053096, "grad_norm": 0.09379900991916656, "learning_rate": 9.522091543906352e-05, "loss": 0.0042, "step": 18330 }, { "epoch": 162.30088495575222, "grad_norm": 0.07346818596124649, "learning_rate": 9.521385848988369e-05, "loss": 0.004, "step": 18340 }, { "epoch": 162.38938053097345, "grad_norm": 0.07661738991737366, "learning_rate": 9.520679659618428e-05, "loss": 0.0045, "step": 18350 }, { "epoch": 162.47787610619469, "grad_norm": 0.07569757103919983, "learning_rate": 9.519972975873754e-05, "loss": 0.0043, "step": 18360 }, { "epoch": 162.56637168141592, "grad_norm": 0.08861418813467026, "learning_rate": 9.519265797831633e-05, "loss": 0.0044, "step": 18370 }, { "epoch": 162.65486725663717, "grad_norm": 0.0929049625992775, "learning_rate": 9.518558125569399e-05, "loss": 0.0041, "step": 18380 }, { "epoch": 162.7433628318584, "grad_norm": 0.07228565961122513, "learning_rate": 9.517849959164442e-05, "loss": 0.0037, "step": 18390 }, { "epoch": 162.83185840707964, "grad_norm": 0.07522029429674149, "learning_rate": 9.517141298694205e-05, "loss": 0.0041, "step": 18400 }, { "epoch": 162.9203539823009, "grad_norm": 0.07942932099103928, "learning_rate": 9.516432144236188e-05, "loss": 0.0036, "step": 18410 }, { "epoch": 163.00884955752213, "grad_norm": 0.07833880931138992, "learning_rate": 9.515722495867941e-05, "loss": 0.0039, "step": 18420 }, { "epoch": 163.09734513274336, "grad_norm": 0.09451162070035934, "learning_rate": 9.515012353667072e-05, "loss": 0.0041, "step": 18430 }, { "epoch": 163.18584070796462, "grad_norm": 0.09868040680885315, "learning_rate": 9.51430171771124e-05, "loss": 0.0048, "step": 18440 }, { "epoch": 163.27433628318585, "grad_norm": 0.0822039395570755, "learning_rate": 9.513590588078159e-05, "loss": 0.0042, "step": 18450 }, { "epoch": 163.36283185840708, "grad_norm": 0.06876639276742935, "learning_rate": 9.512878964845597e-05, "loss": 0.0044, "step": 18460 }, { "epoch": 163.4513274336283, "grad_norm": 0.06805103272199631, "learning_rate": 9.512166848091377e-05, "loss": 0.0044, "step": 18470 }, { "epoch": 163.53982300884957, "grad_norm": 0.09280479699373245, "learning_rate": 9.511454237893376e-05, "loss": 0.004, "step": 18480 }, { "epoch": 163.6283185840708, "grad_norm": 0.08265211433172226, "learning_rate": 9.51074113432952e-05, "loss": 0.0041, "step": 18490 }, { "epoch": 163.71681415929203, "grad_norm": 0.07943312078714371, "learning_rate": 9.510027537477797e-05, "loss": 0.0041, "step": 18500 }, { "epoch": 163.8053097345133, "grad_norm": 0.09830750524997711, "learning_rate": 9.509313447416242e-05, "loss": 0.0042, "step": 18510 }, { "epoch": 163.89380530973452, "grad_norm": 0.07099904865026474, "learning_rate": 9.508598864222949e-05, "loss": 0.0039, "step": 18520 }, { "epoch": 163.98230088495575, "grad_norm": 0.08108550310134888, "learning_rate": 9.507883787976062e-05, "loss": 0.0039, "step": 18530 }, { "epoch": 164.07079646017698, "grad_norm": 0.07927399128675461, "learning_rate": 9.507168218753781e-05, "loss": 0.0039, "step": 18540 }, { "epoch": 164.15929203539824, "grad_norm": 0.07294873893260956, "learning_rate": 9.506452156634362e-05, "loss": 0.004, "step": 18550 }, { "epoch": 164.24778761061947, "grad_norm": 0.06878101825714111, "learning_rate": 9.505735601696109e-05, "loss": 0.0044, "step": 18560 }, { "epoch": 164.3362831858407, "grad_norm": 0.09136318415403366, "learning_rate": 9.505018554017385e-05, "loss": 0.0043, "step": 18570 }, { "epoch": 164.42477876106196, "grad_norm": 0.07686548680067062, "learning_rate": 9.504301013676604e-05, "loss": 0.0036, "step": 18580 }, { "epoch": 164.5132743362832, "grad_norm": 0.0986580178141594, "learning_rate": 9.503582980752238e-05, "loss": 0.0041, "step": 18590 }, { "epoch": 164.60176991150442, "grad_norm": 0.08796390891075134, "learning_rate": 9.502864455322809e-05, "loss": 0.0047, "step": 18600 }, { "epoch": 164.69026548672565, "grad_norm": 0.08720848709344864, "learning_rate": 9.502145437466891e-05, "loss": 0.0046, "step": 18610 }, { "epoch": 164.7787610619469, "grad_norm": 0.08983920514583588, "learning_rate": 9.501425927263116e-05, "loss": 0.0042, "step": 18620 }, { "epoch": 164.86725663716814, "grad_norm": 0.0793861448764801, "learning_rate": 9.500705924790172e-05, "loss": 0.004, "step": 18630 }, { "epoch": 164.95575221238937, "grad_norm": 0.0797121673822403, "learning_rate": 9.499985430126794e-05, "loss": 0.0046, "step": 18640 }, { "epoch": 165.04424778761063, "grad_norm": 0.08993679285049438, "learning_rate": 9.499264443351775e-05, "loss": 0.0043, "step": 18650 }, { "epoch": 165.13274336283186, "grad_norm": 0.08226118981838226, "learning_rate": 9.498542964543961e-05, "loss": 0.0035, "step": 18660 }, { "epoch": 165.2212389380531, "grad_norm": 0.07073994725942612, "learning_rate": 9.497820993782252e-05, "loss": 0.004, "step": 18670 }, { "epoch": 165.30973451327435, "grad_norm": 0.07091126590967178, "learning_rate": 9.497098531145601e-05, "loss": 0.0039, "step": 18680 }, { "epoch": 165.39823008849558, "grad_norm": 0.07996762543916702, "learning_rate": 9.496375576713017e-05, "loss": 0.0044, "step": 18690 }, { "epoch": 165.4867256637168, "grad_norm": 0.11916101723909378, "learning_rate": 9.49565213056356e-05, "loss": 0.0042, "step": 18700 }, { "epoch": 165.57522123893804, "grad_norm": 0.10253284126520157, "learning_rate": 9.494928192776342e-05, "loss": 0.0043, "step": 18710 }, { "epoch": 165.6637168141593, "grad_norm": 0.081699900329113, "learning_rate": 9.494203763430538e-05, "loss": 0.0042, "step": 18720 }, { "epoch": 165.75221238938053, "grad_norm": 0.0912235677242279, "learning_rate": 9.493478842605366e-05, "loss": 0.0044, "step": 18730 }, { "epoch": 165.84070796460176, "grad_norm": 0.08015649020671844, "learning_rate": 9.492753430380105e-05, "loss": 0.0037, "step": 18740 }, { "epoch": 165.92920353982302, "grad_norm": 0.06691010296344757, "learning_rate": 9.492027526834083e-05, "loss": 0.0039, "step": 18750 }, { "epoch": 166.01769911504425, "grad_norm": 0.061063963919878006, "learning_rate": 9.491301132046684e-05, "loss": 0.0041, "step": 18760 }, { "epoch": 166.10619469026548, "grad_norm": 0.0702151283621788, "learning_rate": 9.490574246097345e-05, "loss": 0.0039, "step": 18770 }, { "epoch": 166.1946902654867, "grad_norm": 0.0625680461525917, "learning_rate": 9.48984686906556e-05, "loss": 0.0042, "step": 18780 }, { "epoch": 166.28318584070797, "grad_norm": 0.0910443365573883, "learning_rate": 9.489119001030871e-05, "loss": 0.0038, "step": 18790 }, { "epoch": 166.3716814159292, "grad_norm": 0.06484772264957428, "learning_rate": 9.488390642072878e-05, "loss": 0.0041, "step": 18800 }, { "epoch": 166.46017699115043, "grad_norm": 0.09850802272558212, "learning_rate": 9.48766179227123e-05, "loss": 0.0038, "step": 18810 }, { "epoch": 166.5486725663717, "grad_norm": 0.058781906962394714, "learning_rate": 9.486932451705636e-05, "loss": 0.0038, "step": 18820 }, { "epoch": 166.63716814159292, "grad_norm": 0.09541474282741547, "learning_rate": 9.486202620455857e-05, "loss": 0.0044, "step": 18830 }, { "epoch": 166.72566371681415, "grad_norm": 0.09085864573717117, "learning_rate": 9.485472298601704e-05, "loss": 0.0042, "step": 18840 }, { "epoch": 166.81415929203538, "grad_norm": 0.08480574935674667, "learning_rate": 9.484741486223043e-05, "loss": 0.004, "step": 18850 }, { "epoch": 166.90265486725664, "grad_norm": 0.0717371478676796, "learning_rate": 9.484010183399797e-05, "loss": 0.0042, "step": 18860 }, { "epoch": 166.99115044247787, "grad_norm": 0.08525565266609192, "learning_rate": 9.483278390211938e-05, "loss": 0.0046, "step": 18870 }, { "epoch": 167.0796460176991, "grad_norm": 0.08052633702754974, "learning_rate": 9.482546106739496e-05, "loss": 0.0042, "step": 18880 }, { "epoch": 167.16814159292036, "grad_norm": 0.07514137029647827, "learning_rate": 9.48181333306255e-05, "loss": 0.0044, "step": 18890 }, { "epoch": 167.2566371681416, "grad_norm": 0.05321681126952171, "learning_rate": 9.481080069261237e-05, "loss": 0.0041, "step": 18900 }, { "epoch": 167.34513274336283, "grad_norm": 0.10621104389429092, "learning_rate": 9.480346315415745e-05, "loss": 0.0046, "step": 18910 }, { "epoch": 167.43362831858408, "grad_norm": 0.08053136616945267, "learning_rate": 9.479612071606314e-05, "loss": 0.0039, "step": 18920 }, { "epoch": 167.52212389380531, "grad_norm": 0.08102152496576309, "learning_rate": 9.478877337913244e-05, "loss": 0.004, "step": 18930 }, { "epoch": 167.61061946902655, "grad_norm": 0.06869881600141525, "learning_rate": 9.478142114416881e-05, "loss": 0.0045, "step": 18940 }, { "epoch": 167.69911504424778, "grad_norm": 0.08891229331493378, "learning_rate": 9.47740640119763e-05, "loss": 0.0041, "step": 18950 }, { "epoch": 167.78761061946904, "grad_norm": 0.12049701809883118, "learning_rate": 9.476670198335947e-05, "loss": 0.0045, "step": 18960 }, { "epoch": 167.87610619469027, "grad_norm": 0.09915345907211304, "learning_rate": 9.47593350591234e-05, "loss": 0.004, "step": 18970 }, { "epoch": 167.9646017699115, "grad_norm": 0.08027512580156326, "learning_rate": 9.475196324007376e-05, "loss": 0.0043, "step": 18980 }, { "epoch": 168.05309734513276, "grad_norm": 0.08525636792182922, "learning_rate": 9.474458652701669e-05, "loss": 0.0042, "step": 18990 }, { "epoch": 168.141592920354, "grad_norm": 0.08083896338939667, "learning_rate": 9.473720492075892e-05, "loss": 0.0042, "step": 19000 }, { "epoch": 168.23008849557522, "grad_norm": 0.07597064971923828, "learning_rate": 9.472981842210768e-05, "loss": 0.005, "step": 19010 }, { "epoch": 168.31858407079645, "grad_norm": 0.0775659829378128, "learning_rate": 9.472242703187074e-05, "loss": 0.0037, "step": 19020 }, { "epoch": 168.4070796460177, "grad_norm": 0.07426168024539948, "learning_rate": 9.471503075085643e-05, "loss": 0.0041, "step": 19030 }, { "epoch": 168.49557522123894, "grad_norm": 0.10601849108934402, "learning_rate": 9.470762957987359e-05, "loss": 0.0044, "step": 19040 }, { "epoch": 168.58407079646017, "grad_norm": 0.09191782772541046, "learning_rate": 9.470022351973158e-05, "loss": 0.0041, "step": 19050 }, { "epoch": 168.67256637168143, "grad_norm": 0.08066841214895248, "learning_rate": 9.469281257124034e-05, "loss": 0.0038, "step": 19060 }, { "epoch": 168.76106194690266, "grad_norm": 0.06517547369003296, "learning_rate": 9.46853967352103e-05, "loss": 0.0043, "step": 19070 }, { "epoch": 168.8495575221239, "grad_norm": 0.081117182970047, "learning_rate": 9.467797601245246e-05, "loss": 0.0042, "step": 19080 }, { "epoch": 168.93805309734512, "grad_norm": 0.08300390094518661, "learning_rate": 9.467055040377834e-05, "loss": 0.004, "step": 19090 }, { "epoch": 169.02654867256638, "grad_norm": 0.08992671221494675, "learning_rate": 9.466311990999999e-05, "loss": 0.0038, "step": 19100 }, { "epoch": 169.1150442477876, "grad_norm": 0.11833495646715164, "learning_rate": 9.465568453193e-05, "loss": 0.0041, "step": 19110 }, { "epoch": 169.20353982300884, "grad_norm": 0.10707447677850723, "learning_rate": 9.464824427038148e-05, "loss": 0.0042, "step": 19120 }, { "epoch": 169.2920353982301, "grad_norm": 0.10828002542257309, "learning_rate": 9.46407991261681e-05, "loss": 0.0044, "step": 19130 }, { "epoch": 169.38053097345133, "grad_norm": 0.08549857139587402, "learning_rate": 9.463334910010404e-05, "loss": 0.0039, "step": 19140 }, { "epoch": 169.46902654867256, "grad_norm": 0.0827985554933548, "learning_rate": 9.462589419300403e-05, "loss": 0.004, "step": 19150 }, { "epoch": 169.55752212389382, "grad_norm": 0.06796583533287048, "learning_rate": 9.461843440568333e-05, "loss": 0.0037, "step": 19160 }, { "epoch": 169.64601769911505, "grad_norm": 0.10102774202823639, "learning_rate": 9.461096973895773e-05, "loss": 0.0042, "step": 19170 }, { "epoch": 169.73451327433628, "grad_norm": 0.09978064894676208, "learning_rate": 9.460350019364355e-05, "loss": 0.0042, "step": 19180 }, { "epoch": 169.8230088495575, "grad_norm": 0.08184050023555756, "learning_rate": 9.459602577055764e-05, "loss": 0.0044, "step": 19190 }, { "epoch": 169.91150442477877, "grad_norm": 0.09962663799524307, "learning_rate": 9.45885464705174e-05, "loss": 0.0048, "step": 19200 }, { "epoch": 170.0, "grad_norm": 0.07746028155088425, "learning_rate": 9.458106229434076e-05, "loss": 0.0038, "step": 19210 }, { "epoch": 170.08849557522123, "grad_norm": 0.06921591609716415, "learning_rate": 9.457357324284617e-05, "loss": 0.004, "step": 19220 }, { "epoch": 170.1769911504425, "grad_norm": 0.08217597007751465, "learning_rate": 9.456607931685262e-05, "loss": 0.0037, "step": 19230 }, { "epoch": 170.26548672566372, "grad_norm": 0.08453896641731262, "learning_rate": 9.455858051717965e-05, "loss": 0.0035, "step": 19240 }, { "epoch": 170.35398230088495, "grad_norm": 0.07753022760152817, "learning_rate": 9.45510768446473e-05, "loss": 0.0037, "step": 19250 }, { "epoch": 170.44247787610618, "grad_norm": 0.08236990869045258, "learning_rate": 9.454356830007618e-05, "loss": 0.0044, "step": 19260 }, { "epoch": 170.53097345132744, "grad_norm": 0.0611865371465683, "learning_rate": 9.45360548842874e-05, "loss": 0.0041, "step": 19270 }, { "epoch": 170.61946902654867, "grad_norm": 0.07795462757349014, "learning_rate": 9.452853659810261e-05, "loss": 0.004, "step": 19280 }, { "epoch": 170.7079646017699, "grad_norm": 0.08893118053674698, "learning_rate": 9.452101344234401e-05, "loss": 0.0041, "step": 19290 }, { "epoch": 170.79646017699116, "grad_norm": 0.07845061272382736, "learning_rate": 9.451348541783431e-05, "loss": 0.0038, "step": 19300 }, { "epoch": 170.8849557522124, "grad_norm": 0.0894150659441948, "learning_rate": 9.450595252539678e-05, "loss": 0.0042, "step": 19310 }, { "epoch": 170.97345132743362, "grad_norm": 0.08306597173213959, "learning_rate": 9.449841476585518e-05, "loss": 0.0046, "step": 19320 }, { "epoch": 171.06194690265488, "grad_norm": 0.07859474420547485, "learning_rate": 9.449087214003384e-05, "loss": 0.0037, "step": 19330 }, { "epoch": 171.1504424778761, "grad_norm": 0.09534873068332672, "learning_rate": 9.448332464875765e-05, "loss": 0.0048, "step": 19340 }, { "epoch": 171.23893805309734, "grad_norm": 0.09124890714883804, "learning_rate": 9.447577229285192e-05, "loss": 0.0042, "step": 19350 }, { "epoch": 171.32743362831857, "grad_norm": 0.11949527263641357, "learning_rate": 9.446821507314261e-05, "loss": 0.0043, "step": 19360 }, { "epoch": 171.41592920353983, "grad_norm": 0.09522955119609833, "learning_rate": 9.446065299045617e-05, "loss": 0.0043, "step": 19370 }, { "epoch": 171.50442477876106, "grad_norm": 0.08563883602619171, "learning_rate": 9.445308604561955e-05, "loss": 0.004, "step": 19380 }, { "epoch": 171.5929203539823, "grad_norm": 0.07838291674852371, "learning_rate": 9.444551423946028e-05, "loss": 0.0041, "step": 19390 }, { "epoch": 171.68141592920355, "grad_norm": 0.07998950034379959, "learning_rate": 9.443793757280638e-05, "loss": 0.0036, "step": 19400 }, { "epoch": 171.76991150442478, "grad_norm": 0.0982682853937149, "learning_rate": 9.443035604648646e-05, "loss": 0.0044, "step": 19410 }, { "epoch": 171.858407079646, "grad_norm": 0.06853248178958893, "learning_rate": 9.44227696613296e-05, "loss": 0.004, "step": 19420 }, { "epoch": 171.94690265486724, "grad_norm": 0.09221836179494858, "learning_rate": 9.441517841816542e-05, "loss": 0.0044, "step": 19430 }, { "epoch": 172.0353982300885, "grad_norm": 0.08939971774816513, "learning_rate": 9.440758231782413e-05, "loss": 0.0043, "step": 19440 }, { "epoch": 172.12389380530973, "grad_norm": 0.09037607163190842, "learning_rate": 9.439998136113639e-05, "loss": 0.0039, "step": 19450 }, { "epoch": 172.21238938053096, "grad_norm": 0.08845403790473938, "learning_rate": 9.439237554893344e-05, "loss": 0.004, "step": 19460 }, { "epoch": 172.30088495575222, "grad_norm": 0.09238039702177048, "learning_rate": 9.438476488204705e-05, "loss": 0.004, "step": 19470 }, { "epoch": 172.38938053097345, "grad_norm": 0.07265423238277435, "learning_rate": 9.43771493613095e-05, "loss": 0.0037, "step": 19480 }, { "epoch": 172.47787610619469, "grad_norm": 0.06599529832601547, "learning_rate": 9.436952898755362e-05, "loss": 0.0042, "step": 19490 }, { "epoch": 172.56637168141592, "grad_norm": 0.06784096360206604, "learning_rate": 9.436190376161276e-05, "loss": 0.004, "step": 19500 }, { "epoch": 172.65486725663717, "grad_norm": 0.1020347848534584, "learning_rate": 9.43542736843208e-05, "loss": 0.0039, "step": 19510 }, { "epoch": 172.7433628318584, "grad_norm": 0.06800811737775803, "learning_rate": 9.434663875651216e-05, "loss": 0.0044, "step": 19520 }, { "epoch": 172.83185840707964, "grad_norm": 0.05789966508746147, "learning_rate": 9.433899897902177e-05, "loss": 0.0037, "step": 19530 }, { "epoch": 172.9203539823009, "grad_norm": 0.07562093436717987, "learning_rate": 9.433135435268511e-05, "loss": 0.0039, "step": 19540 }, { "epoch": 173.00884955752213, "grad_norm": 0.10311379283666611, "learning_rate": 9.432370487833819e-05, "loss": 0.0037, "step": 19550 }, { "epoch": 173.09734513274336, "grad_norm": 0.0989622250199318, "learning_rate": 9.431605055681756e-05, "loss": 0.0049, "step": 19560 }, { "epoch": 173.18584070796462, "grad_norm": 0.08348415791988373, "learning_rate": 9.430839138896026e-05, "loss": 0.0042, "step": 19570 }, { "epoch": 173.27433628318585, "grad_norm": 0.0677868127822876, "learning_rate": 9.43007273756039e-05, "loss": 0.0042, "step": 19580 }, { "epoch": 173.36283185840708, "grad_norm": 0.09345083683729172, "learning_rate": 9.429305851758658e-05, "loss": 0.0036, "step": 19590 }, { "epoch": 173.4513274336283, "grad_norm": 0.09111639857292175, "learning_rate": 9.428538481574699e-05, "loss": 0.0036, "step": 19600 }, { "epoch": 173.53982300884957, "grad_norm": 0.07490076869726181, "learning_rate": 9.42777062709243e-05, "loss": 0.0041, "step": 19610 }, { "epoch": 173.6283185840708, "grad_norm": 0.09672997146844864, "learning_rate": 9.427002288395821e-05, "loss": 0.0039, "step": 19620 }, { "epoch": 173.71681415929203, "grad_norm": 0.08480148762464523, "learning_rate": 9.426233465568898e-05, "loss": 0.0039, "step": 19630 }, { "epoch": 173.8053097345133, "grad_norm": 0.08265836536884308, "learning_rate": 9.42546415869574e-05, "loss": 0.0041, "step": 19640 }, { "epoch": 173.89380530973452, "grad_norm": 0.0863218680024147, "learning_rate": 9.424694367860473e-05, "loss": 0.0039, "step": 19650 }, { "epoch": 173.98230088495575, "grad_norm": 0.06450959295034409, "learning_rate": 9.423924093147284e-05, "loss": 0.004, "step": 19660 }, { "epoch": 174.07079646017698, "grad_norm": 0.07350517064332962, "learning_rate": 9.423153334640407e-05, "loss": 0.0034, "step": 19670 }, { "epoch": 174.15929203539824, "grad_norm": 0.08973131328821182, "learning_rate": 9.42238209242413e-05, "loss": 0.0039, "step": 19680 }, { "epoch": 174.24778761061947, "grad_norm": 0.07590024173259735, "learning_rate": 9.421610366582798e-05, "loss": 0.0038, "step": 19690 }, { "epoch": 174.3362831858407, "grad_norm": 0.07915940880775452, "learning_rate": 9.420838157200803e-05, "loss": 0.0043, "step": 19700 }, { "epoch": 174.42477876106196, "grad_norm": 0.0858997255563736, "learning_rate": 9.420065464362594e-05, "loss": 0.0039, "step": 19710 }, { "epoch": 174.5132743362832, "grad_norm": 0.08462238311767578, "learning_rate": 9.419292288152673e-05, "loss": 0.0042, "step": 19720 }, { "epoch": 174.60176991150442, "grad_norm": 0.07310474663972855, "learning_rate": 9.418518628655588e-05, "loss": 0.0038, "step": 19730 }, { "epoch": 174.69026548672565, "grad_norm": 0.0742570012807846, "learning_rate": 9.417744485955951e-05, "loss": 0.0045, "step": 19740 }, { "epoch": 174.7787610619469, "grad_norm": 0.1290748566389084, "learning_rate": 9.41696986013842e-05, "loss": 0.0041, "step": 19750 }, { "epoch": 174.86725663716814, "grad_norm": 0.09959506243467331, "learning_rate": 9.416194751287705e-05, "loss": 0.0037, "step": 19760 }, { "epoch": 174.95575221238937, "grad_norm": 0.10798613727092743, "learning_rate": 9.415419159488572e-05, "loss": 0.0039, "step": 19770 }, { "epoch": 175.04424778761063, "grad_norm": 0.09513808786869049, "learning_rate": 9.414643084825837e-05, "loss": 0.0044, "step": 19780 }, { "epoch": 175.13274336283186, "grad_norm": 0.09733065962791443, "learning_rate": 9.413866527384372e-05, "loss": 0.0043, "step": 19790 }, { "epoch": 175.2212389380531, "grad_norm": 0.0793900415301323, "learning_rate": 9.4130894872491e-05, "loss": 0.0045, "step": 19800 }, { "epoch": 175.30973451327435, "grad_norm": 0.06679559499025345, "learning_rate": 9.412311964504998e-05, "loss": 0.004, "step": 19810 }, { "epoch": 175.39823008849558, "grad_norm": 0.07884012162685394, "learning_rate": 9.411533959237091e-05, "loss": 0.0044, "step": 19820 }, { "epoch": 175.4867256637168, "grad_norm": 0.07481736689805984, "learning_rate": 9.410755471530464e-05, "loss": 0.0042, "step": 19830 }, { "epoch": 175.57522123893804, "grad_norm": 0.08440539985895157, "learning_rate": 9.40997650147025e-05, "loss": 0.0039, "step": 19840 }, { "epoch": 175.6637168141593, "grad_norm": 0.08772435039281845, "learning_rate": 9.409197049141637e-05, "loss": 0.0039, "step": 19850 }, { "epoch": 175.75221238938053, "grad_norm": 0.07907542586326599, "learning_rate": 9.408417114629863e-05, "loss": 0.0045, "step": 19860 }, { "epoch": 175.84070796460176, "grad_norm": 0.08277586102485657, "learning_rate": 9.40763669802022e-05, "loss": 0.0039, "step": 19870 }, { "epoch": 175.92920353982302, "grad_norm": 0.07221365720033646, "learning_rate": 9.406855799398056e-05, "loss": 0.0044, "step": 19880 }, { "epoch": 176.01769911504425, "grad_norm": 0.09694644808769226, "learning_rate": 9.406074418848767e-05, "loss": 0.0041, "step": 19890 }, { "epoch": 176.10619469026548, "grad_norm": 0.061376482248306274, "learning_rate": 9.405292556457805e-05, "loss": 0.0039, "step": 19900 }, { "epoch": 176.1946902654867, "grad_norm": 0.07396472990512848, "learning_rate": 9.404510212310671e-05, "loss": 0.0044, "step": 19910 }, { "epoch": 176.28318584070797, "grad_norm": 0.07590167969465256, "learning_rate": 9.403727386492924e-05, "loss": 0.0037, "step": 19920 }, { "epoch": 176.3716814159292, "grad_norm": 0.07043332606554031, "learning_rate": 9.40294407909017e-05, "loss": 0.0039, "step": 19930 }, { "epoch": 176.46017699115043, "grad_norm": 0.07913874089717865, "learning_rate": 9.40216029018807e-05, "loss": 0.0049, "step": 19940 }, { "epoch": 176.5486725663717, "grad_norm": 0.06852279603481293, "learning_rate": 9.401376019872338e-05, "loss": 0.0038, "step": 19950 }, { "epoch": 176.63716814159292, "grad_norm": 0.10703850537538528, "learning_rate": 9.400591268228746e-05, "loss": 0.0042, "step": 19960 }, { "epoch": 176.72566371681415, "grad_norm": 0.08993610739707947, "learning_rate": 9.399806035343106e-05, "loss": 0.0043, "step": 19970 }, { "epoch": 176.81415929203538, "grad_norm": 0.09097947180271149, "learning_rate": 9.399020321301294e-05, "loss": 0.0039, "step": 19980 }, { "epoch": 176.90265486725664, "grad_norm": 0.10059264302253723, "learning_rate": 9.398234126189234e-05, "loss": 0.004, "step": 19990 }, { "epoch": 176.99115044247787, "grad_norm": 0.06510476022958755, "learning_rate": 9.397447450092902e-05, "loss": 0.0041, "step": 20000 }, { "epoch": 177.0796460176991, "grad_norm": 0.08020340651273727, "learning_rate": 9.39666029309833e-05, "loss": 0.0034, "step": 20010 }, { "epoch": 177.16814159292036, "grad_norm": 0.07128261029720306, "learning_rate": 9.395872655291596e-05, "loss": 0.0039, "step": 20020 }, { "epoch": 177.2566371681416, "grad_norm": 0.07829936593770981, "learning_rate": 9.395084536758838e-05, "loss": 0.0041, "step": 20030 }, { "epoch": 177.34513274336283, "grad_norm": 0.08790192753076553, "learning_rate": 9.394295937586243e-05, "loss": 0.004, "step": 20040 }, { "epoch": 177.43362831858408, "grad_norm": 0.07871206104755402, "learning_rate": 9.393506857860052e-05, "loss": 0.004, "step": 20050 }, { "epoch": 177.52212389380531, "grad_norm": 0.07078137248754501, "learning_rate": 9.392717297666555e-05, "loss": 0.0039, "step": 20060 }, { "epoch": 177.61061946902655, "grad_norm": 0.07591986656188965, "learning_rate": 9.391927257092101e-05, "loss": 0.004, "step": 20070 }, { "epoch": 177.69911504424778, "grad_norm": 0.08053716272115707, "learning_rate": 9.391136736223085e-05, "loss": 0.0044, "step": 20080 }, { "epoch": 177.78761061946904, "grad_norm": 0.07328096777200699, "learning_rate": 9.390345735145956e-05, "loss": 0.0039, "step": 20090 }, { "epoch": 177.87610619469027, "grad_norm": 0.07759694755077362, "learning_rate": 9.389554253947219e-05, "loss": 0.0038, "step": 20100 }, { "epoch": 177.9646017699115, "grad_norm": 0.07374036312103271, "learning_rate": 9.388762292713428e-05, "loss": 0.0043, "step": 20110 }, { "epoch": 178.05309734513276, "grad_norm": 0.09014523029327393, "learning_rate": 9.38796985153119e-05, "loss": 0.0041, "step": 20120 }, { "epoch": 178.141592920354, "grad_norm": 0.1023508608341217, "learning_rate": 9.387176930487169e-05, "loss": 0.0044, "step": 20130 }, { "epoch": 178.23008849557522, "grad_norm": 0.09559142589569092, "learning_rate": 9.386383529668072e-05, "loss": 0.0037, "step": 20140 }, { "epoch": 178.31858407079645, "grad_norm": 0.08172408491373062, "learning_rate": 9.385589649160669e-05, "loss": 0.0043, "step": 20150 }, { "epoch": 178.4070796460177, "grad_norm": 0.06453653424978256, "learning_rate": 9.384795289051775e-05, "loss": 0.0042, "step": 20160 }, { "epoch": 178.49557522123894, "grad_norm": 0.07143473625183105, "learning_rate": 9.384000449428261e-05, "loss": 0.0038, "step": 20170 }, { "epoch": 178.58407079646017, "grad_norm": 0.06573064625263214, "learning_rate": 9.383205130377048e-05, "loss": 0.0039, "step": 20180 }, { "epoch": 178.67256637168143, "grad_norm": 0.06708468496799469, "learning_rate": 9.382409331985114e-05, "loss": 0.0036, "step": 20190 }, { "epoch": 178.76106194690266, "grad_norm": 0.0729638934135437, "learning_rate": 9.381613054339482e-05, "loss": 0.0035, "step": 20200 }, { "epoch": 178.8495575221239, "grad_norm": 0.06747657060623169, "learning_rate": 9.380816297527235e-05, "loss": 0.0041, "step": 20210 }, { "epoch": 178.93805309734512, "grad_norm": 0.06362009048461914, "learning_rate": 9.380019061635506e-05, "loss": 0.0035, "step": 20220 }, { "epoch": 179.02654867256638, "grad_norm": 0.06666436791419983, "learning_rate": 9.379221346751474e-05, "loss": 0.0034, "step": 20230 }, { "epoch": 179.1150442477876, "grad_norm": 0.09586484730243683, "learning_rate": 9.378423152962382e-05, "loss": 0.004, "step": 20240 }, { "epoch": 179.20353982300884, "grad_norm": 0.07238700985908508, "learning_rate": 9.377624480355517e-05, "loss": 0.0036, "step": 20250 }, { "epoch": 179.2920353982301, "grad_norm": 0.07268580049276352, "learning_rate": 9.376825329018219e-05, "loss": 0.0038, "step": 20260 }, { "epoch": 179.38053097345133, "grad_norm": 0.07987114787101746, "learning_rate": 9.376025699037884e-05, "loss": 0.0043, "step": 20270 }, { "epoch": 179.46902654867256, "grad_norm": 0.11818955093622208, "learning_rate": 9.37522559050196e-05, "loss": 0.0039, "step": 20280 }, { "epoch": 179.55752212389382, "grad_norm": 0.08061552047729492, "learning_rate": 9.37442500349794e-05, "loss": 0.0034, "step": 20290 }, { "epoch": 179.64601769911505, "grad_norm": 0.08475670218467712, "learning_rate": 9.373623938113381e-05, "loss": 0.0041, "step": 20300 }, { "epoch": 179.73451327433628, "grad_norm": 0.08405469357967377, "learning_rate": 9.372822394435883e-05, "loss": 0.0043, "step": 20310 }, { "epoch": 179.8230088495575, "grad_norm": 0.06625127792358398, "learning_rate": 9.372020372553102e-05, "loss": 0.0038, "step": 20320 }, { "epoch": 179.91150442477877, "grad_norm": 0.06287863105535507, "learning_rate": 9.371217872552746e-05, "loss": 0.0036, "step": 20330 }, { "epoch": 180.0, "grad_norm": 0.09477771073579788, "learning_rate": 9.370414894522576e-05, "loss": 0.0039, "step": 20340 }, { "epoch": 180.08849557522123, "grad_norm": 0.0731310024857521, "learning_rate": 9.369611438550406e-05, "loss": 0.0037, "step": 20350 }, { "epoch": 180.1769911504425, "grad_norm": 0.07393135130405426, "learning_rate": 9.368807504724095e-05, "loss": 0.0038, "step": 20360 }, { "epoch": 180.26548672566372, "grad_norm": 0.08532673120498657, "learning_rate": 9.368003093131565e-05, "loss": 0.0036, "step": 20370 }, { "epoch": 180.35398230088495, "grad_norm": 0.07947520911693573, "learning_rate": 9.367198203860785e-05, "loss": 0.004, "step": 20380 }, { "epoch": 180.44247787610618, "grad_norm": 0.0858626514673233, "learning_rate": 9.366392836999774e-05, "loss": 0.0035, "step": 20390 }, { "epoch": 180.53097345132744, "grad_norm": 0.11620592325925827, "learning_rate": 9.365586992636607e-05, "loss": 0.0042, "step": 20400 }, { "epoch": 180.61946902654867, "grad_norm": 0.10247030109167099, "learning_rate": 9.364780670859412e-05, "loss": 0.0041, "step": 20410 }, { "epoch": 180.7079646017699, "grad_norm": 0.08185840398073196, "learning_rate": 9.363973871756364e-05, "loss": 0.0044, "step": 20420 }, { "epoch": 180.79646017699116, "grad_norm": 0.1120855063199997, "learning_rate": 9.363166595415696e-05, "loss": 0.0039, "step": 20430 }, { "epoch": 180.8849557522124, "grad_norm": 0.09527669847011566, "learning_rate": 9.362358841925686e-05, "loss": 0.0038, "step": 20440 }, { "epoch": 180.97345132743362, "grad_norm": 0.09735829383134842, "learning_rate": 9.361550611374674e-05, "loss": 0.0041, "step": 20450 }, { "epoch": 181.06194690265488, "grad_norm": 0.07887858897447586, "learning_rate": 9.360741903851043e-05, "loss": 0.0045, "step": 20460 }, { "epoch": 181.1504424778761, "grad_norm": 0.09639515727758408, "learning_rate": 9.359932719443236e-05, "loss": 0.0041, "step": 20470 }, { "epoch": 181.23893805309734, "grad_norm": 0.0962006002664566, "learning_rate": 9.35912305823974e-05, "loss": 0.004, "step": 20480 }, { "epoch": 181.32743362831857, "grad_norm": 0.0888102799654007, "learning_rate": 9.358312920329101e-05, "loss": 0.0038, "step": 20490 }, { "epoch": 181.41592920353983, "grad_norm": 0.0716932937502861, "learning_rate": 9.357502305799914e-05, "loss": 0.0037, "step": 20500 }, { "epoch": 181.50442477876106, "grad_norm": 0.09295344352722168, "learning_rate": 9.356691214740824e-05, "loss": 0.0039, "step": 20510 }, { "epoch": 181.5929203539823, "grad_norm": 0.11772622168064117, "learning_rate": 9.355879647240535e-05, "loss": 0.0037, "step": 20520 }, { "epoch": 181.68141592920355, "grad_norm": 0.10872572660446167, "learning_rate": 9.355067603387798e-05, "loss": 0.0038, "step": 20530 }, { "epoch": 181.76991150442478, "grad_norm": 0.0688805803656578, "learning_rate": 9.354255083271412e-05, "loss": 0.0042, "step": 20540 }, { "epoch": 181.858407079646, "grad_norm": 0.07946322858333588, "learning_rate": 9.353442086980239e-05, "loss": 0.0034, "step": 20550 }, { "epoch": 181.94690265486724, "grad_norm": 0.09623207151889801, "learning_rate": 9.352628614603185e-05, "loss": 0.0047, "step": 20560 }, { "epoch": 182.0353982300885, "grad_norm": 0.07878690958023071, "learning_rate": 9.351814666229209e-05, "loss": 0.0035, "step": 20570 }, { "epoch": 182.12389380530973, "grad_norm": 0.08772572129964828, "learning_rate": 9.351000241947324e-05, "loss": 0.0038, "step": 20580 }, { "epoch": 182.21238938053096, "grad_norm": 0.08872461318969727, "learning_rate": 9.350185341846594e-05, "loss": 0.0038, "step": 20590 }, { "epoch": 182.30088495575222, "grad_norm": 0.08488435298204422, "learning_rate": 9.349369966016134e-05, "loss": 0.0035, "step": 20600 }, { "epoch": 182.38938053097345, "grad_norm": 0.09878759831190109, "learning_rate": 9.348554114545117e-05, "loss": 0.0036, "step": 20610 }, { "epoch": 182.47787610619469, "grad_norm": 0.06845632195472717, "learning_rate": 9.347737787522758e-05, "loss": 0.0041, "step": 20620 }, { "epoch": 182.56637168141592, "grad_norm": 0.06875836849212646, "learning_rate": 9.346920985038332e-05, "loss": 0.0039, "step": 20630 }, { "epoch": 182.65486725663717, "grad_norm": 0.0993909165263176, "learning_rate": 9.346103707181162e-05, "loss": 0.0038, "step": 20640 }, { "epoch": 182.7433628318584, "grad_norm": 0.07137247174978256, "learning_rate": 9.345285954040626e-05, "loss": 0.0038, "step": 20650 }, { "epoch": 182.83185840707964, "grad_norm": 0.07004869729280472, "learning_rate": 9.34446772570615e-05, "loss": 0.0043, "step": 20660 }, { "epoch": 182.9203539823009, "grad_norm": 0.07503717392683029, "learning_rate": 9.343649022267214e-05, "loss": 0.0043, "step": 20670 }, { "epoch": 183.00884955752213, "grad_norm": 0.09222564101219177, "learning_rate": 9.342829843813353e-05, "loss": 0.004, "step": 20680 }, { "epoch": 183.09734513274336, "grad_norm": 0.07718320190906525, "learning_rate": 9.342010190434149e-05, "loss": 0.004, "step": 20690 }, { "epoch": 183.18584070796462, "grad_norm": 0.10429288446903229, "learning_rate": 9.34119006221924e-05, "loss": 0.0044, "step": 20700 }, { "epoch": 183.27433628318585, "grad_norm": 0.1032039001584053, "learning_rate": 9.340369459258313e-05, "loss": 0.0042, "step": 20710 }, { "epoch": 183.36283185840708, "grad_norm": 0.09556618332862854, "learning_rate": 9.339548381641106e-05, "loss": 0.0045, "step": 20720 }, { "epoch": 183.4513274336283, "grad_norm": 0.06444036960601807, "learning_rate": 9.338726829457413e-05, "loss": 0.0042, "step": 20730 }, { "epoch": 183.53982300884957, "grad_norm": 0.06911274790763855, "learning_rate": 9.337904802797078e-05, "loss": 0.0038, "step": 20740 }, { "epoch": 183.6283185840708, "grad_norm": 0.08496007323265076, "learning_rate": 9.337082301749993e-05, "loss": 0.0043, "step": 20750 }, { "epoch": 183.71681415929203, "grad_norm": 0.08296088129281998, "learning_rate": 9.336259326406109e-05, "loss": 0.0039, "step": 20760 }, { "epoch": 183.8053097345133, "grad_norm": 0.09526020288467407, "learning_rate": 9.335435876855427e-05, "loss": 0.0038, "step": 20770 }, { "epoch": 183.89380530973452, "grad_norm": 0.08453704416751862, "learning_rate": 9.334611953187994e-05, "loss": 0.0037, "step": 20780 }, { "epoch": 183.98230088495575, "grad_norm": 0.08331198990345001, "learning_rate": 9.333787555493914e-05, "loss": 0.0041, "step": 20790 }, { "epoch": 184.07079646017698, "grad_norm": 0.07319522649049759, "learning_rate": 9.332962683863345e-05, "loss": 0.0038, "step": 20800 }, { "epoch": 184.15929203539824, "grad_norm": 0.06976374983787537, "learning_rate": 9.332137338386489e-05, "loss": 0.0039, "step": 20810 }, { "epoch": 184.24778761061947, "grad_norm": 0.10796578973531723, "learning_rate": 9.33131151915361e-05, "loss": 0.0042, "step": 20820 }, { "epoch": 184.3362831858407, "grad_norm": 0.07061614841222763, "learning_rate": 9.330485226255012e-05, "loss": 0.0037, "step": 20830 }, { "epoch": 184.42477876106196, "grad_norm": 0.061021968722343445, "learning_rate": 9.329658459781061e-05, "loss": 0.004, "step": 20840 }, { "epoch": 184.5132743362832, "grad_norm": 0.07236694544553757, "learning_rate": 9.328831219822172e-05, "loss": 0.0038, "step": 20850 }, { "epoch": 184.60176991150442, "grad_norm": 0.07362212240695953, "learning_rate": 9.328003506468808e-05, "loss": 0.0038, "step": 20860 }, { "epoch": 184.69026548672565, "grad_norm": 0.054903242737054825, "learning_rate": 9.327175319811488e-05, "loss": 0.0037, "step": 20870 }, { "epoch": 184.7787610619469, "grad_norm": 0.07023978233337402, "learning_rate": 9.326346659940781e-05, "loss": 0.0035, "step": 20880 }, { "epoch": 184.86725663716814, "grad_norm": 0.09310893714427948, "learning_rate": 9.325517526947308e-05, "loss": 0.0038, "step": 20890 }, { "epoch": 184.95575221238937, "grad_norm": 0.0996529757976532, "learning_rate": 9.32468792092174e-05, "loss": 0.0044, "step": 20900 }, { "epoch": 185.04424778761063, "grad_norm": 0.07990477234125137, "learning_rate": 9.323857841954803e-05, "loss": 0.0039, "step": 20910 }, { "epoch": 185.13274336283186, "grad_norm": 0.07692430913448334, "learning_rate": 9.323027290137276e-05, "loss": 0.0034, "step": 20920 }, { "epoch": 185.2212389380531, "grad_norm": 0.0680701732635498, "learning_rate": 9.322196265559981e-05, "loss": 0.0039, "step": 20930 }, { "epoch": 185.30973451327435, "grad_norm": 0.0800628587603569, "learning_rate": 9.321364768313803e-05, "loss": 0.0037, "step": 20940 }, { "epoch": 185.39823008849558, "grad_norm": 0.07966773957014084, "learning_rate": 9.32053279848967e-05, "loss": 0.0039, "step": 20950 }, { "epoch": 185.4867256637168, "grad_norm": 0.10012353956699371, "learning_rate": 9.319700356178567e-05, "loss": 0.0044, "step": 20960 }, { "epoch": 185.57522123893804, "grad_norm": 0.08024784177541733, "learning_rate": 9.318867441471527e-05, "loss": 0.0037, "step": 20970 }, { "epoch": 185.6637168141593, "grad_norm": 0.10217513144016266, "learning_rate": 9.318034054459637e-05, "loss": 0.0037, "step": 20980 }, { "epoch": 185.75221238938053, "grad_norm": 0.07723687589168549, "learning_rate": 9.317200195234034e-05, "loss": 0.0041, "step": 20990 }, { "epoch": 185.84070796460176, "grad_norm": 0.08617885410785675, "learning_rate": 9.316365863885909e-05, "loss": 0.0039, "step": 21000 }, { "epoch": 185.92920353982302, "grad_norm": 0.07418305426836014, "learning_rate": 9.315531060506502e-05, "loss": 0.0037, "step": 21010 }, { "epoch": 186.01769911504425, "grad_norm": 0.07359225302934647, "learning_rate": 9.314695785187108e-05, "loss": 0.0042, "step": 21020 }, { "epoch": 186.10619469026548, "grad_norm": 0.09720060974359512, "learning_rate": 9.313860038019069e-05, "loss": 0.0037, "step": 21030 }, { "epoch": 186.1946902654867, "grad_norm": 0.07933292537927628, "learning_rate": 9.313023819093782e-05, "loss": 0.0041, "step": 21040 }, { "epoch": 186.28318584070797, "grad_norm": 0.0815187320113182, "learning_rate": 9.312187128502695e-05, "loss": 0.0038, "step": 21050 }, { "epoch": 186.3716814159292, "grad_norm": 0.0864122286438942, "learning_rate": 9.311349966337307e-05, "loss": 0.0041, "step": 21060 }, { "epoch": 186.46017699115043, "grad_norm": 0.10588196665048599, "learning_rate": 9.310512332689169e-05, "loss": 0.0038, "step": 21070 }, { "epoch": 186.5486725663717, "grad_norm": 0.08149340003728867, "learning_rate": 9.309674227649883e-05, "loss": 0.0038, "step": 21080 }, { "epoch": 186.63716814159292, "grad_norm": 0.07616796344518661, "learning_rate": 9.308835651311103e-05, "loss": 0.0039, "step": 21090 }, { "epoch": 186.72566371681415, "grad_norm": 0.08726868033409119, "learning_rate": 9.307996603764533e-05, "loss": 0.0035, "step": 21100 }, { "epoch": 186.81415929203538, "grad_norm": 0.06776058673858643, "learning_rate": 9.307157085101932e-05, "loss": 0.0032, "step": 21110 }, { "epoch": 186.90265486725664, "grad_norm": 0.05962707847356796, "learning_rate": 9.306317095415109e-05, "loss": 0.0033, "step": 21120 }, { "epoch": 186.99115044247787, "grad_norm": 0.07708761841058731, "learning_rate": 9.305476634795922e-05, "loss": 0.0038, "step": 21130 }, { "epoch": 187.0796460176991, "grad_norm": 0.0638175830245018, "learning_rate": 9.304635703336284e-05, "loss": 0.0041, "step": 21140 }, { "epoch": 187.16814159292036, "grad_norm": 0.08063431084156036, "learning_rate": 9.303794301128157e-05, "loss": 0.0037, "step": 21150 }, { "epoch": 187.2566371681416, "grad_norm": 0.08112648874521255, "learning_rate": 9.302952428263555e-05, "loss": 0.0037, "step": 21160 }, { "epoch": 187.34513274336283, "grad_norm": 0.0833049938082695, "learning_rate": 9.302110084834545e-05, "loss": 0.0042, "step": 21170 }, { "epoch": 187.43362831858408, "grad_norm": 0.08205411583185196, "learning_rate": 9.301267270933245e-05, "loss": 0.0033, "step": 21180 }, { "epoch": 187.52212389380531, "grad_norm": 0.09973905235528946, "learning_rate": 9.300423986651823e-05, "loss": 0.0042, "step": 21190 }, { "epoch": 187.61061946902655, "grad_norm": 0.06118624284863472, "learning_rate": 9.299580232082501e-05, "loss": 0.0039, "step": 21200 }, { "epoch": 187.69911504424778, "grad_norm": 0.08033032715320587, "learning_rate": 9.298736007317547e-05, "loss": 0.004, "step": 21210 }, { "epoch": 187.78761061946904, "grad_norm": 0.09097793698310852, "learning_rate": 9.297891312449288e-05, "loss": 0.0038, "step": 21220 }, { "epoch": 187.87610619469027, "grad_norm": 0.07428064942359924, "learning_rate": 9.297046147570094e-05, "loss": 0.0043, "step": 21230 }, { "epoch": 187.9646017699115, "grad_norm": 0.08478528261184692, "learning_rate": 9.296200512772396e-05, "loss": 0.0037, "step": 21240 }, { "epoch": 188.05309734513276, "grad_norm": 0.07276212424039841, "learning_rate": 9.295354408148668e-05, "loss": 0.0039, "step": 21250 }, { "epoch": 188.141592920354, "grad_norm": 0.09267200529575348, "learning_rate": 9.294507833791441e-05, "loss": 0.0037, "step": 21260 }, { "epoch": 188.23008849557522, "grad_norm": 0.08443468064069748, "learning_rate": 9.293660789793295e-05, "loss": 0.0037, "step": 21270 }, { "epoch": 188.31858407079645, "grad_norm": 0.08962515741586685, "learning_rate": 9.292813276246858e-05, "loss": 0.0046, "step": 21280 }, { "epoch": 188.4070796460177, "grad_norm": 0.09988413006067276, "learning_rate": 9.291965293244816e-05, "loss": 0.0036, "step": 21290 }, { "epoch": 188.49557522123894, "grad_norm": 0.05909333378076553, "learning_rate": 9.291116840879904e-05, "loss": 0.0037, "step": 21300 }, { "epoch": 188.58407079646017, "grad_norm": 0.061280105262994766, "learning_rate": 9.290267919244904e-05, "loss": 0.004, "step": 21310 }, { "epoch": 188.67256637168143, "grad_norm": 0.06884753704071045, "learning_rate": 9.289418528432655e-05, "loss": 0.0035, "step": 21320 }, { "epoch": 188.76106194690266, "grad_norm": 0.08523071557283401, "learning_rate": 9.288568668536045e-05, "loss": 0.0041, "step": 21330 }, { "epoch": 188.8495575221239, "grad_norm": 0.07894181460142136, "learning_rate": 9.287718339648013e-05, "loss": 0.0035, "step": 21340 }, { "epoch": 188.93805309734512, "grad_norm": 0.10104768723249435, "learning_rate": 9.28686754186155e-05, "loss": 0.0039, "step": 21350 }, { "epoch": 189.02654867256638, "grad_norm": 0.09707627445459366, "learning_rate": 9.286016275269698e-05, "loss": 0.0038, "step": 21360 }, { "epoch": 189.1150442477876, "grad_norm": 0.07473446428775787, "learning_rate": 9.285164539965551e-05, "loss": 0.0036, "step": 21370 }, { "epoch": 189.20353982300884, "grad_norm": 0.06168669834733009, "learning_rate": 9.284312336042251e-05, "loss": 0.0044, "step": 21380 }, { "epoch": 189.2920353982301, "grad_norm": 0.06950633972883224, "learning_rate": 9.283459663592996e-05, "loss": 0.0039, "step": 21390 }, { "epoch": 189.38053097345133, "grad_norm": 0.09130657464265823, "learning_rate": 9.282606522711033e-05, "loss": 0.0038, "step": 21400 }, { "epoch": 189.46902654867256, "grad_norm": 0.06765027344226837, "learning_rate": 9.281752913489657e-05, "loss": 0.004, "step": 21410 }, { "epoch": 189.55752212389382, "grad_norm": 0.09573005884885788, "learning_rate": 9.280898836022222e-05, "loss": 0.0037, "step": 21420 }, { "epoch": 189.64601769911505, "grad_norm": 0.07828065752983093, "learning_rate": 9.280044290402126e-05, "loss": 0.0035, "step": 21430 }, { "epoch": 189.73451327433628, "grad_norm": 0.08700332790613174, "learning_rate": 9.279189276722821e-05, "loss": 0.0038, "step": 21440 }, { "epoch": 189.8230088495575, "grad_norm": 0.07936663925647736, "learning_rate": 9.278333795077812e-05, "loss": 0.0042, "step": 21450 }, { "epoch": 189.91150442477877, "grad_norm": 0.09082474559545517, "learning_rate": 9.27747784556065e-05, "loss": 0.0043, "step": 21460 }, { "epoch": 190.0, "grad_norm": 0.08082649856805801, "learning_rate": 9.276621428264942e-05, "loss": 0.0037, "step": 21470 }, { "epoch": 190.08849557522123, "grad_norm": 0.06546024233102798, "learning_rate": 9.275764543284345e-05, "loss": 0.0038, "step": 21480 }, { "epoch": 190.1769911504425, "grad_norm": 0.07076419144868851, "learning_rate": 9.274907190712566e-05, "loss": 0.004, "step": 21490 }, { "epoch": 190.26548672566372, "grad_norm": 0.07917308062314987, "learning_rate": 9.274049370643363e-05, "loss": 0.0034, "step": 21500 }, { "epoch": 190.35398230088495, "grad_norm": 0.07609212398529053, "learning_rate": 9.273191083170547e-05, "loss": 0.0042, "step": 21510 }, { "epoch": 190.44247787610618, "grad_norm": 0.07720059156417847, "learning_rate": 9.27233232838798e-05, "loss": 0.0034, "step": 21520 }, { "epoch": 190.53097345132744, "grad_norm": 0.08922521770000458, "learning_rate": 9.27147310638957e-05, "loss": 0.0041, "step": 21530 }, { "epoch": 190.61946902654867, "grad_norm": 0.07834940403699875, "learning_rate": 9.270613417269286e-05, "loss": 0.0044, "step": 21540 }, { "epoch": 190.7079646017699, "grad_norm": 0.09716508537530899, "learning_rate": 9.269753261121138e-05, "loss": 0.0037, "step": 21550 }, { "epoch": 190.79646017699116, "grad_norm": 0.08082454651594162, "learning_rate": 9.268892638039194e-05, "loss": 0.0039, "step": 21560 }, { "epoch": 190.8849557522124, "grad_norm": 0.07883698493242264, "learning_rate": 9.268031548117569e-05, "loss": 0.0037, "step": 21570 }, { "epoch": 190.97345132743362, "grad_norm": 0.0638427659869194, "learning_rate": 9.26716999145043e-05, "loss": 0.0039, "step": 21580 }, { "epoch": 191.06194690265488, "grad_norm": 0.0913836881518364, "learning_rate": 9.266307968131998e-05, "loss": 0.0035, "step": 21590 }, { "epoch": 191.1504424778761, "grad_norm": 0.07262829691171646, "learning_rate": 9.26544547825654e-05, "loss": 0.0037, "step": 21600 }, { "epoch": 191.23893805309734, "grad_norm": 0.08343373239040375, "learning_rate": 9.264582521918376e-05, "loss": 0.0043, "step": 21610 }, { "epoch": 191.32743362831857, "grad_norm": 0.09689713269472122, "learning_rate": 9.263719099211881e-05, "loss": 0.004, "step": 21620 }, { "epoch": 191.41592920353983, "grad_norm": 0.06147231534123421, "learning_rate": 9.262855210231476e-05, "loss": 0.0035, "step": 21630 }, { "epoch": 191.50442477876106, "grad_norm": 0.09630943089723587, "learning_rate": 9.261990855071633e-05, "loss": 0.0038, "step": 21640 }, { "epoch": 191.5929203539823, "grad_norm": 0.07932459563016891, "learning_rate": 9.261126033826878e-05, "loss": 0.0035, "step": 21650 }, { "epoch": 191.68141592920355, "grad_norm": 0.09826038032770157, "learning_rate": 9.260260746591786e-05, "loss": 0.0038, "step": 21660 }, { "epoch": 191.76991150442478, "grad_norm": 0.09192106127738953, "learning_rate": 9.259394993460985e-05, "loss": 0.0041, "step": 21670 }, { "epoch": 191.858407079646, "grad_norm": 0.0990438312292099, "learning_rate": 9.258528774529151e-05, "loss": 0.0041, "step": 21680 }, { "epoch": 191.94690265486724, "grad_norm": 0.09524327516555786, "learning_rate": 9.257662089891013e-05, "loss": 0.004, "step": 21690 }, { "epoch": 192.0353982300885, "grad_norm": 0.10537286102771759, "learning_rate": 9.25679493964135e-05, "loss": 0.004, "step": 21700 }, { "epoch": 192.12389380530973, "grad_norm": 0.08624209463596344, "learning_rate": 9.255927323874994e-05, "loss": 0.004, "step": 21710 }, { "epoch": 192.21238938053096, "grad_norm": 0.1009453684091568, "learning_rate": 9.255059242686822e-05, "loss": 0.0039, "step": 21720 }, { "epoch": 192.30088495575222, "grad_norm": 0.055449243634939194, "learning_rate": 9.254190696171769e-05, "loss": 0.0037, "step": 21730 }, { "epoch": 192.38938053097345, "grad_norm": 0.09305421262979507, "learning_rate": 9.25332168442482e-05, "loss": 0.004, "step": 21740 }, { "epoch": 192.47787610619469, "grad_norm": 0.08705069124698639, "learning_rate": 9.252452207541004e-05, "loss": 0.0039, "step": 21750 }, { "epoch": 192.56637168141592, "grad_norm": 0.09536687284708023, "learning_rate": 9.251582265615409e-05, "loss": 0.0038, "step": 21760 }, { "epoch": 192.65486725663717, "grad_norm": 0.07183288037776947, "learning_rate": 9.250711858743169e-05, "loss": 0.0042, "step": 21770 }, { "epoch": 192.7433628318584, "grad_norm": 0.0929826870560646, "learning_rate": 9.24984098701947e-05, "loss": 0.0039, "step": 21780 }, { "epoch": 192.83185840707964, "grad_norm": 0.083900585770607, "learning_rate": 9.248969650539552e-05, "loss": 0.004, "step": 21790 }, { "epoch": 192.9203539823009, "grad_norm": 0.07655060291290283, "learning_rate": 9.2480978493987e-05, "loss": 0.0039, "step": 21800 }, { "epoch": 193.00884955752213, "grad_norm": 0.06679751724004745, "learning_rate": 9.247225583692256e-05, "loss": 0.0036, "step": 21810 }, { "epoch": 193.09734513274336, "grad_norm": 0.08623490482568741, "learning_rate": 9.246352853515607e-05, "loss": 0.0042, "step": 21820 }, { "epoch": 193.18584070796462, "grad_norm": 0.06469805538654327, "learning_rate": 9.245479658964194e-05, "loss": 0.0036, "step": 21830 }, { "epoch": 193.27433628318585, "grad_norm": 0.06493022292852402, "learning_rate": 9.244606000133507e-05, "loss": 0.0038, "step": 21840 }, { "epoch": 193.36283185840708, "grad_norm": 0.07649438083171844, "learning_rate": 9.24373187711909e-05, "loss": 0.0038, "step": 21850 }, { "epoch": 193.4513274336283, "grad_norm": 0.08520516753196716, "learning_rate": 9.242857290016537e-05, "loss": 0.0035, "step": 21860 }, { "epoch": 193.53982300884957, "grad_norm": 0.08253809809684753, "learning_rate": 9.241982238921488e-05, "loss": 0.0038, "step": 21870 }, { "epoch": 193.6283185840708, "grad_norm": 0.09124661237001419, "learning_rate": 9.24110672392964e-05, "loss": 0.0038, "step": 21880 }, { "epoch": 193.71681415929203, "grad_norm": 0.0864466056227684, "learning_rate": 9.240230745136737e-05, "loss": 0.0039, "step": 21890 }, { "epoch": 193.8053097345133, "grad_norm": 0.07378936558961868, "learning_rate": 9.239354302638575e-05, "loss": 0.0034, "step": 21900 }, { "epoch": 193.89380530973452, "grad_norm": 0.10056614130735397, "learning_rate": 9.238477396531e-05, "loss": 0.0038, "step": 21910 }, { "epoch": 193.98230088495575, "grad_norm": 0.09817727655172348, "learning_rate": 9.23760002690991e-05, "loss": 0.004, "step": 21920 }, { "epoch": 194.07079646017698, "grad_norm": 0.07198985666036606, "learning_rate": 9.236722193871252e-05, "loss": 0.0038, "step": 21930 }, { "epoch": 194.15929203539824, "grad_norm": 0.07811857759952545, "learning_rate": 9.235843897511023e-05, "loss": 0.0037, "step": 21940 }, { "epoch": 194.24778761061947, "grad_norm": 0.0575491227209568, "learning_rate": 9.234965137925276e-05, "loss": 0.004, "step": 21950 }, { "epoch": 194.3362831858407, "grad_norm": 0.05630374327301979, "learning_rate": 9.234085915210108e-05, "loss": 0.0035, "step": 21960 }, { "epoch": 194.42477876106196, "grad_norm": 0.08701113611459732, "learning_rate": 9.23320622946167e-05, "loss": 0.0034, "step": 21970 }, { "epoch": 194.5132743362832, "grad_norm": 0.07206215709447861, "learning_rate": 9.232326080776163e-05, "loss": 0.0035, "step": 21980 }, { "epoch": 194.60176991150442, "grad_norm": 0.07162699848413467, "learning_rate": 9.23144546924984e-05, "loss": 0.0041, "step": 21990 }, { "epoch": 194.69026548672565, "grad_norm": 0.08553720265626907, "learning_rate": 9.230564394979e-05, "loss": 0.0036, "step": 22000 }, { "epoch": 194.7787610619469, "grad_norm": 0.07556815445423126, "learning_rate": 9.22968285806e-05, "loss": 0.0034, "step": 22010 }, { "epoch": 194.86725663716814, "grad_norm": 0.0791526660323143, "learning_rate": 9.228800858589242e-05, "loss": 0.0041, "step": 22020 }, { "epoch": 194.95575221238937, "grad_norm": 0.07403822988271713, "learning_rate": 9.227918396663179e-05, "loss": 0.0035, "step": 22030 }, { "epoch": 195.04424778761063, "grad_norm": 0.07902740687131882, "learning_rate": 9.227035472378319e-05, "loss": 0.0038, "step": 22040 }, { "epoch": 195.13274336283186, "grad_norm": 0.09853962063789368, "learning_rate": 9.226152085831213e-05, "loss": 0.0034, "step": 22050 }, { "epoch": 195.2212389380531, "grad_norm": 0.09927576780319214, "learning_rate": 9.22526823711847e-05, "loss": 0.0038, "step": 22060 }, { "epoch": 195.30973451327435, "grad_norm": 0.10307316482067108, "learning_rate": 9.224383926336745e-05, "loss": 0.0039, "step": 22070 }, { "epoch": 195.39823008849558, "grad_norm": 0.07893572002649307, "learning_rate": 9.223499153582744e-05, "loss": 0.0035, "step": 22080 }, { "epoch": 195.4867256637168, "grad_norm": 0.10626024007797241, "learning_rate": 9.222613918953226e-05, "loss": 0.0041, "step": 22090 }, { "epoch": 195.57522123893804, "grad_norm": 0.0987868458032608, "learning_rate": 9.221728222544999e-05, "loss": 0.0039, "step": 22100 }, { "epoch": 195.6637168141593, "grad_norm": 0.08044595271348953, "learning_rate": 9.22084206445492e-05, "loss": 0.0041, "step": 22110 }, { "epoch": 195.75221238938053, "grad_norm": 0.07852774113416672, "learning_rate": 9.2199554447799e-05, "loss": 0.0041, "step": 22120 }, { "epoch": 195.84070796460176, "grad_norm": 0.0665941834449768, "learning_rate": 9.219068363616897e-05, "loss": 0.0042, "step": 22130 }, { "epoch": 195.92920353982302, "grad_norm": 0.07471539080142975, "learning_rate": 9.218180821062919e-05, "loss": 0.0039, "step": 22140 }, { "epoch": 196.01769911504425, "grad_norm": 0.07104301452636719, "learning_rate": 9.21729281721503e-05, "loss": 0.0032, "step": 22150 }, { "epoch": 196.10619469026548, "grad_norm": 0.07556476444005966, "learning_rate": 9.216404352170339e-05, "loss": 0.0036, "step": 22160 }, { "epoch": 196.1946902654867, "grad_norm": 0.09453286975622177, "learning_rate": 9.215515426026007e-05, "loss": 0.0035, "step": 22170 }, { "epoch": 196.28318584070797, "grad_norm": 0.0882367491722107, "learning_rate": 9.214626038879246e-05, "loss": 0.004, "step": 22180 }, { "epoch": 196.3716814159292, "grad_norm": 0.08014094084501266, "learning_rate": 9.21373619082732e-05, "loss": 0.0038, "step": 22190 }, { "epoch": 196.46017699115043, "grad_norm": 0.0734698697924614, "learning_rate": 9.212845881967535e-05, "loss": 0.0031, "step": 22200 }, { "epoch": 196.5486725663717, "grad_norm": 0.08945397287607193, "learning_rate": 9.211955112397262e-05, "loss": 0.0036, "step": 22210 }, { "epoch": 196.63716814159292, "grad_norm": 0.12276508659124374, "learning_rate": 9.211063882213909e-05, "loss": 0.0044, "step": 22220 }, { "epoch": 196.72566371681415, "grad_norm": 0.09880165010690689, "learning_rate": 9.210172191514942e-05, "loss": 0.0032, "step": 22230 }, { "epoch": 196.81415929203538, "grad_norm": 0.0890197828412056, "learning_rate": 9.209280040397874e-05, "loss": 0.0039, "step": 22240 }, { "epoch": 196.90265486725664, "grad_norm": 0.07146730273962021, "learning_rate": 9.208387428960268e-05, "loss": 0.0034, "step": 22250 }, { "epoch": 196.99115044247787, "grad_norm": 0.06619938462972641, "learning_rate": 9.20749435729974e-05, "loss": 0.0037, "step": 22260 }, { "epoch": 197.0796460176991, "grad_norm": 0.0804591104388237, "learning_rate": 9.206600825513957e-05, "loss": 0.0034, "step": 22270 }, { "epoch": 197.16814159292036, "grad_norm": 0.06175504997372627, "learning_rate": 9.20570683370063e-05, "loss": 0.0038, "step": 22280 }, { "epoch": 197.2566371681416, "grad_norm": 0.08729055523872375, "learning_rate": 9.204812381957528e-05, "loss": 0.0041, "step": 22290 }, { "epoch": 197.34513274336283, "grad_norm": 0.0795275941491127, "learning_rate": 9.203917470382465e-05, "loss": 0.0036, "step": 22300 }, { "epoch": 197.43362831858408, "grad_norm": 0.060933854430913925, "learning_rate": 9.203022099073309e-05, "loss": 0.0038, "step": 22310 }, { "epoch": 197.52212389380531, "grad_norm": 0.06471677869558334, "learning_rate": 9.202126268127976e-05, "loss": 0.0036, "step": 22320 }, { "epoch": 197.61061946902655, "grad_norm": 0.07593563944101334, "learning_rate": 9.20122997764443e-05, "loss": 0.0039, "step": 22330 }, { "epoch": 197.69911504424778, "grad_norm": 0.07858465611934662, "learning_rate": 9.200333227720692e-05, "loss": 0.0035, "step": 22340 }, { "epoch": 197.78761061946904, "grad_norm": 0.06509635597467422, "learning_rate": 9.199436018454826e-05, "loss": 0.0038, "step": 22350 }, { "epoch": 197.87610619469027, "grad_norm": 0.06615686416625977, "learning_rate": 9.198538349944952e-05, "loss": 0.0037, "step": 22360 }, { "epoch": 197.9646017699115, "grad_norm": 0.08927623182535172, "learning_rate": 9.197640222289234e-05, "loss": 0.004, "step": 22370 }, { "epoch": 198.05309734513276, "grad_norm": 0.07273422926664352, "learning_rate": 9.196741635585895e-05, "loss": 0.004, "step": 22380 }, { "epoch": 198.141592920354, "grad_norm": 0.08928161859512329, "learning_rate": 9.195842589933199e-05, "loss": 0.0036, "step": 22390 }, { "epoch": 198.23008849557522, "grad_norm": 0.07775069773197174, "learning_rate": 9.194943085429466e-05, "loss": 0.0039, "step": 22400 }, { "epoch": 198.31858407079645, "grad_norm": 0.08344908803701401, "learning_rate": 9.194043122173065e-05, "loss": 0.004, "step": 22410 }, { "epoch": 198.4070796460177, "grad_norm": 0.07033766806125641, "learning_rate": 9.193142700262413e-05, "loss": 0.0041, "step": 22420 }, { "epoch": 198.49557522123894, "grad_norm": 0.08355972915887833, "learning_rate": 9.192241819795979e-05, "loss": 0.0035, "step": 22430 }, { "epoch": 198.58407079646017, "grad_norm": 0.08347007632255554, "learning_rate": 9.191340480872284e-05, "loss": 0.0041, "step": 22440 }, { "epoch": 198.67256637168143, "grad_norm": 0.05052827298641205, "learning_rate": 9.190438683589895e-05, "loss": 0.0035, "step": 22450 }, { "epoch": 198.76106194690266, "grad_norm": 0.08279304206371307, "learning_rate": 9.189536428047432e-05, "loss": 0.0043, "step": 22460 }, { "epoch": 198.8495575221239, "grad_norm": 0.07325541228055954, "learning_rate": 9.188633714343564e-05, "loss": 0.0035, "step": 22470 }, { "epoch": 198.93805309734512, "grad_norm": 0.06422501802444458, "learning_rate": 9.18773054257701e-05, "loss": 0.004, "step": 22480 }, { "epoch": 199.02654867256638, "grad_norm": 0.06307002156972885, "learning_rate": 9.18682691284654e-05, "loss": 0.0034, "step": 22490 }, { "epoch": 199.1150442477876, "grad_norm": 0.07600340247154236, "learning_rate": 9.185922825250974e-05, "loss": 0.0038, "step": 22500 }, { "epoch": 199.20353982300884, "grad_norm": 0.07363784313201904, "learning_rate": 9.185018279889181e-05, "loss": 0.004, "step": 22510 }, { "epoch": 199.2920353982301, "grad_norm": 0.0958881750702858, "learning_rate": 9.184113276860082e-05, "loss": 0.0039, "step": 22520 }, { "epoch": 199.38053097345133, "grad_norm": 0.07303283363580704, "learning_rate": 9.183207816262645e-05, "loss": 0.0039, "step": 22530 }, { "epoch": 199.46902654867256, "grad_norm": 0.09992337971925735, "learning_rate": 9.182301898195891e-05, "loss": 0.0037, "step": 22540 }, { "epoch": 199.55752212389382, "grad_norm": 0.07216878980398178, "learning_rate": 9.181395522758889e-05, "loss": 0.0032, "step": 22550 }, { "epoch": 199.64601769911505, "grad_norm": 0.07426071912050247, "learning_rate": 9.180488690050759e-05, "loss": 0.0038, "step": 22560 }, { "epoch": 199.73451327433628, "grad_norm": 0.07985056191682816, "learning_rate": 9.179581400170671e-05, "loss": 0.0041, "step": 22570 }, { "epoch": 199.8230088495575, "grad_norm": 0.07631205022335052, "learning_rate": 9.178673653217845e-05, "loss": 0.004, "step": 22580 }, { "epoch": 199.91150442477877, "grad_norm": 0.10810162127017975, "learning_rate": 9.177765449291551e-05, "loss": 0.0038, "step": 22590 }, { "epoch": 200.0, "grad_norm": 0.09476763755083084, "learning_rate": 9.176856788491109e-05, "loss": 0.0036, "step": 22600 }, { "epoch": 200.08849557522123, "grad_norm": 0.08056830614805222, "learning_rate": 9.175947670915887e-05, "loss": 0.0038, "step": 22610 }, { "epoch": 200.1769911504425, "grad_norm": 0.086063452064991, "learning_rate": 9.175038096665309e-05, "loss": 0.0038, "step": 22620 }, { "epoch": 200.26548672566372, "grad_norm": 0.08917040377855301, "learning_rate": 9.17412806583884e-05, "loss": 0.0033, "step": 22630 }, { "epoch": 200.35398230088495, "grad_norm": 0.07770542800426483, "learning_rate": 9.173217578536002e-05, "loss": 0.0035, "step": 22640 }, { "epoch": 200.44247787610618, "grad_norm": 0.08158016204833984, "learning_rate": 9.172306634856362e-05, "loss": 0.0039, "step": 22650 }, { "epoch": 200.53097345132744, "grad_norm": 0.07456375658512115, "learning_rate": 9.171395234899545e-05, "loss": 0.0041, "step": 22660 }, { "epoch": 200.61946902654867, "grad_norm": 0.06548357009887695, "learning_rate": 9.170483378765214e-05, "loss": 0.0034, "step": 22670 }, { "epoch": 200.7079646017699, "grad_norm": 0.07070416957139969, "learning_rate": 9.169571066553091e-05, "loss": 0.0035, "step": 22680 }, { "epoch": 200.79646017699116, "grad_norm": 0.0668175145983696, "learning_rate": 9.168658298362946e-05, "loss": 0.004, "step": 22690 }, { "epoch": 200.8849557522124, "grad_norm": 0.07187973707914352, "learning_rate": 9.167745074294598e-05, "loss": 0.004, "step": 22700 }, { "epoch": 200.97345132743362, "grad_norm": 0.0924161821603775, "learning_rate": 9.166831394447913e-05, "loss": 0.0037, "step": 22710 }, { "epoch": 201.06194690265488, "grad_norm": 0.08670619130134583, "learning_rate": 9.165917258922812e-05, "loss": 0.0039, "step": 22720 }, { "epoch": 201.1504424778761, "grad_norm": 0.08757905662059784, "learning_rate": 9.165002667819262e-05, "loss": 0.0037, "step": 22730 }, { "epoch": 201.23893805309734, "grad_norm": 0.09489789605140686, "learning_rate": 9.164087621237282e-05, "loss": 0.0041, "step": 22740 }, { "epoch": 201.32743362831857, "grad_norm": 0.06681405007839203, "learning_rate": 9.163172119276942e-05, "loss": 0.0033, "step": 22750 }, { "epoch": 201.41592920353983, "grad_norm": 0.06225181743502617, "learning_rate": 9.162256162038358e-05, "loss": 0.0039, "step": 22760 }, { "epoch": 201.50442477876106, "grad_norm": 0.07421994209289551, "learning_rate": 9.161339749621698e-05, "loss": 0.0039, "step": 22770 }, { "epoch": 201.5929203539823, "grad_norm": 0.07789818942546844, "learning_rate": 9.160422882127177e-05, "loss": 0.0037, "step": 22780 }, { "epoch": 201.68141592920355, "grad_norm": 0.08075109869241714, "learning_rate": 9.159505559655069e-05, "loss": 0.0039, "step": 22790 }, { "epoch": 201.76991150442478, "grad_norm": 0.08723414689302444, "learning_rate": 9.158587782305684e-05, "loss": 0.0038, "step": 22800 }, { "epoch": 201.858407079646, "grad_norm": 0.06887923926115036, "learning_rate": 9.157669550179391e-05, "loss": 0.0034, "step": 22810 }, { "epoch": 201.94690265486724, "grad_norm": 0.0557122677564621, "learning_rate": 9.156750863376609e-05, "loss": 0.0032, "step": 22820 }, { "epoch": 202.0353982300885, "grad_norm": 0.07838674634695053, "learning_rate": 9.155831721997801e-05, "loss": 0.0038, "step": 22830 }, { "epoch": 202.12389380530973, "grad_norm": 0.09532970190048218, "learning_rate": 9.154912126143484e-05, "loss": 0.0036, "step": 22840 }, { "epoch": 202.21238938053096, "grad_norm": 0.09390272200107574, "learning_rate": 9.153992075914224e-05, "loss": 0.0037, "step": 22850 }, { "epoch": 202.30088495575222, "grad_norm": 0.07773000746965408, "learning_rate": 9.153071571410635e-05, "loss": 0.0036, "step": 22860 }, { "epoch": 202.38938053097345, "grad_norm": 0.07230383902788162, "learning_rate": 9.152150612733384e-05, "loss": 0.0035, "step": 22870 }, { "epoch": 202.47787610619469, "grad_norm": 0.0741966962814331, "learning_rate": 9.151229199983184e-05, "loss": 0.0029, "step": 22880 }, { "epoch": 202.56637168141592, "grad_norm": 0.0682280883193016, "learning_rate": 9.150307333260802e-05, "loss": 0.0035, "step": 22890 }, { "epoch": 202.65486725663717, "grad_norm": 0.06574603915214539, "learning_rate": 9.149385012667048e-05, "loss": 0.0034, "step": 22900 }, { "epoch": 202.7433628318584, "grad_norm": 0.09467878192663193, "learning_rate": 9.148462238302788e-05, "loss": 0.0034, "step": 22910 }, { "epoch": 202.83185840707964, "grad_norm": 0.09866183996200562, "learning_rate": 9.147539010268936e-05, "loss": 0.0037, "step": 22920 }, { "epoch": 202.9203539823009, "grad_norm": 0.0626540556550026, "learning_rate": 9.14661532866645e-05, "loss": 0.0039, "step": 22930 }, { "epoch": 203.00884955752213, "grad_norm": 0.06472469121217728, "learning_rate": 9.145691193596348e-05, "loss": 0.0033, "step": 22940 }, { "epoch": 203.09734513274336, "grad_norm": 0.08047271519899368, "learning_rate": 9.144766605159691e-05, "loss": 0.0037, "step": 22950 }, { "epoch": 203.18584070796462, "grad_norm": 0.08388623595237732, "learning_rate": 9.14384156345759e-05, "loss": 0.0043, "step": 22960 }, { "epoch": 203.27433628318585, "grad_norm": 0.1019647940993309, "learning_rate": 9.142916068591204e-05, "loss": 0.0039, "step": 22970 }, { "epoch": 203.36283185840708, "grad_norm": 0.11025030165910721, "learning_rate": 9.141990120661746e-05, "loss": 0.0039, "step": 22980 }, { "epoch": 203.4513274336283, "grad_norm": 0.08682364970445633, "learning_rate": 9.141063719770475e-05, "loss": 0.0036, "step": 22990 }, { "epoch": 203.53982300884957, "grad_norm": 0.08883027732372284, "learning_rate": 9.140136866018704e-05, "loss": 0.004, "step": 23000 }, { "epoch": 203.6283185840708, "grad_norm": 0.07246774435043335, "learning_rate": 9.139209559507788e-05, "loss": 0.0037, "step": 23010 }, { "epoch": 203.71681415929203, "grad_norm": 0.09462893009185791, "learning_rate": 9.13828180033914e-05, "loss": 0.0039, "step": 23020 }, { "epoch": 203.8053097345133, "grad_norm": 0.09476759284734726, "learning_rate": 9.137353588614212e-05, "loss": 0.0037, "step": 23030 }, { "epoch": 203.89380530973452, "grad_norm": 0.06946789473295212, "learning_rate": 9.136424924434519e-05, "loss": 0.0042, "step": 23040 }, { "epoch": 203.98230088495575, "grad_norm": 0.09068784862756729, "learning_rate": 9.135495807901615e-05, "loss": 0.0035, "step": 23050 }, { "epoch": 204.07079646017698, "grad_norm": 0.064038947224617, "learning_rate": 9.134566239117108e-05, "loss": 0.0035, "step": 23060 }, { "epoch": 204.15929203539824, "grad_norm": 0.09247218817472458, "learning_rate": 9.13363621818265e-05, "loss": 0.0036, "step": 23070 }, { "epoch": 204.24778761061947, "grad_norm": 0.06820853799581528, "learning_rate": 9.132705745199953e-05, "loss": 0.0034, "step": 23080 }, { "epoch": 204.3362831858407, "grad_norm": 0.05287754163146019, "learning_rate": 9.131774820270768e-05, "loss": 0.0033, "step": 23090 }, { "epoch": 204.42477876106196, "grad_norm": 0.0738256573677063, "learning_rate": 9.130843443496901e-05, "loss": 0.0037, "step": 23100 }, { "epoch": 204.5132743362832, "grad_norm": 0.09140079468488693, "learning_rate": 9.129911614980206e-05, "loss": 0.0038, "step": 23110 }, { "epoch": 204.60176991150442, "grad_norm": 0.08064408600330353, "learning_rate": 9.128979334822584e-05, "loss": 0.0035, "step": 23120 }, { "epoch": 204.69026548672565, "grad_norm": 0.06398759037256241, "learning_rate": 9.128046603125992e-05, "loss": 0.0034, "step": 23130 }, { "epoch": 204.7787610619469, "grad_norm": 0.10744472593069077, "learning_rate": 9.12711341999243e-05, "loss": 0.0039, "step": 23140 }, { "epoch": 204.86725663716814, "grad_norm": 0.0664854496717453, "learning_rate": 9.12617978552395e-05, "loss": 0.0038, "step": 23150 }, { "epoch": 204.95575221238937, "grad_norm": 0.07415587455034256, "learning_rate": 9.12524569982265e-05, "loss": 0.004, "step": 23160 }, { "epoch": 205.04424778761063, "grad_norm": 0.09895265102386475, "learning_rate": 9.124311162990684e-05, "loss": 0.0037, "step": 23170 }, { "epoch": 205.13274336283186, "grad_norm": 0.08505862206220627, "learning_rate": 9.12337617513025e-05, "loss": 0.0035, "step": 23180 }, { "epoch": 205.2212389380531, "grad_norm": 0.07412861287593842, "learning_rate": 9.122440736343596e-05, "loss": 0.0038, "step": 23190 }, { "epoch": 205.30973451327435, "grad_norm": 0.10917285829782486, "learning_rate": 9.12150484673302e-05, "loss": 0.0038, "step": 23200 }, { "epoch": 205.39823008849558, "grad_norm": 0.10063585638999939, "learning_rate": 9.120568506400873e-05, "loss": 0.0041, "step": 23210 }, { "epoch": 205.4867256637168, "grad_norm": 0.07997056841850281, "learning_rate": 9.119631715449548e-05, "loss": 0.0038, "step": 23220 }, { "epoch": 205.57522123893804, "grad_norm": 0.06944447010755539, "learning_rate": 9.118694473981493e-05, "loss": 0.0034, "step": 23230 }, { "epoch": 205.6637168141593, "grad_norm": 0.05928461253643036, "learning_rate": 9.117756782099203e-05, "loss": 0.0037, "step": 23240 }, { "epoch": 205.75221238938053, "grad_norm": 0.07960258424282074, "learning_rate": 9.11681863990522e-05, "loss": 0.0033, "step": 23250 }, { "epoch": 205.84070796460176, "grad_norm": 0.08947581797838211, "learning_rate": 9.115880047502142e-05, "loss": 0.0037, "step": 23260 }, { "epoch": 205.92920353982302, "grad_norm": 0.06323418766260147, "learning_rate": 9.114941004992609e-05, "loss": 0.0044, "step": 23270 }, { "epoch": 206.01769911504425, "grad_norm": 0.05908490717411041, "learning_rate": 9.114001512479317e-05, "loss": 0.0036, "step": 23280 }, { "epoch": 206.10619469026548, "grad_norm": 0.07699526101350784, "learning_rate": 9.113061570065003e-05, "loss": 0.0033, "step": 23290 }, { "epoch": 206.1946902654867, "grad_norm": 0.07943914085626602, "learning_rate": 9.112121177852459e-05, "loss": 0.0035, "step": 23300 }, { "epoch": 206.28318584070797, "grad_norm": 0.09817540645599365, "learning_rate": 9.111180335944527e-05, "loss": 0.0036, "step": 23310 }, { "epoch": 206.3716814159292, "grad_norm": 0.08399830758571625, "learning_rate": 9.110239044444093e-05, "loss": 0.004, "step": 23320 }, { "epoch": 206.46017699115043, "grad_norm": 0.07956275343894958, "learning_rate": 9.109297303454099e-05, "loss": 0.0034, "step": 23330 }, { "epoch": 206.5486725663717, "grad_norm": 0.09770376235246658, "learning_rate": 9.108355113077526e-05, "loss": 0.0037, "step": 23340 }, { "epoch": 206.63716814159292, "grad_norm": 0.08544039726257324, "learning_rate": 9.107412473417419e-05, "loss": 0.0038, "step": 23350 }, { "epoch": 206.72566371681415, "grad_norm": 0.10806979984045029, "learning_rate": 9.106469384576858e-05, "loss": 0.0037, "step": 23360 }, { "epoch": 206.81415929203538, "grad_norm": 0.06751487404108047, "learning_rate": 9.105525846658978e-05, "loss": 0.0035, "step": 23370 }, { "epoch": 206.90265486725664, "grad_norm": 0.08069217205047607, "learning_rate": 9.104581859766965e-05, "loss": 0.0033, "step": 23380 }, { "epoch": 206.99115044247787, "grad_norm": 0.06812684237957001, "learning_rate": 9.10363742400405e-05, "loss": 0.0037, "step": 23390 }, { "epoch": 207.0796460176991, "grad_norm": 0.07775121927261353, "learning_rate": 9.102692539473518e-05, "loss": 0.0038, "step": 23400 }, { "epoch": 207.16814159292036, "grad_norm": 0.10755199193954468, "learning_rate": 9.101747206278697e-05, "loss": 0.004, "step": 23410 }, { "epoch": 207.2566371681416, "grad_norm": 0.07915190607309341, "learning_rate": 9.100801424522968e-05, "loss": 0.004, "step": 23420 }, { "epoch": 207.34513274336283, "grad_norm": 0.08215562999248505, "learning_rate": 9.099855194309762e-05, "loss": 0.0041, "step": 23430 }, { "epoch": 207.43362831858408, "grad_norm": 0.08712712675333023, "learning_rate": 9.098908515742554e-05, "loss": 0.0038, "step": 23440 }, { "epoch": 207.52212389380531, "grad_norm": 0.07660869508981705, "learning_rate": 9.097961388924873e-05, "loss": 0.0037, "step": 23450 }, { "epoch": 207.61061946902655, "grad_norm": 0.07111310213804245, "learning_rate": 9.097013813960298e-05, "loss": 0.0034, "step": 23460 }, { "epoch": 207.69911504424778, "grad_norm": 0.08286506682634354, "learning_rate": 9.09606579095245e-05, "loss": 0.0037, "step": 23470 }, { "epoch": 207.78761061946904, "grad_norm": 0.07357780635356903, "learning_rate": 9.095117320005008e-05, "loss": 0.0041, "step": 23480 }, { "epoch": 207.87610619469027, "grad_norm": 0.06829012930393219, "learning_rate": 9.094168401221691e-05, "loss": 0.0039, "step": 23490 }, { "epoch": 207.9646017699115, "grad_norm": 0.0832628533244133, "learning_rate": 9.093219034706273e-05, "loss": 0.0039, "step": 23500 }, { "epoch": 208.05309734513276, "grad_norm": 0.08994574099779129, "learning_rate": 9.092269220562577e-05, "loss": 0.004, "step": 23510 }, { "epoch": 208.141592920354, "grad_norm": 0.07848171889781952, "learning_rate": 9.09131895889447e-05, "loss": 0.0038, "step": 23520 }, { "epoch": 208.23008849557522, "grad_norm": 0.06785155087709427, "learning_rate": 9.090368249805873e-05, "loss": 0.0038, "step": 23530 }, { "epoch": 208.31858407079645, "grad_norm": 0.0899403840303421, "learning_rate": 9.089417093400754e-05, "loss": 0.0041, "step": 23540 }, { "epoch": 208.4070796460177, "grad_norm": 0.07028195261955261, "learning_rate": 9.088465489783131e-05, "loss": 0.003, "step": 23550 }, { "epoch": 208.49557522123894, "grad_norm": 0.0836910754442215, "learning_rate": 9.087513439057068e-05, "loss": 0.0036, "step": 23560 }, { "epoch": 208.58407079646017, "grad_norm": 0.08764295279979706, "learning_rate": 9.08656094132668e-05, "loss": 0.0038, "step": 23570 }, { "epoch": 208.67256637168143, "grad_norm": 0.0974014401435852, "learning_rate": 9.085607996696134e-05, "loss": 0.0038, "step": 23580 }, { "epoch": 208.76106194690266, "grad_norm": 0.07935497909784317, "learning_rate": 9.084654605269639e-05, "loss": 0.0033, "step": 23590 }, { "epoch": 208.8495575221239, "grad_norm": 0.076028011739254, "learning_rate": 9.083700767151457e-05, "loss": 0.0036, "step": 23600 }, { "epoch": 208.93805309734512, "grad_norm": 0.06477377563714981, "learning_rate": 9.082746482445898e-05, "loss": 0.0035, "step": 23610 }, { "epoch": 209.02654867256638, "grad_norm": 0.06287270784378052, "learning_rate": 9.081791751257325e-05, "loss": 0.0036, "step": 23620 }, { "epoch": 209.1150442477876, "grad_norm": 0.07396765053272247, "learning_rate": 9.080836573690142e-05, "loss": 0.0041, "step": 23630 }, { "epoch": 209.20353982300884, "grad_norm": 0.0628039538860321, "learning_rate": 9.079880949848805e-05, "loss": 0.0034, "step": 23640 }, { "epoch": 209.2920353982301, "grad_norm": 0.07508698850870132, "learning_rate": 9.078924879837822e-05, "loss": 0.0038, "step": 23650 }, { "epoch": 209.38053097345133, "grad_norm": 0.0642697811126709, "learning_rate": 9.077968363761747e-05, "loss": 0.0032, "step": 23660 }, { "epoch": 209.46902654867256, "grad_norm": 0.04804707318544388, "learning_rate": 9.077011401725182e-05, "loss": 0.0033, "step": 23670 }, { "epoch": 209.55752212389382, "grad_norm": 0.05412571504712105, "learning_rate": 9.07605399383278e-05, "loss": 0.0037, "step": 23680 }, { "epoch": 209.64601769911505, "grad_norm": 0.08293721079826355, "learning_rate": 9.075096140189243e-05, "loss": 0.0041, "step": 23690 }, { "epoch": 209.73451327433628, "grad_norm": 0.07088792324066162, "learning_rate": 9.074137840899318e-05, "loss": 0.0034, "step": 23700 }, { "epoch": 209.8230088495575, "grad_norm": 0.07065549492835999, "learning_rate": 9.073179096067804e-05, "loss": 0.0033, "step": 23710 }, { "epoch": 209.91150442477877, "grad_norm": 0.06936166435480118, "learning_rate": 9.072219905799549e-05, "loss": 0.0034, "step": 23720 }, { "epoch": 210.0, "grad_norm": 0.09869847446680069, "learning_rate": 9.071260270199447e-05, "loss": 0.0036, "step": 23730 }, { "epoch": 210.08849557522123, "grad_norm": 0.07508743554353714, "learning_rate": 9.070300189372441e-05, "loss": 0.0036, "step": 23740 }, { "epoch": 210.1769911504425, "grad_norm": 0.0818825513124466, "learning_rate": 9.069339663423528e-05, "loss": 0.0039, "step": 23750 }, { "epoch": 210.26548672566372, "grad_norm": 0.07747120410203934, "learning_rate": 9.068378692457747e-05, "loss": 0.0036, "step": 23760 }, { "epoch": 210.35398230088495, "grad_norm": 0.07640978693962097, "learning_rate": 9.067417276580189e-05, "loss": 0.0035, "step": 23770 }, { "epoch": 210.44247787610618, "grad_norm": 0.0787690132856369, "learning_rate": 9.066455415895993e-05, "loss": 0.0034, "step": 23780 }, { "epoch": 210.53097345132744, "grad_norm": 0.08103237301111221, "learning_rate": 9.065493110510346e-05, "loss": 0.0041, "step": 23790 }, { "epoch": 210.61946902654867, "grad_norm": 0.056754954159259796, "learning_rate": 9.064530360528484e-05, "loss": 0.0033, "step": 23800 }, { "epoch": 210.7079646017699, "grad_norm": 0.06702844798564911, "learning_rate": 9.063567166055695e-05, "loss": 0.0031, "step": 23810 }, { "epoch": 210.79646017699116, "grad_norm": 0.06571734696626663, "learning_rate": 9.062603527197308e-05, "loss": 0.0035, "step": 23820 }, { "epoch": 210.8849557522124, "grad_norm": 0.06457842886447906, "learning_rate": 9.06163944405871e-05, "loss": 0.0041, "step": 23830 }, { "epoch": 210.97345132743362, "grad_norm": 0.08263809233903885, "learning_rate": 9.060674916745327e-05, "loss": 0.0037, "step": 23840 }, { "epoch": 211.06194690265488, "grad_norm": 0.10890538990497589, "learning_rate": 9.05970994536264e-05, "loss": 0.0037, "step": 23850 }, { "epoch": 211.1504424778761, "grad_norm": 0.07795419543981552, "learning_rate": 9.05874453001618e-05, "loss": 0.0029, "step": 23860 }, { "epoch": 211.23893805309734, "grad_norm": 0.07065803557634354, "learning_rate": 9.057778670811517e-05, "loss": 0.0037, "step": 23870 }, { "epoch": 211.32743362831857, "grad_norm": 0.06743314862251282, "learning_rate": 9.056812367854281e-05, "loss": 0.0039, "step": 23880 }, { "epoch": 211.41592920353983, "grad_norm": 0.07282949239015579, "learning_rate": 9.055845621250143e-05, "loss": 0.0033, "step": 23890 }, { "epoch": 211.50442477876106, "grad_norm": 0.09314876049757004, "learning_rate": 9.054878431104825e-05, "loss": 0.0039, "step": 23900 }, { "epoch": 211.5929203539823, "grad_norm": 0.05483514070510864, "learning_rate": 9.0539107975241e-05, "loss": 0.0037, "step": 23910 }, { "epoch": 211.68141592920355, "grad_norm": 0.0900578573346138, "learning_rate": 9.052942720613784e-05, "loss": 0.0037, "step": 23920 }, { "epoch": 211.76991150442478, "grad_norm": 0.10148106515407562, "learning_rate": 9.051974200479745e-05, "loss": 0.0042, "step": 23930 }, { "epoch": 211.858407079646, "grad_norm": 0.06995644420385361, "learning_rate": 9.051005237227901e-05, "loss": 0.0036, "step": 23940 }, { "epoch": 211.94690265486724, "grad_norm": 0.0834411159157753, "learning_rate": 9.050035830964215e-05, "loss": 0.0038, "step": 23950 }, { "epoch": 212.0353982300885, "grad_norm": 0.10178662091493607, "learning_rate": 9.049065981794698e-05, "loss": 0.0037, "step": 23960 }, { "epoch": 212.12389380530973, "grad_norm": 0.06525283306837082, "learning_rate": 9.048095689825414e-05, "loss": 0.003, "step": 23970 }, { "epoch": 212.21238938053096, "grad_norm": 0.09843452274799347, "learning_rate": 9.047124955162472e-05, "loss": 0.0041, "step": 23980 }, { "epoch": 212.30088495575222, "grad_norm": 0.08082056045532227, "learning_rate": 9.046153777912028e-05, "loss": 0.0039, "step": 23990 }, { "epoch": 212.38938053097345, "grad_norm": 0.09244649857282639, "learning_rate": 9.045182158180292e-05, "loss": 0.0037, "step": 24000 }, { "epoch": 212.47787610619469, "grad_norm": 0.06746778637170792, "learning_rate": 9.044210096073516e-05, "loss": 0.0036, "step": 24010 }, { "epoch": 212.56637168141592, "grad_norm": 0.09055531024932861, "learning_rate": 9.043237591698004e-05, "loss": 0.0036, "step": 24020 }, { "epoch": 212.65486725663717, "grad_norm": 0.10005653649568558, "learning_rate": 9.04226464516011e-05, "loss": 0.0047, "step": 24030 }, { "epoch": 212.7433628318584, "grad_norm": 0.08051538467407227, "learning_rate": 9.041291256566229e-05, "loss": 0.0035, "step": 24040 }, { "epoch": 212.83185840707964, "grad_norm": 0.08661269396543503, "learning_rate": 9.040317426022814e-05, "loss": 0.004, "step": 24050 }, { "epoch": 212.9203539823009, "grad_norm": 0.06792548298835754, "learning_rate": 9.03934315363636e-05, "loss": 0.0031, "step": 24060 }, { "epoch": 213.00884955752213, "grad_norm": 0.07760316133499146, "learning_rate": 9.038368439513409e-05, "loss": 0.0038, "step": 24070 }, { "epoch": 213.09734513274336, "grad_norm": 0.08928103744983673, "learning_rate": 9.03739328376056e-05, "loss": 0.0035, "step": 24080 }, { "epoch": 213.18584070796462, "grad_norm": 0.07265899330377579, "learning_rate": 9.036417686484451e-05, "loss": 0.0036, "step": 24090 }, { "epoch": 213.27433628318585, "grad_norm": 0.08417605608701706, "learning_rate": 9.035441647791773e-05, "loss": 0.0033, "step": 24100 }, { "epoch": 213.36283185840708, "grad_norm": 0.07183258980512619, "learning_rate": 9.034465167789263e-05, "loss": 0.0032, "step": 24110 }, { "epoch": 213.4513274336283, "grad_norm": 0.06887985020875931, "learning_rate": 9.033488246583706e-05, "loss": 0.0032, "step": 24120 }, { "epoch": 213.53982300884957, "grad_norm": 0.06173054501414299, "learning_rate": 9.032510884281941e-05, "loss": 0.0032, "step": 24130 }, { "epoch": 213.6283185840708, "grad_norm": 0.07124881446361542, "learning_rate": 9.031533080990848e-05, "loss": 0.0035, "step": 24140 }, { "epoch": 213.71681415929203, "grad_norm": 0.07151872664690018, "learning_rate": 9.030554836817358e-05, "loss": 0.0035, "step": 24150 }, { "epoch": 213.8053097345133, "grad_norm": 0.0772918090224266, "learning_rate": 9.029576151868451e-05, "loss": 0.0036, "step": 24160 }, { "epoch": 213.89380530973452, "grad_norm": 0.07966598123311996, "learning_rate": 9.028597026251155e-05, "loss": 0.0035, "step": 24170 }, { "epoch": 213.98230088495575, "grad_norm": 0.07903409004211426, "learning_rate": 9.027617460072547e-05, "loss": 0.0036, "step": 24180 }, { "epoch": 214.07079646017698, "grad_norm": 0.06361192464828491, "learning_rate": 9.026637453439745e-05, "loss": 0.0031, "step": 24190 }, { "epoch": 214.15929203539824, "grad_norm": 0.05868132412433624, "learning_rate": 9.025657006459927e-05, "loss": 0.0034, "step": 24200 }, { "epoch": 214.24778761061947, "grad_norm": 0.06792785972356796, "learning_rate": 9.024676119240311e-05, "loss": 0.0038, "step": 24210 }, { "epoch": 214.3362831858407, "grad_norm": 0.08927278220653534, "learning_rate": 9.023694791888166e-05, "loss": 0.0036, "step": 24220 }, { "epoch": 214.42477876106196, "grad_norm": 0.07022232562303543, "learning_rate": 9.022713024510808e-05, "loss": 0.0033, "step": 24230 }, { "epoch": 214.5132743362832, "grad_norm": 0.05978376418352127, "learning_rate": 9.021730817215601e-05, "loss": 0.0033, "step": 24240 }, { "epoch": 214.60176991150442, "grad_norm": 0.08627578616142273, "learning_rate": 9.02074817010996e-05, "loss": 0.0036, "step": 24250 }, { "epoch": 214.69026548672565, "grad_norm": 0.07120474427938461, "learning_rate": 9.019765083301342e-05, "loss": 0.0041, "step": 24260 }, { "epoch": 214.7787610619469, "grad_norm": 0.0767846554517746, "learning_rate": 9.01878155689726e-05, "loss": 0.0033, "step": 24270 }, { "epoch": 214.86725663716814, "grad_norm": 0.08121646195650101, "learning_rate": 9.017797591005268e-05, "loss": 0.0035, "step": 24280 }, { "epoch": 214.95575221238937, "grad_norm": 0.08982532471418381, "learning_rate": 9.016813185732972e-05, "loss": 0.0033, "step": 24290 }, { "epoch": 215.04424778761063, "grad_norm": 0.07996156811714172, "learning_rate": 9.015828341188027e-05, "loss": 0.0037, "step": 24300 }, { "epoch": 215.13274336283186, "grad_norm": 0.05882789567112923, "learning_rate": 9.01484305747813e-05, "loss": 0.0037, "step": 24310 }, { "epoch": 215.2212389380531, "grad_norm": 0.06116959825158119, "learning_rate": 9.013857334711033e-05, "loss": 0.0033, "step": 24320 }, { "epoch": 215.30973451327435, "grad_norm": 0.0681551918387413, "learning_rate": 9.012871172994534e-05, "loss": 0.0033, "step": 24330 }, { "epoch": 215.39823008849558, "grad_norm": 0.07166259735822678, "learning_rate": 9.011884572436476e-05, "loss": 0.0035, "step": 24340 }, { "epoch": 215.4867256637168, "grad_norm": 0.08145429193973541, "learning_rate": 9.010897533144754e-05, "loss": 0.0034, "step": 24350 }, { "epoch": 215.57522123893804, "grad_norm": 0.07191053777933121, "learning_rate": 9.009910055227306e-05, "loss": 0.0034, "step": 24360 }, { "epoch": 215.6637168141593, "grad_norm": 0.06748538464307785, "learning_rate": 9.008922138792124e-05, "loss": 0.0034, "step": 24370 }, { "epoch": 215.75221238938053, "grad_norm": 0.08400677144527435, "learning_rate": 9.007933783947244e-05, "loss": 0.0039, "step": 24380 }, { "epoch": 215.84070796460176, "grad_norm": 0.07247722148895264, "learning_rate": 9.006944990800752e-05, "loss": 0.0037, "step": 24390 }, { "epoch": 215.92920353982302, "grad_norm": 0.07424608618021011, "learning_rate": 9.005955759460779e-05, "loss": 0.0037, "step": 24400 }, { "epoch": 216.01769911504425, "grad_norm": 0.07552197575569153, "learning_rate": 9.004966090035508e-05, "loss": 0.0035, "step": 24410 }, { "epoch": 216.10619469026548, "grad_norm": 0.07694648206233978, "learning_rate": 9.003975982633166e-05, "loss": 0.0034, "step": 24420 }, { "epoch": 216.1946902654867, "grad_norm": 0.07118372619152069, "learning_rate": 9.00298543736203e-05, "loss": 0.0036, "step": 24430 }, { "epoch": 216.28318584070797, "grad_norm": 0.07741916924715042, "learning_rate": 9.001994454330427e-05, "loss": 0.0036, "step": 24440 }, { "epoch": 216.3716814159292, "grad_norm": 0.08691401779651642, "learning_rate": 9.001003033646727e-05, "loss": 0.0036, "step": 24450 }, { "epoch": 216.46017699115043, "grad_norm": 0.07585687190294266, "learning_rate": 9.00001117541935e-05, "loss": 0.004, "step": 24460 }, { "epoch": 216.5486725663717, "grad_norm": 0.07667946815490723, "learning_rate": 8.999018879756764e-05, "loss": 0.0034, "step": 24470 }, { "epoch": 216.63716814159292, "grad_norm": 0.07903046160936356, "learning_rate": 8.998026146767487e-05, "loss": 0.0035, "step": 24480 }, { "epoch": 216.72566371681415, "grad_norm": 0.09275577962398529, "learning_rate": 8.99703297656008e-05, "loss": 0.0035, "step": 24490 }, { "epoch": 216.81415929203538, "grad_norm": 0.06409694999456406, "learning_rate": 8.996039369243156e-05, "loss": 0.0037, "step": 24500 }, { "epoch": 216.90265486725664, "grad_norm": 0.07093489915132523, "learning_rate": 8.995045324925378e-05, "loss": 0.0037, "step": 24510 }, { "epoch": 216.99115044247787, "grad_norm": 0.07223805040121078, "learning_rate": 8.994050843715448e-05, "loss": 0.0036, "step": 24520 }, { "epoch": 217.0796460176991, "grad_norm": 0.06730297952890396, "learning_rate": 8.993055925722121e-05, "loss": 0.0034, "step": 24530 }, { "epoch": 217.16814159292036, "grad_norm": 0.08530806750059128, "learning_rate": 8.992060571054202e-05, "loss": 0.0032, "step": 24540 }, { "epoch": 217.2566371681416, "grad_norm": 0.08205147832632065, "learning_rate": 8.991064779820542e-05, "loss": 0.0032, "step": 24550 }, { "epoch": 217.34513274336283, "grad_norm": 0.08547929674386978, "learning_rate": 8.990068552130036e-05, "loss": 0.0039, "step": 24560 }, { "epoch": 217.43362831858408, "grad_norm": 0.06534826010465622, "learning_rate": 8.989071888091634e-05, "loss": 0.0036, "step": 24570 }, { "epoch": 217.52212389380531, "grad_norm": 0.06399383395910263, "learning_rate": 8.988074787814329e-05, "loss": 0.0037, "step": 24580 }, { "epoch": 217.61061946902655, "grad_norm": 0.07203923165798187, "learning_rate": 8.987077251407158e-05, "loss": 0.003, "step": 24590 }, { "epoch": 217.69911504424778, "grad_norm": 0.06482212990522385, "learning_rate": 8.986079278979216e-05, "loss": 0.0033, "step": 24600 }, { "epoch": 217.78761061946904, "grad_norm": 0.05813268944621086, "learning_rate": 8.985080870639635e-05, "loss": 0.0035, "step": 24610 }, { "epoch": 217.87610619469027, "grad_norm": 0.07784246653318405, "learning_rate": 8.984082026497603e-05, "loss": 0.0032, "step": 24620 }, { "epoch": 217.9646017699115, "grad_norm": 0.07548565417528152, "learning_rate": 8.98308274666235e-05, "loss": 0.0032, "step": 24630 }, { "epoch": 218.05309734513276, "grad_norm": 0.07725110650062561, "learning_rate": 8.982083031243155e-05, "loss": 0.0039, "step": 24640 }, { "epoch": 218.141592920354, "grad_norm": 0.08021509647369385, "learning_rate": 8.98108288034935e-05, "loss": 0.0034, "step": 24650 }, { "epoch": 218.23008849557522, "grad_norm": 0.07842417061328888, "learning_rate": 8.980082294090305e-05, "loss": 0.0036, "step": 24660 }, { "epoch": 218.31858407079645, "grad_norm": 0.06980028003454208, "learning_rate": 8.979081272575443e-05, "loss": 0.0034, "step": 24670 }, { "epoch": 218.4070796460177, "grad_norm": 0.07422588020563126, "learning_rate": 8.978079815914236e-05, "loss": 0.0034, "step": 24680 }, { "epoch": 218.49557522123894, "grad_norm": 0.07519762963056564, "learning_rate": 8.977077924216202e-05, "loss": 0.0045, "step": 24690 }, { "epoch": 218.58407079646017, "grad_norm": 0.09417594224214554, "learning_rate": 8.976075597590905e-05, "loss": 0.0038, "step": 24700 }, { "epoch": 218.67256637168143, "grad_norm": 0.08058250695466995, "learning_rate": 8.975072836147958e-05, "loss": 0.0033, "step": 24710 }, { "epoch": 218.76106194690266, "grad_norm": 0.07794233411550522, "learning_rate": 8.974069639997025e-05, "loss": 0.0035, "step": 24720 }, { "epoch": 218.8495575221239, "grad_norm": 0.07807698845863342, "learning_rate": 8.973066009247808e-05, "loss": 0.0034, "step": 24730 }, { "epoch": 218.93805309734512, "grad_norm": 0.08997167646884918, "learning_rate": 8.972061944010066e-05, "loss": 0.004, "step": 24740 }, { "epoch": 219.02654867256638, "grad_norm": 0.07925969362258911, "learning_rate": 8.971057444393603e-05, "loss": 0.0032, "step": 24750 }, { "epoch": 219.1150442477876, "grad_norm": 0.09219438582658768, "learning_rate": 8.970052510508268e-05, "loss": 0.0038, "step": 24760 }, { "epoch": 219.20353982300884, "grad_norm": 0.0590081624686718, "learning_rate": 8.969047142463959e-05, "loss": 0.0033, "step": 24770 }, { "epoch": 219.2920353982301, "grad_norm": 0.06942696124315262, "learning_rate": 8.968041340370621e-05, "loss": 0.0035, "step": 24780 }, { "epoch": 219.38053097345133, "grad_norm": 0.08520414680242538, "learning_rate": 8.96703510433825e-05, "loss": 0.0037, "step": 24790 }, { "epoch": 219.46902654867256, "grad_norm": 0.06687764078378677, "learning_rate": 8.966028434476883e-05, "loss": 0.004, "step": 24800 }, { "epoch": 219.55752212389382, "grad_norm": 0.0692441463470459, "learning_rate": 8.96502133089661e-05, "loss": 0.0034, "step": 24810 }, { "epoch": 219.64601769911505, "grad_norm": 0.06612828373908997, "learning_rate": 8.964013793707564e-05, "loss": 0.0037, "step": 24820 }, { "epoch": 219.73451327433628, "grad_norm": 0.05997235327959061, "learning_rate": 8.963005823019932e-05, "loss": 0.0036, "step": 24830 }, { "epoch": 219.8230088495575, "grad_norm": 0.07625973969697952, "learning_rate": 8.961997418943939e-05, "loss": 0.0034, "step": 24840 }, { "epoch": 219.91150442477877, "grad_norm": 0.07219722121953964, "learning_rate": 8.960988581589865e-05, "loss": 0.0041, "step": 24850 }, { "epoch": 220.0, "grad_norm": 0.09107779711484909, "learning_rate": 8.959979311068037e-05, "loss": 0.0038, "step": 24860 }, { "epoch": 220.08849557522123, "grad_norm": 0.0765790119767189, "learning_rate": 8.958969607488823e-05, "loss": 0.003, "step": 24870 }, { "epoch": 220.1769911504425, "grad_norm": 0.1044715866446495, "learning_rate": 8.957959470962647e-05, "loss": 0.0036, "step": 24880 }, { "epoch": 220.26548672566372, "grad_norm": 0.09146273881196976, "learning_rate": 8.956948901599971e-05, "loss": 0.0038, "step": 24890 }, { "epoch": 220.35398230088495, "grad_norm": 0.08030342310667038, "learning_rate": 8.955937899511315e-05, "loss": 0.0042, "step": 24900 }, { "epoch": 220.44247787610618, "grad_norm": 0.07929990440607071, "learning_rate": 8.954926464807238e-05, "loss": 0.0037, "step": 24910 }, { "epoch": 220.53097345132744, "grad_norm": 0.05292513966560364, "learning_rate": 8.953914597598347e-05, "loss": 0.0035, "step": 24920 }, { "epoch": 220.61946902654867, "grad_norm": 0.07680570334196091, "learning_rate": 8.952902297995303e-05, "loss": 0.0036, "step": 24930 }, { "epoch": 220.7079646017699, "grad_norm": 0.07301726937294006, "learning_rate": 8.951889566108804e-05, "loss": 0.0041, "step": 24940 }, { "epoch": 220.79646017699116, "grad_norm": 0.06163802370429039, "learning_rate": 8.950876402049606e-05, "loss": 0.0036, "step": 24950 }, { "epoch": 220.8849557522124, "grad_norm": 0.055513035506010056, "learning_rate": 8.949862805928504e-05, "loss": 0.0036, "step": 24960 }, { "epoch": 220.97345132743362, "grad_norm": 0.06848324835300446, "learning_rate": 8.948848777856343e-05, "loss": 0.0035, "step": 24970 }, { "epoch": 221.06194690265488, "grad_norm": 0.062350645661354065, "learning_rate": 8.947834317944017e-05, "loss": 0.0034, "step": 24980 }, { "epoch": 221.1504424778761, "grad_norm": 0.06463529169559479, "learning_rate": 8.946819426302466e-05, "loss": 0.0039, "step": 24990 }, { "epoch": 221.23893805309734, "grad_norm": 0.06696197390556335, "learning_rate": 8.945804103042676e-05, "loss": 0.0039, "step": 25000 }, { "epoch": 221.32743362831857, "grad_norm": 0.05802430585026741, "learning_rate": 8.944788348275681e-05, "loss": 0.0032, "step": 25010 }, { "epoch": 221.41592920353983, "grad_norm": 0.06944183260202408, "learning_rate": 8.943772162112565e-05, "loss": 0.0035, "step": 25020 }, { "epoch": 221.50442477876106, "grad_norm": 0.05761045590043068, "learning_rate": 8.942755544664454e-05, "loss": 0.0037, "step": 25030 }, { "epoch": 221.5929203539823, "grad_norm": 0.05732056871056557, "learning_rate": 8.941738496042525e-05, "loss": 0.0032, "step": 25040 }, { "epoch": 221.68141592920355, "grad_norm": 0.06501127034425735, "learning_rate": 8.940721016357999e-05, "loss": 0.0036, "step": 25050 }, { "epoch": 221.76991150442478, "grad_norm": 0.07099594175815582, "learning_rate": 8.939703105722148e-05, "loss": 0.0033, "step": 25060 }, { "epoch": 221.858407079646, "grad_norm": 0.07153046876192093, "learning_rate": 8.93868476424629e-05, "loss": 0.0045, "step": 25070 }, { "epoch": 221.94690265486724, "grad_norm": 0.08770574629306793, "learning_rate": 8.937665992041786e-05, "loss": 0.0034, "step": 25080 }, { "epoch": 222.0353982300885, "grad_norm": 0.06495349854230881, "learning_rate": 8.93664678922005e-05, "loss": 0.0039, "step": 25090 }, { "epoch": 222.12389380530973, "grad_norm": 0.06946646422147751, "learning_rate": 8.93562715589254e-05, "loss": 0.0036, "step": 25100 }, { "epoch": 222.21238938053096, "grad_norm": 0.053379300981760025, "learning_rate": 8.934607092170762e-05, "loss": 0.0034, "step": 25110 }, { "epoch": 222.30088495575222, "grad_norm": 0.076665498316288, "learning_rate": 8.933586598166266e-05, "loss": 0.0032, "step": 25120 }, { "epoch": 222.38938053097345, "grad_norm": 0.09874489158391953, "learning_rate": 8.932565673990655e-05, "loss": 0.0039, "step": 25130 }, { "epoch": 222.47787610619469, "grad_norm": 0.07837827503681183, "learning_rate": 8.931544319755574e-05, "loss": 0.0035, "step": 25140 }, { "epoch": 222.56637168141592, "grad_norm": 0.06459379941225052, "learning_rate": 8.930522535572718e-05, "loss": 0.0035, "step": 25150 }, { "epoch": 222.65486725663717, "grad_norm": 0.06393622606992722, "learning_rate": 8.929500321553826e-05, "loss": 0.0035, "step": 25160 }, { "epoch": 222.7433628318584, "grad_norm": 0.06603694707155228, "learning_rate": 8.928477677810686e-05, "loss": 0.003, "step": 25170 }, { "epoch": 222.83185840707964, "grad_norm": 0.06425154209136963, "learning_rate": 8.927454604455137e-05, "loss": 0.0035, "step": 25180 }, { "epoch": 222.9203539823009, "grad_norm": 0.07523582130670547, "learning_rate": 8.926431101599053e-05, "loss": 0.0036, "step": 25190 }, { "epoch": 223.00884955752213, "grad_norm": 0.08968783915042877, "learning_rate": 8.925407169354369e-05, "loss": 0.0032, "step": 25200 }, { "epoch": 223.09734513274336, "grad_norm": 0.0844515711069107, "learning_rate": 8.92438280783306e-05, "loss": 0.0038, "step": 25210 }, { "epoch": 223.18584070796462, "grad_norm": 0.0855516716837883, "learning_rate": 8.923358017147146e-05, "loss": 0.0034, "step": 25220 }, { "epoch": 223.27433628318585, "grad_norm": 0.09453852474689484, "learning_rate": 8.922332797408697e-05, "loss": 0.0034, "step": 25230 }, { "epoch": 223.36283185840708, "grad_norm": 0.06746950000524521, "learning_rate": 8.921307148729831e-05, "loss": 0.0034, "step": 25240 }, { "epoch": 223.4513274336283, "grad_norm": 0.061182230710983276, "learning_rate": 8.920281071222712e-05, "loss": 0.0034, "step": 25250 }, { "epoch": 223.53982300884957, "grad_norm": 0.0714508667588234, "learning_rate": 8.919254564999548e-05, "loss": 0.0031, "step": 25260 }, { "epoch": 223.6283185840708, "grad_norm": 0.05512872338294983, "learning_rate": 8.918227630172598e-05, "loss": 0.0031, "step": 25270 }, { "epoch": 223.71681415929203, "grad_norm": 0.0644640401005745, "learning_rate": 8.917200266854165e-05, "loss": 0.0043, "step": 25280 }, { "epoch": 223.8053097345133, "grad_norm": 0.06658808887004852, "learning_rate": 8.9161724751566e-05, "loss": 0.0036, "step": 25290 }, { "epoch": 223.89380530973452, "grad_norm": 0.07180739939212799, "learning_rate": 8.915144255192302e-05, "loss": 0.0038, "step": 25300 }, { "epoch": 223.98230088495575, "grad_norm": 0.10012784600257874, "learning_rate": 8.914115607073714e-05, "loss": 0.004, "step": 25310 }, { "epoch": 224.07079646017698, "grad_norm": 0.07063000649213791, "learning_rate": 8.913086530913327e-05, "loss": 0.0037, "step": 25320 }, { "epoch": 224.15929203539824, "grad_norm": 0.07779288291931152, "learning_rate": 8.912057026823681e-05, "loss": 0.0037, "step": 25330 }, { "epoch": 224.24778761061947, "grad_norm": 0.07084037363529205, "learning_rate": 8.91102709491736e-05, "loss": 0.0037, "step": 25340 }, { "epoch": 224.3362831858407, "grad_norm": 0.08665444701910019, "learning_rate": 8.909996735306996e-05, "loss": 0.0035, "step": 25350 }, { "epoch": 224.42477876106196, "grad_norm": 0.0680994912981987, "learning_rate": 8.908965948105268e-05, "loss": 0.0037, "step": 25360 }, { "epoch": 224.5132743362832, "grad_norm": 0.07220783084630966, "learning_rate": 8.907934733424901e-05, "loss": 0.0035, "step": 25370 }, { "epoch": 224.60176991150442, "grad_norm": 0.07290265709161758, "learning_rate": 8.906903091378666e-05, "loss": 0.0033, "step": 25380 }, { "epoch": 224.69026548672565, "grad_norm": 0.0652046874165535, "learning_rate": 8.905871022079384e-05, "loss": 0.0032, "step": 25390 }, { "epoch": 224.7787610619469, "grad_norm": 0.05633305013179779, "learning_rate": 8.90483852563992e-05, "loss": 0.0032, "step": 25400 }, { "epoch": 224.86725663716814, "grad_norm": 0.05948828160762787, "learning_rate": 8.903805602173185e-05, "loss": 0.0032, "step": 25410 }, { "epoch": 224.95575221238937, "grad_norm": 0.06098970025777817, "learning_rate": 8.902772251792137e-05, "loss": 0.0033, "step": 25420 }, { "epoch": 225.04424778761063, "grad_norm": 0.05994003266096115, "learning_rate": 8.901738474609786e-05, "loss": 0.0031, "step": 25430 }, { "epoch": 225.13274336283186, "grad_norm": 0.08047087490558624, "learning_rate": 8.900704270739179e-05, "loss": 0.0036, "step": 25440 }, { "epoch": 225.2212389380531, "grad_norm": 0.06435946375131607, "learning_rate": 8.89966964029342e-05, "loss": 0.0032, "step": 25450 }, { "epoch": 225.30973451327435, "grad_norm": 0.07007691264152527, "learning_rate": 8.898634583385652e-05, "loss": 0.0033, "step": 25460 }, { "epoch": 225.39823008849558, "grad_norm": 0.06351058930158615, "learning_rate": 8.897599100129065e-05, "loss": 0.0029, "step": 25470 }, { "epoch": 225.4867256637168, "grad_norm": 0.08369868993759155, "learning_rate": 8.896563190636903e-05, "loss": 0.0035, "step": 25480 }, { "epoch": 225.57522123893804, "grad_norm": 0.069444440305233, "learning_rate": 8.895526855022448e-05, "loss": 0.0031, "step": 25490 }, { "epoch": 225.6637168141593, "grad_norm": 0.08747955411672592, "learning_rate": 8.894490093399033e-05, "loss": 0.0034, "step": 25500 }, { "epoch": 225.75221238938053, "grad_norm": 0.07100962847471237, "learning_rate": 8.893452905880035e-05, "loss": 0.0031, "step": 25510 }, { "epoch": 225.84070796460176, "grad_norm": 0.06962308287620544, "learning_rate": 8.892415292578883e-05, "loss": 0.0036, "step": 25520 }, { "epoch": 225.92920353982302, "grad_norm": 0.08004166930913925, "learning_rate": 8.891377253609046e-05, "loss": 0.0038, "step": 25530 }, { "epoch": 226.01769911504425, "grad_norm": 0.07521668076515198, "learning_rate": 8.890338789084043e-05, "loss": 0.0033, "step": 25540 }, { "epoch": 226.10619469026548, "grad_norm": 0.0931495949625969, "learning_rate": 8.88929989911744e-05, "loss": 0.0035, "step": 25550 }, { "epoch": 226.1946902654867, "grad_norm": 0.08795981854200363, "learning_rate": 8.888260583822847e-05, "loss": 0.0039, "step": 25560 }, { "epoch": 226.28318584070797, "grad_norm": 0.08782129734754562, "learning_rate": 8.887220843313921e-05, "loss": 0.0036, "step": 25570 }, { "epoch": 226.3716814159292, "grad_norm": 0.08343826234340668, "learning_rate": 8.88618067770437e-05, "loss": 0.0038, "step": 25580 }, { "epoch": 226.46017699115043, "grad_norm": 0.07868389040231705, "learning_rate": 8.885140087107942e-05, "loss": 0.0037, "step": 25590 }, { "epoch": 226.5486725663717, "grad_norm": 0.09629865735769272, "learning_rate": 8.884099071638436e-05, "loss": 0.0034, "step": 25600 }, { "epoch": 226.63716814159292, "grad_norm": 0.07448412477970123, "learning_rate": 8.883057631409695e-05, "loss": 0.0042, "step": 25610 }, { "epoch": 226.72566371681415, "grad_norm": 0.061576247215270996, "learning_rate": 8.882015766535608e-05, "loss": 0.0037, "step": 25620 }, { "epoch": 226.81415929203538, "grad_norm": 0.0768008753657341, "learning_rate": 8.880973477130115e-05, "loss": 0.0032, "step": 25630 }, { "epoch": 226.90265486725664, "grad_norm": 0.08236567676067352, "learning_rate": 8.879930763307197e-05, "loss": 0.0035, "step": 25640 }, { "epoch": 226.99115044247787, "grad_norm": 0.0642288476228714, "learning_rate": 8.878887625180884e-05, "loss": 0.0034, "step": 25650 }, { "epoch": 227.0796460176991, "grad_norm": 0.06926168501377106, "learning_rate": 8.877844062865253e-05, "loss": 0.0033, "step": 25660 }, { "epoch": 227.16814159292036, "grad_norm": 0.07114057242870331, "learning_rate": 8.876800076474424e-05, "loss": 0.0036, "step": 25670 }, { "epoch": 227.2566371681416, "grad_norm": 0.0693613663315773, "learning_rate": 8.875755666122568e-05, "loss": 0.0034, "step": 25680 }, { "epoch": 227.34513274336283, "grad_norm": 0.083440400660038, "learning_rate": 8.8747108319239e-05, "loss": 0.0033, "step": 25690 }, { "epoch": 227.43362831858408, "grad_norm": 0.08012443780899048, "learning_rate": 8.87366557399268e-05, "loss": 0.0035, "step": 25700 }, { "epoch": 227.52212389380531, "grad_norm": 0.07514654099941254, "learning_rate": 8.872619892443217e-05, "loss": 0.0039, "step": 25710 }, { "epoch": 227.61061946902655, "grad_norm": 0.08335541933774948, "learning_rate": 8.871573787389865e-05, "loss": 0.0038, "step": 25720 }, { "epoch": 227.69911504424778, "grad_norm": 0.06223238632082939, "learning_rate": 8.870527258947024e-05, "loss": 0.0033, "step": 25730 }, { "epoch": 227.78761061946904, "grad_norm": 0.050697341561317444, "learning_rate": 8.869480307229143e-05, "loss": 0.0033, "step": 25740 }, { "epoch": 227.87610619469027, "grad_norm": 0.0693097785115242, "learning_rate": 8.868432932350712e-05, "loss": 0.0032, "step": 25750 }, { "epoch": 227.9646017699115, "grad_norm": 0.07008224725723267, "learning_rate": 8.867385134426272e-05, "loss": 0.0031, "step": 25760 }, { "epoch": 228.05309734513276, "grad_norm": 0.06786320358514786, "learning_rate": 8.866336913570407e-05, "loss": 0.0033, "step": 25770 }, { "epoch": 228.141592920354, "grad_norm": 0.1065177470445633, "learning_rate": 8.865288269897751e-05, "loss": 0.0039, "step": 25780 }, { "epoch": 228.23008849557522, "grad_norm": 0.09451400488615036, "learning_rate": 8.864239203522981e-05, "loss": 0.0039, "step": 25790 }, { "epoch": 228.31858407079645, "grad_norm": 0.0908544659614563, "learning_rate": 8.863189714560822e-05, "loss": 0.0035, "step": 25800 }, { "epoch": 228.4070796460177, "grad_norm": 0.058955494314432144, "learning_rate": 8.862139803126043e-05, "loss": 0.0033, "step": 25810 }, { "epoch": 228.49557522123894, "grad_norm": 0.08240800350904465, "learning_rate": 8.861089469333463e-05, "loss": 0.0031, "step": 25820 }, { "epoch": 228.58407079646017, "grad_norm": 0.06502421200275421, "learning_rate": 8.860038713297944e-05, "loss": 0.0038, "step": 25830 }, { "epoch": 228.67256637168143, "grad_norm": 0.07117869704961777, "learning_rate": 8.858987535134394e-05, "loss": 0.0032, "step": 25840 }, { "epoch": 228.76106194690266, "grad_norm": 0.0700320452451706, "learning_rate": 8.857935934957769e-05, "loss": 0.0033, "step": 25850 }, { "epoch": 228.8495575221239, "grad_norm": 0.06600068509578705, "learning_rate": 8.856883912883071e-05, "loss": 0.0033, "step": 25860 }, { "epoch": 228.93805309734512, "grad_norm": 0.08924765139818192, "learning_rate": 8.855831469025346e-05, "loss": 0.0039, "step": 25870 }, { "epoch": 229.02654867256638, "grad_norm": 0.07005716115236282, "learning_rate": 8.854778603499689e-05, "loss": 0.0036, "step": 25880 }, { "epoch": 229.1150442477876, "grad_norm": 0.06872308999300003, "learning_rate": 8.85372531642124e-05, "loss": 0.0033, "step": 25890 }, { "epoch": 229.20353982300884, "grad_norm": 0.06391239911317825, "learning_rate": 8.852671607905185e-05, "loss": 0.0034, "step": 25900 }, { "epoch": 229.2920353982301, "grad_norm": 0.055492691695690155, "learning_rate": 8.851617478066754e-05, "loss": 0.0035, "step": 25910 }, { "epoch": 229.38053097345133, "grad_norm": 0.07870180904865265, "learning_rate": 8.850562927021227e-05, "loss": 0.0036, "step": 25920 }, { "epoch": 229.46902654867256, "grad_norm": 0.08819001913070679, "learning_rate": 8.849507954883928e-05, "loss": 0.0034, "step": 25930 }, { "epoch": 229.55752212389382, "grad_norm": 0.08731727302074432, "learning_rate": 8.848452561770226e-05, "loss": 0.0035, "step": 25940 }, { "epoch": 229.64601769911505, "grad_norm": 0.085537388920784, "learning_rate": 8.847396747795538e-05, "loss": 0.0035, "step": 25950 }, { "epoch": 229.73451327433628, "grad_norm": 0.07838968187570572, "learning_rate": 8.846340513075327e-05, "loss": 0.0037, "step": 25960 }, { "epoch": 229.8230088495575, "grad_norm": 0.0721469298005104, "learning_rate": 8.845283857725099e-05, "loss": 0.0032, "step": 25970 }, { "epoch": 229.91150442477877, "grad_norm": 0.07279793173074722, "learning_rate": 8.844226781860409e-05, "loss": 0.0033, "step": 25980 }, { "epoch": 230.0, "grad_norm": 0.07539823651313782, "learning_rate": 8.84316928559686e-05, "loss": 0.0034, "step": 25990 }, { "epoch": 230.08849557522123, "grad_norm": 0.08499997109174728, "learning_rate": 8.842111369050094e-05, "loss": 0.0035, "step": 26000 }, { "epoch": 230.1769911504425, "grad_norm": 0.0595211423933506, "learning_rate": 8.841053032335808e-05, "loss": 0.0034, "step": 26010 }, { "epoch": 230.26548672566372, "grad_norm": 0.09984140843153, "learning_rate": 8.839994275569735e-05, "loss": 0.0032, "step": 26020 }, { "epoch": 230.35398230088495, "grad_norm": 0.07632765918970108, "learning_rate": 8.838935098867662e-05, "loss": 0.0034, "step": 26030 }, { "epoch": 230.44247787610618, "grad_norm": 0.11010551452636719, "learning_rate": 8.837875502345418e-05, "loss": 0.0034, "step": 26040 }, { "epoch": 230.53097345132744, "grad_norm": 0.08206736296415329, "learning_rate": 8.83681548611888e-05, "loss": 0.0035, "step": 26050 }, { "epoch": 230.61946902654867, "grad_norm": 0.07291911542415619, "learning_rate": 8.835755050303969e-05, "loss": 0.0032, "step": 26060 }, { "epoch": 230.7079646017699, "grad_norm": 0.08314281702041626, "learning_rate": 8.834694195016653e-05, "loss": 0.0034, "step": 26070 }, { "epoch": 230.79646017699116, "grad_norm": 0.05861566215753555, "learning_rate": 8.833632920372942e-05, "loss": 0.0032, "step": 26080 }, { "epoch": 230.8849557522124, "grad_norm": 0.07550393044948578, "learning_rate": 8.832571226488903e-05, "loss": 0.0036, "step": 26090 }, { "epoch": 230.97345132743362, "grad_norm": 0.07166876643896103, "learning_rate": 8.831509113480634e-05, "loss": 0.0034, "step": 26100 }, { "epoch": 231.06194690265488, "grad_norm": 0.08739220350980759, "learning_rate": 8.83044658146429e-05, "loss": 0.0037, "step": 26110 }, { "epoch": 231.1504424778761, "grad_norm": 0.06610670685768127, "learning_rate": 8.829383630556067e-05, "loss": 0.0036, "step": 26120 }, { "epoch": 231.23893805309734, "grad_norm": 0.10333293676376343, "learning_rate": 8.828320260872207e-05, "loss": 0.0035, "step": 26130 }, { "epoch": 231.32743362831857, "grad_norm": 0.06637080758810043, "learning_rate": 8.827256472529e-05, "loss": 0.0036, "step": 26140 }, { "epoch": 231.41592920353983, "grad_norm": 0.06333745270967484, "learning_rate": 8.826192265642778e-05, "loss": 0.0037, "step": 26150 }, { "epoch": 231.50442477876106, "grad_norm": 0.07165852934122086, "learning_rate": 8.825127640329923e-05, "loss": 0.0039, "step": 26160 }, { "epoch": 231.5929203539823, "grad_norm": 0.09004877507686615, "learning_rate": 8.824062596706861e-05, "loss": 0.0037, "step": 26170 }, { "epoch": 231.68141592920355, "grad_norm": 0.06592722237110138, "learning_rate": 8.822997134890062e-05, "loss": 0.0033, "step": 26180 }, { "epoch": 231.76991150442478, "grad_norm": 0.06878456473350525, "learning_rate": 8.821931254996044e-05, "loss": 0.0031, "step": 26190 }, { "epoch": 231.858407079646, "grad_norm": 0.07393597811460495, "learning_rate": 8.82086495714137e-05, "loss": 0.0033, "step": 26200 }, { "epoch": 231.94690265486724, "grad_norm": 0.06932399421930313, "learning_rate": 8.81979824144265e-05, "loss": 0.0033, "step": 26210 }, { "epoch": 232.0353982300885, "grad_norm": 0.06092839315533638, "learning_rate": 8.818731108016536e-05, "loss": 0.003, "step": 26220 }, { "epoch": 232.12389380530973, "grad_norm": 0.06663522124290466, "learning_rate": 8.81766355697973e-05, "loss": 0.003, "step": 26230 }, { "epoch": 232.21238938053096, "grad_norm": 0.0485076904296875, "learning_rate": 8.816595588448977e-05, "loss": 0.0032, "step": 26240 }, { "epoch": 232.30088495575222, "grad_norm": 0.07002142816781998, "learning_rate": 8.81552720254107e-05, "loss": 0.0039, "step": 26250 }, { "epoch": 232.38938053097345, "grad_norm": 0.08969194442033768, "learning_rate": 8.814458399372842e-05, "loss": 0.0039, "step": 26260 }, { "epoch": 232.47787610619469, "grad_norm": 0.11944957077503204, "learning_rate": 8.813389179061181e-05, "loss": 0.0037, "step": 26270 }, { "epoch": 232.56637168141592, "grad_norm": 0.07423993945121765, "learning_rate": 8.812319541723012e-05, "loss": 0.0033, "step": 26280 }, { "epoch": 232.65486725663717, "grad_norm": 0.07952113449573517, "learning_rate": 8.811249487475309e-05, "loss": 0.0037, "step": 26290 }, { "epoch": 232.7433628318584, "grad_norm": 0.07490961253643036, "learning_rate": 8.810179016435092e-05, "loss": 0.0035, "step": 26300 }, { "epoch": 232.83185840707964, "grad_norm": 0.0818890929222107, "learning_rate": 8.809108128719428e-05, "loss": 0.0037, "step": 26310 }, { "epoch": 232.9203539823009, "grad_norm": 0.06267788261175156, "learning_rate": 8.808036824445424e-05, "loss": 0.0031, "step": 26320 }, { "epoch": 233.00884955752213, "grad_norm": 0.0744706243276596, "learning_rate": 8.806965103730238e-05, "loss": 0.0031, "step": 26330 }, { "epoch": 233.09734513274336, "grad_norm": 0.07173064351081848, "learning_rate": 8.805892966691074e-05, "loss": 0.0034, "step": 26340 }, { "epoch": 233.18584070796462, "grad_norm": 0.0739995613694191, "learning_rate": 8.804820413445175e-05, "loss": 0.0035, "step": 26350 }, { "epoch": 233.27433628318585, "grad_norm": 0.08190913498401642, "learning_rate": 8.803747444109837e-05, "loss": 0.0032, "step": 26360 }, { "epoch": 233.36283185840708, "grad_norm": 0.08579634130001068, "learning_rate": 8.802674058802399e-05, "loss": 0.0038, "step": 26370 }, { "epoch": 233.4513274336283, "grad_norm": 0.0763777419924736, "learning_rate": 8.801600257640241e-05, "loss": 0.0035, "step": 26380 }, { "epoch": 233.53982300884957, "grad_norm": 0.08644838631153107, "learning_rate": 8.800526040740795e-05, "loss": 0.0034, "step": 26390 }, { "epoch": 233.6283185840708, "grad_norm": 0.0687650814652443, "learning_rate": 8.799451408221535e-05, "loss": 0.0032, "step": 26400 }, { "epoch": 233.71681415929203, "grad_norm": 0.08150175213813782, "learning_rate": 8.798376360199982e-05, "loss": 0.0036, "step": 26410 }, { "epoch": 233.8053097345133, "grad_norm": 0.06787759065628052, "learning_rate": 8.797300896793701e-05, "loss": 0.0039, "step": 26420 }, { "epoch": 233.89380530973452, "grad_norm": 0.06215018033981323, "learning_rate": 8.796225018120302e-05, "loss": 0.0034, "step": 26430 }, { "epoch": 233.98230088495575, "grad_norm": 0.05255233868956566, "learning_rate": 8.795148724297444e-05, "loss": 0.0034, "step": 26440 }, { "epoch": 234.07079646017698, "grad_norm": 0.06377724558115005, "learning_rate": 8.794072015442825e-05, "loss": 0.0029, "step": 26450 }, { "epoch": 234.15929203539824, "grad_norm": 0.06133386120200157, "learning_rate": 8.792994891674198e-05, "loss": 0.0033, "step": 26460 }, { "epoch": 234.24778761061947, "grad_norm": 0.051648613065481186, "learning_rate": 8.79191735310935e-05, "loss": 0.0027, "step": 26470 }, { "epoch": 234.3362831858407, "grad_norm": 0.0750749483704567, "learning_rate": 8.790839399866122e-05, "loss": 0.0037, "step": 26480 }, { "epoch": 234.42477876106196, "grad_norm": 0.05908052995800972, "learning_rate": 8.789761032062397e-05, "loss": 0.0037, "step": 26490 }, { "epoch": 234.5132743362832, "grad_norm": 0.07009205967187881, "learning_rate": 8.788682249816103e-05, "loss": 0.0037, "step": 26500 }, { "epoch": 234.60176991150442, "grad_norm": 0.055598530918359756, "learning_rate": 8.787603053245215e-05, "loss": 0.0031, "step": 26510 }, { "epoch": 234.69026548672565, "grad_norm": 0.0572778582572937, "learning_rate": 8.78652344246775e-05, "loss": 0.003, "step": 26520 }, { "epoch": 234.7787610619469, "grad_norm": 0.06387221068143845, "learning_rate": 8.785443417601776e-05, "loss": 0.003, "step": 26530 }, { "epoch": 234.86725663716814, "grad_norm": 0.07304037362337112, "learning_rate": 8.784362978765401e-05, "loss": 0.0033, "step": 26540 }, { "epoch": 234.95575221238937, "grad_norm": 0.07059606909751892, "learning_rate": 8.783282126076779e-05, "loss": 0.0034, "step": 26550 }, { "epoch": 235.04424778761063, "grad_norm": 0.08289309591054916, "learning_rate": 8.782200859654112e-05, "loss": 0.0035, "step": 26560 }, { "epoch": 235.13274336283186, "grad_norm": 0.07437478005886078, "learning_rate": 8.781119179615646e-05, "loss": 0.0035, "step": 26570 }, { "epoch": 235.2212389380531, "grad_norm": 0.08741918951272964, "learning_rate": 8.780037086079674e-05, "loss": 0.0033, "step": 26580 }, { "epoch": 235.30973451327435, "grad_norm": 0.06672391295433044, "learning_rate": 8.778954579164527e-05, "loss": 0.0033, "step": 26590 }, { "epoch": 235.39823008849558, "grad_norm": 0.07918433099985123, "learning_rate": 8.777871658988588e-05, "loss": 0.0032, "step": 26600 }, { "epoch": 235.4867256637168, "grad_norm": 0.07492371648550034, "learning_rate": 8.776788325670285e-05, "loss": 0.0033, "step": 26610 }, { "epoch": 235.57522123893804, "grad_norm": 0.06537526845932007, "learning_rate": 8.775704579328089e-05, "loss": 0.0038, "step": 26620 }, { "epoch": 235.6637168141593, "grad_norm": 0.06130046397447586, "learning_rate": 8.774620420080517e-05, "loss": 0.0032, "step": 26630 }, { "epoch": 235.75221238938053, "grad_norm": 0.08269962668418884, "learning_rate": 8.773535848046131e-05, "loss": 0.0036, "step": 26640 }, { "epoch": 235.84070796460176, "grad_norm": 0.08128021657466888, "learning_rate": 8.772450863343538e-05, "loss": 0.0033, "step": 26650 }, { "epoch": 235.92920353982302, "grad_norm": 0.09423965215682983, "learning_rate": 8.77136546609139e-05, "loss": 0.0033, "step": 26660 }, { "epoch": 236.01769911504425, "grad_norm": 0.08704504370689392, "learning_rate": 8.770279656408385e-05, "loss": 0.0037, "step": 26670 }, { "epoch": 236.10619469026548, "grad_norm": 0.09186926484107971, "learning_rate": 8.769193434413265e-05, "loss": 0.0035, "step": 26680 }, { "epoch": 236.1946902654867, "grad_norm": 0.07595513761043549, "learning_rate": 8.76810680022482e-05, "loss": 0.0034, "step": 26690 }, { "epoch": 236.28318584070797, "grad_norm": 0.07699780911207199, "learning_rate": 8.767019753961878e-05, "loss": 0.003, "step": 26700 }, { "epoch": 236.3716814159292, "grad_norm": 0.06329695135354996, "learning_rate": 8.765932295743321e-05, "loss": 0.0032, "step": 26710 }, { "epoch": 236.46017699115043, "grad_norm": 0.07473143935203552, "learning_rate": 8.764844425688068e-05, "loss": 0.0034, "step": 26720 }, { "epoch": 236.5486725663717, "grad_norm": 0.08552607148885727, "learning_rate": 8.763756143915092e-05, "loss": 0.0035, "step": 26730 }, { "epoch": 236.63716814159292, "grad_norm": 0.0883508175611496, "learning_rate": 8.7626674505434e-05, "loss": 0.0036, "step": 26740 }, { "epoch": 236.72566371681415, "grad_norm": 0.05940056964755058, "learning_rate": 8.761578345692053e-05, "loss": 0.0036, "step": 26750 }, { "epoch": 236.81415929203538, "grad_norm": 0.06908457726240158, "learning_rate": 8.760488829480156e-05, "loss": 0.0035, "step": 26760 }, { "epoch": 236.90265486725664, "grad_norm": 0.09380361437797546, "learning_rate": 8.759398902026854e-05, "loss": 0.0037, "step": 26770 }, { "epoch": 236.99115044247787, "grad_norm": 0.08764813840389252, "learning_rate": 8.758308563451339e-05, "loss": 0.0039, "step": 26780 }, { "epoch": 237.0796460176991, "grad_norm": 0.0707768052816391, "learning_rate": 8.75721781387285e-05, "loss": 0.0035, "step": 26790 }, { "epoch": 237.16814159292036, "grad_norm": 0.10423657298088074, "learning_rate": 8.75612665341067e-05, "loss": 0.0036, "step": 26800 }, { "epoch": 237.2566371681416, "grad_norm": 0.07636339962482452, "learning_rate": 8.755035082184126e-05, "loss": 0.004, "step": 26810 }, { "epoch": 237.34513274336283, "grad_norm": 0.07745011895895004, "learning_rate": 8.753943100312592e-05, "loss": 0.0032, "step": 26820 }, { "epoch": 237.43362831858408, "grad_norm": 0.06332860887050629, "learning_rate": 8.752850707915484e-05, "loss": 0.003, "step": 26830 }, { "epoch": 237.52212389380531, "grad_norm": 0.06311886012554169, "learning_rate": 8.751757905112264e-05, "loss": 0.0031, "step": 26840 }, { "epoch": 237.61061946902655, "grad_norm": 0.06340576708316803, "learning_rate": 8.75066469202244e-05, "loss": 0.0029, "step": 26850 }, { "epoch": 237.69911504424778, "grad_norm": 0.0955771952867508, "learning_rate": 8.749571068765567e-05, "loss": 0.0037, "step": 26860 }, { "epoch": 237.78761061946904, "grad_norm": 0.0909370556473732, "learning_rate": 8.748477035461238e-05, "loss": 0.0037, "step": 26870 }, { "epoch": 237.87610619469027, "grad_norm": 0.10718273371458054, "learning_rate": 8.747382592229095e-05, "loss": 0.0036, "step": 26880 }, { "epoch": 237.9646017699115, "grad_norm": 0.08805255591869354, "learning_rate": 8.746287739188828e-05, "loss": 0.0032, "step": 26890 }, { "epoch": 238.05309734513276, "grad_norm": 0.07930716127157211, "learning_rate": 8.745192476460165e-05, "loss": 0.0032, "step": 26900 }, { "epoch": 238.141592920354, "grad_norm": 0.05763214826583862, "learning_rate": 8.744096804162882e-05, "loss": 0.0034, "step": 26910 }, { "epoch": 238.23008849557522, "grad_norm": 0.07115606218576431, "learning_rate": 8.743000722416804e-05, "loss": 0.0034, "step": 26920 }, { "epoch": 238.31858407079645, "grad_norm": 0.06727219372987747, "learning_rate": 8.741904231341793e-05, "loss": 0.0032, "step": 26930 }, { "epoch": 238.4070796460177, "grad_norm": 0.07728970795869827, "learning_rate": 8.740807331057762e-05, "loss": 0.0036, "step": 26940 }, { "epoch": 238.49557522123894, "grad_norm": 0.08866860717535019, "learning_rate": 8.739710021684667e-05, "loss": 0.0035, "step": 26950 }, { "epoch": 238.58407079646017, "grad_norm": 0.08448145538568497, "learning_rate": 8.738612303342503e-05, "loss": 0.0038, "step": 26960 }, { "epoch": 238.67256637168143, "grad_norm": 0.06011221185326576, "learning_rate": 8.73751417615132e-05, "loss": 0.0034, "step": 26970 }, { "epoch": 238.76106194690266, "grad_norm": 0.07255960255861282, "learning_rate": 8.736415640231208e-05, "loss": 0.0037, "step": 26980 }, { "epoch": 238.8495575221239, "grad_norm": 0.05211587995290756, "learning_rate": 8.735316695702297e-05, "loss": 0.0035, "step": 26990 }, { "epoch": 238.93805309734512, "grad_norm": 0.07819949835538864, "learning_rate": 8.734217342684769e-05, "loss": 0.0034, "step": 27000 }, { "epoch": 239.02654867256638, "grad_norm": 0.07173926383256912, "learning_rate": 8.733117581298847e-05, "loss": 0.0037, "step": 27010 }, { "epoch": 239.1150442477876, "grad_norm": 0.05779208987951279, "learning_rate": 8.732017411664796e-05, "loss": 0.003, "step": 27020 }, { "epoch": 239.20353982300884, "grad_norm": 0.06899895519018173, "learning_rate": 8.730916833902936e-05, "loss": 0.0029, "step": 27030 }, { "epoch": 239.2920353982301, "grad_norm": 0.06041233614087105, "learning_rate": 8.729815848133618e-05, "loss": 0.0036, "step": 27040 }, { "epoch": 239.38053097345133, "grad_norm": 0.08026457577943802, "learning_rate": 8.728714454477247e-05, "loss": 0.0036, "step": 27050 }, { "epoch": 239.46902654867256, "grad_norm": 0.06257560849189758, "learning_rate": 8.727612653054269e-05, "loss": 0.0035, "step": 27060 }, { "epoch": 239.55752212389382, "grad_norm": 0.07400017231702805, "learning_rate": 8.726510443985176e-05, "loss": 0.0033, "step": 27070 }, { "epoch": 239.64601769911505, "grad_norm": 0.04970429465174675, "learning_rate": 8.725407827390503e-05, "loss": 0.0034, "step": 27080 }, { "epoch": 239.73451327433628, "grad_norm": 0.09084036946296692, "learning_rate": 8.724304803390833e-05, "loss": 0.0033, "step": 27090 }, { "epoch": 239.8230088495575, "grad_norm": 0.05956766754388809, "learning_rate": 8.723201372106788e-05, "loss": 0.0028, "step": 27100 }, { "epoch": 239.91150442477877, "grad_norm": 0.0815809965133667, "learning_rate": 8.722097533659038e-05, "loss": 0.0034, "step": 27110 }, { "epoch": 240.0, "grad_norm": 0.07560472190380096, "learning_rate": 8.720993288168299e-05, "loss": 0.0035, "step": 27120 }, { "epoch": 240.08849557522123, "grad_norm": 0.101240374147892, "learning_rate": 8.719888635755327e-05, "loss": 0.0035, "step": 27130 }, { "epoch": 240.1769911504425, "grad_norm": 0.08339432626962662, "learning_rate": 8.718783576540928e-05, "loss": 0.0034, "step": 27140 }, { "epoch": 240.26548672566372, "grad_norm": 0.07372108101844788, "learning_rate": 8.717678110645948e-05, "loss": 0.0035, "step": 27150 }, { "epoch": 240.35398230088495, "grad_norm": 0.07629609107971191, "learning_rate": 8.716572238191279e-05, "loss": 0.0032, "step": 27160 }, { "epoch": 240.44247787610618, "grad_norm": 0.05740641430020332, "learning_rate": 8.715465959297857e-05, "loss": 0.0032, "step": 27170 }, { "epoch": 240.53097345132744, "grad_norm": 0.06663396954536438, "learning_rate": 8.714359274086665e-05, "loss": 0.0031, "step": 27180 }, { "epoch": 240.61946902654867, "grad_norm": 0.057327695190906525, "learning_rate": 8.713252182678726e-05, "loss": 0.0031, "step": 27190 }, { "epoch": 240.7079646017699, "grad_norm": 0.07196827232837677, "learning_rate": 8.712144685195112e-05, "loss": 0.0036, "step": 27200 }, { "epoch": 240.79646017699116, "grad_norm": 0.07010219246149063, "learning_rate": 8.711036781756936e-05, "loss": 0.0031, "step": 27210 }, { "epoch": 240.8849557522124, "grad_norm": 0.0649804100394249, "learning_rate": 8.709928472485357e-05, "loss": 0.0034, "step": 27220 }, { "epoch": 240.97345132743362, "grad_norm": 0.04239922761917114, "learning_rate": 8.708819757501579e-05, "loss": 0.0032, "step": 27230 }, { "epoch": 241.06194690265488, "grad_norm": 0.07121150940656662, "learning_rate": 8.707710636926846e-05, "loss": 0.0038, "step": 27240 }, { "epoch": 241.1504424778761, "grad_norm": 0.07645291835069656, "learning_rate": 8.706601110882455e-05, "loss": 0.0033, "step": 27250 }, { "epoch": 241.23893805309734, "grad_norm": 0.08237289637327194, "learning_rate": 8.705491179489738e-05, "loss": 0.003, "step": 27260 }, { "epoch": 241.32743362831857, "grad_norm": 0.08581662178039551, "learning_rate": 8.704380842870077e-05, "loss": 0.0034, "step": 27270 }, { "epoch": 241.41592920353983, "grad_norm": 0.07561948150396347, "learning_rate": 8.703270101144895e-05, "loss": 0.0032, "step": 27280 }, { "epoch": 241.50442477876106, "grad_norm": 0.07059737294912338, "learning_rate": 8.702158954435664e-05, "loss": 0.0032, "step": 27290 }, { "epoch": 241.5929203539823, "grad_norm": 0.07296078652143478, "learning_rate": 8.701047402863896e-05, "loss": 0.0033, "step": 27300 }, { "epoch": 241.68141592920355, "grad_norm": 0.07799617201089859, "learning_rate": 8.699935446551148e-05, "loss": 0.0038, "step": 27310 }, { "epoch": 241.76991150442478, "grad_norm": 0.07239911705255508, "learning_rate": 8.698823085619022e-05, "loss": 0.0033, "step": 27320 }, { "epoch": 241.858407079646, "grad_norm": 0.0816146731376648, "learning_rate": 8.697710320189166e-05, "loss": 0.0036, "step": 27330 }, { "epoch": 241.94690265486724, "grad_norm": 0.07602347433567047, "learning_rate": 8.696597150383268e-05, "loss": 0.0032, "step": 27340 }, { "epoch": 242.0353982300885, "grad_norm": 0.06409493833780289, "learning_rate": 8.695483576323063e-05, "loss": 0.0038, "step": 27350 }, { "epoch": 242.12389380530973, "grad_norm": 0.09464302659034729, "learning_rate": 8.69436959813033e-05, "loss": 0.0035, "step": 27360 }, { "epoch": 242.21238938053096, "grad_norm": 0.10575643926858902, "learning_rate": 8.693255215926892e-05, "loss": 0.0034, "step": 27370 }, { "epoch": 242.30088495575222, "grad_norm": 0.07708536833524704, "learning_rate": 8.692140429834617e-05, "loss": 0.003, "step": 27380 }, { "epoch": 242.38938053097345, "grad_norm": 0.10114165395498276, "learning_rate": 8.691025239975415e-05, "loss": 0.0039, "step": 27390 }, { "epoch": 242.47787610619469, "grad_norm": 0.08773376047611237, "learning_rate": 8.689909646471243e-05, "loss": 0.0035, "step": 27400 }, { "epoch": 242.56637168141592, "grad_norm": 0.11537333577871323, "learning_rate": 8.688793649444099e-05, "loss": 0.0037, "step": 27410 }, { "epoch": 242.65486725663717, "grad_norm": 0.07751411944627762, "learning_rate": 8.687677249016029e-05, "loss": 0.0034, "step": 27420 }, { "epoch": 242.7433628318584, "grad_norm": 0.08539415895938873, "learning_rate": 8.686560445309118e-05, "loss": 0.0035, "step": 27430 }, { "epoch": 242.83185840707964, "grad_norm": 0.06872936338186264, "learning_rate": 8.685443238445499e-05, "loss": 0.0032, "step": 27440 }, { "epoch": 242.9203539823009, "grad_norm": 0.06753963232040405, "learning_rate": 8.68432562854735e-05, "loss": 0.0035, "step": 27450 }, { "epoch": 243.00884955752213, "grad_norm": 0.07682932913303375, "learning_rate": 8.683207615736887e-05, "loss": 0.0036, "step": 27460 }, { "epoch": 243.09734513274336, "grad_norm": 0.08534223586320877, "learning_rate": 8.682089200136379e-05, "loss": 0.0033, "step": 27470 }, { "epoch": 243.18584070796462, "grad_norm": 0.07924920320510864, "learning_rate": 8.680970381868132e-05, "loss": 0.0037, "step": 27480 }, { "epoch": 243.27433628318585, "grad_norm": 0.06568797677755356, "learning_rate": 8.679851161054498e-05, "loss": 0.0031, "step": 27490 }, { "epoch": 243.36283185840708, "grad_norm": 0.07763943821191788, "learning_rate": 8.678731537817873e-05, "loss": 0.0031, "step": 27500 }, { "epoch": 243.4513274336283, "grad_norm": 0.05403003469109535, "learning_rate": 8.677611512280697e-05, "loss": 0.0029, "step": 27510 }, { "epoch": 243.53982300884957, "grad_norm": 0.05248148366808891, "learning_rate": 8.676491084565457e-05, "loss": 0.0035, "step": 27520 }, { "epoch": 243.6283185840708, "grad_norm": 0.06397444009780884, "learning_rate": 8.675370254794678e-05, "loss": 0.0038, "step": 27530 }, { "epoch": 243.71681415929203, "grad_norm": 0.08695276081562042, "learning_rate": 8.674249023090935e-05, "loss": 0.0034, "step": 27540 }, { "epoch": 243.8053097345133, "grad_norm": 0.09847042709589005, "learning_rate": 8.673127389576843e-05, "loss": 0.0034, "step": 27550 }, { "epoch": 243.89380530973452, "grad_norm": 0.10192367434501648, "learning_rate": 8.67200535437506e-05, "loss": 0.0039, "step": 27560 }, { "epoch": 243.98230088495575, "grad_norm": 0.0980057641863823, "learning_rate": 8.670882917608296e-05, "loss": 0.0037, "step": 27570 }, { "epoch": 244.07079646017698, "grad_norm": 0.07433212548494339, "learning_rate": 8.669760079399292e-05, "loss": 0.0033, "step": 27580 }, { "epoch": 244.15929203539824, "grad_norm": 0.06901790201663971, "learning_rate": 8.668636839870845e-05, "loss": 0.0035, "step": 27590 }, { "epoch": 244.24778761061947, "grad_norm": 0.0743006020784378, "learning_rate": 8.667513199145789e-05, "loss": 0.003, "step": 27600 }, { "epoch": 244.3362831858407, "grad_norm": 0.08407370746135712, "learning_rate": 8.666389157347002e-05, "loss": 0.0035, "step": 27610 }, { "epoch": 244.42477876106196, "grad_norm": 0.08539833128452301, "learning_rate": 8.66526471459741e-05, "loss": 0.0033, "step": 27620 }, { "epoch": 244.5132743362832, "grad_norm": 0.06554204225540161, "learning_rate": 8.66413987101998e-05, "loss": 0.0033, "step": 27630 }, { "epoch": 244.60176991150442, "grad_norm": 0.055339694023132324, "learning_rate": 8.663014626737723e-05, "loss": 0.0034, "step": 27640 }, { "epoch": 244.69026548672565, "grad_norm": 0.08248431980609894, "learning_rate": 8.661888981873691e-05, "loss": 0.0035, "step": 27650 }, { "epoch": 244.7787610619469, "grad_norm": 0.08687355369329453, "learning_rate": 8.660762936550988e-05, "loss": 0.0038, "step": 27660 }, { "epoch": 244.86725663716814, "grad_norm": 0.10065907984972, "learning_rate": 8.659636490892753e-05, "loss": 0.0034, "step": 27670 }, { "epoch": 244.95575221238937, "grad_norm": 0.0972185805439949, "learning_rate": 8.658509645022174e-05, "loss": 0.0036, "step": 27680 }, { "epoch": 245.04424778761063, "grad_norm": 0.1123584434390068, "learning_rate": 8.657382399062481e-05, "loss": 0.0039, "step": 27690 }, { "epoch": 245.13274336283186, "grad_norm": 0.09343881160020828, "learning_rate": 8.656254753136946e-05, "loss": 0.0039, "step": 27700 }, { "epoch": 245.2212389380531, "grad_norm": 0.09851207584142685, "learning_rate": 8.655126707368891e-05, "loss": 0.0036, "step": 27710 }, { "epoch": 245.30973451327435, "grad_norm": 0.0816434770822525, "learning_rate": 8.653998261881672e-05, "loss": 0.0038, "step": 27720 }, { "epoch": 245.39823008849558, "grad_norm": 0.07943480461835861, "learning_rate": 8.652869416798699e-05, "loss": 0.0032, "step": 27730 }, { "epoch": 245.4867256637168, "grad_norm": 0.06022726371884346, "learning_rate": 8.651740172243417e-05, "loss": 0.0031, "step": 27740 }, { "epoch": 245.57522123893804, "grad_norm": 0.0813436433672905, "learning_rate": 8.65061052833932e-05, "loss": 0.0037, "step": 27750 }, { "epoch": 245.6637168141593, "grad_norm": 0.08012643456459045, "learning_rate": 8.649480485209945e-05, "loss": 0.0033, "step": 27760 }, { "epoch": 245.75221238938053, "grad_norm": 0.07250431925058365, "learning_rate": 8.64835004297887e-05, "loss": 0.0035, "step": 27770 }, { "epoch": 245.84070796460176, "grad_norm": 0.06616228073835373, "learning_rate": 8.64721920176972e-05, "loss": 0.0039, "step": 27780 }, { "epoch": 245.92920353982302, "grad_norm": 0.07948927581310272, "learning_rate": 8.646087961706164e-05, "loss": 0.0032, "step": 27790 }, { "epoch": 246.01769911504425, "grad_norm": 0.07382088154554367, "learning_rate": 8.644956322911908e-05, "loss": 0.0029, "step": 27800 }, { "epoch": 246.10619469026548, "grad_norm": 0.06662662327289581, "learning_rate": 8.643824285510709e-05, "loss": 0.0031, "step": 27810 }, { "epoch": 246.1946902654867, "grad_norm": 0.06371674686670303, "learning_rate": 8.642691849626364e-05, "loss": 0.0031, "step": 27820 }, { "epoch": 246.28318584070797, "grad_norm": 0.06748683750629425, "learning_rate": 8.641559015382717e-05, "loss": 0.0033, "step": 27830 }, { "epoch": 246.3716814159292, "grad_norm": 0.06527277082204819, "learning_rate": 8.640425782903649e-05, "loss": 0.0038, "step": 27840 }, { "epoch": 246.46017699115043, "grad_norm": 0.07685236632823944, "learning_rate": 8.639292152313091e-05, "loss": 0.0035, "step": 27850 }, { "epoch": 246.5486725663717, "grad_norm": 0.09269026666879654, "learning_rate": 8.638158123735015e-05, "loss": 0.0037, "step": 27860 }, { "epoch": 246.63716814159292, "grad_norm": 0.06635969877243042, "learning_rate": 8.637023697293436e-05, "loss": 0.003, "step": 27870 }, { "epoch": 246.72566371681415, "grad_norm": 0.0651502013206482, "learning_rate": 8.635888873112414e-05, "loss": 0.0035, "step": 27880 }, { "epoch": 246.81415929203538, "grad_norm": 0.08585961163043976, "learning_rate": 8.634753651316052e-05, "loss": 0.0031, "step": 27890 }, { "epoch": 246.90265486725664, "grad_norm": 0.07922837883234024, "learning_rate": 8.633618032028496e-05, "loss": 0.0031, "step": 27900 }, { "epoch": 246.99115044247787, "grad_norm": 0.060365960001945496, "learning_rate": 8.632482015373934e-05, "loss": 0.003, "step": 27910 }, { "epoch": 247.0796460176991, "grad_norm": 0.06580059975385666, "learning_rate": 8.6313456014766e-05, "loss": 0.0032, "step": 27920 }, { "epoch": 247.16814159292036, "grad_norm": 0.07637941092252731, "learning_rate": 8.630208790460771e-05, "loss": 0.0036, "step": 27930 }, { "epoch": 247.2566371681416, "grad_norm": 0.0660456046462059, "learning_rate": 8.629071582450768e-05, "loss": 0.0033, "step": 27940 }, { "epoch": 247.34513274336283, "grad_norm": 0.07134925574064255, "learning_rate": 8.62793397757095e-05, "loss": 0.0033, "step": 27950 }, { "epoch": 247.43362831858408, "grad_norm": 0.0717705711722374, "learning_rate": 8.626795975945729e-05, "loss": 0.0034, "step": 27960 }, { "epoch": 247.52212389380531, "grad_norm": 0.06218991056084633, "learning_rate": 8.625657577699551e-05, "loss": 0.0029, "step": 27970 }, { "epoch": 247.61061946902655, "grad_norm": 0.06579019874334335, "learning_rate": 8.624518782956914e-05, "loss": 0.003, "step": 27980 }, { "epoch": 247.69911504424778, "grad_norm": 0.06046956032514572, "learning_rate": 8.62337959184235e-05, "loss": 0.0033, "step": 27990 }, { "epoch": 247.78761061946904, "grad_norm": 0.06730349361896515, "learning_rate": 8.622240004480441e-05, "loss": 0.003, "step": 28000 }, { "epoch": 247.87610619469027, "grad_norm": 0.07440892606973648, "learning_rate": 8.621100020995814e-05, "loss": 0.0033, "step": 28010 }, { "epoch": 247.9646017699115, "grad_norm": 0.07617069780826569, "learning_rate": 8.619959641513132e-05, "loss": 0.0037, "step": 28020 }, { "epoch": 248.05309734513276, "grad_norm": 0.05940552428364754, "learning_rate": 8.618818866157105e-05, "loss": 0.0033, "step": 28030 }, { "epoch": 248.141592920354, "grad_norm": 0.07276551425457001, "learning_rate": 8.617677695052487e-05, "loss": 0.0037, "step": 28040 }, { "epoch": 248.23008849557522, "grad_norm": 0.0971188172698021, "learning_rate": 8.616536128324078e-05, "loss": 0.0036, "step": 28050 }, { "epoch": 248.31858407079645, "grad_norm": 0.08170472085475922, "learning_rate": 8.615394166096712e-05, "loss": 0.0033, "step": 28060 }, { "epoch": 248.4070796460177, "grad_norm": 0.08012739568948746, "learning_rate": 8.614251808495279e-05, "loss": 0.0036, "step": 28070 }, { "epoch": 248.49557522123894, "grad_norm": 0.08446774631738663, "learning_rate": 8.6131090556447e-05, "loss": 0.0029, "step": 28080 }, { "epoch": 248.58407079646017, "grad_norm": 0.10401690006256104, "learning_rate": 8.611965907669947e-05, "loss": 0.0035, "step": 28090 }, { "epoch": 248.67256637168143, "grad_norm": 0.07404624670743942, "learning_rate": 8.610822364696034e-05, "loss": 0.0036, "step": 28100 }, { "epoch": 248.76106194690266, "grad_norm": 0.07010047882795334, "learning_rate": 8.609678426848015e-05, "loss": 0.003, "step": 28110 }, { "epoch": 248.8495575221239, "grad_norm": 0.07056023180484772, "learning_rate": 8.60853409425099e-05, "loss": 0.0033, "step": 28120 }, { "epoch": 248.93805309734512, "grad_norm": 0.08714379370212555, "learning_rate": 8.607389367030104e-05, "loss": 0.0032, "step": 28130 }, { "epoch": 249.02654867256638, "grad_norm": 0.06614648550748825, "learning_rate": 8.606244245310538e-05, "loss": 0.0034, "step": 28140 }, { "epoch": 249.1150442477876, "grad_norm": 0.05907708778977394, "learning_rate": 8.605098729217525e-05, "loss": 0.0032, "step": 28150 }, { "epoch": 249.20353982300884, "grad_norm": 0.05132438242435455, "learning_rate": 8.603952818876335e-05, "loss": 0.004, "step": 28160 }, { "epoch": 249.2920353982301, "grad_norm": 0.064887635409832, "learning_rate": 8.602806514412281e-05, "loss": 0.0037, "step": 28170 }, { "epoch": 249.38053097345133, "grad_norm": 0.06561199575662613, "learning_rate": 8.601659815950726e-05, "loss": 0.0032, "step": 28180 }, { "epoch": 249.46902654867256, "grad_norm": 0.06760241091251373, "learning_rate": 8.600512723617067e-05, "loss": 0.0037, "step": 28190 }, { "epoch": 249.55752212389382, "grad_norm": 0.08649715036153793, "learning_rate": 8.59936523753675e-05, "loss": 0.0032, "step": 28200 }, { "epoch": 249.64601769911505, "grad_norm": 0.06608615070581436, "learning_rate": 8.598217357835264e-05, "loss": 0.0035, "step": 28210 }, { "epoch": 249.73451327433628, "grad_norm": 0.09520736336708069, "learning_rate": 8.597069084638135e-05, "loss": 0.0037, "step": 28220 }, { "epoch": 249.8230088495575, "grad_norm": 0.09637265652418137, "learning_rate": 8.595920418070939e-05, "loss": 0.0033, "step": 28230 }, { "epoch": 249.91150442477877, "grad_norm": 0.07793596386909485, "learning_rate": 8.594771358259295e-05, "loss": 0.0033, "step": 28240 }, { "epoch": 250.0, "grad_norm": 0.06690602004528046, "learning_rate": 8.593621905328858e-05, "loss": 0.0032, "step": 28250 }, { "epoch": 250.08849557522123, "grad_norm": 0.04947846382856369, "learning_rate": 8.592472059405333e-05, "loss": 0.0034, "step": 28260 }, { "epoch": 250.1769911504425, "grad_norm": 0.06876225769519806, "learning_rate": 8.591321820614464e-05, "loss": 0.003, "step": 28270 }, { "epoch": 250.26548672566372, "grad_norm": 0.0631076991558075, "learning_rate": 8.590171189082041e-05, "loss": 0.0032, "step": 28280 }, { "epoch": 250.35398230088495, "grad_norm": 0.06628824770450592, "learning_rate": 8.589020164933894e-05, "loss": 0.0036, "step": 28290 }, { "epoch": 250.44247787610618, "grad_norm": 0.0721016526222229, "learning_rate": 8.587868748295898e-05, "loss": 0.0036, "step": 28300 }, { "epoch": 250.53097345132744, "grad_norm": 0.07504332065582275, "learning_rate": 8.586716939293971e-05, "loss": 0.0034, "step": 28310 }, { "epoch": 250.61946902654867, "grad_norm": 0.0942879244685173, "learning_rate": 8.58556473805407e-05, "loss": 0.0034, "step": 28320 }, { "epoch": 250.7079646017699, "grad_norm": 0.061289042234420776, "learning_rate": 8.584412144702202e-05, "loss": 0.0032, "step": 28330 }, { "epoch": 250.79646017699116, "grad_norm": 0.06808647513389587, "learning_rate": 8.58325915936441e-05, "loss": 0.0038, "step": 28340 }, { "epoch": 250.8849557522124, "grad_norm": 0.060219574719667435, "learning_rate": 8.582105782166783e-05, "loss": 0.0036, "step": 28350 }, { "epoch": 250.97345132743362, "grad_norm": 0.08689513057470322, "learning_rate": 8.580952013235455e-05, "loss": 0.0034, "step": 28360 }, { "epoch": 251.06194690265488, "grad_norm": 0.05943585932254791, "learning_rate": 8.579797852696596e-05, "loss": 0.0036, "step": 28370 }, { "epoch": 251.1504424778761, "grad_norm": 0.05912910774350166, "learning_rate": 8.578643300676428e-05, "loss": 0.0033, "step": 28380 }, { "epoch": 251.23893805309734, "grad_norm": 0.07748554646968842, "learning_rate": 8.577488357301209e-05, "loss": 0.0033, "step": 28390 }, { "epoch": 251.32743362831857, "grad_norm": 0.060374245047569275, "learning_rate": 8.576333022697242e-05, "loss": 0.0036, "step": 28400 }, { "epoch": 251.41592920353983, "grad_norm": 0.0782068744301796, "learning_rate": 8.575177296990873e-05, "loss": 0.0034, "step": 28410 }, { "epoch": 251.50442477876106, "grad_norm": 0.06778387725353241, "learning_rate": 8.574021180308489e-05, "loss": 0.0036, "step": 28420 }, { "epoch": 251.5929203539823, "grad_norm": 0.08144747465848923, "learning_rate": 8.572864672776523e-05, "loss": 0.0034, "step": 28430 }, { "epoch": 251.68141592920355, "grad_norm": 0.05695458874106407, "learning_rate": 8.571707774521447e-05, "loss": 0.0029, "step": 28440 }, { "epoch": 251.76991150442478, "grad_norm": 0.07721395045518875, "learning_rate": 8.57055048566978e-05, "loss": 0.0033, "step": 28450 }, { "epoch": 251.858407079646, "grad_norm": 0.060557857155799866, "learning_rate": 8.569392806348078e-05, "loss": 0.0036, "step": 28460 }, { "epoch": 251.94690265486724, "grad_norm": 0.07758229225873947, "learning_rate": 8.568234736682947e-05, "loss": 0.0032, "step": 28470 }, { "epoch": 252.0353982300885, "grad_norm": 0.059478700160980225, "learning_rate": 8.567076276801029e-05, "loss": 0.0028, "step": 28480 }, { "epoch": 252.12389380530973, "grad_norm": 0.05978113412857056, "learning_rate": 8.565917426829013e-05, "loss": 0.0035, "step": 28490 }, { "epoch": 252.21238938053096, "grad_norm": 0.06888242810964584, "learning_rate": 8.564758186893628e-05, "loss": 0.0028, "step": 28500 }, { "epoch": 252.30088495575222, "grad_norm": 0.06498850882053375, "learning_rate": 8.563598557121649e-05, "loss": 0.0034, "step": 28510 }, { "epoch": 252.38938053097345, "grad_norm": 0.06644155830144882, "learning_rate": 8.562438537639888e-05, "loss": 0.0033, "step": 28520 }, { "epoch": 252.47787610619469, "grad_norm": 0.062438759952783585, "learning_rate": 8.561278128575206e-05, "loss": 0.0035, "step": 28530 }, { "epoch": 252.56637168141592, "grad_norm": 0.060420289635658264, "learning_rate": 8.5601173300545e-05, "loss": 0.0033, "step": 28540 }, { "epoch": 252.65486725663717, "grad_norm": 0.058756664395332336, "learning_rate": 8.558956142204717e-05, "loss": 0.0033, "step": 28550 }, { "epoch": 252.7433628318584, "grad_norm": 0.06439929455518723, "learning_rate": 8.55779456515284e-05, "loss": 0.003, "step": 28560 }, { "epoch": 252.83185840707964, "grad_norm": 0.07532541453838348, "learning_rate": 8.556632599025898e-05, "loss": 0.0035, "step": 28570 }, { "epoch": 252.9203539823009, "grad_norm": 0.09762299805879593, "learning_rate": 8.555470243950964e-05, "loss": 0.0037, "step": 28580 }, { "epoch": 253.00884955752213, "grad_norm": 0.058699801564216614, "learning_rate": 8.554307500055148e-05, "loss": 0.0034, "step": 28590 }, { "epoch": 253.09734513274336, "grad_norm": 0.0543123260140419, "learning_rate": 8.553144367465609e-05, "loss": 0.0028, "step": 28600 }, { "epoch": 253.18584070796462, "grad_norm": 0.06914706528186798, "learning_rate": 8.551980846309544e-05, "loss": 0.0029, "step": 28610 }, { "epoch": 253.27433628318585, "grad_norm": 0.07472000271081924, "learning_rate": 8.550816936714193e-05, "loss": 0.0033, "step": 28620 }, { "epoch": 253.36283185840708, "grad_norm": 0.08924397081136703, "learning_rate": 8.549652638806841e-05, "loss": 0.0034, "step": 28630 }, { "epoch": 253.4513274336283, "grad_norm": 0.09941164404153824, "learning_rate": 8.548487952714812e-05, "loss": 0.0032, "step": 28640 }, { "epoch": 253.53982300884957, "grad_norm": 0.0782296285033226, "learning_rate": 8.547322878565478e-05, "loss": 0.0036, "step": 28650 }, { "epoch": 253.6283185840708, "grad_norm": 0.10000191628932953, "learning_rate": 8.546157416486245e-05, "loss": 0.0035, "step": 28660 }, { "epoch": 253.71681415929203, "grad_norm": 0.07770363241434097, "learning_rate": 8.54499156660457e-05, "loss": 0.0035, "step": 28670 }, { "epoch": 253.8053097345133, "grad_norm": 0.06693620979785919, "learning_rate": 8.543825329047947e-05, "loss": 0.0032, "step": 28680 }, { "epoch": 253.89380530973452, "grad_norm": 0.0782831609249115, "learning_rate": 8.542658703943913e-05, "loss": 0.0028, "step": 28690 }, { "epoch": 253.98230088495575, "grad_norm": 0.054685793817043304, "learning_rate": 8.541491691420051e-05, "loss": 0.0032, "step": 28700 }, { "epoch": 254.07079646017698, "grad_norm": 0.06868727505207062, "learning_rate": 8.54032429160398e-05, "loss": 0.0034, "step": 28710 }, { "epoch": 254.15929203539824, "grad_norm": 0.05635147541761398, "learning_rate": 8.539156504623369e-05, "loss": 0.0029, "step": 28720 }, { "epoch": 254.24778761061947, "grad_norm": 0.06002165004611015, "learning_rate": 8.537988330605923e-05, "loss": 0.0032, "step": 28730 }, { "epoch": 254.3362831858407, "grad_norm": 0.06433845311403275, "learning_rate": 8.536819769679393e-05, "loss": 0.0034, "step": 28740 }, { "epoch": 254.42477876106196, "grad_norm": 0.07224614918231964, "learning_rate": 8.53565082197157e-05, "loss": 0.0031, "step": 28750 }, { "epoch": 254.5132743362832, "grad_norm": 0.06284494698047638, "learning_rate": 8.534481487610289e-05, "loss": 0.004, "step": 28760 }, { "epoch": 254.60176991150442, "grad_norm": 0.06810867041349411, "learning_rate": 8.533311766723428e-05, "loss": 0.0031, "step": 28770 }, { "epoch": 254.69026548672565, "grad_norm": 0.06391815096139908, "learning_rate": 8.532141659438901e-05, "loss": 0.0031, "step": 28780 }, { "epoch": 254.7787610619469, "grad_norm": 0.08181624859571457, "learning_rate": 8.530971165884675e-05, "loss": 0.0031, "step": 28790 }, { "epoch": 254.86725663716814, "grad_norm": 0.0668865218758583, "learning_rate": 8.529800286188752e-05, "loss": 0.0036, "step": 28800 }, { "epoch": 254.95575221238937, "grad_norm": 0.08124502003192902, "learning_rate": 8.528629020479175e-05, "loss": 0.0032, "step": 28810 }, { "epoch": 255.04424778761063, "grad_norm": 0.07732603698968887, "learning_rate": 8.527457368884033e-05, "loss": 0.0034, "step": 28820 }, { "epoch": 255.13274336283186, "grad_norm": 0.0720565915107727, "learning_rate": 8.526285331531458e-05, "loss": 0.0035, "step": 28830 }, { "epoch": 255.2212389380531, "grad_norm": 0.0701301321387291, "learning_rate": 8.525112908549621e-05, "loss": 0.0034, "step": 28840 }, { "epoch": 255.30973451327435, "grad_norm": 0.08057865500450134, "learning_rate": 8.523940100066735e-05, "loss": 0.0032, "step": 28850 }, { "epoch": 255.39823008849558, "grad_norm": 0.06796091049909592, "learning_rate": 8.52276690621106e-05, "loss": 0.0031, "step": 28860 }, { "epoch": 255.4867256637168, "grad_norm": 0.08057154715061188, "learning_rate": 8.521593327110889e-05, "loss": 0.0036, "step": 28870 }, { "epoch": 255.57522123893804, "grad_norm": 0.05897509306669235, "learning_rate": 8.520419362894569e-05, "loss": 0.0026, "step": 28880 }, { "epoch": 255.6637168141593, "grad_norm": 0.08053771406412125, "learning_rate": 8.51924501369048e-05, "loss": 0.0036, "step": 28890 }, { "epoch": 255.75221238938053, "grad_norm": 0.06868140399456024, "learning_rate": 8.518070279627047e-05, "loss": 0.003, "step": 28900 }, { "epoch": 255.84070796460176, "grad_norm": 0.05794830620288849, "learning_rate": 8.516895160832737e-05, "loss": 0.0033, "step": 28910 }, { "epoch": 255.92920353982302, "grad_norm": 0.05384915694594383, "learning_rate": 8.515719657436061e-05, "loss": 0.0031, "step": 28920 }, { "epoch": 256.01769911504425, "grad_norm": 0.06673529744148254, "learning_rate": 8.514543769565568e-05, "loss": 0.0029, "step": 28930 }, { "epoch": 256.1061946902655, "grad_norm": 0.058856233954429626, "learning_rate": 8.513367497349853e-05, "loss": 0.0035, "step": 28940 }, { "epoch": 256.1946902654867, "grad_norm": 0.05967045947909355, "learning_rate": 8.51219084091755e-05, "loss": 0.0036, "step": 28950 }, { "epoch": 256.283185840708, "grad_norm": 0.057497162371873856, "learning_rate": 8.511013800397338e-05, "loss": 0.0031, "step": 28960 }, { "epoch": 256.37168141592923, "grad_norm": 0.06020091101527214, "learning_rate": 8.509836375917937e-05, "loss": 0.0031, "step": 28970 }, { "epoch": 256.46017699115043, "grad_norm": 0.0761679857969284, "learning_rate": 8.508658567608104e-05, "loss": 0.003, "step": 28980 }, { "epoch": 256.5486725663717, "grad_norm": 0.06888946890830994, "learning_rate": 8.507480375596647e-05, "loss": 0.0033, "step": 28990 }, { "epoch": 256.6371681415929, "grad_norm": 0.07728346437215805, "learning_rate": 8.506301800012408e-05, "loss": 0.003, "step": 29000 }, { "epoch": 256.72566371681415, "grad_norm": 0.06941675394773483, "learning_rate": 8.505122840984278e-05, "loss": 0.003, "step": 29010 }, { "epoch": 256.8141592920354, "grad_norm": 0.0655403658747673, "learning_rate": 8.503943498641182e-05, "loss": 0.0036, "step": 29020 }, { "epoch": 256.9026548672566, "grad_norm": 0.07446739822626114, "learning_rate": 8.502763773112095e-05, "loss": 0.0035, "step": 29030 }, { "epoch": 256.9911504424779, "grad_norm": 0.08550873398780823, "learning_rate": 8.501583664526026e-05, "loss": 0.0029, "step": 29040 }, { "epoch": 257.07964601769913, "grad_norm": 0.07925383001565933, "learning_rate": 8.500403173012032e-05, "loss": 0.0033, "step": 29050 }, { "epoch": 257.16814159292034, "grad_norm": 0.06312176585197449, "learning_rate": 8.499222298699211e-05, "loss": 0.0032, "step": 29060 }, { "epoch": 257.2566371681416, "grad_norm": 0.06199989095330238, "learning_rate": 8.498041041716701e-05, "loss": 0.0031, "step": 29070 }, { "epoch": 257.34513274336285, "grad_norm": 0.07045721262693405, "learning_rate": 8.496859402193681e-05, "loss": 0.0029, "step": 29080 }, { "epoch": 257.43362831858406, "grad_norm": 0.05736267566680908, "learning_rate": 8.495677380259374e-05, "loss": 0.0026, "step": 29090 }, { "epoch": 257.5221238938053, "grad_norm": 0.05556948855519295, "learning_rate": 8.494494976043045e-05, "loss": 0.0031, "step": 29100 }, { "epoch": 257.6106194690266, "grad_norm": 0.060114696621894836, "learning_rate": 8.493312189673998e-05, "loss": 0.0034, "step": 29110 }, { "epoch": 257.6991150442478, "grad_norm": 0.06774993240833282, "learning_rate": 8.492129021281584e-05, "loss": 0.0032, "step": 29120 }, { "epoch": 257.78761061946904, "grad_norm": 0.0722884088754654, "learning_rate": 8.490945470995188e-05, "loss": 0.0033, "step": 29130 }, { "epoch": 257.87610619469024, "grad_norm": 0.08727022260427475, "learning_rate": 8.489761538944247e-05, "loss": 0.0031, "step": 29140 }, { "epoch": 257.9646017699115, "grad_norm": 0.08892770111560822, "learning_rate": 8.48857722525823e-05, "loss": 0.0031, "step": 29150 }, { "epoch": 258.05309734513276, "grad_norm": 0.07251773774623871, "learning_rate": 8.487392530066652e-05, "loss": 0.0034, "step": 29160 }, { "epoch": 258.14159292035396, "grad_norm": 0.07943584769964218, "learning_rate": 8.486207453499069e-05, "loss": 0.0033, "step": 29170 }, { "epoch": 258.2300884955752, "grad_norm": 0.08051355183124542, "learning_rate": 8.485021995685082e-05, "loss": 0.0035, "step": 29180 }, { "epoch": 258.3185840707965, "grad_norm": 0.09371034055948257, "learning_rate": 8.483836156754328e-05, "loss": 0.0034, "step": 29190 }, { "epoch": 258.4070796460177, "grad_norm": 0.07337057590484619, "learning_rate": 8.482649936836491e-05, "loss": 0.0038, "step": 29200 }, { "epoch": 258.49557522123894, "grad_norm": 0.057822827249765396, "learning_rate": 8.481463336061293e-05, "loss": 0.0035, "step": 29210 }, { "epoch": 258.5840707964602, "grad_norm": 0.07943731546401978, "learning_rate": 8.480276354558496e-05, "loss": 0.0036, "step": 29220 }, { "epoch": 258.6725663716814, "grad_norm": 0.07211572676897049, "learning_rate": 8.479088992457913e-05, "loss": 0.0036, "step": 29230 }, { "epoch": 258.76106194690266, "grad_norm": 0.08156314492225647, "learning_rate": 8.477901249889387e-05, "loss": 0.0029, "step": 29240 }, { "epoch": 258.8495575221239, "grad_norm": 0.08635958284139633, "learning_rate": 8.47671312698281e-05, "loss": 0.0036, "step": 29250 }, { "epoch": 258.9380530973451, "grad_norm": 0.06871464103460312, "learning_rate": 8.475524623868112e-05, "loss": 0.0038, "step": 29260 }, { "epoch": 259.0265486725664, "grad_norm": 0.10013473778963089, "learning_rate": 8.474335740675266e-05, "loss": 0.0033, "step": 29270 }, { "epoch": 259.11504424778764, "grad_norm": 0.06679364293813705, "learning_rate": 8.473146477534289e-05, "loss": 0.0027, "step": 29280 }, { "epoch": 259.20353982300884, "grad_norm": 0.05437986180186272, "learning_rate": 8.471956834575232e-05, "loss": 0.0036, "step": 29290 }, { "epoch": 259.2920353982301, "grad_norm": 0.07190488278865814, "learning_rate": 8.470766811928197e-05, "loss": 0.0029, "step": 29300 }, { "epoch": 259.3805309734513, "grad_norm": 0.06288619339466095, "learning_rate": 8.469576409723323e-05, "loss": 0.003, "step": 29310 }, { "epoch": 259.46902654867256, "grad_norm": 0.07875878363847733, "learning_rate": 8.468385628090788e-05, "loss": 0.0035, "step": 29320 }, { "epoch": 259.5575221238938, "grad_norm": 0.08202241361141205, "learning_rate": 8.467194467160815e-05, "loss": 0.0034, "step": 29330 }, { "epoch": 259.646017699115, "grad_norm": 0.10377220809459686, "learning_rate": 8.466002927063667e-05, "loss": 0.0031, "step": 29340 }, { "epoch": 259.7345132743363, "grad_norm": 0.07662896811962128, "learning_rate": 8.464811007929651e-05, "loss": 0.0035, "step": 29350 }, { "epoch": 259.82300884955754, "grad_norm": 0.07223580032587051, "learning_rate": 8.463618709889114e-05, "loss": 0.0031, "step": 29360 }, { "epoch": 259.91150442477874, "grad_norm": 0.059547651559114456, "learning_rate": 8.462426033072442e-05, "loss": 0.0031, "step": 29370 }, { "epoch": 260.0, "grad_norm": 0.09325040131807327, "learning_rate": 8.461232977610061e-05, "loss": 0.003, "step": 29380 }, { "epoch": 260.08849557522126, "grad_norm": 0.07164835184812546, "learning_rate": 8.46003954363245e-05, "loss": 0.003, "step": 29390 }, { "epoch": 260.17699115044246, "grad_norm": 0.078645259141922, "learning_rate": 8.458845731270115e-05, "loss": 0.0039, "step": 29400 }, { "epoch": 260.2654867256637, "grad_norm": 0.07282843440771103, "learning_rate": 8.45765154065361e-05, "loss": 0.0034, "step": 29410 }, { "epoch": 260.353982300885, "grad_norm": 0.09319477528333664, "learning_rate": 8.456456971913532e-05, "loss": 0.0037, "step": 29420 }, { "epoch": 260.4424778761062, "grad_norm": 0.10506348311901093, "learning_rate": 8.455262025180517e-05, "loss": 0.0038, "step": 29430 }, { "epoch": 260.53097345132744, "grad_norm": 0.0758274719119072, "learning_rate": 8.454066700585242e-05, "loss": 0.0036, "step": 29440 }, { "epoch": 260.6194690265487, "grad_norm": 0.057735513895750046, "learning_rate": 8.452870998258423e-05, "loss": 0.0034, "step": 29450 }, { "epoch": 260.7079646017699, "grad_norm": 0.05758112296462059, "learning_rate": 8.451674918330825e-05, "loss": 0.0029, "step": 29460 }, { "epoch": 260.79646017699116, "grad_norm": 0.06381456553936005, "learning_rate": 8.450478460933246e-05, "loss": 0.0034, "step": 29470 }, { "epoch": 260.88495575221236, "grad_norm": 0.05322045087814331, "learning_rate": 8.449281626196532e-05, "loss": 0.0028, "step": 29480 }, { "epoch": 260.9734513274336, "grad_norm": 0.059817325323820114, "learning_rate": 8.448084414251564e-05, "loss": 0.0033, "step": 29490 }, { "epoch": 261.0619469026549, "grad_norm": 0.06666016578674316, "learning_rate": 8.446886825229271e-05, "loss": 0.0034, "step": 29500 }, { "epoch": 261.1504424778761, "grad_norm": 0.06523222476243973, "learning_rate": 8.445688859260615e-05, "loss": 0.003, "step": 29510 }, { "epoch": 261.23893805309734, "grad_norm": 0.06009840965270996, "learning_rate": 8.444490516476606e-05, "loss": 0.0029, "step": 29520 }, { "epoch": 261.3274336283186, "grad_norm": 0.0665576308965683, "learning_rate": 8.443291797008293e-05, "loss": 0.003, "step": 29530 }, { "epoch": 261.4159292035398, "grad_norm": 0.07083910703659058, "learning_rate": 8.442092700986765e-05, "loss": 0.0029, "step": 29540 }, { "epoch": 261.50442477876106, "grad_norm": 0.06016381457448006, "learning_rate": 8.440893228543156e-05, "loss": 0.004, "step": 29550 }, { "epoch": 261.5929203539823, "grad_norm": 0.08724050968885422, "learning_rate": 8.439693379808638e-05, "loss": 0.003, "step": 29560 }, { "epoch": 261.6814159292035, "grad_norm": 0.0781373605132103, "learning_rate": 8.43849315491442e-05, "loss": 0.003, "step": 29570 }, { "epoch": 261.7699115044248, "grad_norm": 0.07976669818162918, "learning_rate": 8.437292553991763e-05, "loss": 0.0034, "step": 29580 }, { "epoch": 261.85840707964604, "grad_norm": 0.07916860282421112, "learning_rate": 8.436091577171959e-05, "loss": 0.0032, "step": 29590 }, { "epoch": 261.94690265486724, "grad_norm": 0.059990692883729935, "learning_rate": 8.434890224586347e-05, "loss": 0.003, "step": 29600 }, { "epoch": 262.0353982300885, "grad_norm": 0.0840144157409668, "learning_rate": 8.433688496366303e-05, "loss": 0.0032, "step": 29610 }, { "epoch": 262.12389380530976, "grad_norm": 0.07952767610549927, "learning_rate": 8.432486392643248e-05, "loss": 0.0034, "step": 29620 }, { "epoch": 262.21238938053096, "grad_norm": 0.07261509448289871, "learning_rate": 8.431283913548643e-05, "loss": 0.0041, "step": 29630 }, { "epoch": 262.3008849557522, "grad_norm": 0.06740890443325043, "learning_rate": 8.430081059213985e-05, "loss": 0.0034, "step": 29640 }, { "epoch": 262.3893805309734, "grad_norm": 0.08193156868219376, "learning_rate": 8.428877829770823e-05, "loss": 0.0029, "step": 29650 }, { "epoch": 262.4778761061947, "grad_norm": 0.07831292599439621, "learning_rate": 8.427674225350735e-05, "loss": 0.0027, "step": 29660 }, { "epoch": 262.56637168141594, "grad_norm": 0.04897677153348923, "learning_rate": 8.426470246085347e-05, "loss": 0.003, "step": 29670 }, { "epoch": 262.65486725663715, "grad_norm": 0.05944943428039551, "learning_rate": 8.425265892106324e-05, "loss": 0.0035, "step": 29680 }, { "epoch": 262.7433628318584, "grad_norm": 0.08495219796895981, "learning_rate": 8.424061163545374e-05, "loss": 0.0034, "step": 29690 }, { "epoch": 262.83185840707966, "grad_norm": 0.054963093250989914, "learning_rate": 8.422856060534243e-05, "loss": 0.0029, "step": 29700 }, { "epoch": 262.92035398230087, "grad_norm": 0.06401991844177246, "learning_rate": 8.421650583204718e-05, "loss": 0.0033, "step": 29710 }, { "epoch": 263.0088495575221, "grad_norm": 0.07772617787122726, "learning_rate": 8.420444731688633e-05, "loss": 0.003, "step": 29720 }, { "epoch": 263.0973451327434, "grad_norm": 0.07619351148605347, "learning_rate": 8.419238506117852e-05, "loss": 0.0032, "step": 29730 }, { "epoch": 263.1858407079646, "grad_norm": 0.060291655361652374, "learning_rate": 8.418031906624289e-05, "loss": 0.0032, "step": 29740 }, { "epoch": 263.27433628318585, "grad_norm": 0.06297405064105988, "learning_rate": 8.416824933339898e-05, "loss": 0.0033, "step": 29750 }, { "epoch": 263.3628318584071, "grad_norm": 0.07559987902641296, "learning_rate": 8.415617586396667e-05, "loss": 0.0033, "step": 29760 }, { "epoch": 263.4513274336283, "grad_norm": 0.061636850237846375, "learning_rate": 8.414409865926632e-05, "loss": 0.0027, "step": 29770 }, { "epoch": 263.53982300884957, "grad_norm": 0.06979813426733017, "learning_rate": 8.413201772061867e-05, "loss": 0.0035, "step": 29780 }, { "epoch": 263.62831858407077, "grad_norm": 0.07966155558824539, "learning_rate": 8.411993304934488e-05, "loss": 0.0038, "step": 29790 }, { "epoch": 263.716814159292, "grad_norm": 0.08211088925600052, "learning_rate": 8.410784464676654e-05, "loss": 0.0029, "step": 29800 }, { "epoch": 263.8053097345133, "grad_norm": 0.07818065583705902, "learning_rate": 8.409575251420556e-05, "loss": 0.0029, "step": 29810 }, { "epoch": 263.8938053097345, "grad_norm": 0.09522417187690735, "learning_rate": 8.408365665298435e-05, "loss": 0.0034, "step": 29820 }, { "epoch": 263.98230088495575, "grad_norm": 0.08845879137516022, "learning_rate": 8.40715570644257e-05, "loss": 0.0034, "step": 29830 }, { "epoch": 264.070796460177, "grad_norm": 0.07381059974431992, "learning_rate": 8.40594537498528e-05, "loss": 0.0031, "step": 29840 }, { "epoch": 264.1592920353982, "grad_norm": 0.07607392221689224, "learning_rate": 8.404734671058924e-05, "loss": 0.0036, "step": 29850 }, { "epoch": 264.24778761061947, "grad_norm": 0.07502534985542297, "learning_rate": 8.403523594795902e-05, "loss": 0.0031, "step": 29860 }, { "epoch": 264.3362831858407, "grad_norm": 0.09197360277175903, "learning_rate": 8.402312146328659e-05, "loss": 0.0033, "step": 29870 }, { "epoch": 264.42477876106193, "grad_norm": 0.0915946438908577, "learning_rate": 8.401100325789675e-05, "loss": 0.0033, "step": 29880 }, { "epoch": 264.5132743362832, "grad_norm": 0.05980079993605614, "learning_rate": 8.399888133311472e-05, "loss": 0.0032, "step": 29890 }, { "epoch": 264.60176991150445, "grad_norm": 0.07528424263000488, "learning_rate": 8.398675569026613e-05, "loss": 0.0029, "step": 29900 }, { "epoch": 264.69026548672565, "grad_norm": 0.06225307658314705, "learning_rate": 8.397462633067705e-05, "loss": 0.0034, "step": 29910 }, { "epoch": 264.7787610619469, "grad_norm": 0.06493095308542252, "learning_rate": 8.396249325567392e-05, "loss": 0.0031, "step": 29920 }, { "epoch": 264.86725663716817, "grad_norm": 0.06745022535324097, "learning_rate": 8.395035646658357e-05, "loss": 0.0032, "step": 29930 }, { "epoch": 264.95575221238937, "grad_norm": 0.06481379270553589, "learning_rate": 8.39382159647333e-05, "loss": 0.0032, "step": 29940 }, { "epoch": 265.04424778761063, "grad_norm": 0.06339843571186066, "learning_rate": 8.392607175145075e-05, "loss": 0.003, "step": 29950 }, { "epoch": 265.13274336283183, "grad_norm": 0.06295178085565567, "learning_rate": 8.3913923828064e-05, "loss": 0.0031, "step": 29960 }, { "epoch": 265.2212389380531, "grad_norm": 0.0703488364815712, "learning_rate": 8.390177219590152e-05, "loss": 0.0031, "step": 29970 }, { "epoch": 265.30973451327435, "grad_norm": 0.05964634194970131, "learning_rate": 8.388961685629222e-05, "loss": 0.0027, "step": 29980 }, { "epoch": 265.39823008849555, "grad_norm": 0.05641529709100723, "learning_rate": 8.387745781056536e-05, "loss": 0.0032, "step": 29990 }, { "epoch": 265.4867256637168, "grad_norm": 0.07229544967412949, "learning_rate": 8.386529506005065e-05, "loss": 0.0032, "step": 30000 }, { "epoch": 265.57522123893807, "grad_norm": 0.0780099630355835, "learning_rate": 8.38531286060782e-05, "loss": 0.0033, "step": 30010 }, { "epoch": 265.6637168141593, "grad_norm": 0.08707484602928162, "learning_rate": 8.384095844997849e-05, "loss": 0.0034, "step": 30020 }, { "epoch": 265.75221238938053, "grad_norm": 0.0598570741713047, "learning_rate": 8.382878459308245e-05, "loss": 0.0035, "step": 30030 }, { "epoch": 265.8407079646018, "grad_norm": 0.06374727934598923, "learning_rate": 8.381660703672138e-05, "loss": 0.0029, "step": 30040 }, { "epoch": 265.929203539823, "grad_norm": 0.07101389020681381, "learning_rate": 8.380442578222702e-05, "loss": 0.0031, "step": 30050 }, { "epoch": 266.01769911504425, "grad_norm": 0.05621369183063507, "learning_rate": 8.379224083093146e-05, "loss": 0.003, "step": 30060 }, { "epoch": 266.1061946902655, "grad_norm": 0.0776141881942749, "learning_rate": 8.378005218416727e-05, "loss": 0.0031, "step": 30070 }, { "epoch": 266.1946902654867, "grad_norm": 0.0523408018052578, "learning_rate": 8.376785984326735e-05, "loss": 0.0032, "step": 30080 }, { "epoch": 266.283185840708, "grad_norm": 0.07195472717285156, "learning_rate": 8.375566380956506e-05, "loss": 0.0029, "step": 30090 }, { "epoch": 266.37168141592923, "grad_norm": 0.07572682946920395, "learning_rate": 8.374346408439411e-05, "loss": 0.0031, "step": 30100 }, { "epoch": 266.46017699115043, "grad_norm": 0.06902444362640381, "learning_rate": 8.373126066908868e-05, "loss": 0.0032, "step": 30110 }, { "epoch": 266.5486725663717, "grad_norm": 0.08613183349370956, "learning_rate": 8.371905356498326e-05, "loss": 0.0028, "step": 30120 }, { "epoch": 266.6371681415929, "grad_norm": 0.054517943412065506, "learning_rate": 8.370684277341288e-05, "loss": 0.0029, "step": 30130 }, { "epoch": 266.72566371681415, "grad_norm": 0.07605861127376556, "learning_rate": 8.369462829571282e-05, "loss": 0.003, "step": 30140 }, { "epoch": 266.8141592920354, "grad_norm": 0.09095414727926254, "learning_rate": 8.36824101332189e-05, "loss": 0.0032, "step": 30150 }, { "epoch": 266.9026548672566, "grad_norm": 0.06632758677005768, "learning_rate": 8.367018828726721e-05, "loss": 0.0029, "step": 30160 }, { "epoch": 266.9911504424779, "grad_norm": 0.0696689784526825, "learning_rate": 8.365796275919438e-05, "loss": 0.0033, "step": 30170 }, { "epoch": 267.07964601769913, "grad_norm": 0.07149624079465866, "learning_rate": 8.364573355033734e-05, "loss": 0.0031, "step": 30180 }, { "epoch": 267.16814159292034, "grad_norm": 0.06670485436916351, "learning_rate": 8.363350066203346e-05, "loss": 0.0032, "step": 30190 }, { "epoch": 267.2566371681416, "grad_norm": 0.07828574627637863, "learning_rate": 8.362126409562053e-05, "loss": 0.003, "step": 30200 }, { "epoch": 267.34513274336285, "grad_norm": 0.07752962410449982, "learning_rate": 8.360902385243667e-05, "loss": 0.0032, "step": 30210 }, { "epoch": 267.43362831858406, "grad_norm": 0.06218593195080757, "learning_rate": 8.359677993382052e-05, "loss": 0.0029, "step": 30220 }, { "epoch": 267.5221238938053, "grad_norm": 0.05761337652802467, "learning_rate": 8.358453234111103e-05, "loss": 0.0028, "step": 30230 }, { "epoch": 267.6106194690266, "grad_norm": 0.07663553953170776, "learning_rate": 8.357228107564756e-05, "loss": 0.003, "step": 30240 }, { "epoch": 267.6991150442478, "grad_norm": 0.07718848437070847, "learning_rate": 8.356002613876993e-05, "loss": 0.0034, "step": 30250 }, { "epoch": 267.78761061946904, "grad_norm": 0.07494930922985077, "learning_rate": 8.35477675318183e-05, "loss": 0.003, "step": 30260 }, { "epoch": 267.87610619469024, "grad_norm": 0.05851662531495094, "learning_rate": 8.353550525613323e-05, "loss": 0.0032, "step": 30270 }, { "epoch": 267.9646017699115, "grad_norm": 0.06654157489538193, "learning_rate": 8.352323931305572e-05, "loss": 0.0034, "step": 30280 }, { "epoch": 268.05309734513276, "grad_norm": 0.06057656556367874, "learning_rate": 8.351096970392717e-05, "loss": 0.0033, "step": 30290 }, { "epoch": 268.14159292035396, "grad_norm": 0.059937115758657455, "learning_rate": 8.349869643008937e-05, "loss": 0.003, "step": 30300 }, { "epoch": 268.2300884955752, "grad_norm": 0.06216247379779816, "learning_rate": 8.348641949288449e-05, "loss": 0.003, "step": 30310 }, { "epoch": 268.3185840707965, "grad_norm": 0.060827940702438354, "learning_rate": 8.34741388936551e-05, "loss": 0.0029, "step": 30320 }, { "epoch": 268.4070796460177, "grad_norm": 0.058094192296266556, "learning_rate": 8.346185463374423e-05, "loss": 0.0035, "step": 30330 }, { "epoch": 268.49557522123894, "grad_norm": 0.0735178142786026, "learning_rate": 8.344956671449524e-05, "loss": 0.0031, "step": 30340 }, { "epoch": 268.5840707964602, "grad_norm": 0.07003750652074814, "learning_rate": 8.343727513725192e-05, "loss": 0.0033, "step": 30350 }, { "epoch": 268.6725663716814, "grad_norm": 0.08763961493968964, "learning_rate": 8.342497990335847e-05, "loss": 0.0033, "step": 30360 }, { "epoch": 268.76106194690266, "grad_norm": 0.07420606911182404, "learning_rate": 8.341268101415946e-05, "loss": 0.003, "step": 30370 }, { "epoch": 268.8495575221239, "grad_norm": 0.046375174075365067, "learning_rate": 8.34003784709999e-05, "loss": 0.0032, "step": 30380 }, { "epoch": 268.9380530973451, "grad_norm": 0.07440797984600067, "learning_rate": 8.338807227522516e-05, "loss": 0.0029, "step": 30390 }, { "epoch": 269.0265486725664, "grad_norm": 0.07360599935054779, "learning_rate": 8.337576242818103e-05, "loss": 0.0031, "step": 30400 }, { "epoch": 269.11504424778764, "grad_norm": 0.057932909578084946, "learning_rate": 8.336344893121372e-05, "loss": 0.003, "step": 30410 }, { "epoch": 269.20353982300884, "grad_norm": 0.07335598021745682, "learning_rate": 8.335113178566977e-05, "loss": 0.0037, "step": 30420 }, { "epoch": 269.2920353982301, "grad_norm": 0.06752992421388626, "learning_rate": 8.33388109928962e-05, "loss": 0.0029, "step": 30430 }, { "epoch": 269.3805309734513, "grad_norm": 0.06935624778270721, "learning_rate": 8.33264865542404e-05, "loss": 0.0031, "step": 30440 }, { "epoch": 269.46902654867256, "grad_norm": 0.06100798398256302, "learning_rate": 8.331415847105013e-05, "loss": 0.0033, "step": 30450 }, { "epoch": 269.5575221238938, "grad_norm": 0.0690862163901329, "learning_rate": 8.330182674467357e-05, "loss": 0.0036, "step": 30460 }, { "epoch": 269.646017699115, "grad_norm": 0.07491207867860794, "learning_rate": 8.32894913764593e-05, "loss": 0.0034, "step": 30470 }, { "epoch": 269.7345132743363, "grad_norm": 0.06744735687971115, "learning_rate": 8.327715236775633e-05, "loss": 0.0036, "step": 30480 }, { "epoch": 269.82300884955754, "grad_norm": 0.06967149674892426, "learning_rate": 8.326480971991398e-05, "loss": 0.0029, "step": 30490 }, { "epoch": 269.91150442477874, "grad_norm": 0.05289319530129433, "learning_rate": 8.325246343428206e-05, "loss": 0.0027, "step": 30500 }, { "epoch": 270.0, "grad_norm": 0.07259471714496613, "learning_rate": 8.324011351221072e-05, "loss": 0.003, "step": 30510 }, { "epoch": 270.08849557522126, "grad_norm": 0.06832297891378403, "learning_rate": 8.322775995505057e-05, "loss": 0.0031, "step": 30520 }, { "epoch": 270.17699115044246, "grad_norm": 0.06663279235363007, "learning_rate": 8.321540276415254e-05, "loss": 0.0034, "step": 30530 }, { "epoch": 270.2654867256637, "grad_norm": 0.09455490112304688, "learning_rate": 8.320304194086798e-05, "loss": 0.0037, "step": 30540 }, { "epoch": 270.353982300885, "grad_norm": 0.06952641159296036, "learning_rate": 8.31906774865487e-05, "loss": 0.0033, "step": 30550 }, { "epoch": 270.4424778761062, "grad_norm": 0.07877687364816666, "learning_rate": 8.317830940254682e-05, "loss": 0.0033, "step": 30560 }, { "epoch": 270.53097345132744, "grad_norm": 0.08070841431617737, "learning_rate": 8.316593769021491e-05, "loss": 0.0029, "step": 30570 }, { "epoch": 270.6194690265487, "grad_norm": 0.08183667808771133, "learning_rate": 8.315356235090592e-05, "loss": 0.0033, "step": 30580 }, { "epoch": 270.7079646017699, "grad_norm": 0.06732363253831863, "learning_rate": 8.314118338597319e-05, "loss": 0.0032, "step": 30590 }, { "epoch": 270.79646017699116, "grad_norm": 0.07502687722444534, "learning_rate": 8.312880079677048e-05, "loss": 0.0033, "step": 30600 }, { "epoch": 270.88495575221236, "grad_norm": 0.06330551952123642, "learning_rate": 8.311641458465191e-05, "loss": 0.003, "step": 30610 }, { "epoch": 270.9734513274336, "grad_norm": 0.07829611003398895, "learning_rate": 8.310402475097205e-05, "loss": 0.0034, "step": 30620 }, { "epoch": 271.0619469026549, "grad_norm": 0.06705602258443832, "learning_rate": 8.309163129708581e-05, "loss": 0.0026, "step": 30630 }, { "epoch": 271.1504424778761, "grad_norm": 0.08152607828378677, "learning_rate": 8.307923422434852e-05, "loss": 0.0032, "step": 30640 }, { "epoch": 271.23893805309734, "grad_norm": 0.09381666034460068, "learning_rate": 8.30668335341159e-05, "loss": 0.0029, "step": 30650 }, { "epoch": 271.3274336283186, "grad_norm": 0.0892510935664177, "learning_rate": 8.305442922774408e-05, "loss": 0.0034, "step": 30660 }, { "epoch": 271.4159292035398, "grad_norm": 0.09322934597730637, "learning_rate": 8.304202130658959e-05, "loss": 0.0031, "step": 30670 }, { "epoch": 271.50442477876106, "grad_norm": 0.08108902722597122, "learning_rate": 8.302960977200931e-05, "loss": 0.0031, "step": 30680 }, { "epoch": 271.5929203539823, "grad_norm": 0.07257851958274841, "learning_rate": 8.301719462536058e-05, "loss": 0.0032, "step": 30690 }, { "epoch": 271.6814159292035, "grad_norm": 0.08072934299707413, "learning_rate": 8.300477586800108e-05, "loss": 0.0031, "step": 30700 }, { "epoch": 271.7699115044248, "grad_norm": 0.0765933245420456, "learning_rate": 8.299235350128892e-05, "loss": 0.0032, "step": 30710 }, { "epoch": 271.85840707964604, "grad_norm": 0.07900026440620422, "learning_rate": 8.297992752658258e-05, "loss": 0.0031, "step": 30720 }, { "epoch": 271.94690265486724, "grad_norm": 0.07524921745061874, "learning_rate": 8.296749794524094e-05, "loss": 0.0032, "step": 30730 }, { "epoch": 272.0353982300885, "grad_norm": 0.07634364813566208, "learning_rate": 8.295506475862332e-05, "loss": 0.0035, "step": 30740 }, { "epoch": 272.12389380530976, "grad_norm": 0.07774457335472107, "learning_rate": 8.294262796808933e-05, "loss": 0.0032, "step": 30750 }, { "epoch": 272.21238938053096, "grad_norm": 0.071512870490551, "learning_rate": 8.293018757499909e-05, "loss": 0.003, "step": 30760 }, { "epoch": 272.3008849557522, "grad_norm": 0.06830643862485886, "learning_rate": 8.291774358071305e-05, "loss": 0.0032, "step": 30770 }, { "epoch": 272.3893805309734, "grad_norm": 0.08222298324108124, "learning_rate": 8.290529598659205e-05, "loss": 0.0029, "step": 30780 }, { "epoch": 272.4778761061947, "grad_norm": 0.06324493885040283, "learning_rate": 8.289284479399738e-05, "loss": 0.0031, "step": 30790 }, { "epoch": 272.56637168141594, "grad_norm": 0.07166349142789841, "learning_rate": 8.288039000429064e-05, "loss": 0.0028, "step": 30800 }, { "epoch": 272.65486725663715, "grad_norm": 0.07703594863414764, "learning_rate": 8.286793161883388e-05, "loss": 0.0032, "step": 30810 }, { "epoch": 272.7433628318584, "grad_norm": 0.07149528712034225, "learning_rate": 8.285546963898954e-05, "loss": 0.0032, "step": 30820 }, { "epoch": 272.83185840707966, "grad_norm": 0.0826297476887703, "learning_rate": 8.284300406612044e-05, "loss": 0.0031, "step": 30830 }, { "epoch": 272.92035398230087, "grad_norm": 0.06846296042203903, "learning_rate": 8.283053490158978e-05, "loss": 0.0029, "step": 30840 }, { "epoch": 273.0088495575221, "grad_norm": 0.06689579784870148, "learning_rate": 8.28180621467612e-05, "loss": 0.0028, "step": 30850 }, { "epoch": 273.0973451327434, "grad_norm": 0.09323889017105103, "learning_rate": 8.280558580299867e-05, "loss": 0.0031, "step": 30860 }, { "epoch": 273.1858407079646, "grad_norm": 0.06295816600322723, "learning_rate": 8.279310587166661e-05, "loss": 0.003, "step": 30870 }, { "epoch": 273.27433628318585, "grad_norm": 0.07861677557229996, "learning_rate": 8.278062235412978e-05, "loss": 0.003, "step": 30880 }, { "epoch": 273.3628318584071, "grad_norm": 0.06405859440565109, "learning_rate": 8.276813525175339e-05, "loss": 0.0029, "step": 30890 }, { "epoch": 273.4513274336283, "grad_norm": 0.05851652845740318, "learning_rate": 8.2755644565903e-05, "loss": 0.003, "step": 30900 }, { "epoch": 273.53982300884957, "grad_norm": 0.0560171939432621, "learning_rate": 8.274315029794454e-05, "loss": 0.0032, "step": 30910 }, { "epoch": 273.62831858407077, "grad_norm": 0.077827088534832, "learning_rate": 8.273065244924443e-05, "loss": 0.0033, "step": 30920 }, { "epoch": 273.716814159292, "grad_norm": 0.056711237877607346, "learning_rate": 8.271815102116936e-05, "loss": 0.0029, "step": 30930 }, { "epoch": 273.8053097345133, "grad_norm": 0.06462927907705307, "learning_rate": 8.270564601508648e-05, "loss": 0.003, "step": 30940 }, { "epoch": 273.8938053097345, "grad_norm": 0.06514304131269455, "learning_rate": 8.269313743236333e-05, "loss": 0.0031, "step": 30950 }, { "epoch": 273.98230088495575, "grad_norm": 0.07211273908615112, "learning_rate": 8.268062527436783e-05, "loss": 0.0027, "step": 30960 }, { "epoch": 274.070796460177, "grad_norm": 0.06832178682088852, "learning_rate": 8.266810954246828e-05, "loss": 0.0031, "step": 30970 }, { "epoch": 274.1592920353982, "grad_norm": 0.05487721413373947, "learning_rate": 8.265559023803338e-05, "loss": 0.0025, "step": 30980 }, { "epoch": 274.24778761061947, "grad_norm": 0.06871311366558075, "learning_rate": 8.264306736243223e-05, "loss": 0.0031, "step": 30990 }, { "epoch": 274.3362831858407, "grad_norm": 0.07978742569684982, "learning_rate": 8.263054091703432e-05, "loss": 0.0029, "step": 31000 }, { "epoch": 274.42477876106193, "grad_norm": 0.06752348691225052, "learning_rate": 8.26180109032095e-05, "loss": 0.0033, "step": 31010 }, { "epoch": 274.5132743362832, "grad_norm": 0.057503972202539444, "learning_rate": 8.260547732232807e-05, "loss": 0.003, "step": 31020 }, { "epoch": 274.60176991150445, "grad_norm": 0.05510031804442406, "learning_rate": 8.259294017576063e-05, "loss": 0.0028, "step": 31030 }, { "epoch": 274.69026548672565, "grad_norm": 0.0641007274389267, "learning_rate": 8.258039946487828e-05, "loss": 0.003, "step": 31040 }, { "epoch": 274.7787610619469, "grad_norm": 0.0658622533082962, "learning_rate": 8.256785519105241e-05, "loss": 0.0028, "step": 31050 }, { "epoch": 274.86725663716817, "grad_norm": 0.0844467282295227, "learning_rate": 8.255530735565488e-05, "loss": 0.0032, "step": 31060 }, { "epoch": 274.95575221238937, "grad_norm": 0.08807449042797089, "learning_rate": 8.254275596005787e-05, "loss": 0.0031, "step": 31070 }, { "epoch": 275.04424778761063, "grad_norm": 0.07292008399963379, "learning_rate": 8.2530201005634e-05, "loss": 0.003, "step": 31080 }, { "epoch": 275.13274336283183, "grad_norm": 0.055302031338214874, "learning_rate": 8.251764249375626e-05, "loss": 0.003, "step": 31090 }, { "epoch": 275.2212389380531, "grad_norm": 0.07376070320606232, "learning_rate": 8.250508042579803e-05, "loss": 0.0031, "step": 31100 }, { "epoch": 275.30973451327435, "grad_norm": 0.07500667870044708, "learning_rate": 8.249251480313307e-05, "loss": 0.003, "step": 31110 }, { "epoch": 275.39823008849555, "grad_norm": 0.0885225236415863, "learning_rate": 8.247994562713555e-05, "loss": 0.0033, "step": 31120 }, { "epoch": 275.4867256637168, "grad_norm": 0.070461206138134, "learning_rate": 8.246737289917998e-05, "loss": 0.003, "step": 31130 }, { "epoch": 275.57522123893807, "grad_norm": 0.09478802233934402, "learning_rate": 8.245479662064135e-05, "loss": 0.0033, "step": 31140 }, { "epoch": 275.6637168141593, "grad_norm": 0.09381696581840515, "learning_rate": 8.244221679289496e-05, "loss": 0.003, "step": 31150 }, { "epoch": 275.75221238938053, "grad_norm": 0.08834349364042282, "learning_rate": 8.242963341731652e-05, "loss": 0.003, "step": 31160 }, { "epoch": 275.8407079646018, "grad_norm": 0.08704610913991928, "learning_rate": 8.24170464952821e-05, "loss": 0.0033, "step": 31170 }, { "epoch": 275.929203539823, "grad_norm": 0.09262850135564804, "learning_rate": 8.240445602816824e-05, "loss": 0.003, "step": 31180 }, { "epoch": 276.01769911504425, "grad_norm": 0.06599617004394531, "learning_rate": 8.239186201735179e-05, "loss": 0.0032, "step": 31190 }, { "epoch": 276.1061946902655, "grad_norm": 0.06624548882246017, "learning_rate": 8.237926446420998e-05, "loss": 0.0031, "step": 31200 }, { "epoch": 276.1946902654867, "grad_norm": 0.06383863091468811, "learning_rate": 8.236666337012052e-05, "loss": 0.0035, "step": 31210 }, { "epoch": 276.283185840708, "grad_norm": 0.054547809064388275, "learning_rate": 8.23540587364614e-05, "loss": 0.0032, "step": 31220 }, { "epoch": 276.37168141592923, "grad_norm": 0.09256267547607422, "learning_rate": 8.234145056461107e-05, "loss": 0.0033, "step": 31230 }, { "epoch": 276.46017699115043, "grad_norm": 0.08659355342388153, "learning_rate": 8.232883885594832e-05, "loss": 0.0033, "step": 31240 }, { "epoch": 276.5486725663717, "grad_norm": 0.08453293144702911, "learning_rate": 8.231622361185236e-05, "loss": 0.0029, "step": 31250 }, { "epoch": 276.6371681415929, "grad_norm": 0.08448074758052826, "learning_rate": 8.230360483370278e-05, "loss": 0.0033, "step": 31260 }, { "epoch": 276.72566371681415, "grad_norm": 0.06000560149550438, "learning_rate": 8.229098252287953e-05, "loss": 0.0029, "step": 31270 }, { "epoch": 276.8141592920354, "grad_norm": 0.06474163383245468, "learning_rate": 8.2278356680763e-05, "loss": 0.003, "step": 31280 }, { "epoch": 276.9026548672566, "grad_norm": 0.08055142313241959, "learning_rate": 8.22657273087339e-05, "loss": 0.0032, "step": 31290 }, { "epoch": 276.9911504424779, "grad_norm": 0.06472662091255188, "learning_rate": 8.225309440817336e-05, "loss": 0.0029, "step": 31300 }, { "epoch": 277.07964601769913, "grad_norm": 0.09503763169050217, "learning_rate": 8.224045798046293e-05, "loss": 0.003, "step": 31310 }, { "epoch": 277.16814159292034, "grad_norm": 0.1044878140091896, "learning_rate": 8.22278180269845e-05, "loss": 0.0031, "step": 31320 }, { "epoch": 277.2566371681416, "grad_norm": 0.09791107475757599, "learning_rate": 8.221517454912031e-05, "loss": 0.0032, "step": 31330 }, { "epoch": 277.34513274336285, "grad_norm": 0.08896005153656006, "learning_rate": 8.220252754825308e-05, "loss": 0.0032, "step": 31340 }, { "epoch": 277.43362831858406, "grad_norm": 0.09640409052371979, "learning_rate": 8.218987702576586e-05, "loss": 0.003, "step": 31350 }, { "epoch": 277.5221238938053, "grad_norm": 0.08072144538164139, "learning_rate": 8.217722298304207e-05, "loss": 0.0031, "step": 31360 }, { "epoch": 277.6106194690266, "grad_norm": 0.06399810314178467, "learning_rate": 8.216456542146557e-05, "loss": 0.0033, "step": 31370 }, { "epoch": 277.6991150442478, "grad_norm": 0.07878907769918442, "learning_rate": 8.215190434242055e-05, "loss": 0.0028, "step": 31380 }, { "epoch": 277.78761061946904, "grad_norm": 0.07264313846826553, "learning_rate": 8.213923974729161e-05, "loss": 0.0032, "step": 31390 }, { "epoch": 277.87610619469024, "grad_norm": 0.07526564598083496, "learning_rate": 8.212657163746373e-05, "loss": 0.0031, "step": 31400 }, { "epoch": 277.9646017699115, "grad_norm": 0.06251968443393707, "learning_rate": 8.211390001432227e-05, "loss": 0.0031, "step": 31410 }, { "epoch": 278.05309734513276, "grad_norm": 0.09143874794244766, "learning_rate": 8.210122487925297e-05, "loss": 0.003, "step": 31420 }, { "epoch": 278.14159292035396, "grad_norm": 0.08104745298624039, "learning_rate": 8.208854623364202e-05, "loss": 0.0028, "step": 31430 }, { "epoch": 278.2300884955752, "grad_norm": 0.05914970114827156, "learning_rate": 8.207586407887589e-05, "loss": 0.0033, "step": 31440 }, { "epoch": 278.3185840707965, "grad_norm": 0.06940729171037674, "learning_rate": 8.206317841634148e-05, "loss": 0.0033, "step": 31450 }, { "epoch": 278.4070796460177, "grad_norm": 0.06395614147186279, "learning_rate": 8.205048924742609e-05, "loss": 0.0032, "step": 31460 }, { "epoch": 278.49557522123894, "grad_norm": 0.08829939365386963, "learning_rate": 8.203779657351738e-05, "loss": 0.0032, "step": 31470 }, { "epoch": 278.5840707964602, "grad_norm": 0.07255679368972778, "learning_rate": 8.20251003960034e-05, "loss": 0.0033, "step": 31480 }, { "epoch": 278.6725663716814, "grad_norm": 0.06963211297988892, "learning_rate": 8.201240071627258e-05, "loss": 0.0031, "step": 31490 }, { "epoch": 278.76106194690266, "grad_norm": 0.07635489851236343, "learning_rate": 8.199969753571377e-05, "loss": 0.003, "step": 31500 }, { "epoch": 278.8495575221239, "grad_norm": 0.0690208226442337, "learning_rate": 8.198699085571615e-05, "loss": 0.0033, "step": 31510 }, { "epoch": 278.9380530973451, "grad_norm": 0.06556485593318939, "learning_rate": 8.197428067766928e-05, "loss": 0.003, "step": 31520 }, { "epoch": 279.0265486725664, "grad_norm": 0.07142473012208939, "learning_rate": 8.196156700296316e-05, "loss": 0.0033, "step": 31530 }, { "epoch": 279.11504424778764, "grad_norm": 0.08324439823627472, "learning_rate": 8.194884983298814e-05, "loss": 0.0031, "step": 31540 }, { "epoch": 279.20353982300884, "grad_norm": 0.06740289181470871, "learning_rate": 8.193612916913491e-05, "loss": 0.003, "step": 31550 }, { "epoch": 279.2920353982301, "grad_norm": 0.07061455398797989, "learning_rate": 8.192340501279463e-05, "loss": 0.0033, "step": 31560 }, { "epoch": 279.3805309734513, "grad_norm": 0.0771554633975029, "learning_rate": 8.191067736535876e-05, "loss": 0.0029, "step": 31570 }, { "epoch": 279.46902654867256, "grad_norm": 0.07043633610010147, "learning_rate": 8.18979462282192e-05, "loss": 0.0035, "step": 31580 }, { "epoch": 279.5575221238938, "grad_norm": 0.05164299160242081, "learning_rate": 8.188521160276819e-05, "loss": 0.0031, "step": 31590 }, { "epoch": 279.646017699115, "grad_norm": 0.06932879239320755, "learning_rate": 8.187247349039837e-05, "loss": 0.0031, "step": 31600 }, { "epoch": 279.7345132743363, "grad_norm": 0.06877367943525314, "learning_rate": 8.185973189250278e-05, "loss": 0.0034, "step": 31610 }, { "epoch": 279.82300884955754, "grad_norm": 0.09597580879926682, "learning_rate": 8.184698681047482e-05, "loss": 0.003, "step": 31620 }, { "epoch": 279.91150442477874, "grad_norm": 0.0605299286544323, "learning_rate": 8.183423824570827e-05, "loss": 0.0031, "step": 31630 }, { "epoch": 280.0, "grad_norm": 0.08081115782260895, "learning_rate": 8.182148619959725e-05, "loss": 0.0033, "step": 31640 }, { "epoch": 280.08849557522126, "grad_norm": 0.05291002243757248, "learning_rate": 8.180873067353636e-05, "loss": 0.0033, "step": 31650 }, { "epoch": 280.17699115044246, "grad_norm": 0.08321083337068558, "learning_rate": 8.179597166892052e-05, "loss": 0.0036, "step": 31660 }, { "epoch": 280.2654867256637, "grad_norm": 0.08418931066989899, "learning_rate": 8.178320918714501e-05, "loss": 0.0034, "step": 31670 }, { "epoch": 280.353982300885, "grad_norm": 0.08587847650051117, "learning_rate": 8.177044322960554e-05, "loss": 0.0032, "step": 31680 }, { "epoch": 280.4424778761062, "grad_norm": 0.07803655415773392, "learning_rate": 8.175767379769816e-05, "loss": 0.0031, "step": 31690 }, { "epoch": 280.53097345132744, "grad_norm": 0.06262493878602982, "learning_rate": 8.174490089281932e-05, "loss": 0.0034, "step": 31700 }, { "epoch": 280.6194690265487, "grad_norm": 0.06814960390329361, "learning_rate": 8.173212451636584e-05, "loss": 0.0033, "step": 31710 }, { "epoch": 280.7079646017699, "grad_norm": 0.05747896805405617, "learning_rate": 8.171934466973493e-05, "loss": 0.0033, "step": 31720 }, { "epoch": 280.79646017699116, "grad_norm": 0.07480241358280182, "learning_rate": 8.170656135432418e-05, "loss": 0.0028, "step": 31730 }, { "epoch": 280.88495575221236, "grad_norm": 0.0657065361738205, "learning_rate": 8.169377457153155e-05, "loss": 0.003, "step": 31740 }, { "epoch": 280.9734513274336, "grad_norm": 0.06275679916143417, "learning_rate": 8.168098432275539e-05, "loss": 0.0029, "step": 31750 }, { "epoch": 281.0619469026549, "grad_norm": 0.07578925043344498, "learning_rate": 8.166819060939442e-05, "loss": 0.0031, "step": 31760 }, { "epoch": 281.1504424778761, "grad_norm": 0.09182662516832352, "learning_rate": 8.165539343284772e-05, "loss": 0.0033, "step": 31770 }, { "epoch": 281.23893805309734, "grad_norm": 0.08031430095434189, "learning_rate": 8.16425927945148e-05, "loss": 0.003, "step": 31780 }, { "epoch": 281.3274336283186, "grad_norm": 0.07726019620895386, "learning_rate": 8.162978869579551e-05, "loss": 0.0029, "step": 31790 }, { "epoch": 281.4159292035398, "grad_norm": 0.057472456246614456, "learning_rate": 8.161698113809007e-05, "loss": 0.0038, "step": 31800 }, { "epoch": 281.50442477876106, "grad_norm": 0.07451364398002625, "learning_rate": 8.160417012279911e-05, "loss": 0.0029, "step": 31810 }, { "epoch": 281.5929203539823, "grad_norm": 0.07824298739433289, "learning_rate": 8.159135565132363e-05, "loss": 0.0033, "step": 31820 }, { "epoch": 281.6814159292035, "grad_norm": 0.05895956605672836, "learning_rate": 8.157853772506498e-05, "loss": 0.0033, "step": 31830 }, { "epoch": 281.7699115044248, "grad_norm": 0.07417383044958115, "learning_rate": 8.156571634542494e-05, "loss": 0.0032, "step": 31840 }, { "epoch": 281.85840707964604, "grad_norm": 0.06643228232860565, "learning_rate": 8.15528915138056e-05, "loss": 0.0031, "step": 31850 }, { "epoch": 281.94690265486724, "grad_norm": 0.059420060366392136, "learning_rate": 8.154006323160949e-05, "loss": 0.0029, "step": 31860 }, { "epoch": 282.0353982300885, "grad_norm": 0.07440023124217987, "learning_rate": 8.152723150023949e-05, "loss": 0.0028, "step": 31870 }, { "epoch": 282.12389380530976, "grad_norm": 0.06191738694906235, "learning_rate": 8.151439632109886e-05, "loss": 0.003, "step": 31880 }, { "epoch": 282.21238938053096, "grad_norm": 0.05560492351651192, "learning_rate": 8.150155769559122e-05, "loss": 0.003, "step": 31890 }, { "epoch": 282.3008849557522, "grad_norm": 0.0729057639837265, "learning_rate": 8.148871562512058e-05, "loss": 0.0028, "step": 31900 }, { "epoch": 282.3893805309734, "grad_norm": 0.0758318305015564, "learning_rate": 8.147587011109136e-05, "loss": 0.0031, "step": 31910 }, { "epoch": 282.4778761061947, "grad_norm": 0.0750914141535759, "learning_rate": 8.14630211549083e-05, "loss": 0.003, "step": 31920 }, { "epoch": 282.56637168141594, "grad_norm": 0.06709709763526917, "learning_rate": 8.145016875797655e-05, "loss": 0.0034, "step": 31930 }, { "epoch": 282.65486725663715, "grad_norm": 0.08869866281747818, "learning_rate": 8.143731292170164e-05, "loss": 0.0031, "step": 31940 }, { "epoch": 282.7433628318584, "grad_norm": 0.0770382359623909, "learning_rate": 8.142445364748944e-05, "loss": 0.0033, "step": 31950 }, { "epoch": 282.83185840707966, "grad_norm": 0.0635785236954689, "learning_rate": 8.141159093674624e-05, "loss": 0.0034, "step": 31960 }, { "epoch": 282.92035398230087, "grad_norm": 0.05893268063664436, "learning_rate": 8.139872479087869e-05, "loss": 0.0031, "step": 31970 }, { "epoch": 283.0088495575221, "grad_norm": 0.06397490948438644, "learning_rate": 8.13858552112938e-05, "loss": 0.0031, "step": 31980 }, { "epoch": 283.0973451327434, "grad_norm": 0.06711224466562271, "learning_rate": 8.137298219939895e-05, "loss": 0.0033, "step": 31990 }, { "epoch": 283.1858407079646, "grad_norm": 0.07149463891983032, "learning_rate": 8.136010575660196e-05, "loss": 0.0031, "step": 32000 } ], "logging_steps": 10, "max_steps": 100000, "num_input_tokens_seen": 0, "num_train_epochs": 885, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 512, "trial_name": null, "trial_params": null }